lib/libc/arch/mips64/gen/cacheflush.c
27
cacheflush(void *addr, int nbytes, int cache)
lib/libc/arch/mips64/gen/cacheflush.c
33
args.which = cache;
lib/libc/stdlib/malloc.c
2635
struct smallcache *cache;
lib/libc/stdlib/malloc.c
2640
cache = &d->smallcache[i];
lib/libc/stdlib/malloc.c
2641
if (cache->length != 0)
lib/libc/stdlib/malloc.c
2642
ulog("%zu(%u): %u = %zu\n", i + 1, cache->max,
lib/libc/stdlib/malloc.c
2643
cache->length, cache->length * (i + 1));
lib/libc/stdlib/malloc.c
2644
total += cache->length * (i + 1);
lib/libc/stdlib/malloc.c
801
struct smallcache *cache;
lib/libc/stdlib/malloc.c
855
cache = &d->smallcache[psz - 1];
lib/libc/stdlib/malloc.c
856
if (cache->length == cache->max) {
lib/libc/stdlib/malloc.c
859
i = getrbyte(d) & (cache->max - 1);
lib/libc/stdlib/malloc.c
860
r = cache->pages[i];
lib/libc/stdlib/malloc.c
868
cache->length--;
lib/libc/stdlib/malloc.c
870
i = cache->length;
lib/libc/stdlib/malloc.c
879
cache->pages[i] = p;
lib/libc/stdlib/malloc.c
880
cache->length++;
lib/libc/stdlib/malloc.c
889
struct smallcache *cache;
lib/libc/stdlib/malloc.c
926
cache = &d->smallcache[psz - 1];
lib/libc/stdlib/malloc.c
927
if (cache->length > 0) {
lib/libc/stdlib/malloc.c
929
if (cache->length == 1)
lib/libc/stdlib/malloc.c
930
p = cache->pages[--cache->length];
lib/libc/stdlib/malloc.c
932
i = getrbyte(d) % cache->length;
lib/libc/stdlib/malloc.c
933
p = cache->pages[i];
lib/libc/stdlib/malloc.c
934
cache->pages[i] = cache->pages[--cache->length];
lib/libc/stdlib/malloc.c
951
p = MMAP(cache->max * sz, d->mmap_flag);
lib/libc/stdlib/malloc.c
953
STATS_ADD(d->malloc_used, cache->max * sz);
lib/libc/stdlib/malloc.c
954
cache->length = cache->max - 1;
lib/libc/stdlib/malloc.c
955
for (i = 0; i < cache->max - 1; i++) {
lib/libc/stdlib/malloc.c
957
cache->pages[i] = q;
lib/libc/stdlib/malloc.c
959
*(uintptr_t*)&cache->pages[i] |= 1UL;
lib/libc/stdlib/malloc.c
962
mprotect(p, (cache->max - 1) * sz,
lib/libc/stdlib/malloc.c
964
p = (char*)p + (cache->max - 1) * sz;
lib/libssl/ssl_sess.c
1122
struct lhash_st_SSL_SESSION *cache;
lib/libssl/ssl_sess.c
1132
(void)lh_SSL_SESSION_delete(p->cache, s);
lib/libssl/ssl_sess.c
1157
tp.cache = s->sessions;
lib/libssl/ssl_sess.c
1158
if (tp.cache == NULL)
lib/libssl/ssl_sess.c
1163
lh_SSL_SESSION_doall_arg(tp.cache, timeout_LHASH_DOALL_ARG,
regress/lib/libfuse/fuse-opt-parse.c
119
assert(data.cache == 0);
regress/lib/libfuse/fuse-opt-parse.c
155
assert(data.cache == 0);
regress/lib/libfuse/fuse-opt-parse.c
208
assert(data.cache == 0);
regress/lib/libfuse/fuse-opt-parse.c
267
assert(data.cache == 0);
regress/lib/libfuse/fuse-opt-parse.c
34
int cache;
regress/lib/libfuse/fuse-opt-parse.c
56
DATA_OPT("cache=yes", cache, 1),
regress/lib/libfuse/fuse-opt-parse.c
57
DATA_OPT("cache=no", cache, 0),
sbin/bioctl/bioctl.c
392
char *status, *cache;
sbin/bioctl/bioctl.c
470
cache = BIOC_CVWRITEBACK_S;
sbin/bioctl/bioctl.c
473
cache = BIOC_CVWRITETHROUGH_S;
sbin/bioctl/bioctl.c
477
cache = BIOC_CVUNKNOWN_S;
sbin/bioctl/bioctl.c
509
bv.bv_level, percent, seconds, cache);
sbin/bioctl/bioctl.c
513
bv.bv_level, percent, seconds, cache);
sbin/iked/sntrup761.c
1147
static void HashConfirm(unsigned char *h,const unsigned char *r,const unsigned char *pk,const unsigned char *cache)
sbin/iked/sntrup761.c
1154
for (i = 0;i < Hash_bytes;++i) x[Hash_bytes+i] = cache[i];
sbin/iked/sntrup761.c
1160
for (i = 0;i < Hash_bytes;++i) x[Inputs_bytes+i] = cache[i];
sbin/iked/sntrup761.c
1200
static void Hide(unsigned char *c,unsigned char *r_enc,const Inputs r,const unsigned char *pk,const unsigned char *cache)
sbin/iked/sntrup761.c
1204
HashConfirm(c,r_enc,pk,cache);
sbin/iked/sntrup761.c
1212
unsigned char cache[Hash_bytes];
sbin/iked/sntrup761.c
1214
Hash_prefix(cache,4,pk,PublicKeys_bytes);
sbin/iked/sntrup761.c
1216
Hide(c,r_enc,r,pk,cache);
sbin/iked/sntrup761.c
1235
const unsigned char *cache = rho + Inputs_bytes;
sbin/iked/sntrup761.c
1243
Hide(cnew,r_enc,r,pk,cache);
sys/arch/arm/arm/pmap7.c
2495
pmap_map_section(vaddr_t l1pt, vaddr_t va, paddr_t pa, int prot, int cache)
sys/arch/arm/arm/pmap7.c
2500
switch (cache) {
sys/arch/arm/arm/pmap7.c
2526
pmap_map_entry(vaddr_t l1pt, vaddr_t va, paddr_t pa, int prot, int cache)
sys/arch/arm/arm/pmap7.c
2532
switch (cache) {
sys/arch/arm/arm/pmap7.c
2604
int prot, int cache)
sys/arch/arm/arm/pmap7.c
2618
"prot=0x%x cache=%d\n", pa, va, size, resid, prot, cache);
sys/arch/arm/arm/pmap7.c
2621
switch (cache) {
sys/arch/arm64/arm64/bus_dma.c
498
int curseg, pmap_flags, cache;
sys/arch/arm64/arm64/bus_dma.c
512
cache = PMAP_CACHE_WB;
sys/arch/arm64/arm64/bus_dma.c
515
cache = PMAP_CACHE_CI;
sys/arch/arm64/arm64/bus_dma.c
525
cache);
sys/arch/arm64/arm64/bus_space.c
194
int cache = PMAP_CACHE_DEV_NGNRNE;
sys/arch/arm64/arm64/bus_space.c
197
cache = PMAP_CACHE_WB;
sys/arch/arm64/arm64/bus_space.c
199
cache = PMAP_CACHE_CI;
sys/arch/arm64/arm64/bus_space.c
201
cache = PMAP_CACHE_DEV_NGNRE;
sys/arch/arm64/arm64/bus_space.c
213
pmap_kenter_cache(va, pa, PROT_READ | PROT_WRITE, cache);
sys/arch/arm64/arm64/machdep.c
1165
int cache = PMAP_CACHE_DEV_NGNRNE;
sys/arch/arm64/arm64/machdep.c
1168
cache = PMAP_CACHE_CI;
sys/arch/arm64/arm64/machdep.c
1178
pmap_kenter_cache(va, pa, PROT_READ | PROT_WRITE, cache);
sys/arch/arm64/arm64/pmap.c
606
int cache = PMAP_CACHE_WB;
sys/arch/arm64/arm64/pmap.c
609
cache = PMAP_CACHE_CI;
sys/arch/arm64/arm64/pmap.c
611
cache = PMAP_CACHE_DEV_NGNRNE;
sys/arch/arm64/arm64/pmap.c
619
(pted->pted_va & PMAP_CACHE_BITS) == cache) {
sys/arch/arm64/arm64/pmap.c
662
pmap_fill_pte(pm, va, pa, pted, prot, flags, cache);
sys/arch/arm64/arm64/pmap.c
759
_pmap_kenter_pa(vaddr_t va, paddr_t pa, vm_prot_t prot, int flags, int cache)
sys/arch/arm64/arm64/pmap.c
778
pmap_fill_pte(pm, va, pa, pted, prot, flags, cache);
sys/arch/arm64/arm64/pmap.c
785
if (pg && (cache == PMAP_CACHE_CI || cache == PMAP_CACHE_DEV_NGNRNE))
sys/arch/arm64/arm64/pmap.c
792
int cache = PMAP_CACHE_WB;
sys/arch/arm64/arm64/pmap.c
795
cache = PMAP_CACHE_CI;
sys/arch/arm64/arm64/pmap.c
797
cache = PMAP_CACHE_DEV_NGNRNE;
sys/arch/arm64/arm64/pmap.c
799
_pmap_kenter_pa(va, pa, prot, prot, cache);
sys/arch/arm64/arm64/pmap.c
857
vm_prot_t prot, int flags, int cache)
sys/arch/arm64/arm64/pmap.c
862
switch (cache) {
sys/arch/arm64/arm64/pmap.c
876
pted->pted_va |= cache;
sys/arch/arm64/dev/smmu.c
1084
vm_prot_t prot, int flags, int cache)
sys/arch/arm64/dev/smmu.c
1090
switch (cache) {
sys/arch/arm64/dev/smmu.c
1105
pted |= cache;
sys/arch/arm64/dev/smmu.c
1218
int flags, int cache)
sys/arch/arm64/dev/smmu.c
1228
smmu_map(dom, va, pa, prot, flags, cache);
sys/arch/arm64/dev/smmu.c
1235
int flags, int cache)
sys/arch/arm64/dev/smmu.c
1246
pted = smmu_fill_pte(dom, va, pa, prot, flags, cache);
sys/arch/mips64/mips64/cache_loongson2.c
123
cache(IndexInvalidate_I, 0, sva);
sys/arch/mips64/mips64/cache_loongson2.c
130
cache(IndexWBInvalidate_D, 0, sva);
sys/arch/mips64/mips64/cache_loongson2.c
131
cache(IndexWBInvalidate_D, 1, sva);
sys/arch/mips64/mips64/cache_loongson2.c
132
cache(IndexWBInvalidate_D, 2, sva);
sys/arch/mips64/mips64/cache_loongson2.c
133
cache(IndexWBInvalidate_D, 3, sva);
sys/arch/mips64/mips64/cache_loongson2.c
140
cache(IndexWBInvalidate_S, 0, sva);
sys/arch/mips64/mips64/cache_loongson2.c
141
cache(IndexWBInvalidate_S, 1, sva);
sys/arch/mips64/mips64/cache_loongson2.c
142
cache(IndexWBInvalidate_S, 2, sva);
sys/arch/mips64/mips64/cache_loongson2.c
143
cache(IndexWBInvalidate_S, 3, sva);
sys/arch/mips64/mips64/cache_loongson2.c
166
cache(IndexInvalidate_I, 0, sva);
sys/arch/mips64/mips64/cache_loongson2.c
198
cache(IndexInvalidate_I, 0, sva);
sys/arch/mips64/mips64/cache_loongson2.c
223
cache(HitWBInvalidate_D, 0, va);
sys/arch/mips64/mips64/cache_loongson2.c
225
cache(HitWBInvalidate_S, 0, va);
sys/arch/mips64/mips64/cache_loongson2.c
241
cache(HitWBInvalidate_D, 0, va);
sys/arch/mips64/mips64/cache_loongson2.c
253
cache(HitWBInvalidate_S, 0, va);
sys/arch/mips64/mips64/cache_loongson2.c
287
cache(HitInvalidate_D, 0, va);
sys/arch/mips64/mips64/cache_loongson2.c
299
cache(HitInvalidate_S, 0, va);
sys/arch/mips64/mips64/cache_loongson2.c
346
cache(HitWBInvalidate_D, 0, va);
sys/arch/mips64/mips64/cache_loongson2.c
347
cache(HitWBInvalidate_S, 0, va);
sys/arch/mips64/mips64/cache_loongson2.c
352
cache(HitWBInvalidate_D, 0, va + sz - LS2F_CACHE_LINE);
sys/arch/mips64/mips64/cache_loongson2.c
353
cache(HitWBInvalidate_S, 0, va + sz - LS2F_CACHE_LINE);
sys/arch/mips64/mips64/cache_mips64r2.c
144
cache(HitWBInvalidate_D, va);
sys/arch/mips64/mips64/cache_mips64r2.c
156
cache(HitWBInvalidate_S, va);
sys/arch/mips64/mips64/cache_mips64r2.c
168
cache(HitWBInvalidate_T, va);
sys/arch/mips64/mips64/cache_mips64r2.c
180
cache(HitInvalidate_D, va);
sys/arch/mips64/mips64/cache_mips64r2.c
192
cache(HitInvalidate_S, va);
sys/arch/mips64/mips64/cache_mips64r2.c
204
cache(HitInvalidate_T, va);
sys/arch/mips64/mips64/cache_mips64r2.c
220
cache(IndexInvalidate_I, sva);
sys/arch/mips64/mips64/cache_mips64r2.c
227
cache(IndexWBInvalidate_D, sva);
sys/arch/mips64/mips64/cache_mips64r2.c
235
cache(IndexWBInvalidate_S, sva);
sys/arch/mips64/mips64/cache_mips64r2.c
244
cache(IndexWBInvalidate_T, sva);
sys/arch/mips64/mips64/cache_mips64r2.c
273
cache(IndexInvalidate_I, iva);
sys/arch/mips64/mips64/cache_mips64r2.c
315
cache(IndexWBInvalidate_D, iva);
sys/arch/mips64/mips64/cache_mips64r2.c
398
cache(HitWBInvalidate_D, va);
sys/arch/mips64/mips64/cache_mips64r2.c
404
cache(HitWBInvalidate_D, va + sz);
sys/arch/mips64/mips64/cache_mips64r2.c
434
cache(HitWBInvalidate_S, va);
sys/arch/mips64/mips64/cache_mips64r2.c
440
cache(HitWBInvalidate_S, va + sz);
sys/arch/mips64/mips64/cache_mips64r2.c
471
cache(HitWBInvalidate_S, va);
sys/arch/mips64/mips64/cache_mips64r2.c
477
cache(HitWBInvalidate_S, va + sz);
sys/arch/powerpc/powerpc/pmap.c
541
int cache, error = 0;
sys/arch/powerpc/powerpc/pmap.c
578
cache = PMAP_CACHE_WT;
sys/arch/powerpc/powerpc/pmap.c
580
cache = PMAP_CACHE_WB;
sys/arch/powerpc/powerpc/pmap.c
582
cache = PMAP_CACHE_CI;
sys/arch/powerpc/powerpc/pmap.c
586
pmap_fill_pte64(pm, va, pa, pted, prot, cache);
sys/arch/powerpc/powerpc/pmap.c
588
pmap_fill_pte32(pm, va, pa, pted, prot, cache);
sys/arch/powerpc/powerpc/pmap.c
716
int cache;
sys/arch/powerpc/powerpc/pmap.c
743
cache = PMAP_CACHE_WT;
sys/arch/powerpc/powerpc/pmap.c
745
cache = PMAP_CACHE_WB;
sys/arch/powerpc/powerpc/pmap.c
747
cache = PMAP_CACHE_CI;
sys/arch/powerpc/powerpc/pmap.c
751
pmap_fill_pte64(pm, va, pa, pted, prot, cache);
sys/arch/powerpc/powerpc/pmap.c
753
pmap_fill_pte32(pm, va, pa, pted, prot, cache);
sys/arch/powerpc/powerpc/pmap.c
878
vm_prot_t prot, int cache)
sys/arch/powerpc/powerpc/pmap.c
892
if (cache == PMAP_CACHE_WB)
sys/arch/powerpc/powerpc/pmap.c
894
else if (cache == PMAP_CACHE_WT)
sys/arch/powerpc/powerpc/pmap.c
922
vm_prot_t prot, int cache)
sys/arch/powerpc/powerpc/pmap.c
934
if (cache == PMAP_CACHE_WB)
sys/arch/powerpc/powerpc/pmap.c
936
else if (cache == PMAP_CACHE_WT)
sys/arch/powerpc64/powerpc64/bus_space.c
175
int cache = (flags & BUS_SPACE_MAP_CACHEABLE) ? 0 : PMAP_NOCACHE;
sys/arch/powerpc64/powerpc64/bus_space.c
187
pmap_kenter_pa(va, pa | cache, PROT_READ | PROT_WRITE);
sys/arch/powerpc64/powerpc64/cpu.c
150
uint32_t cache;
sys/arch/powerpc64/powerpc64/cpu.c
176
cache = OF_getpropint(node, "l2-cache", 0);
sys/arch/powerpc64/powerpc64/cpu.c
177
node = OF_getnodebyphandle(cache);
sys/arch/powerpc64/powerpc64/pmap.c
1065
int cache = PMAP_CACHE_WB;
sys/arch/powerpc64/powerpc64/pmap.c
1070
cache = PMAP_CACHE_CI;
sys/arch/powerpc64/powerpc64/pmap.c
1084
if (old_acwimgn == pmap_acwimgn(prot, cache) &&
sys/arch/powerpc64/powerpc64/pmap.c
1118
pmap_fill_pte(pm, va, pa, pted, prot, cache);
sys/arch/powerpc64/powerpc64/pmap.c
1131
if ((pg->pg_flags & PG_DEV) == 0 && cache != PMAP_CACHE_WB)
sys/arch/powerpc64/powerpc64/pmap.c
1329
int cache = (pa & PMAP_NOCACHE) ? PMAP_CACHE_CI : PMAP_CACHE_WB;
sys/arch/powerpc64/powerpc64/pmap.c
1340
pmap_fill_pte(pm, va, pa, &pted, prot, cache);
sys/arch/powerpc64/powerpc64/pmap.c
713
pmap_acwimgn(vm_prot_t prot, int cache)
sys/arch/powerpc64/powerpc64/pmap.c
720
if (cache == PMAP_CACHE_WB)
sys/arch/powerpc64/powerpc64/pmap.c
733
vm_prot_t prot, int cache)
sys/arch/powerpc64/powerpc64/pmap.c
754
pte->pte_lo |= pmap_acwimgn(prot, cache);
sys/arch/riscv64/riscv64/bus_dma.c
505
int curseg, pmap_flags, cache;
sys/arch/riscv64/riscv64/bus_dma.c
519
cache = PMAP_CACHE_WB;
sys/arch/riscv64/riscv64/bus_dma.c
522
cache = PMAP_CACHE_CI;
sys/arch/riscv64/riscv64/bus_dma.c
533
cache);
sys/arch/riscv64/riscv64/bus_space.c
219
int cache = flags & BUS_SPACE_MAP_CACHEABLE ?
sys/arch/riscv64/riscv64/bus_space.c
232
pmap_kenter_cache(va, pa, PROT_READ | PROT_WRITE, cache);
sys/arch/riscv64/riscv64/cpu.c
340
uint32_t cache;
sys/arch/riscv64/riscv64/cpu.c
374
cache = OF_getpropint(node, "next-level-cache", 0);
sys/arch/riscv64/riscv64/cpu.c
375
node = OF_getnodebyphandle(cache);
sys/arch/riscv64/riscv64/pmap.c
536
int cache = PMAP_CACHE_WB;
sys/arch/riscv64/riscv64/pmap.c
539
cache = PMAP_CACHE_CI;
sys/arch/riscv64/riscv64/pmap.c
541
cache = PMAP_CACHE_DEV;
sys/arch/riscv64/riscv64/pmap.c
549
(pted->pted_va & PMAP_CACHE_BITS) == cache) {
sys/arch/riscv64/riscv64/pmap.c
593
pmap_fill_pte(pm, va, pa, pted, prot, flags, cache);
sys/arch/riscv64/riscv64/pmap.c
690
_pmap_kenter_pa(vaddr_t va, paddr_t pa, vm_prot_t prot, int flags, int cache)
sys/arch/riscv64/riscv64/pmap.c
697
if (pg && cache == PMAP_CACHE_CI) {
sys/arch/riscv64/riscv64/pmap.c
717
pmap_fill_pte(pm, va, pa, pted, prot, flags, cache);
sys/arch/riscv64/riscv64/pmap.c
731
int cache = PMAP_CACHE_WB;
sys/arch/riscv64/riscv64/pmap.c
734
cache = PMAP_CACHE_CI;
sys/arch/riscv64/riscv64/pmap.c
736
cache = PMAP_CACHE_DEV;
sys/arch/riscv64/riscv64/pmap.c
738
_pmap_kenter_pa(va, pa, prot, prot, cache);
sys/arch/riscv64/riscv64/pmap.c
801
vm_prot_t prot, int flags, int cache)
sys/arch/riscv64/riscv64/pmap.c
806
switch (cache) {
sys/arch/riscv64/riscv64/pmap.c
818
pted->pted_va |= cache;
sys/arch/sparc64/dev/creator.c
535
int32_t cache = sc->sc_fifo_cache;
sys/arch/sparc64/dev/creator.c
537
if (cache < n) {
sys/arch/sparc64/dev/creator.c
539
cache = FBC_READ(sc, FFB_FBC_UCSR);
sys/arch/sparc64/dev/creator.c
540
cache = (cache & FBC_UCSR_FIFO_MASK) - 8;
sys/arch/sparc64/dev/creator.c
541
} while (cache < n);
sys/arch/sparc64/dev/creator.c
543
sc->sc_fifo_cache = cache - n;
sys/arch/sparc64/include/pte.h
158
#define SUN4U_TSB_DATA(g,sz,pa,priv,write,cache,aliased,valid,ie) \
sys/arch/sparc64/include/pte.h
161
((cache)?((aliased)?SUN4U_TLB_CP:SUN4U_TLB_CACHE_MASK):SUN4U_TLB_E)|\
sys/arch/sparc64/include/pte.h
197
#define SUN4V_TSB_DATA(g,sz,pa,priv,write,cache,aliased,valid,ie) \
sys/arch/sparc64/include/pte.h
200
((cache)?((aliased)?SUN4V_TLB_CP:SUN4V_TLB_CACHE_MASK):SUN4V_TLB_E)|\
sys/arch/sparc64/sparc64/pmap.c
151
#define TSB_DATA(g,sz,pa,priv,write,cache,aliased,valid,ie) \
sys/arch/sparc64/sparc64/pmap.c
153
SUN4V_TSB_DATA(g,sz,pa,priv,write,cache,aliased,valid,ie) : \
sys/arch/sparc64/sparc64/pmap.c
154
SUN4U_TSB_DATA(g,sz,pa,priv,write,cache,aliased,valid,ie))
sys/dev/ic/cissreg.h
253
u_int8_t cache;
sys/dev/ic/sti.c
467
r->cache ? " cache" : "", r->btlb ? " btlb" : "",
sys/dev/ic/sti.c
482
BUS_SPACE_MAP_LINEAR | (r->cache ?
sys/dev/ic/stireg.h
177
u_int cache : 1; /* map in cache */
sys/dev/pci/arc.c
321
u_int8_t cache;
sys/dev/pci/drm/amd/amdkfd/kfd_crat.c
1161
static int kfd_parse_subtype_cache(struct crat_subtype_cache *cache,
sys/dev/pci/drm/amd/amdkfd/kfd_crat.c
1169
id = cache->processor_id_low;
sys/dev/pci/drm/amd/amdkfd/kfd_crat.c
1195
props->cache_level = cache->cache_level;
sys/dev/pci/drm/amd/amdkfd/kfd_crat.c
1196
props->cache_size = cache->cache_size;
sys/dev/pci/drm/amd/amdkfd/kfd_crat.c
1197
props->cacheline_size = cache->cache_line_size;
sys/dev/pci/drm/amd/amdkfd/kfd_crat.c
1198
props->cachelines_per_tag = cache->lines_per_tag;
sys/dev/pci/drm/amd/amdkfd/kfd_crat.c
1199
props->cache_assoc = cache->associativity;
sys/dev/pci/drm/amd/amdkfd/kfd_crat.c
1200
props->cache_latency = cache->cache_latency;
sys/dev/pci/drm/amd/amdkfd/kfd_crat.c
1202
memcpy(props->sibling_map, cache->sibling_map,
sys/dev/pci/drm/amd/amdkfd/kfd_crat.c
1208
if (cache->flags & CRAT_CACHE_FLAGS_DATA_CACHE)
sys/dev/pci/drm/amd/amdkfd/kfd_crat.c
1210
if (cache->flags & CRAT_CACHE_FLAGS_INST_CACHE)
sys/dev/pci/drm/amd/amdkfd/kfd_crat.c
1212
if (cache->flags & CRAT_CACHE_FLAGS_CPU_CACHE)
sys/dev/pci/drm/amd/amdkfd/kfd_crat.c
1214
if (cache->flags & CRAT_CACHE_FLAGS_SIMD_CACHE)
sys/dev/pci/drm/amd/amdkfd/kfd_crat.c
1312
struct crat_subtype_cache *cache;
sys/dev/pci/drm/amd/amdkfd/kfd_crat.c
1326
cache = (struct crat_subtype_cache *)sub_type_hdr;
sys/dev/pci/drm/amd/amdkfd/kfd_crat.c
1327
ret = kfd_parse_subtype_cache(cache, device_list);
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
115
struct kfd_cache_properties *cache;
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
1154
struct kfd_cache_properties *cache;
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
1171
list_for_each_entry(cache, &dev->cache_props, list)
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
1172
cache->gpu = dev->gpu;
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
130
cache = container_of(dev->cache_props.next,
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
132
list_del(&cache->list);
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
133
kfree(cache);
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
330
struct kfd_cache_properties *cache;
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
334
cache = container_of(attr, struct kfd_cache_properties, attr);
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
335
if (cache->gpu && kfd_devcgroup_check_permission(cache->gpu))
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
338
cache->processor_id_low);
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
339
sysfs_show_32bit_prop(buffer, offs, "level", cache->cache_level);
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
340
sysfs_show_32bit_prop(buffer, offs, "size", cache->cache_size);
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
342
cache->cacheline_size);
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
344
cache->cachelines_per_tag);
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
345
sysfs_show_32bit_prop(buffer, offs, "association", cache->cache_assoc);
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
346
sysfs_show_32bit_prop(buffer, offs, "latency", cache->cache_latency);
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
347
sysfs_show_32bit_prop(buffer, offs, "type", cache->cache_type);
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
350
for (i = 0; i < cache->sibling_map_size; i++)
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
351
for (j = 0; j < sizeof(cache->sibling_map[0])*8; j++)
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
354
(cache->sibling_map[i] >> j) & 1);
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
570
struct kfd_cache_properties *cache;
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
599
list_for_each_entry(cache, &dev->cache_props, list)
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
600
if (cache->kobj) {
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
601
kfd_remove_sysfs_file(cache->kobj,
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
602
&cache->attr);
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
603
cache->kobj = NULL;
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
646
struct kfd_cache_properties *cache;
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
734
list_for_each_entry(cache, &dev->cache_props, list) {
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
735
cache->kobj = kzalloc(sizeof(struct kobject), GFP_KERNEL);
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
736
if (!cache->kobj)
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
738
ret = kobject_init_and_add(cache->kobj, &cache_type,
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
741
kobject_put(cache->kobj);
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
745
cache->attr.name = "properties";
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
746
cache->attr.mode = KFD_SYSFS_FILE_MODE;
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
747
sysfs_attr_init(&cache->attr);
sys/dev/pci/drm/amd/amdkfd/kfd_topology.c
748
ret = sysfs_create_file(cache->kobj, &cache->attr);
sys/dev/pci/drm/amd/pm/legacy-dpm/kv_dpm.c
405
u32 cache = 0;
sys/dev/pci/drm/amd/pm/legacy-dpm/kv_dpm.c
412
cache |= ((config_regs->value << config_regs->shift) & config_regs->mask);
sys/dev/pci/drm/amd/pm/legacy-dpm/kv_dpm.c
428
data |= cache;
sys/dev/pci/drm/amd/pm/legacy-dpm/kv_dpm.c
429
cache = 0;
sys/dev/pci/drm/amd/pm/powerplay/hwmgr/smu7_powertune.c
901
uint32_t cache = 0;
sys/dev/pci/drm/amd/pm/powerplay/hwmgr/smu7_powertune.c
908
cache |= ((config_regs->value << config_regs->shift) & config_regs->mask);
sys/dev/pci/drm/amd/pm/powerplay/hwmgr/smu7_powertune.c
930
data |= cache;
sys/dev/pci/drm/amd/pm/powerplay/hwmgr/smu7_powertune.c
949
cache = 0;
sys/dev/pci/drm/amd/pm/swsmu/amdgpu_smu.c
3902
if (!table && tables[table_id].cache.size)
sys/dev/pci/drm/amd/pm/swsmu/amdgpu_smu.c
3903
return tables[table_id].cache.size;
sys/dev/pci/drm/amd/pm/swsmu/amdgpu_smu.c
3906
memcpy(table, tables[table_id].cache.buffer,
sys/dev/pci/drm/amd/pm/swsmu/amdgpu_smu.c
3907
tables[table_id].cache.size);
sys/dev/pci/drm/amd/pm/swsmu/amdgpu_smu.c
3908
return tables[table_id].cache.size;
sys/dev/pci/drm/amd/pm/swsmu/inc/amdgpu_smu.h
1684
table->cache.last_cache_time = time;
sys/dev/pci/drm/amd/pm/swsmu/inc/amdgpu_smu.h
1689
if (!table->cache.buffer || !table->cache.last_cache_time ||
sys/dev/pci/drm/amd/pm/swsmu/inc/amdgpu_smu.h
1690
!table->cache.interval || !table->cache.size ||
sys/dev/pci/drm/amd/pm/swsmu/inc/amdgpu_smu.h
1692
table->cache.last_cache_time +
sys/dev/pci/drm/amd/pm/swsmu/inc/amdgpu_smu.h
1693
msecs_to_jiffies(table->cache.interval)))
sys/dev/pci/drm/amd/pm/swsmu/inc/amdgpu_smu.h
1706
tables[table_id].cache.buffer = kzalloc(size, GFP_KERNEL);
sys/dev/pci/drm/amd/pm/swsmu/inc/amdgpu_smu.h
1707
if (!tables[table_id].cache.buffer)
sys/dev/pci/drm/amd/pm/swsmu/inc/amdgpu_smu.h
1710
tables[table_id].cache.last_cache_time = 0;
sys/dev/pci/drm/amd/pm/swsmu/inc/amdgpu_smu.h
1711
tables[table_id].cache.interval = cache_interval;
sys/dev/pci/drm/amd/pm/swsmu/inc/amdgpu_smu.h
1712
tables[table_id].cache.size = size;
sys/dev/pci/drm/amd/pm/swsmu/inc/amdgpu_smu.h
1723
if (tables[table_id].cache.buffer) {
sys/dev/pci/drm/amd/pm/swsmu/inc/amdgpu_smu.h
1724
kfree(tables[table_id].cache.buffer);
sys/dev/pci/drm/amd/pm/swsmu/inc/amdgpu_smu.h
1725
tables[table_id].cache.buffer = NULL;
sys/dev/pci/drm/amd/pm/swsmu/inc/amdgpu_smu.h
1726
tables[table_id].cache.last_cache_time = 0;
sys/dev/pci/drm/amd/pm/swsmu/inc/amdgpu_smu.h
1727
tables[table_id].cache.interval = 0;
sys/dev/pci/drm/amd/pm/swsmu/inc/amdgpu_smu.h
268
struct smu_table_cache cache;
sys/dev/pci/drm/amd/pm/swsmu/smu13/smu_v13_0_12_ppt.c
150
struct smu_table_cache *cache;
sys/dev/pci/drm/amd/pm/swsmu/smu13/smu_v13_0_12_ppt.c
164
cache = &(tables[SMU_TABLE_BASEBOARD_TEMP_METRICS].cache);
sys/dev/pci/drm/amd/pm/swsmu/smu13/smu_v13_0_12_ppt.c
166
(struct amdgpu_baseboard_temp_metrics_v1_0 *) cache->buffer;
sys/dev/pci/drm/amd/pm/swsmu/smu13/smu_v13_0_12_ppt.c
176
cache = &(tables[SMU_TABLE_GPUBOARD_TEMP_METRICS].cache);
sys/dev/pci/drm/amd/pm/swsmu/smu13/smu_v13_0_12_ppt.c
177
gpuboard_temp_metrics = (struct amdgpu_gpuboard_temp_metrics_v1_0 *)cache->buffer;
sys/dev/pci/drm/amd/pm/swsmu/smu13/smu_v13_0_12_ppt.c
454
memcpy(sys_table->cache.buffer, table->cpu_addr,
sys/dev/pci/drm/amd/pm/swsmu/smu13/smu_v13_0_12_ppt.c
612
metrics = (SystemMetricsTable_t *)sys_table->cache.buffer;
sys/dev/pci/drm/amd/pm/swsmu/smu13/smu_v13_0_12_ppt.c
651
data_table->cache.buffer;
sys/dev/pci/drm/amd/pm/swsmu/smu13/smu_v13_0_12_ppt.c
658
data_table->cache.buffer;
sys/dev/pci/drm/amd/pm/swsmu/smu13/smu_v13_0_12_ppt.c
667
metrics = (SystemMetricsTable_t *)sys_table->cache.buffer;
sys/dev/pci/drm/amd/pm/swsmu/smu13/smu_v13_0_12_ppt.c
714
memcpy(table, data_table->cache.buffer, size);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1102
static void reloc_cache_init(struct reloc_cache *cache,
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1105
cache->page = -1;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1106
cache->vaddr = 0;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1108
cache->graphics_ver = GRAPHICS_VER(i915);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1109
cache->has_llc = HAS_LLC(i915);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1110
cache->use_64bit_reloc = HAS_64BIT_RELOC(i915);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1111
cache->has_fence = cache->graphics_ver < 4;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1112
cache->needs_unfenced = INTEL_INFO(i915)->unfenced_needs_alignment;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1113
cache->node.flags = 0;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1128
static struct i915_ggtt *cache_to_ggtt(struct reloc_cache *cache)
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1131
container_of(cache, struct i915_execbuffer, reloc_cache)->i915;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1135
static void reloc_cache_unmap(struct reloc_cache *cache)
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1139
if (!cache->vaddr)
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1142
vaddr = unmask_page(cache->vaddr);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1143
if (cache->vaddr & KMAP)
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1149
static void reloc_cache_remap(struct reloc_cache *cache,
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1154
if (!cache->vaddr)
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1157
if (cache->vaddr & KMAP) {
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1158
struct vm_page *page = i915_gem_object_get_page(obj, cache->page);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1161
cache->vaddr = unmask_flags(cache->vaddr) |
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1164
struct i915_ggtt *ggtt = cache_to_ggtt(cache);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1167
offset = cache->node.start;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1168
if (!drm_mm_node_allocated(&cache->node))
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1169
offset += cache->page << PAGE_SHIFT;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1171
cache->vaddr = (unsigned long)
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1176
static void reloc_cache_reset(struct reloc_cache *cache, struct i915_execbuffer *eb)
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1180
if (!cache->vaddr)
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1183
vaddr = unmask_page(cache->vaddr);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1184
if (cache->vaddr & KMAP) {
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1186
(struct drm_i915_gem_object *)cache->node.mm;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1187
if (cache->vaddr & CLFLUSH_AFTER)
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1193
struct i915_ggtt *ggtt = cache_to_ggtt(cache);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1198
if (drm_mm_node_allocated(&cache->node)) {
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1200
cache->node.start,
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1201
cache->node.size);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1203
drm_mm_remove_node(&cache->node);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1206
i915_vma_unpin((struct i915_vma *)cache->node.mm);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1210
cache->vaddr = 0;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1211
cache->page = -1;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1215
struct reloc_cache *cache,
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1221
if (cache->vaddr) {
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1222
kunmap_local(unmask_page(cache->vaddr));
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1234
cache->vaddr = flushes | KMAP;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1235
cache->node.mm = (void *)obj;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1245
cache->vaddr = unmask_flags(cache->vaddr) | (unsigned long)vaddr;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1246
cache->page = pageno;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1256
struct reloc_cache *cache = &eb->reloc_cache;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1257
struct i915_ggtt *ggtt = cache_to_ggtt(cache);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1261
if (cache->vaddr) {
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1263
io_mapping_unmap_atomic((void __force __iomem *) unmask_page(cache->vaddr));
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1271
if (use_cpu_reloc(cache, obj))
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1297
memset(&cache->node, 0, sizeof(cache->node));
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1300
(&ggtt->vm.mm, &cache->node,
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1308
cache->node.start = i915_ggtt_offset(vma);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1309
cache->node.mm = (void *)vma;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1313
offset = cache->node.start;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1314
if (drm_mm_node_allocated(&cache->node)) {
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1327
cache->page = page;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1328
cache->vaddr = (unsigned long)vaddr;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1337
struct reloc_cache *cache = &eb->reloc_cache;
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1340
if (cache->page == page) {
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1341
vaddr = unmask_page(cache->vaddr);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1344
if ((cache->vaddr & KMAP) == 0)
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
1347
vaddr = reloc_kmap(vma->obj, cache, page);
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
631
static int use_cpu_reloc(const struct reloc_cache *cache,
sys/dev/pci/drm/i915/gem/i915_gem_execbuffer.c
649
return (cache->has_llc ||
sys/dev/pci/drm/i915/gt/uc/intel_guc_capture.c
1610
struct __guc_capture_ads_cache *cache;
sys/dev/pci/drm/i915/gt/uc/intel_guc_capture.c
1615
cache = &gc->ads_cache[i][j][k];
sys/dev/pci/drm/i915/gt/uc/intel_guc_capture.c
1616
if (cache->is_valid)
sys/dev/pci/drm/i915/gt/uc/intel_guc_capture.c
1617
kfree(cache->ptr);
sys/dev/pci/drm/i915/gt/uc/intel_guc_capture.c
477
struct __guc_capture_ads_cache *cache = &gc->ads_cache[owner][type][classid];
sys/dev/pci/drm/i915/gt/uc/intel_guc_capture.c
485
if (cache->is_valid) {
sys/dev/pci/drm/i915/gt/uc/intel_guc_capture.c
486
*size = cache->size;
sys/dev/pci/drm/i915/gt/uc/intel_guc_capture.c
487
return cache->status;
sys/dev/pci/drm/i915/gt/uc/intel_guc_capture.c
527
struct __guc_capture_ads_cache *cache = &gc->ads_cache[owner][type][classid];
sys/dev/pci/drm/i915/gt/uc/intel_guc_capture.c
536
if (cache->is_valid) {
sys/dev/pci/drm/i915/gt/uc/intel_guc_capture.c
537
*outptr = cache->ptr;
sys/dev/pci/drm/i915/gt/uc/intel_guc_capture.c
538
return cache->status;
sys/dev/pci/drm/i915/gt/uc/intel_guc_capture.c
549
cache->is_valid = true;
sys/dev/pci/drm/i915/gt/uc/intel_guc_capture.c
550
cache->ptr = NULL;
sys/dev/pci/drm/i915/gt/uc/intel_guc_capture.c
551
cache->size = 0;
sys/dev/pci/drm/i915/gt/uc/intel_guc_capture.c
552
cache->status = ret;
sys/dev/pci/drm/i915/gt/uc/intel_guc_capture.c
573
cache->is_valid = true;
sys/dev/pci/drm/i915/gt/uc/intel_guc_capture.c
574
cache->ptr = caplist;
sys/dev/pci/drm/i915/gt/uc/intel_guc_capture.c
575
cache->size = size;
sys/dev/pci/drm/i915/gt/uc/intel_guc_capture.c
576
cache->status = 0;
sys/dev/pci/drm/i915/i915_active.c
142
if (!ref->cache)
sys/dev/pci/drm/i915/i915_active.c
143
ref->cache = fetch_node(ref->tree.rb_node);
sys/dev/pci/drm/i915/i915_active.c
146
if (ref->cache) {
sys/dev/pci/drm/i915/i915_active.c
148
rb_erase(&ref->cache->node, &ref->tree);
sys/dev/pci/drm/i915/i915_active.c
152
rb_link_node(&ref->cache->node, NULL, &ref->tree.rb_node);
sys/dev/pci/drm/i915/i915_active.c
153
rb_insert_color(&ref->cache->node, &ref->tree);
sys/dev/pci/drm/i915/i915_active.c
154
GEM_BUG_ON(ref->tree.rb_node != &ref->cache->node);
sys/dev/pci/drm/i915/i915_active.c
157
ref->cache->timeline = 0; /* needs cmpxchg(u64) */
sys/dev/pci/drm/i915/i915_active.c
249
it = READ_ONCE(ref->cache);
sys/dev/pci/drm/i915/i915_active.c
282
WRITE_ONCE(ref->cache, it);
sys/dev/pci/drm/i915/i915_active.c
339
WRITE_ONCE(ref->cache, node);
sys/dev/pci/drm/i915/i915_active.c
360
ref->cache = NULL;
sys/dev/pci/drm/i915/i915_active.c
758
if (ref->cache)
sys/dev/pci/drm/i915/i915_active.c
760
kmem_cache_free(slab_cache, ref->cache);
sys/dev/pci/drm/i915/i915_active.c
762
pool_put(&slab_cache, ref->cache);
sys/dev/pci/drm/i915/i915_active.c
787
if (ref->cache && is_idle_barrier(ref->cache, idx)) {
sys/dev/pci/drm/i915/i915_active.c
788
p = &ref->cache->node;
sys/dev/pci/drm/i915/i915_active.c
847
if (p == &ref->cache->node)
sys/dev/pci/drm/i915/i915_active.c
848
WRITE_ONCE(ref->cache, NULL);
sys/dev/pci/drm/i915/i915_active_types.h
30
struct active_node *cache;
sys/dev/pci/drm/i915/i915_scheduler.c
139
struct sched_cache *cache)
sys/dev/pci/drm/i915/i915_scheduler.c
154
memset(cache, 0, sizeof(*cache));
sys/dev/pci/drm/i915/i915_scheduler.c
170
struct sched_cache cache;
sys/dev/pci/drm/i915/i915_scheduler.c
241
memset(&cache, 0, sizeof(cache));
sys/dev/pci/drm/i915/i915_scheduler.c
246
sched_engine = lock_sched_engine(node, sched_engine, &cache);
sys/dev/pci/drm/i915/i915_scheduler.c
254
sched_engine = lock_sched_engine(node, sched_engine, &cache);
sys/dev/pci/drm/i915/i915_scheduler.c
282
if (!cache.priolist)
sys/dev/pci/drm/i915/i915_scheduler.c
283
cache.priolist =
sys/dev/pci/drm/i915/i915_scheduler.c
286
list_move_tail(&node->link, cache.priolist);
sys/dev/pci/drm/include/drm/ttm/ttm_resource.h
367
} cache;
sys/dev/pci/drm/radeon/ci_dpm.c
543
u32 cache = 0;
sys/dev/pci/drm/radeon/ci_dpm.c
550
cache |= ((config_regs->value << config_regs->shift) & config_regs->mask);
sys/dev/pci/drm/radeon/ci_dpm.c
566
data |= cache;
sys/dev/pci/drm/radeon/ci_dpm.c
579
cache = 0;
sys/dev/pci/drm/radeon/kv_dpm.c
162
u32 cache = 0;
sys/dev/pci/drm/radeon/kv_dpm.c
169
cache |= ((config_regs->value << config_regs->shift) & config_regs->mask);
sys/dev/pci/drm/radeon/kv_dpm.c
185
data |= cache;
sys/dev/pci/drm/radeon/kv_dpm.c
186
cache = 0;
sys/dev/pci/drm/ttm/ttm_resource.c
724
while (i >= iter_io->cache.end) {
sys/dev/pci/drm/ttm/ttm_resource.c
725
iter_io->cache.sg = iter_io->cache.sg ?
sys/dev/pci/drm/ttm/ttm_resource.c
726
sg_next(iter_io->cache.sg) : iter_io->st->sgl;
sys/dev/pci/drm/ttm/ttm_resource.c
727
iter_io->cache.i = iter_io->cache.end;
sys/dev/pci/drm/ttm/ttm_resource.c
728
iter_io->cache.end += sg_dma_len(iter_io->cache.sg) >>
sys/dev/pci/drm/ttm/ttm_resource.c
730
iter_io->cache.offs = sg_dma_address(iter_io->cache.sg) -
sys/dev/pci/drm/ttm/ttm_resource.c
734
if (i < iter_io->cache.i) {
sys/dev/pci/drm/ttm/ttm_resource.c
735
iter_io->cache.end = 0;
sys/dev/pci/drm/ttm/ttm_resource.c
736
iter_io->cache.sg = NULL;
sys/dev/pci/drm/ttm/ttm_resource.c
740
addr = io_mapping_map_local_wc(iter_io->iomap, iter_io->cache.offs +
sys/dev/pci/drm/ttm/ttm_resource.c
741
(((resource_size_t)i - iter_io->cache.i)
sys/dev/pci/drm/ttm/ttm_resource.c
779
memset(&iter_io->cache, 0, sizeof(iter_io->cache));
sys/dev/pci/if_aq_pci.c
3587
uint32_t cache;
sys/dev/pci/if_aq_pci.c
3589
cache = AQ_READ_REG(sc, RX_DMA_DESC_CACHE_INIT_REG);
sys/dev/pci/if_aq_pci.c
3591
(cache & RX_DMA_DESC_CACHE_INIT) ^ RX_DMA_DESC_CACHE_INIT);
sys/kern/vfs_bio.c
1417
bufcache *cache, struct bufqueue *queue, int64_t *queuepages);
sys/kern/vfs_bio.c
1458
struct bufcache *cache = &cleancache[cachenum];
sys/kern/vfs_bio.c
1464
while ((bp = TAILQ_FIRST(&cache->coldqueue)) ||
sys/kern/vfs_bio.c
1465
(bp = TAILQ_FIRST(&cache->warmqueue)) ||
sys/kern/vfs_bio.c
1466
(bp = TAILQ_FIRST(&cache->hotqueue))) {
sys/kern/vfs_bio.c
1474
KASSERT(bp->cache == cachenum);
sys/kern/vfs_bio.c
1491
if (bp->cache == 0 && buf_flip_high(bp) == -1) {
sys/kern/vfs_bio.c
1500
queue = &cache->coldqueue;
sys/kern/vfs_bio.c
1502
queue = &cache->warmqueue;
sys/kern/vfs_bio.c
1503
cache->warmbufpages -= pages;
sys/kern/vfs_bio.c
1505
queue = &cache->hotqueue;
sys/kern/vfs_bio.c
1506
cache->hotbufpages -= pages;
sys/kern/vfs_bio.c
1509
cache->cachepages -= pages;
sys/kern/vfs_bio.c
1512
bp->cache++;
sys/kern/vfs_bio.c
1513
newcache= &cleancache[bp->cache];
sys/kern/vfs_bio.c
1542
struct bufcache *cache = &cleancache[DMA_CACHE];
sys/kern/vfs_bio.c
1549
((bp = TAILQ_FIRST(&cache->coldqueue)) ||
sys/kern/vfs_bio.c
1550
(bp = TAILQ_FIRST(&cache->warmqueue)) ||
sys/kern/vfs_bio.c
1551
(bp = TAILQ_FIRST(&cache->hotqueue)))) {
sys/kern/vfs_bio.c
1559
KASSERT(bp->cache == DMA_CACHE);
sys/kern/vfs_bio.c
1586
if (bp->cache == 0 && buf_flip_high(bp) == -1) {
sys/kern/vfs_bio.c
1597
queue = &cache->coldqueue;
sys/kern/vfs_bio.c
1599
queue = &cache->warmqueue;
sys/kern/vfs_bio.c
1600
cache->warmbufpages -= pages;
sys/kern/vfs_bio.c
1602
queue = &cache->hotqueue;
sys/kern/vfs_bio.c
1603
cache->hotbufpages -= pages;
sys/kern/vfs_bio.c
1606
cache->cachepages -= pages;
sys/kern/vfs_bio.c
1609
bp->cache++;
sys/kern/vfs_bio.c
1610
newcache= &cleancache[bp->cache];
sys/kern/vfs_bio.c
1662
KASSERT(bp->cache >= DMA_CACHE);
sys/kern/vfs_bio.c
1663
KASSERT((bp->cache < NUM_CACHES));
sys/kern/vfs_bio.c
1667
TRACEPOINT(vfs, bufcache_take, bp->b_flags, bp->cache, pages);
sys/kern/vfs_bio.c
1669
struct bufcache *cache = &cleancache[bp->cache];
sys/kern/vfs_bio.c
1672
queue = &cache->coldqueue;
sys/kern/vfs_bio.c
1674
queue = &cache->warmqueue;
sys/kern/vfs_bio.c
1675
cache->warmbufpages -= pages;
sys/kern/vfs_bio.c
1677
queue = &cache->hotqueue;
sys/kern/vfs_bio.c
1678
cache->hotbufpages -= pages;
sys/kern/vfs_bio.c
1681
cache->cachepages -= pages;
sys/kern/vfs_bio.c
1692
chillbufs(struct bufcache *cache, struct bufqueue *queue, int64_t *queuepages)
sys/kern/vfs_bio.c
1703
if (queue == &cache->hotqueue)
sys/kern/vfs_bio.c
1704
limit = min(cache->cachepages / 20, 4096);
sys/kern/vfs_bio.c
1705
else if (queue == &cache->warmqueue)
sys/kern/vfs_bio.c
1706
limit = (cache->cachepages / 2);
sys/kern/vfs_bio.c
1719
TAILQ_INSERT_TAIL(&cache->coldqueue, bp, b_freelist);
sys/kern/vfs_bio.c
1730
struct bufcache *cache = &cleancache[bp->cache];
sys/kern/vfs_bio.c
1735
TRACEPOINT(vfs, bufcache_rel, bp->b_flags, bp->cache, pages);
sys/kern/vfs_bio.c
1738
if (ISSET(bp->b_flags, B_DMA) && bp->cache > 0)
sys/kern/vfs_bio.c
1740
bp->cache);
sys/kern/vfs_bio.c
1741
else if ((!ISSET(bp->b_flags, B_DMA)) && bp->cache == 0)
sys/kern/vfs_bio.c
1743
bp->cache);
sys/kern/vfs_bio.c
1751
queue = &cache->warmqueue;
sys/kern/vfs_bio.c
1752
queuepages = &cache->warmbufpages;
sys/kern/vfs_bio.c
1754
queue = &cache->hotqueue;
sys/kern/vfs_bio.c
1755
queuepages = &cache->hotbufpages;
sys/kern/vfs_bio.c
1759
cache->cachepages += pages;
sys/kern/vfs_bio.c
1760
chillbufs(cache, queue, queuepages);
sys/kern/vfs_bio.c
367
KASSERT(bp->cache == DMA_CACHE);
sys/kern/vfs_bio.c
393
KASSERT(bp->cache < NUM_CACHES);
sys/kern/vfs_bio.c
405
if (bp->cache > DMA_CACHE) {
sys/kern/vfs_bio.c
408
bp->cache = DMA_CACHE;
sys/sys/buf.h
123
int cache; /* which cache are we in */
usr.bin/ssh/sntrup761.c
2083
static void HashConfirm(unsigned char *h, const unsigned char *r, const unsigned char *cache) {
usr.bin/ssh/sntrup761.c
2087
for (i = 0; i < Hash_bytes; ++i) x[Hash_bytes + i] = cache[i];
usr.bin/ssh/sntrup761.c
2109
static void Hide(unsigned char *c, unsigned char *r_enc, const Inputs r, const unsigned char *pk, const unsigned char *cache) {
usr.bin/ssh/sntrup761.c
2112
HashConfirm(c + crypto_kem_sntrup761_CIPHERTEXTBYTES - Confirm_bytes, r_enc, cache);
usr.bin/ssh/sntrup761.c
2117
unsigned char r_enc[Small_bytes], cache[Hash_bytes];
usr.bin/ssh/sntrup761.c
2118
Hash_prefix(cache, 4, pk, crypto_kem_sntrup761_PUBLICKEYBYTES);
usr.bin/ssh/sntrup761.c
2120
Hide(c, r_enc, r, pk, cache);
usr.bin/ssh/sntrup761.c
2135
const unsigned char *cache = rho + Small_bytes;
usr.bin/ssh/sntrup761.c
2140
Hide(cnew, r_enc, r, pk, cache);
usr.bin/ssh/umac.c
212
UINT8 cache[AES_BLOCK_LEN]; /* Previous AES output is saved */
usr.bin/ssh/umac.c
226
aes_encryption(pc->nonce, pc->cache, pc->prf_key);
usr.bin/ssh/umac.c
260
aes_encryption(pc->nonce, pc->cache, pc->prf_key);
usr.bin/ssh/umac.c
264
*((UINT32 *)buf) ^= ((UINT32 *)pc->cache)[ndx];
usr.bin/ssh/umac.c
266
*((UINT64 *)buf) ^= ((UINT64 *)pc->cache)[ndx];
usr.bin/ssh/umac.c
268
((UINT64 *)buf)[0] ^= ((UINT64 *)pc->cache)[0];
usr.bin/ssh/umac.c
269
((UINT32 *)buf)[2] ^= ((UINT32 *)pc->cache)[2];
usr.bin/ssh/umac.c
271
((UINT64 *)buf)[0] ^= ((UINT64 *)pc->cache)[0];
usr.bin/ssh/umac.c
272
((UINT64 *)buf)[1] ^= ((UINT64 *)pc->cache)[1];
usr.bin/systat/pool.c
488
if (pool_get_cache(np, &pc->cache) < 0 ||
usr.bin/systat/pool.c
562
kpc = &pc->cache;
usr.bin/systat/pool.c
599
if (pc1->cache.pr_len < pc2->cache.pr_len)
usr.bin/systat/pool.c
601
if (pc1->cache.pr_len > pc2->cache.pr_len)
usr.bin/systat/pool.c
614
if (pc1->cache.pr_nitems < pc2->cache.pr_nitems)
usr.bin/systat/pool.c
616
if (pc1->cache.pr_nitems > pc2->cache.pr_nitems)
usr.bin/systat/pool.c
629
if (pc1->cache.pr_ngc < pc2->cache.pr_ngc)
usr.bin/systat/pool.c
631
if (pc1->cache.pr_ngc > pc2->cache.pr_ngc)
usr.bin/systat/pool.c
68
struct kinfo_pool_cache cache;
usr.bin/usbhidctl/usbhid.c
153
hidtestrule(struct Susbvar *var, struct usagedata *cache)
usr.bin/usbhidctl/usbhid.c
165
usage_id = cache->usage_id;
usr.bin/usbhidctl/usbhid.c
180
if (cache->isfinal && strind != varlen)
usr.bin/usbhidctl/usbhid.c
213
if (cache->page_name == NULL) {
usr.bin/usbhidctl/usbhid.c
214
cache->page_name = hid_usage_page(HID_PAGE(usage_id));
usr.bin/usbhidctl/usbhid.c
215
cache->page_len = strlen(cache->page_name);
usr.bin/usbhidctl/usbhid.c
222
if (cache->page_len !=
usr.bin/usbhidctl/usbhid.c
224
memcmp(cache->page_name,
usr.bin/usbhidctl/usbhid.c
226
cache->page_len) != 0)
usr.bin/usbhidctl/usbhid.c
247
if (cache->usage_name == NULL) {
usr.bin/usbhidctl/usbhid.c
248
cache->usage_name = hid_usage_in_page(usage_id);
usr.bin/usbhidctl/usbhid.c
249
cache->usage_len = strlen(cache->usage_name);
usr.bin/usbhidctl/usbhid.c
255
if (cache->usage_len != (size_t)(strind - matchindex) ||
usr.bin/usbhidctl/usbhid.c
256
memcmp(cache->usage_name, &varname[matchindex],
usr.bin/usbhidctl/usbhid.c
257
cache->usage_len) != 0)
usr.bin/usbhidctl/usbhid.c
262
if (cache->isfinal)
usr.bin/usbhidctl/usbhid.c
341
struct usagedata cache;
usr.bin/usbhidctl/usbhid.c
343
cache.isfinal = (colind == collen);
usr.bin/usbhidctl/usbhid.c
344
if (cache.isfinal)
usr.bin/usbhidctl/usbhid.c
345
cache.usage_id = item->usage;
usr.bin/usbhidctl/usbhid.c
347
cache.usage_id = collist[colind];
usr.bin/usbhidctl/usbhid.c
349
cache.usage_name = NULL;
usr.bin/usbhidctl/usbhid.c
350
cache.page_name = NULL;
usr.bin/usbhidctl/usbhid.c
369
matchres = hidtestrule(var, &cache);
usr.bin/yacc/reader.c
1039
cache = malloc(cache_size);
usr.bin/yacc/reader.c
1040
if (cache == NULL)
usr.bin/yacc/reader.c
114
cache = realloc(cache, cache_size);
usr.bin/yacc/reader.c
115
if (cache == NULL)
usr.bin/yacc/reader.c
118
cache[cinc] = c;
usr.bin/yacc/reader.c
1239
assert(cache);
usr.bin/yacc/reader.c
1240
snprintf(cache, cache_size, "$$%d", ++gensym);
usr.bin/yacc/reader.c
1241
bp = make_bucket(cache);
usr.bin/yacc/reader.c
297
if (strcmp(cache, "token") == 0 || strcmp(cache, "term") == 0)
usr.bin/yacc/reader.c
299
if (strcmp(cache, "type") == 0)
usr.bin/yacc/reader.c
301
if (strcmp(cache, "left") == 0)
usr.bin/yacc/reader.c
303
if (strcmp(cache, "right") == 0)
usr.bin/yacc/reader.c
305
if (strcmp(cache, "nonassoc") == 0 || strcmp(cache, "binary") == 0)
usr.bin/yacc/reader.c
307
if (strcmp(cache, "start") == 0)
usr.bin/yacc/reader.c
309
if (strcmp(cache, "union") == 0)
usr.bin/yacc/reader.c
311
if (strcmp(cache, "ident") == 0)
usr.bin/yacc/reader.c
313
if (strcmp(cache, "expect") == 0)
usr.bin/yacc/reader.c
46
char *cache;
usr.bin/yacc/reader.c
716
memcpy(s, cache, n);
usr.bin/yacc/reader.c
726
if (c == '\\' || c == cache[0]) {
usr.bin/yacc/reader.c
770
bp = lookup(cache);
usr.bin/yacc/reader.c
811
if (is_reserved(cache))
usr.bin/yacc/reader.c
812
used_reserved(cache);
usr.bin/yacc/reader.c
814
return (lookup(cache));
usr.bin/yacc/reader.c
864
if (strcmp(cache, tag_table[i]) == 0)
usr.bin/yacc/reader.c
877
strlcpy(s, cache, cinc);
usr.sbin/amd/amd/afs_ops.c
225
char *cache;
usr.sbin/amd/amd/afs_ops.c
227
cache = mf->mf_fo->opt_cache;
usr.sbin/amd/amd/afs_ops.c
229
cache = "none";
usr.sbin/amd/amd/afs_ops.c
230
mf->mf_private = mapc_find(mf->mf_info, cache);
usr.sbin/dhcpd/confpars.c
200
struct tree_cache *cache;
usr.sbin/dhcpd/confpars.c
368
cache = tree_cache(tree);
usr.sbin/dhcpd/confpars.c
369
if (!tree_evaluate (cache))
usr.sbin/dhcpd/confpars.c
373
cache->value, group->next_server.len);
usr.sbin/dhcpd/confpars.c
407
cache = parse_fixed_addr_param(cfile);
usr.sbin/dhcpd/confpars.c
409
host_decl->fixed_addr = cache;
usr.sbin/smtpd/mta_session.c
1463
int cache)
usr.sbin/smtpd/mta_session.c
1507
if (cache)
usr.sbin/unbound/dnscrypt/dnscrypt.c
109
dnsc_shared_secret_cache_insert(struct slabhash *cache,
usr.sbin/unbound/dnscrypt/dnscrypt.c
128
slabhash_insert(cache,
usr.sbin/unbound/dnscrypt/dnscrypt.c
143
dnsc_shared_secrets_lookup(struct slabhash* cache,
usr.sbin/unbound/dnscrypt/dnscrypt.c
147
return slabhash_lookup(cache, hash, key, 0);
usr.sbin/unbound/dnscrypt/dnscrypt.c
179
dnsc_nonce_cache_insert(struct slabhash *cache,
usr.sbin/unbound/dnscrypt/dnscrypt.c
198
slabhash_insert(cache,
usr.sbin/unbound/dnscrypt/dnscrypt.c
215
dnsc_nonces_lookup(struct slabhash* cache,
usr.sbin/unbound/dnscrypt/dnscrypt.c
228
return slabhash_lookup(cache, hash, &k, 0);