Symbol: shadow
arch/alpha/include/asm/mce.h
26
unsigned long shadow[8]; /* Shadow reg. 8-14, 25 */
arch/alpha/kernel/core_mcpcia.c
459
i, i+1, frame->shadow[i],
arch/alpha/kernel/core_mcpcia.c
460
frame->shadow[i+1]);
arch/m68k/fpsp040/fpsp.h
137
.set FPSR_SHADOW,LV-64 | fpsr shadow reg
arch/mips/include/asm/sn/ioc3.h
19
u32 shadow;
arch/powerpc/platforms/ps3/spu.c
112
struct spe_shadow __iomem *shadow;
arch/powerpc/platforms/ps3/spu.c
124
unsigned long problem, unsigned long ls, unsigned long shadow,
arch/powerpc/platforms/ps3/spu.c
131
pr_debug("%s:%d: shadow: %lxh\n", func, line, shadow);
arch/powerpc/platforms/ps3/spu.c
180
iounmap(spu_pdata(spu)->shadow);
arch/powerpc/platforms/ps3/spu.c
194
spu_pdata(spu)->shadow = ioremap_prot(spu_pdata(spu)->shadow_addr,
arch/powerpc/platforms/ps3/spu.c
197
if (!spu_pdata(spu)->shadow) {
arch/powerpc/platforms/ps3/spu.c
231
(unsigned long)spu_pdata(spu)->shadow);
arch/powerpc/platforms/ps3/spu.c
367
while (in_be64(&spu_pdata(spu)->shadow->spe_execution_status)
arch/powerpc/platforms/ps3/spu.c
513
return in_be64(&spu_pdata(spu)->shadow->mfc_dar_RW);
arch/powerpc/platforms/ps3/spu.c
523
return in_be64(&spu_pdata(spu)->shadow->mfc_dsisr_RW);
arch/s390/kvm/trace-s390.h
340
TP_PROTO(unsigned long start, unsigned long end, unsigned int shadow),
arch/s390/kvm/trace-s390.h
341
TP_ARGS(start, end, shadow),
arch/s390/kvm/trace-s390.h
346
__field(unsigned int, shadow)
arch/s390/kvm/trace-s390.h
352
__entry->shadow = shadow;
arch/s390/kvm/trace-s390.h
356
__entry->start, __entry->end, __entry->shadow)
arch/x86/include/uapi/asm/kvm.h
352
__u8 shadow;
arch/x86/kvm/vmx/vmx.c
3056
struct vmcs *alloc_vmcs_cpu(bool shadow, int cpu, gfp_t flags)
arch/x86/kvm/vmx/vmx.c
3074
if (shadow)
arch/x86/kvm/vmx/vmx.h
669
struct vmcs *alloc_vmcs_cpu(bool shadow, int cpu, gfp_t flags);
arch/x86/kvm/vmx/vmx.h
674
static inline struct vmcs *alloc_vmcs(bool shadow)
arch/x86/kvm/vmx/vmx.h
676
return alloc_vmcs_cpu(shadow, raw_smp_processor_id(),
arch/x86/kvm/x86.c
5600
events->interrupt.shadow = kvm_x86_call(get_interrupt_shadow)(vcpu);
arch/x86/kvm/x86.c
5681
events->interrupt.shadow);
arch/x86/mm/kasan_init_64.c
320
unsigned long shadow = (unsigned long)kasan_mem_to_shadow((void *)va);
arch/x86/mm/kasan_init_64.c
322
return round_down(shadow, PAGE_SIZE);
arch/x86/mm/kasan_init_64.c
327
unsigned long shadow = (unsigned long)kasan_mem_to_shadow((void *)va);
arch/x86/mm/kasan_init_64.c
329
return round_up(shadow, PAGE_SIZE);
arch/x86/xen/enlighten_pv.c
619
struct desc_struct *shadow = &per_cpu(shadow_tls_desc, cpu).desc[i];
arch/x86/xen/enlighten_pv.c
624
if (desc_equal(shadow, &t->tls_array[i]))
arch/x86/xen/enlighten_pv.c
627
*shadow = t->tls_array[i];
block/sed-opal.c
1970
struct opal_shadow_mbr *shadow = data;
block/sed-opal.c
1972
return generic_table_write_data(dev, shadow->data, shadow->offset,
block/sed-opal.c
1973
shadow->size, opaluid[OPAL_MBR]);
drivers/block/xen-blkfront.c
1249
if (!rinfo->shadow[i].request)
drivers/block/xen-blkfront.c
1252
segs = rinfo->shadow[i].req.operation == BLKIF_OP_INDIRECT ?
drivers/block/xen-blkfront.c
1253
rinfo->shadow[i].req.u.indirect.nr_segments :
drivers/block/xen-blkfront.c
1254
rinfo->shadow[i].req.u.rw.nr_segments;
drivers/block/xen-blkfront.c
1256
persistent_gnt = rinfo->shadow[i].grants_used[j];
drivers/block/xen-blkfront.c
1263
if (rinfo->shadow[i].req.operation != BLKIF_OP_INDIRECT)
drivers/block/xen-blkfront.c
1271
persistent_gnt = rinfo->shadow[i].indirect_grants[j];
drivers/block/xen-blkfront.c
1278
kvfree(rinfo->shadow[i].grants_used);
drivers/block/xen-blkfront.c
1279
rinfo->shadow[i].grants_used = NULL;
drivers/block/xen-blkfront.c
1280
kvfree(rinfo->shadow[i].indirect_grants);
drivers/block/xen-blkfront.c
1281
rinfo->shadow[i].indirect_grants = NULL;
drivers/block/xen-blkfront.c
1282
kvfree(rinfo->shadow[i].sg);
drivers/block/xen-blkfront.c
1283
rinfo->shadow[i].sg = NULL;
drivers/block/xen-blkfront.c
1391
struct blk_shadow *s = &rinfo->shadow[*id];
drivers/block/xen-blkfront.c
1401
struct blk_shadow *s2 = &rinfo->shadow[s->associated_id];
drivers/block/xen-blkfront.c
1557
if (rinfo->shadow[id].status != REQ_WAITING) {
drivers/block/xen-blkfront.c
1563
rinfo->shadow[id].status = REQ_PROCESSING;
drivers/block/xen-blkfront.c
1564
req = rinfo->shadow[id].request;
drivers/block/xen-blkfront.c
1566
op = rinfo->shadow[id].req.operation;
drivers/block/xen-blkfront.c
1568
op = rinfo->shadow[id].req.u.indirect.indirect_op;
drivers/block/xen-blkfront.c
1622
rinfo->shadow[id].req.u.rw.nr_segments == 0)) {
drivers/block/xen-blkfront.c
1878
rinfo->shadow[j].req.u.rw.id = j + 1;
drivers/block/xen-blkfront.c
1879
rinfo->shadow[BLK_RING_SIZE(info)-1].req.u.rw.id = 0x0fffffff;
drivers/block/xen-blkfront.c
190
struct blk_shadow shadow[];
drivers/block/xen-blkfront.c
1910
info->rinfo_size = struct_size(info->rinfo, shadow,
drivers/block/xen-blkfront.c
2075
struct blk_shadow *shadow = rinfo->shadow;
drivers/block/xen-blkfront.c
2079
if (!shadow[j].request)
drivers/block/xen-blkfront.c
2085
if (req_op(shadow[j].request) == REQ_OP_FLUSH ||
drivers/block/xen-blkfront.c
2086
req_op(shadow[j].request) == REQ_OP_DISCARD ||
drivers/block/xen-blkfront.c
2087
req_op(shadow[j].request) == REQ_OP_SECURE_ERASE ||
drivers/block/xen-blkfront.c
2088
shadow[j].request->cmd_flags & REQ_FUA) {
drivers/block/xen-blkfront.c
2096
list_add(&shadow[j].request->queuelist, &info->requests);
drivers/block/xen-blkfront.c
2099
merge_bio.head = shadow[j].request->bio;
drivers/block/xen-blkfront.c
2100
merge_bio.tail = shadow[j].request->biotail;
drivers/block/xen-blkfront.c
2102
shadow[j].request->bio = NULL;
drivers/block/xen-blkfront.c
2103
blk_mq_end_request(shadow[j].request, BLK_STS_OK);
drivers/block/xen-blkfront.c
2209
rinfo->shadow[i].grants_used =
drivers/block/xen-blkfront.c
2210
kvzalloc_objs(rinfo->shadow[i].grants_used[0], grants);
drivers/block/xen-blkfront.c
2211
rinfo->shadow[i].sg = kvzalloc_objs(rinfo->shadow[i].sg[0],
drivers/block/xen-blkfront.c
2214
rinfo->shadow[i].indirect_grants =
drivers/block/xen-blkfront.c
2215
kvzalloc_objs(rinfo->shadow[i].indirect_grants[0],
drivers/block/xen-blkfront.c
2217
if ((rinfo->shadow[i].grants_used == NULL) ||
drivers/block/xen-blkfront.c
2218
(rinfo->shadow[i].sg == NULL) ||
drivers/block/xen-blkfront.c
2220
(rinfo->shadow[i].indirect_grants == NULL)))
drivers/block/xen-blkfront.c
2222
sg_init_table(rinfo->shadow[i].sg, psegs);
drivers/block/xen-blkfront.c
2231
kvfree(rinfo->shadow[i].grants_used);
drivers/block/xen-blkfront.c
2232
rinfo->shadow[i].grants_used = NULL;
drivers/block/xen-blkfront.c
2233
kvfree(rinfo->shadow[i].sg);
drivers/block/xen-blkfront.c
2234
rinfo->shadow[i].sg = NULL;
drivers/block/xen-blkfront.c
2235
kvfree(rinfo->shadow[i].indirect_grants);
drivers/block/xen-blkfront.c
2236
rinfo->shadow[i].indirect_grants = NULL;
drivers/block/xen-blkfront.c
291
rinfo->shadow_free = rinfo->shadow[free].req.u.rw.id;
drivers/block/xen-blkfront.c
292
rinfo->shadow[free].req.u.rw.id = 0x0fffffee; /* debug */
drivers/block/xen-blkfront.c
299
if (rinfo->shadow[id].req.u.rw.id != id)
drivers/block/xen-blkfront.c
301
if (rinfo->shadow[id].request == NULL)
drivers/block/xen-blkfront.c
303
rinfo->shadow[id].req.u.rw.id = rinfo->shadow_free;
drivers/block/xen-blkfront.c
304
rinfo->shadow[id].request = NULL;
drivers/block/xen-blkfront.c
543
rinfo->shadow[id].request = req;
drivers/block/xen-blkfront.c
544
rinfo->shadow[id].status = REQ_PROCESSING;
drivers/block/xen-blkfront.c
545
rinfo->shadow[id].associated_id = NO_ASSOCIATED_ID;
drivers/block/xen-blkfront.c
547
rinfo->shadow[id].req.u.rw.id = id;
drivers/block/xen-blkfront.c
560
ring_req = &rinfo->shadow[id].req;
drivers/block/xen-blkfront.c
573
rinfo->shadow[id].status = REQ_WAITING;
drivers/block/xen-blkfront.c
611
struct blk_shadow *shadow = &rinfo->shadow[setup->id];
drivers/block/xen-blkfront.c
630
shadow->indirect_grants[n] = gnt_list_entry;
drivers/block/xen-blkfront.c
641
shadow->grants_used[setup->grant_idx] = gnt_list_entry;
drivers/block/xen-blkfront.c
752
ring_req = &rinfo->shadow[id].req;
drivers/block/xen-blkfront.c
754
num_sg = blk_rq_map_sg(req, rinfo->shadow[id].sg);
drivers/block/xen-blkfront.c
757
for_each_sg(rinfo->shadow[id].sg, sg, num_sg, i)
drivers/block/xen-blkfront.c
764
rinfo->shadow[id].num_sg = num_sg;
drivers/block/xen-blkfront.c
808
extra_ring_req = &rinfo->shadow[extra_id].req;
drivers/block/xen-blkfront.c
814
rinfo->shadow[extra_id].num_sg = 0;
drivers/block/xen-blkfront.c
819
rinfo->shadow[extra_id].associated_id = id;
drivers/block/xen-blkfront.c
820
rinfo->shadow[id].associated_id = extra_id;
drivers/block/xen-blkfront.c
831
for_each_sg(rinfo->shadow[id].sg, sg, num_sg, i) {
drivers/block/xen-blkfront.c
853
rinfo->shadow[id].status = REQ_WAITING;
drivers/block/xen-blkfront.c
856
rinfo->shadow[extra_id].status = REQ_WAITING;
drivers/crypto/hisilicon/sec/sec_algs.c
541
void sec_alg_callback(struct sec_bd_info *resp, void *shadow)
drivers/crypto/hisilicon/sec/sec_algs.c
543
struct sec_request *sec_req = shadow;
drivers/crypto/hisilicon/sec/sec_drv.c
708
queue->shadow[queue->expected]);
drivers/crypto/hisilicon/sec/sec_drv.c
709
queue->shadow[queue->expected] = NULL;
drivers/crypto/hisilicon/sec/sec_drv.c
872
queue->shadow[write] = ctx;
drivers/crypto/hisilicon/sec/sec_drv.h
351
void *shadow[SEC_QUEUE_LEN];
drivers/gpio/gpio-janz-ttl.c
100
*shadow |= BIT(offset);
drivers/gpio/gpio-janz-ttl.c
102
*shadow &= ~BIT(offset);
drivers/gpio/gpio-janz-ttl.c
104
iowrite16be(*shadow, port);
drivers/gpio/gpio-janz-ttl.c
60
u8 *shadow;
drivers/gpio/gpio-janz-ttl.c
64
shadow = &mod->porta_shadow;
drivers/gpio/gpio-janz-ttl.c
66
shadow = &mod->portb_shadow;
drivers/gpio/gpio-janz-ttl.c
69
shadow = &mod->portc_shadow;
drivers/gpio/gpio-janz-ttl.c
74
ret = *shadow & BIT(offset);
drivers/gpio/gpio-janz-ttl.c
83
u8 *shadow;
drivers/gpio/gpio-janz-ttl.c
87
shadow = &mod->porta_shadow;
drivers/gpio/gpio-janz-ttl.c
90
shadow = &mod->portb_shadow;
drivers/gpio/gpio-janz-ttl.c
94
shadow = &mod->portc_shadow;
drivers/gpio/gpio-latch.c
160
priv->shadow = devm_bitmap_zalloc(dev, n_latches * priv->n_latched_gpios,
drivers/gpio/gpio-latch.c
162
if (!priv->shadow)
drivers/gpio/gpio-latch.c
58
unsigned long *shadow;
drivers/gpio/gpio-latch.c
80
assign_bit(offset, priv->shadow, val);
drivers/gpio/gpio-latch.c
85
priv->shadow));
drivers/gpio/gpio-mm-lantiq.c
104
u32 shadow;
drivers/gpio/gpio-mm-lantiq.c
129
if (!of_property_read_u32(pdev->dev.of_node, "lantiq,shadow", &shadow))
drivers/gpio/gpio-mm-lantiq.c
130
chip->shadow = shadow;
drivers/gpio/gpio-mm-lantiq.c
31
u16 shadow; /* shadow the latches state */
drivers/gpio/gpio-mm-lantiq.c
47
__raw_writew(chip->shadow, chip->regs);
drivers/gpio/gpio-mm-lantiq.c
65
chip->shadow |= (1 << offset);
drivers/gpio/gpio-mm-lantiq.c
67
chip->shadow &= ~(1 << offset);
drivers/gpio/gpio-stp-xway.c
121
chip->shadow |= BIT(gpio);
drivers/gpio/gpio-stp-xway.c
123
chip->shadow &= ~BIT(gpio);
drivers/gpio/gpio-stp-xway.c
124
xway_stp_w32(chip->virt, chip->shadow, XWAY_STP_CPU0);
drivers/gpio/gpio-stp-xway.c
235
u32 shadow, groups, dsl, phy;
drivers/gpio/gpio-stp-xway.c
258
if (!of_property_read_u32(pdev->dev.of_node, "lantiq,shadow", &shadow))
drivers/gpio/gpio-stp-xway.c
259
chip->shadow = shadow;
drivers/gpio/gpio-stp-xway.c
84
u32 shadow; /* shadow the shift registers state */
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
590
struct drm_amdgpu_cs_chunk_cp_gfx_shadow *shadow = chunk->kdata;
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
593
if (shadow->flags & ~AMDGPU_CS_CHUNK_CP_GFX_SHADOW_FLAGS_INIT_SHADOW)
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
597
p->jobs[i]->shadow_va = shadow->shadow_va;
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
598
p->jobs[i]->csa_va = shadow->csa_va;
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
599
p->jobs[i]->gds_va = shadow->gds_va;
drivers/gpu/drm/amd/amdgpu/amdgpu_cs.c
601
shadow->flags & AMDGPU_CS_CHUNK_CP_GFX_SHADOW_FLAGS_INIT_SHADOW;
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
384
struct amdgpu_gfx_shadow_info shadow = {};
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
386
adev->gfx.funcs->get_gfx_shadow_info(adev, &shadow, true);
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
387
meta->shadow_size = shadow.shadow_size;
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
388
meta->shadow_alignment = shadow.shadow_alignment;
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
389
meta->csa_size = shadow.csa_size;
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
390
meta->csa_alignment = shadow.csa_alignment;
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
404
struct amdgpu_gfx_shadow_info shadow = {};
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
406
adev->gfx.funcs->get_gfx_shadow_info(adev, &shadow, true);
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
407
meta->eop_size = shadow.eop_size;
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
408
meta->eop_alignment = shadow.eop_alignment;
drivers/gpu/drm/drm_fbdev_dma.c
81
void *shadow = info->screen_buffer;
drivers/gpu/drm/drm_fbdev_dma.c
89
vfree(shadow);
drivers/gpu/drm/drm_fbdev_ttm.c
45
void *shadow = info->screen_buffer;
drivers/gpu/drm/drm_fbdev_ttm.c
52
vfree(shadow);
drivers/gpu/drm/exynos/exynos_mixer.c
356
u32 base, shadow;
drivers/gpu/drm/exynos/exynos_mixer.c
368
shadow = mixer_reg_read(ctx, MXR_CFG_S);
drivers/gpu/drm/exynos/exynos_mixer.c
369
if (base != shadow)
drivers/gpu/drm/exynos/exynos_mixer.c
373
shadow = mixer_reg_read(ctx, MXR_GRAPHIC_BASE_S(0));
drivers/gpu/drm/exynos/exynos_mixer.c
374
if (base != shadow)
drivers/gpu/drm/exynos/exynos_mixer.c
378
shadow = mixer_reg_read(ctx, MXR_GRAPHIC_BASE_S(1));
drivers/gpu/drm/exynos/exynos_mixer.c
379
if (base != shadow)
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2269
struct i915_vma *shadow, *trampoline, *batch;
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2313
shadow = shadow_batch_pin(eb, pool->obj, eb->context->vm, PIN_USER);
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2314
if (IS_ERR(shadow))
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2315
return PTR_ERR(shadow);
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2318
i915_gem_object_set_readonly(shadow->obj);
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2319
shadow->private = pool;
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2323
trampoline = shadow;
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2325
shadow = shadow_batch_pin(eb, pool->obj,
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2328
if (IS_ERR(shadow))
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2329
return PTR_ERR(shadow);
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2331
shadow->private = pool;
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2336
batch = eb_dispatch_secure(eb, shadow);
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2340
err = dma_resv_reserve_fences(shadow->obj->base.resv, 1);
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2348
shadow, trampoline);
drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c
2353
eb->batches[0]->vma = i915_vma_get(shadow);
drivers/gpu/drm/i915/gvt/cmd_parser.c
3172
struct intel_context *ce = vgpu->submission.shadow[ring_id];
drivers/gpu/drm/i915/gvt/gvt.h
153
struct intel_context *shadow[I915_NUM_ENGINES];
drivers/gpu/drm/i915/gvt/mmio_context.c
531
!is_inhibit_context(s->shadow[engine->id]))
drivers/gpu/drm/i915/gvt/scheduler.c
1335
i915_context_ppgtt_root_restore(s, i915_vm_to_ppgtt(s->shadow[0]->vm));
drivers/gpu/drm/i915/gvt/scheduler.c
1337
intel_context_put(s->shadow[id]);
drivers/gpu/drm/i915/gvt/scheduler.c
1410
s->shadow[i] = ERR_PTR(-EINVAL);
drivers/gpu/drm/i915/gvt/scheduler.c
1426
s->shadow[i] = ce;
drivers/gpu/drm/i915/gvt/scheduler.c
1454
if (IS_ERR(s->shadow[i]))
drivers/gpu/drm/i915/gvt/scheduler.c
1457
intel_context_put(s->shadow[i]);
drivers/gpu/drm/i915/gvt/scheduler.c
1530
intel_context_unpin(s->shadow[workload->engine->id]);
drivers/gpu/drm/i915/gvt/scheduler.c
1776
ret = intel_context_pin(s->shadow[engine->id]);
drivers/gpu/drm/i915/gvt/scheduler.c
476
rq = i915_request_create(s->shadow[workload->engine->id]);
drivers/gpu/drm/i915/gvt/scheduler.c
502
if (workload->shadow)
drivers/gpu/drm/i915/gvt/scheduler.c
506
shadow_context_descriptor_update(s->shadow[workload->engine->id],
drivers/gpu/drm/i915/gvt/scheduler.c
520
workload->shadow = true;
drivers/gpu/drm/i915/gvt/scheduler.c
760
set_context_ppgtt_from_shadow(workload, s->shadow[workload->engine->id]);
drivers/gpu/drm/i915/gvt/scheduler.h
91
bool shadow; /* if workload has done shadow of guest request */
drivers/gpu/drm/i915/i915_cmd_parser.c
1444
struct i915_vma *shadow,
drivers/gpu/drm/i915/i915_cmd_parser.c
1461
cmd = copy_batch(shadow->obj, batch->obj,
drivers/gpu/drm/i915/i915_cmd_parser.c
1474
shadow_addr = gen8_canonical_addr(i915_vma_offset(shadow));
drivers/gpu/drm/i915/i915_cmd_parser.c
1551
cmd = page_mask_bits(shadow->obj->mm.mapping);
drivers/gpu/drm/i915/i915_cmd_parser.c
1572
i915_gem_object_flush_map(shadow->obj);
drivers/gpu/drm/i915/i915_cmd_parser.c
1576
i915_gem_object_unpin_map(shadow->obj);
drivers/gpu/drm/i915/i915_cmd_parser.h
22
struct i915_vma *shadow,
drivers/gpu/drm/msm/adreno/a5xx_gpu.c
1433
a5xx_gpu->shadow[i] = 0;
drivers/gpu/drm/msm/adreno/a5xx_gpu.c
1688
return a5xx_gpu->shadow[ring->id];
drivers/gpu/drm/msm/adreno/a5xx_gpu.c
622
a5xx_gpu->shadow = msm_gem_kernel_new(gpu->dev,
drivers/gpu/drm/msm/adreno/a5xx_gpu.c
628
if (IS_ERR(a5xx_gpu->shadow))
drivers/gpu/drm/msm/adreno/a5xx_gpu.c
629
return PTR_ERR(a5xx_gpu->shadow);
drivers/gpu/drm/msm/adreno/a5xx_gpu.h
45
uint32_t *shadow;
drivers/gpu/drm/msm/adreno/a6xx_gpu.c
1172
a6xx_gpu->shadow = msm_gem_kernel_new(gpu->dev,
drivers/gpu/drm/msm/adreno/a6xx_gpu.c
1178
if (IS_ERR(a6xx_gpu->shadow))
drivers/gpu/drm/msm/adreno/a6xx_gpu.c
1179
return PTR_ERR(a6xx_gpu->shadow);
drivers/gpu/drm/msm/adreno/a6xx_gpu.c
1529
a6xx_gpu->shadow[i] = 0;
drivers/gpu/drm/msm/adreno/a6xx_gpu.c
2371
a6xx_gpu->shadow[i] = 0;
drivers/gpu/drm/msm/adreno/a6xx_gpu.c
2410
a6xx_gpu->shadow[i] = 0;
drivers/gpu/drm/msm/adreno/a6xx_gpu.c
2535
return a6xx_gpu->shadow[ring->id];
drivers/gpu/drm/msm/adreno/a6xx_gpu.h
96
uint32_t *shadow;
drivers/gpu/drm/msm/adreno/a8xx_gpu.c
652
a6xx_gpu->shadow[i] = 0;
drivers/gpu/drm/nouveau/include/nvkm/subdev/pci.h
39
void nvkm_pci_rom_shadow(struct nvkm_pci *, bool shadow);
drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadow.c
110
shadow_method(struct nvkm_bios *bios, struct shadow *mthd, const char *name)
drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadow.c
177
struct shadow mthds[] = {
drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadow.c
39
shadow_fetch(struct nvkm_bios *bios, struct shadow *mthd, u32 upto)
drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadow.c
52
shadow_image(struct nvkm_bios *bios, int idx, u32 offset, struct shadow *mthd)
drivers/gpu/drm/nouveau/nvkm/subdev/mxm/base.c
215
struct mxm_shadow_h *shadow = _mxm_shadow;
drivers/gpu/drm/nouveau/nvkm/subdev/mxm/base.c
217
nvkm_debug(&mxm->subdev, "checking %s\n", shadow->name);
drivers/gpu/drm/nouveau/nvkm/subdev/mxm/base.c
218
if (shadow->exec(mxm, version)) {
drivers/gpu/drm/nouveau/nvkm/subdev/mxm/base.c
224
} while ((++shadow)->name);
drivers/gpu/drm/nouveau/nvkm/subdev/pci/base.c
66
nvkm_pci_rom_shadow(struct nvkm_pci *pci, bool shadow)
drivers/gpu/drm/nouveau/nvkm/subdev/pci/base.c
69
if (shadow)
drivers/gpu/drm/qxl/qxl_display.c
445
is_primary = qobj->shadow ? qobj->shadow->is_primary : qobj->is_primary;
drivers/gpu/drm/qxl/qxl_display.c
680
primary = bo->shadow ? bo->shadow : bo;
drivers/gpu/drm/qxl/qxl_display.c
707
if (bo->shadow)
drivers/gpu/drm/qxl/qxl_display.c
708
bo = bo->shadow;
drivers/gpu/drm/qxl/qxl_display.c
842
if (user_bo->shadow != qdev->dumb_shadow_bo) {
drivers/gpu/drm/qxl/qxl_display.c
843
if (user_bo->shadow) {
drivers/gpu/drm/qxl/qxl_display.c
844
qxl_bo_unpin(user_bo->shadow);
drivers/gpu/drm/qxl/qxl_display.c
846
(&user_bo->shadow->tbo.base);
drivers/gpu/drm/qxl/qxl_display.c
847
user_bo->shadow = NULL;
drivers/gpu/drm/qxl/qxl_display.c
850
user_bo->shadow = qdev->dumb_shadow_bo;
drivers/gpu/drm/qxl/qxl_display.c
851
qxl_bo_pin(user_bo->shadow);
drivers/gpu/drm/qxl/qxl_display.c
910
if (old_state->fb != plane->state->fb && user_bo->shadow) {
drivers/gpu/drm/qxl/qxl_display.c
911
qxl_bo_unpin(user_bo->shadow);
drivers/gpu/drm/qxl/qxl_display.c
912
drm_gem_object_put(&user_bo->shadow->tbo.base);
drivers/gpu/drm/qxl/qxl_display.c
913
user_bo->shadow = NULL;
drivers/gpu/drm/qxl/qxl_drv.h
88
struct qxl_bo *shadow;
drivers/gpu/drm/xe/xe_sa.c
100
return ERR_CAST(shadow);
drivers/gpu/drm/xe/xe_sa.c
102
sa_manager->shadow = shadow;
drivers/gpu/drm/xe/xe_sa.c
126
xe_assert(xe, sa_manager->shadow);
drivers/gpu/drm/xe/xe_sa.c
129
swap(sa_manager->bo, sa_manager->shadow);
drivers/gpu/drm/xe/xe_sa.c
147
xe_assert(xe, sa_manager->shadow);
drivers/gpu/drm/xe/xe_sa.c
150
xe_map_memcpy_to(xe, &sa_manager->shadow->vmap,
drivers/gpu/drm/xe/xe_sa.c
32
sa_manager->shadow = NULL;
drivers/gpu/drm/xe/xe_sa.c
86
struct xe_bo *shadow;
drivers/gpu/drm/xe/xe_sa.c
92
shadow = xe_managed_bo_create_pin_map(xe, tile, size,
drivers/gpu/drm/xe/xe_sa.c
97
if (IS_ERR(shadow)) {
drivers/gpu/drm/xe/xe_sa.c
99
size / SZ_1K, shadow);
drivers/gpu/drm/xe/xe_sa_types.h
15
struct xe_bo *shadow;
drivers/gpu/drm/xe/xe_sriov_vf_ccs.c
166
xe_map_memset(xe, &sa_manager->shadow->vmap, offset, MI_NOOP,
drivers/gpu/drm/xe/xe_sriov_vf_ccs.c
171
xe_map_wr(xe, &sa_manager->shadow->vmap, offset, u32, MI_BATCH_BUFFER_END);
drivers/hv/vmbus_drv.c
2462
struct resource *iter, *shadow;
drivers/hv/vmbus_drv.c
2506
shadow = __request_region(iter, start, size, NULL,
drivers/hv/vmbus_drv.c
2508
if (!shadow)
drivers/hv/vmbus_drv.c
2513
shadow->name = (char *)*new;
drivers/infiniband/hw/irdma/verbs.c
2288
info->shadow_area_pa = srqmr->shadow;
drivers/infiniband/hw/irdma/verbs.c
2575
info.shadow_area_pa = cqmr->shadow;
drivers/infiniband/hw/irdma/verbs.c
2936
qpmr->shadow = (dma_addr_t)arr[total];
drivers/infiniband/hw/irdma/verbs.c
2962
srqmr->shadow = (dma_addr_t)arr[req->rq_pages];
drivers/infiniband/hw/irdma/verbs.c
2976
cqmr->shadow = (dma_addr_t)arr[req->cq_pages];
drivers/infiniband/hw/irdma/verbs.c
589
init_info->shadow_area_pa = qpmr->shadow;
drivers/infiniband/hw/irdma/verbs.h
67
dma_addr_t shadow;
drivers/infiniband/hw/irdma/verbs.h
73
dma_addr_t shadow;
drivers/infiniband/hw/irdma/verbs.h
79
dma_addr_t shadow;
drivers/irqchip/irq-econet-en751221.c
219
u32 shadow = shadows[i + 1];
drivers/irqchip/irq-econet-en751221.c
222
if (shadow > IRQ_COUNT) {
drivers/irqchip/irq-econet-en751221.c
224
node, field, i + 1, shadow);
drivers/irqchip/irq-econet-en751221.c
239
if (econet_intc.interrupt_shadows[shadow] != NOT_PERCPU) {
drivers/irqchip/irq-econet-en751221.c
241
node, field, i + 1, shadow);
drivers/irqchip/irq-econet-en751221.c
245
econet_intc.interrupt_shadows[target] = shadow;
drivers/irqchip/irq-econet-en751221.c
246
econet_intc.interrupt_shadows[shadow] = IS_SHADOW;
drivers/irqchip/irq-econet-en751221.c
74
u8 shadow;
drivers/irqchip/irq-econet-en751221.c
83
shadow = econet_intc.interrupt_shadows[hwirq];
drivers/irqchip/irq-econet-en751221.c
84
if (WARN_ON_ONCE(shadow == IS_SHADOW))
drivers/irqchip/irq-econet-en751221.c
86
else if (shadow != NOT_PERCPU && smp_processor_id() == 1)
drivers/irqchip/irq-econet-en751221.c
87
hwirq = shadow;
drivers/media/dvb-frontends/itd1000.c
382
memset(state->shadow, 0xff, sizeof(state->shadow));
drivers/media/dvb-frontends/itd1000.c
384
state->shadow[i] = itd1000_read_reg(state, i);
drivers/media/dvb-frontends/itd1000.c
77
itd1000_write_regs(state, (reg - 1) & 0xff, &state->shadow[(reg - 1) & 0xff], 1);
drivers/media/dvb-frontends/itd1000.c
90
state->shadow[r] = tmp;
drivers/media/dvb-frontends/itd1000_priv.h
20
u8 shadow[256];
drivers/media/dvb-frontends/s5h1420.c
108
state->shadow[reg] = data;
drivers/media/dvb-frontends/s5h1420.c
48
u8 shadow[256];
drivers/media/dvb-frontends/s5h1420.c
76
b[1] = state->shadow[(reg - 1) & 0xff];
drivers/media/dvb-frontends/s5h1420.c
895
memset(state->shadow, 0xff, sizeof(state->shadow));
drivers/media/dvb-frontends/s5h1420.c
898
state->shadow[i] = s5h1420_readreg(state, i);
drivers/media/i2c/tvaudio.c
1011
if(chip->shadow.bytes[MAXREGS-2] & 0x20) /* DSR.RSSF=1 */
drivers/media/i2c/tvaudio.c
126
audiocmd shadow;
drivers/media/i2c/tvaudio.c
1343
int s1 = chip->shadow.bytes[TDA8425_S1+1] & 0xe1;
drivers/media/i2c/tvaudio.c
163
chip->shadow.bytes[1] = val;
drivers/media/i2c/tvaudio.c
173
if (subaddr + 1 >= ARRAY_SIZE(chip->shadow.bytes)) {
drivers/media/i2c/tvaudio.c
182
chip->shadow.bytes[subaddr+1] = val;
drivers/media/i2c/tvaudio.c
1987
chip->shadow.count = desc->registers+1;
drivers/media/i2c/tvaudio.c
204
val = (chip->shadow.bytes[1] & ~mask) | (val & mask);
drivers/media/i2c/tvaudio.c
206
if (subaddr + 1 >= ARRAY_SIZE(chip->shadow.bytes)) {
drivers/media/i2c/tvaudio.c
213
val = (chip->shadow.bytes[subaddr+1] & ~mask) | (val & mask);
drivers/media/i2c/tvaudio.c
281
if (cmd->count + cmd->bytes[0] - 1 >= ARRAY_SIZE(chip->shadow.bytes)) {
drivers/media/i2c/tvaudio.c
296
chip->shadow.bytes[i+cmd->bytes[0]] = cmd->bytes[i];
drivers/media/i2c/tvaudio.c
441
int t = chip->shadow.bytes[TDA9840_SW + 1] & ~0x7e;
drivers/media/i2c/tvaudio.c
606
int c6 = chip->shadow.bytes[TDA985x_C6+1] & 0x3f;
drivers/media/i2c/tvaudio.c
772
int sw_data = chip->shadow.bytes[TDA9873_SW+1] & ~ TDA9873_TR_MASK;
drivers/media/i2c/tvaudio.c
783
TDA9873_SW+1, chip->shadow.bytes[TDA9873_SW+1]);
drivers/media/i2c/tvaudio.c
975
chip->shadow.bytes[MAXREGS-2] = dsr;
drivers/media/i2c/tvaudio.c
976
chip->shadow.bytes[MAXREGS-1] = nsr;
drivers/media/pci/cx88/cx88.h
346
u32 shadow[SHADOW_MAX];
drivers/media/pci/cx88/cx88.h
597
#define cx_sread(sreg) (core->shadow[sreg])
drivers/media/pci/cx88/cx88.h
599
(core->shadow[sreg] = value, \
drivers/media/pci/cx88/cx88.h
600
writel(core->shadow[sreg], core->lmmio + ((reg) >> 2)))
drivers/media/pci/cx88/cx88.h
602
(core->shadow[sreg] = (core->shadow[sreg] & ~(mask)) | \
drivers/media/pci/cx88/cx88.h
604
writel(core->shadow[sreg], \
drivers/media/platform/allegro-dvt/allegro-core.c
1976
struct allegro_m2m_buffer *shadow = to_allegro_m2m_buffer(b);
drivers/media/platform/allegro-dvt/allegro-core.c
1979
list_add_tail(&shadow->head, list);
drivers/media/platform/allegro-dvt/allegro-core.c
1989
struct allegro_m2m_buffer *shadow, *tmp;
drivers/media/platform/allegro-dvt/allegro-core.c
1993
list_for_each_entry_safe(shadow, tmp, list, head) {
drivers/media/platform/allegro-dvt/allegro-core.c
1994
if (handle == ptr_to_u64(&shadow->buf.vb)) {
drivers/media/platform/allegro-dvt/allegro-core.c
1995
buffer = &shadow->buf.vb;
drivers/media/platform/allegro-dvt/allegro-core.c
1996
list_del_init(&shadow->head);
drivers/media/platform/allegro-dvt/allegro-core.c
2918
struct allegro_m2m_buffer *shadow, *tmp;
drivers/media/platform/allegro-dvt/allegro-core.c
2926
list_for_each_entry_safe(shadow, tmp,
drivers/media/platform/allegro-dvt/allegro-core.c
2928
list_del(&shadow->head);
drivers/media/platform/allegro-dvt/allegro-core.c
2929
v4l2_m2m_buf_done(&shadow->buf.vb, VB2_BUF_STATE_ERROR);
drivers/media/platform/allegro-dvt/allegro-core.c
2937
list_for_each_entry_safe(shadow, tmp,
drivers/media/platform/allegro-dvt/allegro-core.c
2939
list_del(&shadow->head);
drivers/media/platform/allegro-dvt/allegro-core.c
2940
v4l2_m2m_buf_done(&shadow->buf.vb, VB2_BUF_STATE_ERROR);
drivers/media/platform/ti/omap3isp/isppreview.c
672
preview_params_lock(struct isp_prev_device *prev, u32 update, bool shadow)
drivers/media/platform/ti/omap3isp/isppreview.c
676
if (shadow) {
drivers/media/platform/ti/omap3isp/isppreview.c
693
preview_params_unlock(struct isp_prev_device *prev, u32 update, bool shadow)
drivers/media/platform/ti/omap3isp/isppreview.c
697
if (shadow) {
drivers/mfd/cs42l43.c
597
static int cs42l43_mcu_stage_2_3(struct cs42l43 *cs42l43, bool shadow)
drivers/mfd/cs42l43.c
603
if (shadow)
drivers/mfd/cs42l43.c
742
bool patched, shadow;
drivers/mfd/cs42l43.c
777
shadow = (mcu_rev >= CS42L43_MCU_SHADOW_REGS_REQUIRED_REV) ||
drivers/mfd/cs42l43.c
815
return cs42l43_mcu_stage_2_3(cs42l43, shadow);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1240
rx->shadow[idx].addr_low =
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1242
rx->shadow[idx].addr_high =
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1259
&rx->shadow[idx - 7]);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
127
struct mcp_kreq_ether_recv *shadow; /* host shadow of recv ring */
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1989
bytes = rx_ring_entries * sizeof(*ss->rx_small.shadow);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1990
ss->rx_small.shadow = kzalloc(bytes, GFP_KERNEL);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1991
if (ss->rx_small.shadow == NULL)
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1994
bytes = rx_ring_entries * sizeof(*ss->rx_big.shadow);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1995
ss->rx_big.shadow = kzalloc(bytes, GFP_KERNEL);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1996
if (ss->rx_big.shadow == NULL)
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2074
kfree(ss->rx_big.shadow);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2077
kfree(ss->rx_small.shadow);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2151
kfree(ss->rx_big.shadow);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2153
kfree(ss->rx_small.shadow);
drivers/net/ethernet/qlogic/qed/qed_vf.c
1476
struct qed_bulletin_content shadow;
drivers/net/ethernet/qlogic/qed/qed_vf.c
1483
memcpy(&shadow, p_iov->bulletin.p_virt, p_iov->bulletin.size);
drivers/net/ethernet/qlogic/qed/qed_vf.c
1486
if (shadow.version == p_iov->bulletin_shadow.version)
drivers/net/ethernet/qlogic/qed/qed_vf.c
1490
crc = crc32(0, (u8 *)&shadow + crc_size,
drivers/net/ethernet/qlogic/qed/qed_vf.c
1492
if (crc != shadow.crc)
drivers/net/ethernet/qlogic/qed/qed_vf.c
1496
memcpy(&p_iov->bulletin_shadow, &shadow, p_iov->bulletin.size);
drivers/net/ethernet/qlogic/qed/qed_vf.c
1499
"Read a bulletin update %08x\n", shadow.version);
drivers/net/ethernet/ti/icssm/icssm_prueth.h
99
bool shadow;
drivers/net/phy/bcm-phy-lib.c
242
int bcm_phy_read_shadow(struct phy_device *phydev, u16 shadow)
drivers/net/phy/bcm-phy-lib.c
244
phy_write(phydev, MII_BCM54XX_SHD, MII_BCM54XX_SHD_VAL(shadow));
drivers/net/phy/bcm-phy-lib.c
249
int bcm_phy_write_shadow(struct phy_device *phydev, u16 shadow,
drivers/net/phy/bcm-phy-lib.c
254
MII_BCM54XX_SHD_VAL(shadow) |
drivers/net/phy/bcm-phy-lib.c
533
static u64 bcm_phy_get_stat(struct phy_device *phydev, u64 *shadow,
drivers/net/phy/bcm-phy-lib.c
549
shadow[i] += val;
drivers/net/phy/bcm-phy-lib.c
550
ret = shadow[i];
drivers/net/phy/bcm-phy-lib.c
556
void bcm_phy_get_stats(struct phy_device *phydev, u64 *shadow,
drivers/net/phy/bcm-phy-lib.c
562
data[i] = bcm_phy_get_stat(phydev, shadow, i);
drivers/net/phy/bcm-phy-lib.h
59
int bcm_phy_write_shadow(struct phy_device *phydev, u16 shadow,
drivers/net/phy/bcm-phy-lib.h
61
int bcm_phy_read_shadow(struct phy_device *phydev, u16 shadow);
drivers/net/phy/bcm-phy-lib.h
86
void bcm_phy_get_stats(struct phy_device *phydev, u64 *shadow,
drivers/pinctrl/renesas/gpio.c
113
chip->regs[i].shadow = gpio_read_data_reg(chip, dreg);
drivers/pinctrl/renesas/gpio.c
158
reg->shadow |= BIT(pos);
drivers/pinctrl/renesas/gpio.c
160
reg->shadow &= ~BIT(pos);
drivers/pinctrl/renesas/gpio.c
162
gpio_write_data_reg(chip, reg->info, reg->shadow);
drivers/pinctrl/renesas/gpio.c
20
u32 shadow;
drivers/ps3/ps3-lpm.c
1133
lpm_priv->shadow.pm_control = PS3_LPM_SHADOW_REG_INIT;
drivers/ps3/ps3-lpm.c
1134
lpm_priv->shadow.pm_start_stop = PS3_LPM_SHADOW_REG_INIT;
drivers/ps3/ps3-lpm.c
1135
lpm_priv->shadow.group_control = PS3_LPM_SHADOW_REG_INIT;
drivers/ps3/ps3-lpm.c
1136
lpm_priv->shadow.debug_bus_control = PS3_LPM_SHADOW_REG_INIT;
drivers/ps3/ps3-lpm.c
128
struct ps3_lpm_shadow_regs shadow;
drivers/ps3/ps3-lpm.c
394
return lpm_priv->shadow.pm_control;
drivers/ps3/ps3-lpm.c
398
return lpm_priv->shadow.pm_start_stop;
drivers/ps3/ps3-lpm.c
409
return lpm_priv->shadow.group_control;
drivers/ps3/ps3-lpm.c
411
return lpm_priv->shadow.debug_bus_control;
drivers/ps3/ps3-lpm.c
446
if (val != lpm_priv->shadow.group_control)
drivers/ps3/ps3-lpm.c
451
lpm_priv->shadow.group_control = val;
drivers/ps3/ps3-lpm.c
454
if (val != lpm_priv->shadow.debug_bus_control)
drivers/ps3/ps3-lpm.c
459
lpm_priv->shadow.debug_bus_control = val;
drivers/ps3/ps3-lpm.c
465
if (val != lpm_priv->shadow.pm_control)
drivers/ps3/ps3-lpm.c
471
lpm_priv->shadow.pm_control = val;
drivers/ps3/ps3-lpm.c
478
if (val != lpm_priv->shadow.pm_start_stop)
drivers/ps3/ps3-lpm.c
483
lpm_priv->shadow.pm_start_stop = val;
drivers/ps3/ps3-lpm.c
847
if (!(lpm_priv->shadow.pm_start_stop &
drivers/scsi/xen-scsifront.c
121
struct vscsifrnt_shadow *shadow[VSCSIIF_MAX_REQS];
drivers/scsi/xen-scsifront.c
163
info->shadow[id] = NULL;
drivers/scsi/xen-scsifront.c
182
struct vscsifrnt_shadow *shadow)
drivers/scsi/xen-scsifront.c
186
struct scsi_cmnd *sc = shadow->sc;
drivers/scsi/xen-scsifront.c
197
info->shadow[id] = shadow;
drivers/scsi/xen-scsifront.c
198
shadow->rqid = id;
drivers/scsi/xen-scsifront.c
204
ring_req->act = shadow->act;
drivers/scsi/xen-scsifront.c
205
ring_req->ref_rqid = shadow->ref_rqid;
drivers/scsi/xen-scsifront.c
206
ring_req->nr_segments = shadow->nr_segments;
drivers/scsi/xen-scsifront.c
220
for (i = 0; i < (shadow->nr_segments & ~VSCSIIF_SG_GRANT); i++)
drivers/scsi/xen-scsifront.c
221
ring_req->seg[i] = shadow->seg[i];
drivers/scsi/xen-scsifront.c
223
shadow->inflight = true;
drivers/scsi/xen-scsifront.c
240
struct vscsifrnt_shadow *shadow)
drivers/scsi/xen-scsifront.c
244
if (shadow->sc->sc_data_direction == DMA_NONE)
drivers/scsi/xen-scsifront.c
247
for (i = 0; i < shadow->nr_grants; i++) {
drivers/scsi/xen-scsifront.c
248
if (unlikely(!gnttab_try_end_foreign_access(shadow->gref[i]))) {
drivers/scsi/xen-scsifront.c
254
kfree(shadow->sg);
drivers/scsi/xen-scsifront.c
302
struct vscsifrnt_shadow *shadow;
drivers/scsi/xen-scsifront.c
308
shadow = info->shadow[id];
drivers/scsi/xen-scsifront.c
309
sc = shadow->sc;
drivers/scsi/xen-scsifront.c
313
scsifront_gnttab_done(info, shadow);
drivers/scsi/xen-scsifront.c
336
struct vscsifrnt_shadow *shadow = info->shadow[id];
drivers/scsi/xen-scsifront.c
340
shadow->wait_reset = 1;
drivers/scsi/xen-scsifront.c
341
switch (shadow->rslt_reset) {
drivers/scsi/xen-scsifront.c
344
shadow->rslt_reset = SUCCESS;
drivers/scsi/xen-scsifront.c
346
shadow->rslt_reset = FAILED;
drivers/scsi/xen-scsifront.c
351
kfree(shadow);
drivers/scsi/xen-scsifront.c
361
wake_up(&shadow->wq_reset);
drivers/scsi/xen-scsifront.c
367
struct vscsifrnt_shadow *shadow;
drivers/scsi/xen-scsifront.c
370
!info->shadow[ring_rsp->rqid]->inflight) {
drivers/scsi/xen-scsifront.c
374
shadow = info->shadow[ring_rsp->rqid];
drivers/scsi/xen-scsifront.c
375
shadow->inflight = false;
drivers/scsi/xen-scsifront.c
377
if (shadow->act == VSCSIIF_ACT_SCSI_CDB)
drivers/scsi/xen-scsifront.c
472
struct vscsifrnt_shadow *shadow)
drivers/scsi/xen-scsifront.c
497
shadow->sg = kzalloc_objs(struct scsiif_request_segment,
drivers/scsi/xen-scsifront.c
499
if (!shadow->sg)
drivers/scsi/xen-scsifront.c
502
seg = shadow->sg ? : shadow->seg;
drivers/scsi/xen-scsifront.c
507
kfree(shadow->sg);
drivers/scsi/xen-scsifront.c
526
shadow->gref[ref_cnt] = ref;
drivers/scsi/xen-scsifront.c
527
shadow->seg[ref_cnt].gref = ref;
drivers/scsi/xen-scsifront.c
528
shadow->seg[ref_cnt].offset = (uint16_t)off;
drivers/scsi/xen-scsifront.c
529
shadow->seg[ref_cnt].length = (uint16_t)bytes;
drivers/scsi/xen-scsifront.c
562
shadow->gref[ref_cnt] = ref;
drivers/scsi/xen-scsifront.c
577
shadow->nr_segments = VSCSIIF_SG_GRANT | seg_grants;
drivers/scsi/xen-scsifront.c
579
shadow->nr_segments = (uint8_t)ref_cnt;
drivers/scsi/xen-scsifront.c
580
shadow->nr_grants = ref_cnt;
drivers/scsi/xen-scsifront.c
610
struct vscsifrnt_shadow *shadow = scsi_cmd_priv(sc);
drivers/scsi/xen-scsifront.c
619
shadow->sc = sc;
drivers/scsi/xen-scsifront.c
620
shadow->act = VSCSIIF_ACT_SCSI_CDB;
drivers/scsi/xen-scsifront.c
628
err = map_data_for_request(info, sc, shadow);
drivers/scsi/xen-scsifront.c
640
if (scsifront_do_request(info, shadow)) {
drivers/scsi/xen-scsifront.c
641
scsifront_gnttab_done(info, shadow);
drivers/scsi/xen-scsifront.c
666
struct vscsifrnt_shadow *shadow, *s = scsi_cmd_priv(sc);
drivers/scsi/xen-scsifront.c
672
shadow = kzalloc_obj(*shadow, GFP_NOIO);
drivers/scsi/xen-scsifront.c
673
if (!shadow)
drivers/scsi/xen-scsifront.c
676
shadow->act = act;
drivers/scsi/xen-scsifront.c
677
shadow->rslt_reset = RSLT_RESET_WAITING;
drivers/scsi/xen-scsifront.c
678
shadow->sc = sc;
drivers/scsi/xen-scsifront.c
679
shadow->ref_rqid = s->rqid;
drivers/scsi/xen-scsifront.c
680
init_waitqueue_head(&shadow->wq_reset);
drivers/scsi/xen-scsifront.c
688
if (!scsifront_do_request(info, shadow))
drivers/scsi/xen-scsifront.c
702
err = wait_event_interruptible(shadow->wq_reset, shadow->wait_reset);
drivers/scsi/xen-scsifront.c
706
err = shadow->rslt_reset;
drivers/scsi/xen-scsifront.c
707
scsifront_put_rqid(info, shadow->rqid);
drivers/scsi/xen-scsifront.c
708
kfree(shadow);
drivers/scsi/xen-scsifront.c
711
shadow->rslt_reset = RSLT_RESET_ERR;
drivers/scsi/xen-scsifront.c
722
kfree(shadow);
drivers/usb/host/xen-hcd.c
1490
info->shadow[i].req.id = i + 1;
drivers/usb/host/xen-hcd.c
1491
info->shadow[i].urb = NULL;
drivers/usb/host/xen-hcd.c
1492
info->shadow[i].in_flight = false;
drivers/usb/host/xen-hcd.c
1494
info->shadow[XENUSB_URB_RING_SIZE - 1].req.id = 0x0fff;
drivers/usb/host/xen-hcd.c
565
info->shadow_free = info->shadow[free].req.id;
drivers/usb/host/xen-hcd.c
566
info->shadow[free].req.id = 0x0fff; /* debug */
drivers/usb/host/xen-hcd.c
573
info->shadow[id].req.id = info->shadow_free;
drivers/usb/host/xen-hcd.c
574
info->shadow[id].urb = NULL;
drivers/usb/host/xen-hcd.c
718
struct usb_shadow *shadow = info->shadow + id;
drivers/usb/host/xen-hcd.c
722
if (!shadow->in_flight) {
drivers/usb/host/xen-hcd.c
726
shadow->in_flight = false;
drivers/usb/host/xen-hcd.c
728
nr_segs = shadow->req.nr_buffer_segs;
drivers/usb/host/xen-hcd.c
730
if (xenusb_pipeisoc(shadow->req.pipe))
drivers/usb/host/xen-hcd.c
731
nr_segs += shadow->req.u.isoc.nr_frame_desc_segs;
drivers/usb/host/xen-hcd.c
734
if (!gnttab_try_end_foreign_access(shadow->req.seg[i].gref))
drivers/usb/host/xen-hcd.c
738
shadow->req.nr_buffer_segs = 0;
drivers/usb/host/xen-hcd.c
739
shadow->req.u.isoc.nr_frame_desc_segs = 0;
drivers/usb/host/xen-hcd.c
789
req = &info->shadow[id].req;
drivers/usb/host/xen-hcd.c
807
*req = info->shadow[id].req;
drivers/usb/host/xen-hcd.c
810
info->shadow[id].urb = urb;
drivers/usb/host/xen-hcd.c
811
info->shadow[id].in_flight = true;
drivers/usb/host/xen-hcd.c
84
struct usb_shadow shadow[XENUSB_URB_RING_SIZE];
drivers/usb/host/xen-hcd.c
862
info->shadow[req_id].urb = NULL;
drivers/usb/host/xen-hcd.c
983
if (likely(xenusb_pipesubmit(info->shadow[id].req.pipe))) {
drivers/usb/host/xen-hcd.c
987
xenhcd_res_to_urb(info, &res, info->shadow[id].urb);
drivers/video/fbdev/udlfb.c
1214
if (shadow)
drivers/video/fbdev/udlfb.c
1628
dev_dbg(&intf->dev, "shadow enable=%d\n", shadow);
drivers/video/fbdev/udlfb.c
1947
module_param(shadow, bool, S_IWUSR | S_IRUSR | S_IWGRP | S_IRGRP);
drivers/video/fbdev/udlfb.c
1948
MODULE_PARM_DESC(shadow, "Shadow vid mem. Disable to save mem but lose perf");
drivers/video/fbdev/udlfb.c
71
static bool shadow = true; /* Optionally disable shadow framebuffer */
fs/nilfs2/dat.c
32
struct nilfs_shadow_map shadow;
fs/nilfs2/dat.c
523
err = nilfs_mdt_setup_shadow_map(dat, &di->shadow);
fs/nilfs2/mdt.c
480
struct nilfs_shadow_map *shadow = mdi->mi_shadow;
fs/nilfs2/mdt.c
485
if (shadow) {
fs/nilfs2/mdt.c
486
struct inode *s_inode = shadow->inode;
fs/nilfs2/mdt.c
488
shadow->inode = NULL;
fs/nilfs2/mdt.c
524
struct nilfs_shadow_map *shadow)
fs/nilfs2/mdt.c
529
INIT_LIST_HEAD(&shadow->frozen_buffers);
fs/nilfs2/mdt.c
535
shadow->inode = s_inode;
fs/nilfs2/mdt.c
536
mi->mi_shadow = shadow;
fs/nilfs2/mdt.c
550
struct nilfs_shadow_map *shadow = mi->mi_shadow;
fs/nilfs2/mdt.c
551
struct inode *s_inode = shadow->inode;
fs/nilfs2/mdt.c
563
nilfs_bmap_save(ii->i_bmap, &shadow->bmap_store);
fs/nilfs2/mdt.c
570
struct nilfs_shadow_map *shadow = NILFS_MDT(inode)->mi_shadow;
fs/nilfs2/mdt.c
575
folio = filemap_grab_folio(shadow->inode->i_mapping,
fs/nilfs2/mdt.c
591
&shadow->frozen_buffers);
fs/nilfs2/mdt.c
605
struct nilfs_shadow_map *shadow = NILFS_MDT(inode)->mi_shadow;
fs/nilfs2/mdt.c
610
folio = filemap_lock_folio(shadow->inode->i_mapping,
fs/nilfs2/mdt.c
625
static void nilfs_release_frozen_buffers(struct nilfs_shadow_map *shadow)
fs/nilfs2/mdt.c
627
struct list_head *head = &shadow->frozen_buffers;
fs/nilfs2/mdt.c
646
struct nilfs_shadow_map *shadow = mi->mi_shadow;
fs/nilfs2/mdt.c
654
nilfs_copy_back_pages(inode->i_mapping, shadow->inode->i_mapping);
fs/nilfs2/mdt.c
658
NILFS_I(shadow->inode)->i_assoc_inode->i_mapping);
fs/nilfs2/mdt.c
660
nilfs_bmap_restore(ii->i_bmap, &shadow->bmap_store);
fs/nilfs2/mdt.c
672
struct nilfs_shadow_map *shadow = mi->mi_shadow;
fs/nilfs2/mdt.c
673
struct inode *shadow_btnc_inode = NILFS_I(shadow->inode)->i_assoc_inode;
fs/nilfs2/mdt.c
676
nilfs_release_frozen_buffers(shadow);
fs/nilfs2/mdt.c
677
truncate_inode_pages(shadow->inode->i_mapping, 0);
fs/nilfs2/mdt.h
85
struct nilfs_shadow_map *shadow);
fs/xfs/xfs_log_cil.c
552
struct xfs_log_vec *shadow = lip->li_lv_shadow;
fs/xfs/xfs_log_cil.c
563
if (shadow->lv_buf_used == XFS_LOG_VEC_ORDERED) {
fs/xfs/xfs_log_cil.c
565
lv = shadow;
fs/xfs/xfs_log_cil.c
568
ASSERT(shadow->lv_alloc_size == lv->lv_alloc_size);
fs/xfs/xfs_log_cil.c
574
if (!shadow->lv_niovecs)
fs/xfs/xfs_log_cil.c
578
if (lv && shadow->lv_alloc_size <= lv->lv_alloc_size) {
fs/xfs/xfs_log_cil.c
588
lv->lv_niovecs = shadow->lv_niovecs;
fs/xfs/xfs_log_cil.c
597
lv = shadow;
include/linux/jbd2.h
328
BUFFER_FNS(Shadow, shadow)
include/linux/pagemap.h
1290
void __filemap_remove_folio(struct folio *folio, void *shadow);
include/linux/scx200_gpio.h
13
#define __SCx200_GPIO_SHADOW unsigned long *shadow = scx200_gpio_shadow+bank
include/linux/scx200_gpio.h
16
#define __SCx200_GPIO_OUT __asm__ __volatile__("outsl":"=mS" (shadow):"d" (ioaddr), "0" (shadow))
include/linux/scx200_gpio.h
46
set_bit(index, shadow); /* __set_bit()? */
include/linux/scx200_gpio.h
57
clear_bit(index, shadow); /* __clear_bit()? */
include/linux/scx200_gpio.h
69
set_bit(index, shadow);
include/linux/scx200_gpio.h
71
clear_bit(index, shadow);
include/linux/scx200_gpio.h
81
change_bit(index, shadow);
include/linux/swap.h
315
bool workingset_test_recent(void *shadow, bool file, bool *workingset,
include/linux/swap.h
319
void workingset_refault(struct folio *folio, void *shadow);
kernel/livepatch/shadow.c
234
static void klp_shadow_free_struct(struct klp_shadow *shadow,
kernel/livepatch/shadow.c
237
hash_del_rcu(&shadow->node);
kernel/livepatch/shadow.c
239
dtor(shadow->obj, shadow->data);
kernel/livepatch/shadow.c
240
kfree_rcu(shadow, rcu_head);
kernel/livepatch/shadow.c
255
struct klp_shadow *shadow;
kernel/livepatch/shadow.c
261
hash_for_each_possible(klp_shadow_hash, shadow, node,
kernel/livepatch/shadow.c
264
if (klp_shadow_match(shadow, obj, id)) {
kernel/livepatch/shadow.c
265
klp_shadow_free_struct(shadow, dtor);
kernel/livepatch/shadow.c
285
struct klp_shadow *shadow;
kernel/livepatch/shadow.c
292
hash_for_each(klp_shadow_hash, i, shadow, node) {
kernel/livepatch/shadow.c
293
if (klp_shadow_match(shadow, shadow->obj, id))
kernel/livepatch/shadow.c
294
klp_shadow_free_struct(shadow, dtor);
kernel/livepatch/shadow.c
70
static inline bool klp_shadow_match(struct klp_shadow *shadow, void *obj,
kernel/livepatch/shadow.c
73
return shadow->obj == obj && shadow->id == id;
kernel/livepatch/shadow.c
85
struct klp_shadow *shadow;
kernel/livepatch/shadow.c
89
hash_for_each_possible_rcu(klp_shadow_hash, shadow, node,
kernel/livepatch/shadow.c
92
if (klp_shadow_match(shadow, obj, id)) {
kernel/livepatch/shadow.c
94
return shadow->data;
mm/filemap.c
130
struct folio *folio, void *shadow)
mm/filemap.c
142
xas_store(&xas, shadow);
mm/filemap.c
222
void __filemap_remove_folio(struct folio *folio, void *shadow)
mm/filemap.c
228
page_cache_delete(mapping, folio, shadow);
mm/filemap.c
4611
void *shadow = (void *)folio;
mm/filemap.c
4635
shadow = swap_cache_get_shadow(swp);
mm/filemap.c
4636
if (!shadow)
mm/filemap.c
4640
if (workingset_test_recent(shadow, true, &workingset, false))
mm/filemap.c
952
void *shadow = NULL;
mm/filemap.c
966
ret = __filemap_add_folio(mapping, folio, index, gfp, &shadow);
mm/filemap.c
980
if (!(gfp & __GFP_WRITE) && shadow)
mm/filemap.c
981
workingset_refault(folio, shadow);
mm/kasan/report_generic.c
49
u8 *shadow;
mm/kasan/report_generic.c
60
shadow = (u8 *)kasan_mem_to_shadow(object);
mm/kasan/report_generic.c
62
if (*shadow == 0)
mm/kasan/report_generic.c
64
else if (*shadow >= 1 && *shadow <= KASAN_GRANULE_SIZE - 1)
mm/kasan/report_generic.c
65
return size + *shadow;
mm/kasan/report_generic.c
68
shadow++;
mm/kasan/report_sw_tags.c
51
u8 *shadow;
mm/kasan/report_sw_tags.c
62
shadow = (u8 *)kasan_mem_to_shadow(object);
mm/kasan/report_sw_tags.c
64
if (*shadow != KASAN_TAG_INVALID)
mm/kasan/report_sw_tags.c
68
shadow++;
mm/kasan/report_sw_tags.c
81
u8 *shadow = (u8 *)kasan_mem_to_shadow(addr);
mm/kasan/report_sw_tags.c
83
pr_err("Pointer tag: [%02x], memory tag: [%02x]\n", addr_tag, *shadow);
mm/kasan/shadow.c
157
u8 *shadow = (u8 *)kasan_mem_to_shadow(addr + size);
mm/kasan/shadow.c
158
*shadow = size & KASAN_GRANULE_MASK;
mm/kasan/sw_tags.c
115
for (shadow = shadow_first; shadow <= shadow_last; shadow++) {
mm/kasan/sw_tags.c
116
if (*shadow != tag) {
mm/kasan/sw_tags.c
79
u8 *shadow_first, *shadow_last, *shadow;
mm/kmsan/core.c
259
unsigned char *shadow = NULL;
mm/kmsan/core.c
270
shadow = kmsan_get_metadata((void *)(addr64 + pos),
mm/kmsan/core.c
272
if (!shadow) {
mm/kmsan/core.c
288
if (!shadow[i]) {
mm/kmsan/hooks.c
154
struct page *shadow, *origin;
mm/kmsan/hooks.c
164
shadow = alloc_pages(gfp_mask, 1);
mm/kmsan/hooks.c
166
if (!shadow || !origin) {
mm/kmsan/hooks.c
172
vmalloc_shadow(start + off + PAGE_SIZE), prot, &shadow,
mm/kmsan/hooks.c
178
shadow = NULL;
mm/kmsan/hooks.c
202
if (shadow)
mm/kmsan/hooks.c
203
__free_pages(shadow, 1);
mm/kmsan/hooks.c
222
struct page *shadow, *origin;
mm/kmsan/hooks.c
234
shadow = kmsan_vmalloc_to_page_or_null((void *)v_shadow);
mm/kmsan/hooks.c
238
if (shadow)
mm/kmsan/hooks.c
239
__free_pages(shadow, 1);
mm/kmsan/init.c
118
struct page *shadow, *origin;
mm/kmsan/init.c
120
if (!held_back[order].shadow) {
mm/kmsan/init.c
121
held_back[order].shadow = page;
mm/kmsan/init.c
128
shadow = held_back[order].shadow;
mm/kmsan/init.c
130
kmsan_setup_meta(page, shadow, origin, order);
mm/kmsan/init.c
132
held_back[order].shadow = NULL;
mm/kmsan/init.c
170
struct page *page, *shadow, *origin;
mm/kmsan/init.c
174
shadow = smallstack_pop(&collect);
mm/kmsan/init.c
176
kmsan_setup_meta(page, shadow, origin, collect.order);
mm/kmsan/init.c
218
if (held_back[i].shadow)
mm/kmsan/init.c
219
smallstack_push(&collect, held_back[i].shadow);
mm/kmsan/init.c
222
held_back[i].shadow = NULL;
mm/kmsan/init.c
99
struct page *shadow, *origin;
mm/kmsan/instrumentation.c
139
static inline void get_param0_metadata(u64 *shadow,
mm/kmsan/instrumentation.c
144
*shadow = *(u64 *)(ctx->cstate.param_tls);
mm/kmsan/instrumentation.c
148
static inline void set_retval_metadata(u64 shadow, depot_stack_handle_t origin)
mm/kmsan/instrumentation.c
152
*(u64 *)(ctx->cstate.retval_tls) = shadow;
mm/kmsan/instrumentation.c
162
u64 shadow;
mm/kmsan/instrumentation.c
164
get_param0_metadata(&shadow, &origin);
mm/kmsan/instrumentation.c
176
set_retval_metadata(shadow, origin);
mm/kmsan/instrumentation.c
187
u64 shadow;
mm/kmsan/instrumentation.c
189
get_param0_metadata(&shadow, &origin);
mm/kmsan/instrumentation.c
203
set_retval_metadata(shadow, origin);
mm/kmsan/instrumentation.c
214
u64 shadow;
mm/kmsan/instrumentation.c
216
get_param0_metadata(&shadow, &origin);
mm/kmsan/instrumentation.c
229
set_retval_metadata(shadow, origin);
mm/kmsan/kmsan.h
169
void kmsan_setup_meta(struct page *page, struct page *shadow,
mm/kmsan/kmsan.h
42
void *shadow, *origin;
mm/kmsan/shadow.c
101
ret.shadow = shadow;
mm/kmsan/shadow.c
108
ret.shadow = dummy_store_page;
mm/kmsan/shadow.c
112
ret.shadow = dummy_load_page;
mm/kmsan/shadow.c
173
struct page *shadow, *origin;
mm/kmsan/shadow.c
180
shadow = shadow_page_for(page);
mm/kmsan/shadow.c
184
__memset(page_address(shadow), 0, PAGE_SIZE * pages);
mm/kmsan/shadow.c
193
__memset(page_address(shadow), -1, PAGE_SIZE * pages);
mm/kmsan/shadow.c
278
void *shadow, *origin;
mm/kmsan/shadow.c
284
shadow = memblock_alloc_or_panic(size, PAGE_SIZE);
mm/kmsan/shadow.c
289
shadow_p = virt_to_page((char *)shadow + addr);
mm/kmsan/shadow.c
298
void kmsan_setup_meta(struct page *page, struct page *shadow,
mm/kmsan/shadow.c
302
set_no_shadow_origin_page(&shadow[i]);
mm/kmsan/shadow.c
304
shadow_page_for(&page[i]) = &shadow[i];
mm/kmsan/shadow.c
85
void *shadow;
mm/kmsan/shadow.c
97
shadow = kmsan_get_metadata(address, KMSAN_META_SHADOW);
mm/kmsan/shadow.c
98
if (!shadow)
mm/swap.h
286
struct folio *folio, swp_entry_t entry, void *shadow);
mm/swap.h
469
struct folio *folio, swp_entry_t entry, void *shadow)
mm/swap_state.c
182
void *shadow = NULL;
mm/swap_state.c
210
shadow = swp_tb_to_shadow(old_tb);
mm/swap_state.c
216
*shadowp = shadow;
mm/swap_state.c
238
swp_entry_t entry, void *shadow)
mm/swap_state.c
252
new_tb = shadow_swp_to_tb(shadow);
mm/swap_state.c
492
void *shadow;
mm/swap_state.c
502
ret = swap_cache_add_folio(folio, entry, &shadow);
mm/swap_state.c
522
if (shadow)
mm/swap_state.c
523
workingset_refault(folio, shadow);
mm/swap_table.h
40
static inline unsigned long shadow_swp_to_tb(void *shadow)
mm/swap_table.h
44
VM_WARN_ON_ONCE(shadow && !xa_is_value(shadow));
mm/swap_table.h
45
return (unsigned long)shadow;
mm/vmscan.c
710
void *shadow = NULL;
mm/vmscan.c
761
shadow = workingset_eviction(folio, target_memcg);
mm/vmscan.c
763
__swap_cache_del_folio(ci, folio, swap, shadow);
mm/vmscan.c
787
shadow = workingset_eviction(folio, target_memcg);
mm/vmscan.c
788
__filemap_remove_folio(folio, shadow);
mm/workingset.c
210
static void unpack_shadow(void *shadow, int *memcgidp, pg_data_t **pgdat,
mm/workingset.c
213
unsigned long entry = xa_to_value(shadow);
mm/workingset.c
264
static bool lru_gen_test_recent(void *shadow, struct lruvec **lruvec,
mm/workingset.c
272
unpack_shadow(shadow, &memcg_id, &pgdat, token, workingset);
mm/workingset.c
283
static void lru_gen_refault(struct folio *folio, void *shadow)
mm/workingset.c
296
recent = lru_gen_test_recent(shadow, &lruvec, &token, &workingset);
mm/workingset.c
333
static bool lru_gen_test_recent(void *shadow, struct lruvec **lruvec,
mm/workingset.c
339
static void lru_gen_refault(struct folio *folio, void *shadow)
mm/workingset.c
418
bool workingset_test_recent(void *shadow, bool file, bool *workingset,
mm/workingset.c
434
recent = lru_gen_test_recent(shadow, &eviction_lruvec, &eviction, workingset);
mm/workingset.c
440
unpack_shadow(shadow, &memcgid, &pgdat, &eviction, workingset);
mm/workingset.c
534
void workingset_refault(struct folio *folio, void *shadow)
mm/workingset.c
546
lru_gen_refault(folio, shadow);
mm/workingset.c
566
if (!workingset_test_recent(shadow, file, &workingset, true))
scripts/kconfig/lxdialog/dialog.h
83
struct dialog_color shadow;
scripts/kconfig/lxdialog/util.c
158
init_one_color(&dlg.shadow);
scripts/kconfig/lxdialog/util.c
475
wattrset(win, dlg.shadow.atr);
scripts/kconfig/lxdialog/util.c
48
DLG_COLOR(shadow, COLOR_BLACK, COLOR_BLACK, true);
scripts/kconfig/lxdialog/util.c
77
DLG_COLOR(shadow, COLOR_BLACK, COLOR_BLACK, false);
tools/arch/x86/include/uapi/asm/kvm.h
352
__u8 shadow;