Symbol: mapping
lib/libc/gen/strtofflags.c
125
if (strcmp(p, mapping[i].name + 2) == 0) {
lib/libc/gen/strtofflags.c
126
if (mapping[i].invert) {
lib/libc/gen/strtofflags.c
128
*clrp |= mapping[i].flag;
lib/libc/gen/strtofflags.c
131
*setp |= mapping[i].flag;
lib/libc/gen/strtofflags.c
134
} else if (strcmp(p, mapping[i].name) == 0) {
lib/libc/gen/strtofflags.c
135
if (mapping[i].invert) {
lib/libc/gen/strtofflags.c
137
*setp |= mapping[i].flag;
lib/libc/gen/strtofflags.c
140
*clrp |= mapping[i].flag;
lib/libc/gen/strtofflags.c
44
} mapping[] = {
lib/libc/gen/strtofflags.c
70
#define nmappings (sizeof(mapping) / sizeof(mapping[0]))
lib/libc/gen/strtofflags.c
91
if (setflags & mapping[i].flag) {
lib/libc/gen/strtofflags.c
94
for (sp = mapping[i].invert ? mapping[i].name :
lib/libc/gen/strtofflags.c
95
mapping[i].name + 2; *sp; *dp++ = *sp++) ;
lib/libc/gen/strtofflags.c
96
setflags &= ~mapping[i].flag;
lib/libcrypto/x509/x509_policy.c
475
POLICY_MAPPING mapping;
lib/libcrypto/x509/x509_policy.c
476
mapping.issuerDomainPolicy = node->policy;
lib/libcrypto/x509/x509_policy.c
477
if (sk_POLICY_MAPPING_find(mappings, &mapping) < 0)
lib/libcrypto/x509/x509_policy.c
509
POLICY_MAPPING *mapping;
lib/libcrypto/x509/x509_policy.c
534
mapping = sk_POLICY_MAPPING_value(mappings, i);
lib/libcrypto/x509/x509_policy.c
535
if (is_any_policy(mapping->issuerDomainPolicy) ||
lib/libcrypto/x509/x509_policy.c
536
is_any_policy(mapping->subjectDomainPolicy))
lib/libcrypto/x509/x509_policy.c
556
mapping = sk_POLICY_MAPPING_value(mappings, i);
lib/libcrypto/x509/x509_policy.c
562
OBJ_cmp(mapping->issuerDomainPolicy,
lib/libcrypto/x509/x509_policy.c
565
last_policy = mapping->issuerDomainPolicy;
lib/libcrypto/x509/x509_policy.c
570
mapping->issuerDomainPolicy);
lib/libcrypto/x509/x509_policy.c
575
mapping->issuerDomainPolicy);
lib/libcrypto/x509/x509_policy.c
614
mapping = POLICY_MAPPING_new();
lib/libcrypto/x509/x509_policy.c
615
if (mapping == NULL)
lib/libcrypto/x509/x509_policy.c
617
mapping->issuerDomainPolicy = OBJ_dup(node->policy);
lib/libcrypto/x509/x509_policy.c
618
mapping->subjectDomainPolicy = OBJ_dup(node->policy);
lib/libcrypto/x509/x509_policy.c
619
if (mapping->issuerDomainPolicy == NULL ||
lib/libcrypto/x509/x509_policy.c
620
mapping->subjectDomainPolicy == NULL ||
lib/libcrypto/x509/x509_policy.c
621
!sk_POLICY_MAPPING_push(mappings, mapping)) {
lib/libcrypto/x509/x509_policy.c
622
POLICY_MAPPING_free(mapping);
lib/libcrypto/x509/x509_policy.c
640
mapping = sk_POLICY_MAPPING_value(mappings, i);
lib/libcrypto/x509/x509_policy.c
649
mapping->issuerDomainPolicy) == NULL)
lib/libcrypto/x509/x509_policy.c
654
OBJ_cmp(last_node->policy, mapping->subjectDomainPolicy) !=
lib/libcrypto/x509/x509_policy.c
657
mapping->subjectDomainPolicy);
lib/libcrypto/x509/x509_policy.c
666
mapping->issuerDomainPolicy))
lib/libcrypto/x509/x509_policy.c
668
mapping->issuerDomainPolicy = NULL;
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
1785
struct amdgpu_bo_va_mapping *mapping;
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
1790
mapping = amdgpu_vm_bo_lookup_mapping(vm, addr);
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
1791
if (!mapping || !mapping->bo_va || !mapping->bo_va->base.bo)
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
1794
*bo = mapping->bo_va->base.bo;
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
1795
*map = mapping;
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.h
86
struct amdgpu_bo_va_mapping **mapping);
sys/dev/pci/drm/amd/amdgpu/amdgpu_gart.c
142
p[x].mapping = adev->mman.bdev.dev_mapping;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gart.c
252
p[x].mapping = NULL;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
1188
struct amdgpu_bo_va_mapping *mapping;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
1201
amdgpu_vm_bo_va_for_each_valid_mapping(bo_va, mapping) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
1203
vm_entries[num_mappings].addr = mapping->start * AMDGPU_GPU_PAGE_SIZE;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
1204
vm_entries[num_mappings].size = (mapping->last - mapping->start + 1) * AMDGPU_GPU_PAGE_SIZE;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
1205
vm_entries[num_mappings].offset = mapping->offset;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
1206
vm_entries[num_mappings].flags = mapping->flags;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
1211
amdgpu_vm_bo_va_for_each_invalid_mapping(bo_va, mapping) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
1213
vm_entries[num_mappings].addr = mapping->start * AMDGPU_GPU_PAGE_SIZE;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
1214
vm_entries[num_mappings].size = (mapping->last - mapping->start + 1) * AMDGPU_GPU_PAGE_SIZE;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
1215
vm_entries[num_mappings].offset = mapping->offset;
sys/dev/pci/drm/amd/amdgpu/amdgpu_gem.c
1216
vm_entries[num_mappings].flags = mapping->flags;
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
243
struct amdgpu_bo_va_mapping *mapping),
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
244
TP_ARGS(bo_va, mapping),
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
255
__entry->start = mapping->start;
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
256
__entry->last = mapping->last;
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
257
__entry->offset = mapping->offset;
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
258
__entry->flags = mapping->flags;
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
267
struct amdgpu_bo_va_mapping *mapping),
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
268
TP_ARGS(bo_va, mapping),
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
279
__entry->start = mapping->start;
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
280
__entry->last = mapping->last;
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
281
__entry->offset = mapping->offset;
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
282
__entry->flags = mapping->flags;
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
290
TP_PROTO(struct amdgpu_bo_va_mapping *mapping),
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
291
TP_ARGS(mapping),
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
299
__entry->soffset = mapping->start;
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
300
__entry->eoffset = mapping->last + 1;
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
301
__entry->flags = mapping->flags;
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
308
TP_PROTO(struct amdgpu_bo_va_mapping *mapping),
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
309
TP_ARGS(mapping)
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
313
TP_PROTO(struct amdgpu_bo_va_mapping *mapping),
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
314
TP_ARGS(mapping)
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
318
TP_PROTO(struct amdgpu_bo_va_mapping *mapping),
sys/dev/pci/drm/amd/amdgpu/amdgpu_trace.h
319
TP_ARGS(mapping)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ttm.c
1202
ttm->pages[i]->mapping = bdev->dev_mapping;
sys/dev/pci/drm/amd/amdgpu/amdgpu_ttm.c
2719
if (p->mapping != adev->mman.bdev.dev_mapping)
sys/dev/pci/drm/amd/amdgpu/amdgpu_ttm.c
2774
if (p->mapping != adev->mman.bdev.dev_mapping)
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq_fence.c
393
struct amdgpu_bo_va_mapping *mapping;
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq_fence.c
405
mapping = amdgpu_vm_bo_lookup_mapping(queue->vm, addr >> PAGE_SHIFT);
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq_fence.c
406
if (!mapping) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_userq_fence.c
412
bo = amdgpu_bo_ref(mapping->bo_va->base.bo);
sys/dev/pci/drm/amd/amdgpu/amdgpu_uvd.c
582
struct amdgpu_bo_va_mapping *mapping;
sys/dev/pci/drm/amd/amdgpu/amdgpu_uvd.c
588
r = amdgpu_cs_find_mapping(ctx->parser, addr, &bo, &mapping);
sys/dev/pci/drm/amd/amdgpu/amdgpu_uvd.c
914
struct amdgpu_bo_va_mapping *mapping;
sys/dev/pci/drm/amd/amdgpu/amdgpu_uvd.c
921
r = amdgpu_cs_find_mapping(ctx->parser, addr, &bo, &mapping);
sys/dev/pci/drm/amd/amdgpu/amdgpu_uvd.c
929
end = (mapping->last + 1 - mapping->start);
sys/dev/pci/drm/amd/amdgpu/amdgpu_uvd.c
932
addr -= mapping->start * AMDGPU_GPU_PAGE_SIZE;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vce.c
606
struct amdgpu_bo_va_mapping *mapping;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vce.c
623
r = amdgpu_cs_find_mapping(p, addr, &bo, &mapping);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vce.c
654
struct amdgpu_bo_va_mapping *mapping;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vce.c
666
r = amdgpu_cs_find_mapping(p, addr, &bo, &mapping);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vce.c
674
(mapping->last + 1) * AMDGPU_GPU_PAGE_SIZE) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_vce.c
680
addr -= mapping->start * AMDGPU_GPU_PAGE_SIZE;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1323
struct amdgpu_bo_va_mapping *mapping;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1411
list_for_each_entry(mapping, &bo_va->invalids, list) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1417
if (!(mapping->flags & AMDGPU_VM_PAGE_READABLE))
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1419
if (!(mapping->flags & AMDGPU_VM_PAGE_WRITEABLE))
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1423
amdgpu_gmc_get_vm_pte(adev, vm, bo, mapping->flags,
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1426
trace_amdgpu_vm_bo_update(mapping);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1429
!uncached, &sync, mapping->start,
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1430
mapping->last, update_flags,
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1431
mapping->offset, vram_base, mem,
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1457
list_for_each_entry(mapping, &bo_va->valids, list)
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1458
trace_amdgpu_vm_bo_mapping(mapping);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1562
struct amdgpu_bo_va_mapping *mapping,
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1565
if (mapping->flags & AMDGPU_VM_PAGE_PRT)
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1567
kfree(mapping);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1610
struct amdgpu_bo_va_mapping *mapping;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1627
mapping = list_first_entry(&vm->freed,
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1629
list_del(&mapping->list);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1632
&sync, mapping->start, mapping->last,
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1634
amdgpu_vm_free_mapping(adev, vm, mapping, f);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1835
struct amdgpu_bo_va_mapping *mapping)
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1840
mapping->bo_va = bo_va;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1841
list_add(&mapping->list, &bo_va->invalids);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1842
amdgpu_vm_it_insert(mapping, &vm->va);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1844
if (mapping->flags & AMDGPU_VM_PAGE_PRT)
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1850
trace_amdgpu_vm_bo_map(bo_va, mapping);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1906
struct amdgpu_bo_va_mapping *mapping, *tmp;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1928
mapping = kmalloc(sizeof(*mapping), GFP_KERNEL);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1929
if (!mapping)
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1932
mapping->start = saddr;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1933
mapping->last = eaddr;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1934
mapping->offset = offset;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1935
mapping->flags = flags;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1937
amdgpu_vm_bo_insert_map(adev, bo_va, mapping);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1965
struct amdgpu_bo_va_mapping *mapping;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1975
mapping = kmalloc(sizeof(*mapping), GFP_KERNEL);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1976
if (!mapping)
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1981
kfree(mapping);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1988
mapping->start = saddr;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1989
mapping->last = eaddr;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1990
mapping->offset = offset;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1991
mapping->flags = flags;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
1993
amdgpu_vm_bo_insert_map(adev, bo_va, mapping);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2016
struct amdgpu_bo_va_mapping *mapping;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2022
list_for_each_entry(mapping, &bo_va->valids, list) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2023
if (mapping->start == saddr)
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2027
if (&mapping->list == &bo_va->valids) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2030
list_for_each_entry(mapping, &bo_va->invalids, list) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2031
if (mapping->start == saddr)
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2035
if (&mapping->list == &bo_va->invalids)
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2039
list_del(&mapping->list);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2040
amdgpu_vm_it_remove(mapping, &vm->va);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2041
mapping->bo_va = NULL;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2042
trace_amdgpu_vm_bo_unmap(bo_va, mapping);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2045
list_add(&mapping->list, &vm->freed);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2047
amdgpu_vm_free_mapping(adev, vm, mapping,
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2201
struct amdgpu_bo_va_mapping *mapping;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2206
for (mapping = amdgpu_vm_it_iter_first(&vm->va, 0, U64_MAX); mapping;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2207
mapping = amdgpu_vm_it_iter_next(mapping, 0, U64_MAX)) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2208
if (mapping->bo_va && mapping->bo_va->base.bo) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2211
bo = mapping->bo_va->base.bo;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2217
trace_amdgpu_vm_bo_cs(mapping);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2234
struct amdgpu_bo_va_mapping *mapping, *next;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2261
list_for_each_entry_safe(mapping, next, &bo_va->valids, list) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2262
list_del(&mapping->list);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2263
amdgpu_vm_it_remove(mapping, &vm->va);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2264
mapping->bo_va = NULL;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2265
trace_amdgpu_vm_bo_unmap(bo_va, mapping);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2266
list_add(&mapping->list, &vm->freed);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2268
list_for_each_entry_safe(mapping, next, &bo_va->invalids, list) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2269
list_del(&mapping->list);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2270
amdgpu_vm_it_remove(mapping, &vm->va);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2271
amdgpu_vm_free_mapping(adev, vm, mapping,
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2841
struct amdgpu_bo_va_mapping *mapping, *tmp;
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2864
list_for_each_entry_safe(mapping, tmp, &vm->freed, list) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2865
if (mapping->flags & AMDGPU_VM_PAGE_PRT && prt_fini_needed) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2870
list_del(&mapping->list);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2871
amdgpu_vm_free_mapping(adev, vm, mapping, NULL);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2884
rbtree_postorder_for_each_entry_safe(mapping, tmp,
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2889
list_del(&mapping->list);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.c
2890
kfree(mapping);
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.h
706
#define amdgpu_vm_bo_va_for_each_valid_mapping(bo_va, mapping) \
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.h
707
list_for_each_entry(mapping, &(bo_va)->valids, list)
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.h
708
#define amdgpu_vm_bo_va_for_each_invalid_mapping(bo_va, mapping) \
sys/dev/pci/drm/amd/amdgpu/amdgpu_vm.h
709
list_for_each_entry(mapping, &(bo_va)->invalids, list)
sys/dev/pci/drm/amd/amdgpu/vcn_v1_0.c
2063
struct amdgpu_bo_va_mapping *mapping;
sys/dev/pci/drm/amd/amdgpu/vcn_v1_0.c
2073
mapping = amdgpu_vm_bo_lookup_mapping(vm, addr/AMDGPU_GPU_PAGE_SIZE);
sys/dev/pci/drm/amd/amdgpu/vcn_v1_0.c
2074
if (!mapping || !mapping->bo_va || !mapping->bo_va->base.bo)
sys/dev/pci/drm/amd/amdgpu/vcn_v1_0.c
2077
bo = mapping->bo_va->base.bo;
sys/dev/pci/drm/amd/amdkfd/kfd_queue.c
199
struct amdgpu_bo_va_mapping *mapping;
sys/dev/pci/drm/amd/amdkfd/kfd_queue.c
206
mapping = amdgpu_vm_bo_lookup_mapping(vm, user_addr);
sys/dev/pci/drm/amd/amdkfd/kfd_queue.c
207
if (!mapping)
sys/dev/pci/drm/amd/amdkfd/kfd_queue.c
210
if (user_addr != mapping->start ||
sys/dev/pci/drm/amd/amdkfd/kfd_queue.c
211
(size != 0 && user_addr + size - 1 != mapping->last)) {
sys/dev/pci/drm/amd/amdkfd/kfd_queue.c
213
expected_size, mapping->start << AMDGPU_GPU_PAGE_SHIFT,
sys/dev/pci/drm/amd/amdkfd/kfd_queue.c
214
(mapping->last - mapping->start + 1) << AMDGPU_GPU_PAGE_SHIFT);
sys/dev/pci/drm/amd/amdkfd/kfd_queue.c
218
*pbo = amdgpu_bo_ref(mapping->bo_va->base.bo);
sys/dev/pci/drm/amd/amdkfd/kfd_queue.c
219
mapping->bo_va->queue_refcount++;
sys/dev/pci/drm/amd/amdkfd/kfd_svm.c
2818
struct amdgpu_bo_va_mapping *mapping;
sys/dev/pci/drm/amd/amdkfd/kfd_svm.c
2839
mapping = container_of((struct rb_node *)node,
sys/dev/pci/drm/amd/amdkfd/kfd_svm.c
2841
bo = mapping->bo_va->base.bo;
sys/dev/pci/drm/amd/amdkfd/kfd_svm.c
3367
struct amdgpu_bo_va_mapping *mapping;
sys/dev/pci/drm/amd/amdkfd/kfd_svm.c
3387
mapping = container_of((struct rb_node *)node,
sys/dev/pci/drm/amd/amdkfd/kfd_svm.c
3390
*bo_s = mapping->start;
sys/dev/pci/drm/amd/amdkfd/kfd_svm.c
3391
*bo_l = mapping->last;
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
100
if (mapping->disp_cfg_to_stream_id_valid[i] && mapping->disp_cfg_to_stream_id[i] == stream_id)
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
1015
struct dml2_dml_to_dc_pipe_mapping *mapping,
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
1021
populate_odm_factors(ctx, disp_cfg, mapping, state, ctx->pipe_combine_scratch.odm_factors);
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
1023
populate_mpc_factors_for_stream(ctx, disp_cfg, mapping, state,
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
1035
bool dml2_map_dc_pipes(struct dml2_context *ctx, struct dc_state *state, const struct dml_display_cfg_st *disp_cfg, struct dml2_dml_to_dc_pipe_mapping *mapping, const struct dc_state *existing_state)
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
1052
ctx, state, disp_cfg, mapping, existing_state);
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
1079
stream_disp_cfg_index = find_disp_cfg_idx_by_stream_id(mapping, stream_id);
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
1121
plane_disp_cfg_index = find_disp_cfg_idx_by_plane_id(mapping, plane_id);
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
1154
if (!validate_pipe_assignment(ctx, state, disp_cfg, mapping))
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
179
static bool validate_pipe_assignment(const struct dml2_context *ctx, const struct dc_state *state, const struct dml_display_cfg_st *disp_cfg, const struct dml2_dml_to_dc_pipe_mapping *mapping)
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
794
struct dml2_dml_to_dc_pipe_mapping *mapping,
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
806
cfg_idx = find_disp_cfg_idx_by_plane_id(mapping, plane_id);
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
82
static int find_disp_cfg_idx_by_plane_id(struct dml2_dml_to_dc_pipe_mapping *mapping, unsigned int plane_id)
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
834
cfg_idx = find_disp_cfg_idx_by_plane_id(mapping, plane_id);
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
852
struct dml2_dml_to_dc_pipe_mapping *mapping,
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
859
mapping, stream->stream_id);
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
87
if (mapping->disp_cfg_to_plane_id_valid[i] && mapping->disp_cfg_to_plane_id[i] == plane_id)
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
881
mapping, main_stream->stream_id);
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
884
mapping, stream->stream_id);
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
923
struct dml2_dml_to_dc_pipe_mapping *mapping,
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
935
get_target_mpc_factor(ctx, state, disp_cfg, mapping, status, state->streams[stream_idx], i) : 1;
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
941
struct dml2_dml_to_dc_pipe_mapping *mapping,
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
95
static int find_disp_cfg_idx_by_stream_id(struct dml2_dml_to_dc_pipe_mapping *mapping, unsigned int stream_id)
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.c
950
ctx, state, disp_cfg, mapping, state->streams[i]);
sys/dev/pci/drm/amd/display/dc/dml2/dml2_dc_resource_mgmt.h
50
bool dml2_map_dc_pipes(struct dml2_context *ctx, struct dc_state *state, const struct dml_display_cfg_st *disp_cfg, struct dml2_dml_to_dc_pipe_mapping *mapping, const struct dc_state *existing_state);
sys/dev/pci/drm/amd/display/include/grph_object_ctrl_defs.h
246
} mapping;
sys/dev/pci/drm/amd/pm/swsmu/smu_cmn.c
519
struct cmn2asic_mapping mapping;
sys/dev/pci/drm/amd/pm/swsmu/smu_cmn.c
542
mapping = smu->clock_map[index];
sys/dev/pci/drm/amd/pm/swsmu/smu_cmn.c
543
if (!mapping.valid_mapping)
sys/dev/pci/drm/amd/pm/swsmu/smu_cmn.c
546
return mapping.map_to;
sys/dev/pci/drm/amd/pm/swsmu/smu_cmn.c
553
mapping = smu->feature_map[index];
sys/dev/pci/drm/amd/pm/swsmu/smu_cmn.c
554
if (!mapping.valid_mapping)
sys/dev/pci/drm/amd/pm/swsmu/smu_cmn.c
557
return mapping.map_to;
sys/dev/pci/drm/amd/pm/swsmu/smu_cmn.c
564
mapping = smu->table_map[index];
sys/dev/pci/drm/amd/pm/swsmu/smu_cmn.c
565
if (!mapping.valid_mapping)
sys/dev/pci/drm/amd/pm/swsmu/smu_cmn.c
568
return mapping.map_to;
sys/dev/pci/drm/amd/pm/swsmu/smu_cmn.c
575
mapping = smu->pwr_src_map[index];
sys/dev/pci/drm/amd/pm/swsmu/smu_cmn.c
576
if (!mapping.valid_mapping)
sys/dev/pci/drm/amd/pm/swsmu/smu_cmn.c
579
return mapping.map_to;
sys/dev/pci/drm/amd/pm/swsmu/smu_cmn.c
586
mapping = smu->workload_map[index];
sys/dev/pci/drm/amd/pm/swsmu/smu_cmn.c
587
if (!mapping.valid_mapping)
sys/dev/pci/drm/amd/pm/swsmu/smu_cmn.c
590
return mapping.map_to;
sys/dev/pci/drm/dma-resv.c
796
struct address_space mapping;
sys/dev/pci/drm/dma-resv.c
803
address_space_init_once(&mapping);
sys/dev/pci/drm/dma-resv.c
812
i_mmap_lock_write(&mapping);
sys/dev/pci/drm/dma-resv.c
813
i_mmap_unlock_write(&mapping);
sys/dev/pci/drm/drm_gem.c
815
struct address_space *mapping;
sys/dev/pci/drm/drm_gem.c
825
mapping = obj->filp->f_mapping;
sys/dev/pci/drm/drm_gem.c
839
mapping_set_unevictable(mapping);
sys/dev/pci/drm/drm_gem.c
844
folio = shmem_read_folio_gfp(mapping, i,
sys/dev/pci/drm/drm_gem.c
845
mapping_gfp_mask(mapping));
sys/dev/pci/drm/drm_gem.c
857
BUG_ON(mapping_gfp_constraint(mapping, __GFP_DMA32) &&
sys/dev/pci/drm/drm_gem.c
864
mapping_clear_unevictable(mapping);
sys/dev/pci/drm/drm_gem.c
895
struct address_space *mapping;
sys/dev/pci/drm/drm_gem.c
898
mapping = file_inode(obj->filp)->i_mapping;
sys/dev/pci/drm/drm_gem.c
899
mapping_clear_unevictable(mapping);
sys/dev/pci/drm/i915/display/intel_bios.c
1237
struct sdvo_device_mapping *mapping;
sys/dev/pci/drm/i915/display/intel_bios.c
1260
mapping = &display->vbt.sdvo_mappings[child->dvo_port - 1];
sys/dev/pci/drm/i915/display/intel_bios.c
1261
if (!mapping->initialized) {
sys/dev/pci/drm/i915/display/intel_bios.c
1262
mapping->dvo_port = child->dvo_port;
sys/dev/pci/drm/i915/display/intel_bios.c
1263
mapping->target_addr = child->target_addr;
sys/dev/pci/drm/i915/display/intel_bios.c
1264
mapping->dvo_wiring = child->dvo_wiring;
sys/dev/pci/drm/i915/display/intel_bios.c
1265
mapping->ddc_pin = child->ddc_pin;
sys/dev/pci/drm/i915/display/intel_bios.c
1266
mapping->i2c_pin = child->i2c_pin;
sys/dev/pci/drm/i915/display/intel_bios.c
1267
mapping->initialized = 1;
sys/dev/pci/drm/i915/display/intel_bios.c
1270
mapping->dvo_port, mapping->target_addr,
sys/dev/pci/drm/i915/display/intel_bios.c
1271
mapping->dvo_wiring, mapping->ddc_pin,
sys/dev/pci/drm/i915/display/intel_bios.c
1272
mapping->i2c_pin);
sys/dev/pci/drm/i915/display/intel_bios.c
2185
static const u8 mapping[] = { 1, 3, 7 }; /* See VBT spec */
sys/dev/pci/drm/i915/display/intel_bios.c
2187
if (val >= ARRAY_SIZE(mapping)) {
sys/dev/pci/drm/i915/display/intel_bios.c
2192
return mapping[val];
sys/dev/pci/drm/i915/display/intel_sdvo.c
2619
const struct sdvo_device_mapping *mapping;
sys/dev/pci/drm/i915/display/intel_sdvo.c
2623
mapping = &display->vbt.sdvo_mappings[0];
sys/dev/pci/drm/i915/display/intel_sdvo.c
2625
mapping = &display->vbt.sdvo_mappings[1];
sys/dev/pci/drm/i915/display/intel_sdvo.c
2627
if (mapping->initialized)
sys/dev/pci/drm/i915/display/intel_sdvo.c
2628
ddc_bus = (mapping->ddc_pin & 0xf0) >> 4;
sys/dev/pci/drm/i915/display/intel_sdvo.c
2642
const struct sdvo_device_mapping *mapping;
sys/dev/pci/drm/i915/display/intel_sdvo.c
2646
mapping = &display->vbt.sdvo_mappings[0];
sys/dev/pci/drm/i915/display/intel_sdvo.c
2648
mapping = &display->vbt.sdvo_mappings[1];
sys/dev/pci/drm/i915/display/intel_sdvo.c
2650
if (mapping->initialized &&
sys/dev/pci/drm/i915/display/intel_sdvo.c
2651
intel_gmbus_is_valid_pin(display, mapping->i2c_pin))
sys/dev/pci/drm/i915/display/intel_sdvo.c
2652
pin = mapping->i2c_pin;
sys/dev/pci/drm/i915/gem/i915_gem_object.h
849
struct address_space *mapping,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
851
void shmem_sg_free_table(struct sg_table *st, struct address_space *mapping,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
856
struct address_space *mapping,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
859
void shmem_sg_free_table(struct sg_table *st, struct address_space *mapping,
sys/dev/pci/drm/i915/gem/i915_gem_object.h
863
void __shmem_writeback(size_t size, struct address_space *mapping);
sys/dev/pci/drm/i915/gem/i915_gem_object_types.h
669
void *mapping;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
228
if (obj->mm.mapping) {
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
229
unmap_object(obj, page_mask_bits(obj->mm.mapping));
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
230
obj->mm.mapping = NULL;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
457
ptr = page_unpack_bits(obj->mm.mapping, &has_type);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
551
ptr = page_unpack_bits(obj->mm.mapping, &has_type);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
560
ptr = obj->mm.mapping = NULL;
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
579
obj->mm.mapping = page_pack_bits(ptr, type);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
618
ptr = page_unpack_bits(obj->mm.mapping, &has_type);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
631
GEM_BUG_ON(!obj->mm.mapping);
sys/dev/pci/drm/i915/gem/i915_gem_pages.c
639
unmap_object(obj, page_mask_bits(fetch_and_zero(&obj->mm.mapping)));
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
152
struct address_space *mapping = obj->base.filp->f_mapping;
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
161
page = shmem_read_mapping_page(mapping, i);
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
24
struct address_space *mapping = obj->base.filp->f_mapping;
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
315
if (obj->mm.mapping || i915_gem_object_has_pinned_pages(obj))
sys/dev/pci/drm/i915/gem/i915_gem_phys.c
90
page = shmem_read_mapping_page(mapping, i);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
114
mapping_set_unevictable(mapping);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
115
noreclaim = mapping_gfp_constraint(mapping, ~__GFP_RECLAIM);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
131
folio = shmem_read_folio_gfp(mapping, i, gfp);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
153
gfp = mapping_gfp_mask(mapping);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
228
shmem_sg_free_table(st, mapping, false, false, obj);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
230
mapping_clear_unevictable(mapping);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
254
struct address_space *mapping = obj->base.filp->f_mapping;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
274
ret = shmem_sg_alloc_table(i915, st, obj->base.size, mem, mapping,
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
292
shmem_sg_free_table(st, mapping, false, false);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
321
shmem_sg_free_table(st, mapping, false, false);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
35
void shmem_sg_free_table(struct sg_table *st, struct address_space *mapping,
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
366
void __shmem_writeback(size_t size, struct address_space *mapping)
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
385
while ((folio = writeback_iter(mapping, &wbc, folio, &error))) {
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
45
mapping_clear_unevictable(mapping);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
676
struct address_space *mapping;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
697
mapping = obj->base.filp->f_mapping;
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
698
mapping_set_gfp_mask(mapping, mask);
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
699
GEM_BUG_ON(!(mapping_gfp_mask(mapping) & __GFP_RECLAIM));
sys/dev/pci/drm/i915/gem/i915_gem_shmem.c
80
struct address_space *mapping,
sys/dev/pci/drm/i915/gem/i915_gem_shrinker.c
201
!is_vmalloc_addr(obj->mm.mapping))
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
201
struct address_space *mapping;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
210
mapping = filp->f_mapping;
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
211
mapping_set_gfp_mask(mapping, mask);
sys/dev/pci/drm/i915/gem/i915_gem_ttm.c
212
GEM_BUG_ON(!(mapping_gfp_mask(mapping) & __GFP_RECLAIM));
sys/dev/pci/drm/i915/gt/intel_gtt.c
322
return page_unpack_bits(p->mm.mapping, &type);
sys/dev/pci/drm/i915/gt/selftest_timeline.c
841
page_unmask_bits(tl->hwsp_ggtt->obj->mm.mapping));
sys/dev/pci/drm/i915/gt/selftest_tlb.c
184
cs = page_mask_bits(batch->mm.mapping);
sys/dev/pci/drm/i915/gt/selftest_tlb.c
27
memset64(page_mask_bits(vma->obj->mm.mapping) +
sys/dev/pci/drm/i915/i915_cmd_parser.c
1556
cmd = page_mask_bits(shadow->obj->mm.mapping);
sys/dev/pci/drm/i915/i915_debugfs.c
97
return obj->mm.mapping ? 'M' : ' ';
sys/dev/pci/drm/i915/i915_gem.c
278
gtt_user_read(struct io_mapping *mapping,
sys/dev/pci/drm/i915/i915_gem.c
286
vaddr = io_mapping_map_atomic_wc(mapping, base);
sys/dev/pci/drm/i915/i915_gem.c
292
vaddr = io_mapping_map_wc(mapping, base, PAGE_SIZE);
sys/dev/pci/drm/i915/i915_gem.c
513
ggtt_write(struct io_mapping *mapping,
sys/dev/pci/drm/i915/i915_gem.c
521
vaddr = io_mapping_map_atomic_wc(mapping, base);
sys/dev/pci/drm/i915/i915_gem.c
526
vaddr = io_mapping_map_wc(mapping, base, PAGE_SIZE);
sys/dev/pci/drm/include/drm/ttm/ttm_device.h
301
struct device *dev, struct address_space *mapping,
sys/dev/pci/drm/ttm/ttm_backup.c
112
struct address_space *mapping = backup->f_mapping;
sys/dev/pci/drm/ttm/ttm_backup.c
117
to_folio = shmem_read_folio_gfp(mapping, idx, alloc_gfp);
sys/dev/pci/drm/ttm/ttm_backup.c
62
struct address_space *mapping = backup->f_mapping;
sys/dev/pci/drm/ttm/ttm_backup.c
66
from_folio = shmem_read_folio(mapping, idx);
sys/dev/pci/drm/ttm/ttm_device.c
208
struct device *dev, struct address_space *mapping,
sys/dev/pci/drm/ttm/ttm_device.c
243
bdev->dev_mapping = mapping;
usr.bin/tic/dump_entry.c
1406
one_one_mapping(const char *mapping)
usr.bin/tic/dump_entry.c
1410
if (VALID_STRING(mapping)) {
usr.bin/tic/dump_entry.c
1412
while (mapping[n] != '\0' && mapping[n + 1] != '\0') {
usr.bin/tic/dump_entry.c
1413
if (isLine(mapping[n]) &&
usr.bin/tic/dump_entry.c
1414
mapping[n] != mapping[n + 1]) {
usr.sbin/ldomctl/config.c
2336
uint64_t mapping[3];
usr.sbin/ldomctl/config.c
2370
memcpy(mapping, buf, sizeof(mapping));
usr.sbin/ldomctl/config.c
2371
mapping[1] = device->num_msi_eqs;
usr.sbin/ldomctl/config.c
2373
(void *)mapping, sizeof(mapping));
usr.sbin/smtpd/aliases.c
119
ret = table_lookup(mapping, K_ALIAS, buf, &lk);
usr.sbin/smtpd/aliases.c
129
ret = table_lookup(mapping, K_ALIAS, buf, &lk);
usr.sbin/smtpd/aliases.c
140
ret = table_lookup(mapping, K_ALIAS, buf, &lk);
usr.sbin/smtpd/aliases.c
150
ret = table_lookup(mapping, K_ALIAS, buf, &lk);
usr.sbin/smtpd/aliases.c
163
ret = table_lookup(mapping, K_ALIAS, buf, &lk);
usr.sbin/smtpd/aliases.c
170
ret = table_lookup(mapping, K_ALIAS, "@", &lk);
usr.sbin/smtpd/aliases.c
37
struct table *mapping = NULL;
usr.sbin/smtpd/aliases.c
41
mapping = table_find(env, dsp->u.local.table_alias);
usr.sbin/smtpd/aliases.c
48
ret = table_lookup(mapping, K_ALIAS, buf, &lk);
usr.sbin/smtpd/aliases.c
57
ret = table_lookup(mapping, K_ALIAS, buf, &lk);
usr.sbin/smtpd/aliases.c
93
struct table *mapping = NULL;
usr.sbin/smtpd/aliases.c
96
mapping = table_find(env, dsp->u.local.table_virtual);