Symbol: mem_id
arch/um/drivers/mconsole_kern.c
413
.id = mem_id,
drivers/accel/habanalabs/common/command_buffer.c
260
.mem_id = HL_MMAP_TYPE_CB,
drivers/accel/habanalabs/common/habanalabs.h
935
u64 mem_id;
drivers/accel/habanalabs/common/memory.c
2154
.mem_id = HL_MMAP_TYPE_TS_BUFF,
drivers/accel/habanalabs/common/memory_mgr.c
171
buf->handle = (((u64)rc | buf->behavior->mem_id) << PAGE_SHIFT);
drivers/accel/habanalabs/common/memory_mgr.c
324
static void hl_mem_mgr_fini_stats_inc(u64 mem_id, struct hl_mem_mgr_fini_stats *stats)
drivers/accel/habanalabs/common/memory_mgr.c
329
switch (mem_id) {
drivers/accel/habanalabs/common/memory_mgr.c
355
u64 mem_id;
drivers/accel/habanalabs/common/memory_mgr.c
364
mem_id = buf->behavior->mem_id;
drivers/accel/habanalabs/common/memory_mgr.c
369
hl_mem_mgr_fini_stats_inc(mem_id, stats);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c
791
s8 mem_id = KFD_XCP_MEM_ID(adev, xcp_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c
794
if (adev->gmc.num_mem_partitions && xcp_id >= 0 && mem_id >= 0) {
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c
806
tmp = adev->gmc.mem_partitions[mem_id].size;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
366
(adev)->xcp_mgr->xcp[(xcp_id)].mem_id : -1)
drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
120
int8_t mem_id = KFD_XCP_MEM_ID(adev, abo->xcp_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
122
if (adev->gmc.mem_partitions && mem_id >= 0) {
drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
123
places[c].fpfn = adev->gmc.mem_partitions[mem_id].range.fpfn;
drivers/gpu/drm/amd/amdgpu/amdgpu_object.c
128
places[c].lpfn = adev->gmc.mem_partitions[mem_id].range.lpfn + 1;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
448
int8_t mem_id;
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
146
uint8_t mem_id;
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
174
xcp_mgr, &xcp_mgr->xcp[i], &mem_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
178
xcp_mgr->xcp[i].mem_id = mem_id;
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
460
fpriv->vm.mem_id = fpriv->xcp_id == AMDGPU_XCP_NO_PARTITION ? -1 :
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.c
461
adev->xcp_mgr->xcp[fpriv->xcp_id].mem_id;
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.h
107
uint8_t mem_id;
drivers/gpu/drm/amd/amdgpu/amdgpu_xcp.h
145
struct amdgpu_xcp *xcp, uint8_t *mem_id);
drivers/gpu/drm/amd/amdgpu/aqua_vanjaram.c
441
int xcc_id, uint8_t *mem_id)
drivers/gpu/drm/amd/amdgpu/aqua_vanjaram.c
444
*mem_id = xcc_id / adev->gfx.num_xcc_per_xcp;
drivers/gpu/drm/amd/amdgpu/aqua_vanjaram.c
445
*mem_id /= adev->xcp_mgr->num_xcp_per_mem_partition;
drivers/gpu/drm/amd/amdgpu/aqua_vanjaram.c
451
struct amdgpu_xcp *xcp, uint8_t *mem_id)
drivers/gpu/drm/amd/amdgpu/aqua_vanjaram.c
468
*mem_id = 0;
drivers/gpu/drm/amd/amdgpu/aqua_vanjaram.c
478
return __aqua_vanjaram_get_xcp_mem_id(adev, xcc_id, mem_id);
drivers/gpu/drm/amd/amdgpu/aqua_vanjaram.c
488
*mem_id = i;
drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c
1158
KFD_XCP_MEM_ID(adev, bo->xcp_id) == vm->mem_id);
drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c
1273
if (adev->gmc.is_app_apu && vm->mem_id >= 0) {
drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c
1274
local_node = adev->gmc.mem_partitions[vm->mem_id].numa.node;
drivers/gpu/drm/amd/amdgpu/gmc_v9_0.c
1289
vm->mem_id, local_node, nid);
drivers/gpu/drm/amd/amdgpu/soc_v1_0.c
719
int xcc_id, uint8_t *mem_id)
drivers/gpu/drm/amd/amdgpu/soc_v1_0.c
722
*mem_id = xcc_id / adev->gfx.num_xcc_per_xcp;
drivers/gpu/drm/amd/amdgpu/soc_v1_0.c
723
*mem_id /= adev->xcp_mgr->num_xcp_per_mem_partition;
drivers/gpu/drm/amd/amdgpu/soc_v1_0.c
729
struct amdgpu_xcp *xcp, uint8_t *mem_id)
drivers/gpu/drm/amd/amdgpu/soc_v1_0.c
746
*mem_id = 0;
drivers/gpu/drm/amd/amdgpu/soc_v1_0.c
756
return __soc_v1_0_get_xcp_mem_id(adev, xcc_id, mem_id);
drivers/gpu/drm/amd/amdgpu/soc_v1_0.c
766
*mem_id = i;
drivers/gpu/drm/amd/amdkfd/kfd_device.c
889
node->node_id, node->xcp->mem_id,
drivers/gpu/drm/amd/amdkfd/kfd_svm.c
1274
(!bo_node->xcp || !node->xcp || bo_node->xcp->mem_id == node->xcp->mem_id))
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1008
req = &alive_path->reqs[fail_req->permit->mem_id];
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1117
buf_id = req->permit->mem_id;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1247
buf_id = req->permit->mem_id;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1430
permit->mem_id = i;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
154
if (WARN_ON(!test_bit(permit->mem_id, clt->permits_map)))
drivers/infiniband/ulp/rtrs/rtrs-clt.c
89
WARN_ON(permit->mem_id != bit);
drivers/infiniband/ulp/rtrs/rtrs-clt.c
99
clear_bit_unlock(permit->mem_id, clt->permits_map);
drivers/infiniband/ulp/rtrs/rtrs-clt.c
992
req = &clt_path->reqs[permit->mem_id];
drivers/infiniband/ulp/rtrs/rtrs-clt.h
87
unsigned int mem_id;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_ioctl.h
78
uint32_t mem_id;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2460
if (t.mem_id == MEM_CM)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2462
else if (t.mem_id == MEM_PMRX)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2464
else if (t.mem_id == MEM_PMTX)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
315
if (t->mem_id == MEM_CM)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
317
else if (t->mem_id == MEM_PMRX)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
319
else if (t->mem_id == MEM_PMTX)
drivers/net/ipa/ipa_mem.c
127
static bool ipa_mem_id_valid(struct ipa *ipa, enum ipa_mem_id mem_id)
drivers/net/ipa/ipa_mem.c
131
switch (mem_id) {
drivers/net/ipa/ipa_mem.c
188
static bool ipa_mem_id_required(struct ipa *ipa, enum ipa_mem_id mem_id)
drivers/net/ipa/ipa_mem.c
190
switch (mem_id) {
drivers/net/ipa/ipa_mem.c
222
enum ipa_mem_id mem_id = mem->id;
drivers/net/ipa/ipa_mem.c
227
if (!ipa_mem_id_valid(ipa, mem_id)) {
drivers/net/ipa/ipa_mem.c
228
dev_err(dev, "region id %u not valid\n", mem_id);
drivers/net/ipa/ipa_mem.c
233
dev_err(dev, "empty memory region %u\n", mem_id);
drivers/net/ipa/ipa_mem.c
238
size_multiple = mem_id == IPA_MEM_MODEM ? 4 : 8;
drivers/net/ipa/ipa_mem.c
241
mem_id, size_multiple);
drivers/net/ipa/ipa_mem.c
243
dev_err(dev, "region %u offset not 8-byte aligned\n", mem_id);
drivers/net/ipa/ipa_mem.c
246
mem_id, mem->canary_count);
drivers/net/ipa/ipa_mem.c
247
else if (mem_id == IPA_MEM_END_MARKER && mem->size)
drivers/net/ipa/ipa_mem.c
260
enum ipa_mem_id mem_id;
drivers/net/ipa/ipa_mem.c
283
for_each_clear_bit(mem_id, regions, IPA_MEM_COUNT) {
drivers/net/ipa/ipa_mem.c
284
if (ipa_mem_id_required(ipa, mem_id))
drivers/net/ipa/ipa_mem.c
286
mem_id);
drivers/net/ipa/ipa_mem.c
31
const struct ipa_mem *ipa_mem_find(struct ipa *ipa, enum ipa_mem_id mem_id)
drivers/net/ipa/ipa_mem.c
38
if (mem->id == mem_id)
drivers/net/ipa/ipa_mem.c
47
ipa_mem_zero_region_add(struct gsi_trans *trans, enum ipa_mem_id mem_id)
drivers/net/ipa/ipa_mem.c
50
const struct ipa_mem *mem = ipa_mem_find(ipa, mem_id);
drivers/net/ipa/ipa_mem.h
98
const struct ipa_mem *ipa_mem_find(struct ipa *ipa, enum ipa_mem_id mem_id);
drivers/net/ipa/ipa_table.c
147
enum ipa_mem_id mem_id;
drivers/net/ipa/ipa_table.c
149
mem_id = filter ? hashed ? ipv6 ? IPA_MEM_V6_FILTER_HASHED
drivers/net/ipa/ipa_table.c
158
return ipa_mem_find(ipa, mem_id);
drivers/soc/ti/pruss.c
112
int pruss_request_mem_region(struct pruss *pruss, enum pruss_mem mem_id,
drivers/soc/ti/pruss.c
115
if (!pruss || !region || mem_id >= PRUSS_MEM_MAX)
drivers/soc/ti/pruss.c
120
if (pruss->mem_in_use[mem_id]) {
drivers/soc/ti/pruss.c
125
*region = pruss->mem_regions[mem_id];
drivers/soc/ti/pruss.c
126
pruss->mem_in_use[mem_id] = region;
include/linux/pruss_driver.h
110
int pruss_request_mem_region(struct pruss *pruss, enum pruss_mem mem_id,
include/linux/pruss_driver.h
132
enum pruss_mem mem_id,
include/trace/events/xdp.h
312
__field(u32, mem_id)
include/trace/events/xdp.h
319
__entry->mem_id = xa->mem.id;
include/trace/events/xdp.h
325
__entry->mem_id,
include/trace/events/xdp.h
340
__field(u32, mem_id)
include/trace/events/xdp.h
349
__entry->mem_id = xa->mem.id;
include/trace/events/xdp.h
358
__entry->mem_id,
net/core/xdp.c
56
u32 mem_id = *(u32 *)arg->key;
net/core/xdp.c
58
return xa->mem.id != mem_id;