Symbol: tcm
arch/arm/include/asm/vmlinux.lds.h
163
*(.tcm.text) \
arch/arm/include/asm/vmlinux.lds.h
164
*(.tcm.rodata) \
arch/arm/include/asm/vmlinux.lds.h
173
*(.tcm.data) \
drivers/gpu/drm/omapdrm/omap_dmm_priv.h
122
struct tcm *tcm;
drivers/gpu/drm/omapdrm/omap_dmm_priv.h
133
struct tcm *tcm;
drivers/gpu/drm/omapdrm/omap_dmm_priv.h
179
struct tcm **tcm;
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
1106
if (block->area.tcm == omap_dmm->tcm[lut_idx]) {
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
1172
.tcm = NULL,
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
1179
area.tcm = omap_dmm->tcm[i];
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
307
static struct dmm_txn *dmm_txn_init(struct dmm *dmm, struct tcm *tcm)
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
333
engine->tcm = tcm;
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
365
pat->area.y0 += engine->tcm->y_offset;
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
366
pat->area.y1 += engine->tcm->y_offset;
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
37
static struct tcm *containers[TILFMT_NFORMATS];
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
370
.lut_id = engine->tcm->lut_id,
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
483
txn = dmm_txn_init(omap_dmm, area->tcm);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
602
if (block->area.tcm)
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
749
if (omap_dmm->tcm && omap_dmm->tcm[i])
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
750
omap_dmm->tcm[i]->deinit(omap_dmm->tcm[i]);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
751
kfree(omap_dmm->tcm);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
907
omap_dmm->tcm = kzalloc_objs(*omap_dmm->tcm, omap_dmm->num_lut);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
908
if (!omap_dmm->tcm) {
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
918
omap_dmm->tcm[i] = sita_init(omap_dmm->container_width,
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
921
if (!omap_dmm->tcm[i]) {
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
927
omap_dmm->tcm[i]->lut_id = i;
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
933
containers[TILFMT_8BIT] = omap_dmm->tcm[0];
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
934
containers[TILFMT_16BIT] = omap_dmm->tcm[0];
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
935
containers[TILFMT_32BIT] = omap_dmm->tcm[0];
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
941
containers[TILFMT_PAGE] = omap_dmm->tcm[1];
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
942
omap_dmm->tcm[1]->y_offset = OMAP5_LUT_OFFSET;
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
943
omap_dmm->tcm[1]->lut_id = 0;
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
945
containers[TILFMT_PAGE] = omap_dmm->tcm[0];
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
949
.tcm = NULL,
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
974
area.tcm = omap_dmm->tcm[i];
drivers/gpu/drm/omapdrm/tcm-sita.c
154
static s32 sita_reserve_1d(struct tcm *tcm, u32 num_slots,
drivers/gpu/drm/omapdrm/tcm-sita.c
160
spin_lock(&(tcm->lock));
drivers/gpu/drm/omapdrm/tcm-sita.c
161
ret = r2l_b2t_1d(num_slots, &pos, tcm->bitmap, tcm->map_size);
drivers/gpu/drm/omapdrm/tcm-sita.c
163
area->p0.x = pos % tcm->width;
drivers/gpu/drm/omapdrm/tcm-sita.c
164
area->p0.y = pos / tcm->width;
drivers/gpu/drm/omapdrm/tcm-sita.c
165
area->p1.x = (pos + num_slots - 1) % tcm->width;
drivers/gpu/drm/omapdrm/tcm-sita.c
166
area->p1.y = (pos + num_slots - 1) / tcm->width;
drivers/gpu/drm/omapdrm/tcm-sita.c
168
spin_unlock(&(tcm->lock));
drivers/gpu/drm/omapdrm/tcm-sita.c
173
static s32 sita_reserve_2d(struct tcm *tcm, u16 h, u16 w, u16 align,
drivers/gpu/drm/omapdrm/tcm-sita.c
180
spin_lock(&(tcm->lock));
drivers/gpu/drm/omapdrm/tcm-sita.c
181
ret = l2r_t2b(w, h, align, offset, &pos, slot_bytes, tcm->bitmap,
drivers/gpu/drm/omapdrm/tcm-sita.c
182
tcm->map_size, tcm->width);
drivers/gpu/drm/omapdrm/tcm-sita.c
185
area->p0.x = pos % tcm->width;
drivers/gpu/drm/omapdrm/tcm-sita.c
186
area->p0.y = pos / tcm->width;
drivers/gpu/drm/omapdrm/tcm-sita.c
190
spin_unlock(&(tcm->lock));
drivers/gpu/drm/omapdrm/tcm-sita.c
195
static void sita_deinit(struct tcm *tcm)
drivers/gpu/drm/omapdrm/tcm-sita.c
197
kfree(tcm);
drivers/gpu/drm/omapdrm/tcm-sita.c
200
static s32 sita_free(struct tcm *tcm, struct tcm_area *area)
drivers/gpu/drm/omapdrm/tcm-sita.c
205
pos = area->p0.x + area->p0.y * tcm->width;
drivers/gpu/drm/omapdrm/tcm-sita.c
210
w = area->p1.x + area->p1.y * tcm->width - pos + 1;
drivers/gpu/drm/omapdrm/tcm-sita.c
214
spin_lock(&(tcm->lock));
drivers/gpu/drm/omapdrm/tcm-sita.c
215
free_slots(pos, w, h, tcm->bitmap, tcm->width);
drivers/gpu/drm/omapdrm/tcm-sita.c
216
spin_unlock(&(tcm->lock));
drivers/gpu/drm/omapdrm/tcm-sita.c
220
struct tcm *sita_init(u16 width, u16 height)
drivers/gpu/drm/omapdrm/tcm-sita.c
222
struct tcm *tcm;
drivers/gpu/drm/omapdrm/tcm-sita.c
228
tcm = kzalloc(sizeof(*tcm) + map_size, GFP_KERNEL);
drivers/gpu/drm/omapdrm/tcm-sita.c
229
if (!tcm)
drivers/gpu/drm/omapdrm/tcm-sita.c
233
tcm->height = height;
drivers/gpu/drm/omapdrm/tcm-sita.c
234
tcm->width = width;
drivers/gpu/drm/omapdrm/tcm-sita.c
235
tcm->reserve_2d = sita_reserve_2d;
drivers/gpu/drm/omapdrm/tcm-sita.c
236
tcm->reserve_1d = sita_reserve_1d;
drivers/gpu/drm/omapdrm/tcm-sita.c
237
tcm->free = sita_free;
drivers/gpu/drm/omapdrm/tcm-sita.c
238
tcm->deinit = sita_deinit;
drivers/gpu/drm/omapdrm/tcm-sita.c
240
spin_lock_init(&tcm->lock);
drivers/gpu/drm/omapdrm/tcm-sita.c
241
tcm->bitmap = (unsigned long *)(tcm + 1);
drivers/gpu/drm/omapdrm/tcm-sita.c
242
bitmap_clear(tcm->bitmap, 0, width*height);
drivers/gpu/drm/omapdrm/tcm-sita.c
244
tcm->map_size = width*height;
drivers/gpu/drm/omapdrm/tcm-sita.c
246
return tcm;
drivers/gpu/drm/omapdrm/tcm.h
104
static inline void tcm_deinit(struct tcm *tcm)
drivers/gpu/drm/omapdrm/tcm.h
106
if (tcm)
drivers/gpu/drm/omapdrm/tcm.h
107
tcm->deinit(tcm);
drivers/gpu/drm/omapdrm/tcm.h
131
static inline s32 tcm_reserve_2d(struct tcm *tcm, u16 width, u16 height,
drivers/gpu/drm/omapdrm/tcm.h
136
s32 res = tcm == NULL ? -ENODEV :
drivers/gpu/drm/omapdrm/tcm.h
140
(height > tcm->height || width > tcm->width) ? -ENOMEM : 0;
drivers/gpu/drm/omapdrm/tcm.h
144
res = tcm->reserve_2d(tcm, height, width, align, offset,
drivers/gpu/drm/omapdrm/tcm.h
146
area->tcm = res ? NULL : tcm;
drivers/gpu/drm/omapdrm/tcm.h
165
static inline s32 tcm_reserve_1d(struct tcm *tcm, u32 slots,
drivers/gpu/drm/omapdrm/tcm.h
169
s32 res = tcm == NULL ? -ENODEV :
drivers/gpu/drm/omapdrm/tcm.h
171
slots > (tcm->width * (u32) tcm->height) ? -ENOMEM : 0;
drivers/gpu/drm/omapdrm/tcm.h
175
res = tcm->reserve_1d(tcm, slots, area);
drivers/gpu/drm/omapdrm/tcm.h
176
area->tcm = res ? NULL : tcm;
drivers/gpu/drm/omapdrm/tcm.h
199
if (area && area->tcm) {
drivers/gpu/drm/omapdrm/tcm.h
200
res = area->tcm->free(area->tcm, area);
drivers/gpu/drm/omapdrm/tcm.h
202
area->tcm = NULL;
drivers/gpu/drm/omapdrm/tcm.h
227
if (slice->tcm && !slice->is2d &&
drivers/gpu/drm/omapdrm/tcm.h
229
(slice->p0.x || (slice->p1.x != slice->tcm->width - 1))) {
drivers/gpu/drm/omapdrm/tcm.h
231
slice->p1.x = slice->tcm->width - 1;
drivers/gpu/drm/omapdrm/tcm.h
238
parent->tcm = NULL;
drivers/gpu/drm/omapdrm/tcm.h
245
return area && area->tcm &&
drivers/gpu/drm/omapdrm/tcm.h
247
area->p1.x < area->tcm->width &&
drivers/gpu/drm/omapdrm/tcm.h
248
area->p1.y < area->tcm->height &&
drivers/gpu/drm/omapdrm/tcm.h
252
area->p0.x < area->tcm->width &&
drivers/gpu/drm/omapdrm/tcm.h
253
area->p0.x + area->p0.y * area->tcm->width <=
drivers/gpu/drm/omapdrm/tcm.h
254
area->p1.x + area->p1.y * area->tcm->width) ||
drivers/gpu/drm/omapdrm/tcm.h
269
i = p->x + p->y * a->tcm->width;
drivers/gpu/drm/omapdrm/tcm.h
270
return i >= a->p0.x + a->p0.y * a->tcm->width &&
drivers/gpu/drm/omapdrm/tcm.h
271
i <= a->p1.x + a->p1.y * a->tcm->width;
drivers/gpu/drm/omapdrm/tcm.h
293
area->tcm->width;
drivers/gpu/drm/omapdrm/tcm.h
308
a->p1.x = (a->p0.x + num_pg - 1) % a->tcm->width;
drivers/gpu/drm/omapdrm/tcm.h
309
a->p1.y = a->p0.y + ((a->p0.x + num_pg - 1) / a->tcm->width);
drivers/gpu/drm/omapdrm/tcm.h
328
var.tcm; tcm_slice(&safe, &var))
drivers/gpu/drm/omapdrm/tcm.h
40
struct tcm;
drivers/gpu/drm/omapdrm/tcm.h
51
struct tcm *tcm; /* parent */
drivers/gpu/drm/omapdrm/tcm.h
67
s32 (*reserve_2d)(struct tcm *tcm, u16 height, u16 width, u16 align,
drivers/gpu/drm/omapdrm/tcm.h
70
s32 (*reserve_1d)(struct tcm *tcm, u32 slots, struct tcm_area *area);
drivers/gpu/drm/omapdrm/tcm.h
71
s32 (*free)(struct tcm *tcm, struct tcm_area *area);
drivers/gpu/drm/omapdrm/tcm.h
72
void (*deinit)(struct tcm *tcm);
drivers/gpu/drm/omapdrm/tcm.h
91
struct tcm *sita_init(u16 width, u16 height);
drivers/media/platform/mediatek/vpu/mtk_vpu.c
105
void __iomem *tcm;
drivers/media/platform/mediatek/vpu/mtk_vpu.c
468
return (__force void *)(dtcm_dmem_addr + vpu->reg.tcm +
drivers/media/platform/mediatek/vpu/mtk_vpu.c
545
dest = (__force void *)vpu->reg.tcm;
drivers/media/platform/mediatek/vpu/mtk_vpu.c
769
vpu->recv_buf = vpu->reg.tcm + VPU_DTCM_OFFSET;
drivers/media/platform/mediatek/vpu/mtk_vpu.c
825
vpu->reg.tcm = devm_platform_ioremap_resource_byname(pdev, "tcm");
drivers/media/platform/mediatek/vpu/mtk_vpu.c
826
if (IS_ERR((__force void *)vpu->reg.tcm))
drivers/media/platform/mediatek/vpu/mtk_vpu.c
827
return PTR_ERR((__force void *)vpu->reg.tcm);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c
1237
memcpy_fromio(&ringinfo, devinfo->tcm + devinfo->shared.ring_info_addr,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c
1311
memcpy_toio(devinfo->tcm + devinfo->shared.ring_info_addr,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c
1686
memcpy_toio(devinfo->tcm + address, randbuf, BRCMF_RANDOM_SEED_LENGTH);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c
1707
memcpy_toio(devinfo->tcm + devinfo->ci->rambase,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c
1722
memcpy_toio(devinfo->tcm + address, nvram, nvram_len);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c
1738
memcpy_toio(devinfo->tcm + address, &footer,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c
1812
devinfo->tcm = ioremap(bar1_addr, bar1_size);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c
1814
if (!devinfo->regs || !devinfo->tcm) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c
1816
devinfo->tcm);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c
1822
devinfo->tcm, (unsigned long long)bar1_addr,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c
1831
if (devinfo->tcm)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c
1832
iounmap(devinfo->tcm);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c
341
void __iomem *tcm;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c
500
void __iomem *address = devinfo->tcm + mem_offset;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c
509
void __iomem *address = devinfo->tcm + mem_offset;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c
519
void __iomem *address = devinfo->tcm + mem_offset;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c
547
void __iomem *address = devinfo->tcm + mem_offset;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c
557
void __iomem *address = devinfo->tcm + mem_offset;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c
566
void __iomem *addr = devinfo->tcm + devinfo->ci->rambase + mem_offset;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c
576
void __iomem *addr = devinfo->tcm + devinfo->ci->rambase + mem_offset;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/pcie.c
586
void __iomem *address = devinfo->tcm + mem_offset;
drivers/net/wireless/intel/iwlwifi/mvm/coex.c
358
data->primary_load = mvm->tcm.result.load[mvmvif->id];
drivers/net/wireless/intel/iwlwifi/mvm/coex.c
360
data->secondary_load = mvm->tcm.result.load[mvmvif->id];
drivers/net/wireless/intel/iwlwifi/mvm/coex.c
376
data->primary_load = mvm->tcm.result.load[mvmvif->id];
drivers/net/wireless/intel/iwlwifi/mvm/coex.c
378
data->secondary_load = mvm->tcm.result.load[mvmvif->id];
drivers/net/wireless/intel/iwlwifi/mvm/debugfs-vif.c
212
mvm->tcm.result.load[mvmvif->id]);
drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c
1037
struct iwl_mvm_tcm_mac *mdata = &mvm->tcm.data[macid];
drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c
3419
mdata = &mvm->tcm.data[iwl_mvm_vif_from_mac80211(vif)->id];
drivers/net/wireless/intel/iwlwifi/mvm/mvm.h
1143
struct iwl_mvm_tcm tcm;
drivers/net/wireless/intel/iwlwifi/mvm/ops.c
1370
spin_lock_init(&mvm->tcm.lock);
drivers/net/wireless/intel/iwlwifi/mvm/ops.c
1371
INIT_DELAYED_WORK(&mvm->tcm.work, iwl_mvm_tcm_work);
drivers/net/wireless/intel/iwlwifi/mvm/ops.c
1372
mvm->tcm.ts = jiffies;
drivers/net/wireless/intel/iwlwifi/mvm/ops.c
1373
mvm->tcm.ll_ts = jiffies;
drivers/net/wireless/intel/iwlwifi/mvm/ops.c
1374
mvm->tcm.uapsd_nonagg_ts = jiffies;
drivers/net/wireless/intel/iwlwifi/mvm/ops.c
1597
cancel_delayed_work_sync(&mvm->tcm.work);
drivers/net/wireless/intel/iwlwifi/mvm/rx.c
226
if (time_after(jiffies, mvm->tcm.ts + MVM_TCM_PERIOD))
drivers/net/wireless/intel/iwlwifi/mvm/rx.c
227
schedule_delayed_work(&mvm->tcm.work, 0);
drivers/net/wireless/intel/iwlwifi/mvm/rx.c
228
mdata = &mvm->tcm.data[mac];
drivers/net/wireless/intel/iwlwifi/mvm/rx.c
451
if (!mvm->tcm.paused && len >= sizeof(*hdr) &&
drivers/net/wireless/intel/iwlwifi/mvm/rx.c
742
spin_lock(&mvm->tcm.lock);
drivers/net/wireless/intel/iwlwifi/mvm/rx.c
744
struct iwl_mvm_tcm_mac *mdata = &mvm->tcm.data[i];
drivers/net/wireless/intel/iwlwifi/mvm/rx.c
757
spin_unlock(&mvm->tcm.lock);
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
2288
if (!mvm->tcm.paused && len >= sizeof(*hdr) &&
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
2291
time_after(jiffies, mvm->tcm.ts + MVM_TCM_PERIOD))
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
2292
schedule_delayed_work(&mvm->tcm.work, 0);
drivers/net/wireless/intel/iwlwifi/mvm/scan.c
182
return mvm->tcm.result.global_load;
drivers/net/wireless/intel/iwlwifi/mvm/scan.c
188
return mvm->tcm.result.band_load[band];
drivers/net/wireless/intel/iwlwifi/mvm/tx.c
1039
mdata = &mvm->tcm.data[mac];
drivers/net/wireless/intel/iwlwifi/mvm/tx.c
1041
if (mvm->tcm.paused)
drivers/net/wireless/intel/iwlwifi/mvm/tx.c
1044
if (time_after(jiffies, mvm->tcm.ts + MVM_TCM_PERIOD))
drivers/net/wireless/intel/iwlwifi/mvm/tx.c
1045
schedule_delayed_work(&mvm->tcm.work, 0);
drivers/net/wireless/intel/iwlwifi/mvm/tx.c
1060
mdata = &mvm->tcm.data[mac];
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1000
mvm->tcm.result.airtime[mac] = airtime;
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1008
mvm->tcm.result.low_latency[mac] = true;
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1010
mvm->tcm.result.low_latency[mac] = false;
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1017
low_latency |= mvm->tcm.result.low_latency[mac];
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1019
if (!mvm->tcm.result.low_latency[mac] && handle_uapsd)
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1030
mvm->tcm.result.global_load = load;
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1034
mvm->tcm.result.band_load[i] = band_load;
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1070
time_after(ts, mvm->tcm.uapsd_nonagg_ts +
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1073
spin_lock(&mvm->tcm.lock);
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1074
if (mvm->tcm.paused || !time_after(ts, mvm->tcm.ts + MVM_TCM_PERIOD)) {
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1075
spin_unlock(&mvm->tcm.lock);
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1078
spin_unlock(&mvm->tcm.lock);
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1086
spin_lock(&mvm->tcm.lock);
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1088
if (!mvm->tcm.paused && time_after(ts, mvm->tcm.ts + MVM_TCM_PERIOD)) {
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1095
mvm->tcm.ts = ts;
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1097
schedule_delayed_work(&mvm->tcm.work, work_delay);
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1099
spin_unlock(&mvm->tcm.lock);
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1108
tcm.work);
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1115
spin_lock_bh(&mvm->tcm.lock);
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1116
mvm->tcm.paused = true;
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1117
spin_unlock_bh(&mvm->tcm.lock);
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1119
cancel_delayed_work_sync(&mvm->tcm.work);
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1127
spin_lock_bh(&mvm->tcm.lock);
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1128
mvm->tcm.ts = jiffies;
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1129
mvm->tcm.ll_ts = jiffies;
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1131
struct iwl_mvm_tcm_mac *mdata = &mvm->tcm.data[mac];
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1138
if (mvm->tcm.result.low_latency[mac])
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1143
mvm->tcm.paused = false;
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1149
if (mvm->tcm.result.global_load > IWL_MVM_TRAFFIC_LOW)
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1150
schedule_delayed_work(&mvm->tcm.work, MVM_TCM_PERIOD);
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1152
schedule_delayed_work(&mvm->tcm.work, MVM_LL_PERIOD);
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1154
spin_unlock_bh(&mvm->tcm.lock);
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
826
low_latency = mvm->tcm.result.low_latency[mvmvif->id];
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
828
if (!mvm->tcm.result.change[mvmvif->id] &&
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
866
if (mvm->tcm.data[mvmvif->id].opened_rx_ba_sessions)
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
897
if (mvm->tcm.data[mvmvif->id].uapsd_nonagg_detect.detected)
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
900
mvm->tcm.data[mvmvif->id].uapsd_nonagg_detect.detected = true;
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
911
u64 bytes = mvm->tcm.data[mac].uapsd_nonagg_detect.rx_bytes;
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
916
rate = ewma_rate_read(&mvm->tcm.data[mac].uapsd_nonagg_detect.rate);
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
918
if (!rate || mvm->tcm.data[mac].opened_rx_ba_sessions ||
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
919
mvm->tcm.data[mac].uapsd_nonagg_detect.detected)
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
966
unsigned int elapsed = jiffies_to_msecs(ts - mvm->tcm.ts);
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
968
jiffies_to_msecs(ts - mvm->tcm.uapsd_nonagg_ts);
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
975
bool handle_ll = time_after(ts, mvm->tcm.ll_ts + MVM_LL_PERIOD);
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
978
mvm->tcm.ll_ts = ts;
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
980
mvm->tcm.uapsd_nonagg_ts = ts;
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
982
mvm->tcm.result.elapsed = elapsed;
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
990
struct iwl_mvm_tcm_mac *mdata = &mvm->tcm.data[mac];
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
998
mvm->tcm.result.change[mac] = load != mvm->tcm.result.load[mac];
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
999
mvm->tcm.result.load[mac] = load;
drivers/remoteproc/xlnx_r5_remoteproc.c
1040
struct mem_bank_data *tcm;
drivers/remoteproc/xlnx_r5_remoteproc.c
1075
tcm = devm_kzalloc(dev, sizeof(struct mem_bank_data),
drivers/remoteproc/xlnx_r5_remoteproc.c
1077
if (!tcm)
drivers/remoteproc/xlnx_r5_remoteproc.c
1080
r5_core->tcm_banks[j] = tcm;
drivers/remoteproc/xlnx_r5_remoteproc.c
1092
tcm->pm_domain_id = out_args.args[0];
drivers/remoteproc/xlnx_r5_remoteproc.c
1107
tcm->da = (u32)abs_addr;
drivers/remoteproc/xlnx_r5_remoteproc.c
1108
tcm->size = (u32)size;
drivers/remoteproc/xlnx_r5_remoteproc.c
1117
tcm->addr = (u32)res->start;
drivers/remoteproc/xlnx_r5_remoteproc.c
1118
tcm->bank_name = (char *)res->name;
drivers/remoteproc/xlnx_r5_remoteproc.c
1119
res = devm_request_mem_region(dev, tcm->addr, tcm->size,
drivers/remoteproc/xlnx_r5_remoteproc.c
1120
tcm->bank_name);
drivers/thunderbolt/tb.c
113
struct tb_cm *tcm = tb_priv(sw->tb);
drivers/thunderbolt/tb.c
1291
struct tb_cm *tcm = tb_priv(port->sw->tb);
drivers/thunderbolt/tb.c
130
list_add(&port->list, &tcm->dp_resources);
drivers/thunderbolt/tb.c
132
list_add_tail(&port->list, &tcm->dp_resources);
drivers/thunderbolt/tb.c
1364
if (!tcm->hotplug_active) {
drivers/thunderbolt/tb.c
140
struct tb_cm *tcm = tb_priv(sw->tb);
drivers/thunderbolt/tb.c
1418
if (tcm->hotplug_active && tb_tunnel_usb3(sw->tb, sw))
drivers/thunderbolt/tb.c
149
list_for_each_entry_safe(port, tmp, &tcm->dp_resources, list) {
drivers/thunderbolt/tb.c
1517
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
1522
for (i = 0; i < ARRAY_SIZE(tcm->groups); i++) {
drivers/thunderbolt/tb.c
1523
struct tb_bandwidth_group *group = &tcm->groups[i];
drivers/thunderbolt/tb.c
1580
static void tb_init_bandwidth_groups(struct tb_cm *tcm)
drivers/thunderbolt/tb.c
1584
for (i = 0; i < ARRAY_SIZE(tcm->groups); i++) {
drivers/thunderbolt/tb.c
1585
struct tb_bandwidth_group *group = &tcm->groups[i];
drivers/thunderbolt/tb.c
1587
group->tb = tcm_to_tb(tcm);
drivers/thunderbolt/tb.c
159
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
1607
static struct tb_bandwidth_group *tb_find_free_bandwidth_group(struct tb_cm *tcm)
drivers/thunderbolt/tb.c
1611
for (i = 0; i < ARRAY_SIZE(tcm->groups); i++) {
drivers/thunderbolt/tb.c
1612
struct tb_bandwidth_group *group = &tcm->groups[i];
drivers/thunderbolt/tb.c
162
list_for_each_entry(p, &tcm->dp_resources, list) {
drivers/thunderbolt/tb.c
1622
tb_attach_bandwidth_group(struct tb_cm *tcm, struct tb_port *in,
drivers/thunderbolt/tb.c
1634
list_for_each_entry(tunnel, &tcm->tunnel_list, list) {
drivers/thunderbolt/tb.c
1649
group = tb_find_free_bandwidth_group(tcm);
drivers/thunderbolt/tb.c
1658
static void tb_discover_bandwidth_group(struct tb_cm *tcm, struct tb_port *in,
drivers/thunderbolt/tb.c
1665
for (i = 0; i < ARRAY_SIZE(tcm->groups); i++) {
drivers/thunderbolt/tb.c
1666
if (tcm->groups[i].index == index) {
drivers/thunderbolt/tb.c
1667
tb_bandwidth_group_attach_port(&tcm->groups[i], in);
drivers/thunderbolt/tb.c
1673
tb_attach_bandwidth_group(tcm, in, out);
drivers/thunderbolt/tb.c
169
list_add_tail(&port->list, &tcm->dp_resources);
drivers/thunderbolt/tb.c
1696
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
1699
tb_switch_discover_tunnels(tb->root_switch, &tcm->tunnel_list, true);
drivers/thunderbolt/tb.c
1701
list_for_each_entry(tunnel, &tcm->tunnel_list, list) {
drivers/thunderbolt/tb.c
1717
tb_discover_bandwidth_group(tcm, in, out);
drivers/thunderbolt/tb.c
174
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
177
list_for_each_entry(tunnel, &tcm->tunnel_list, list) {
drivers/thunderbolt/tb.c
1777
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
1781
list_for_each_entry_safe(tunnel, n, &tcm->tunnel_list, list) {
drivers/thunderbolt/tb.c
186
struct tb_cm *tcm = tb_priv(sw->tb);
drivers/thunderbolt/tb.c
1866
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
1871
list_for_each_entry(port, &tcm->dp_resources, list) {
drivers/thunderbolt/tb.c
1975
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
1985
list_for_each_entry(tunnel, &tcm->tunnel_list, list) {
drivers/thunderbolt/tb.c
2008
if (!tb_attach_bandwidth_group(tcm, in, out))
drivers/thunderbolt/tb.c
2036
list_add_tail(&tunnel->list, &tcm->tunnel_list);
drivers/thunderbolt/tb.c
2065
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
2081
list_for_each_entry(port, &tcm->dp_resources, list) {
drivers/thunderbolt/tb.c
211
list_for_each_entry(tunnel, &tcm->tunnel_list, list) {
drivers/thunderbolt/tb.c
2211
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
2217
list_for_each_entry(p, &tcm->dp_resources, list) {
drivers/thunderbolt/tb.c
2224
list_add_tail(&port->list, &tcm->dp_resources);
drivers/thunderbolt/tb.c
2233
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
2240
list_for_each_entry_safe_reverse(tunnel, n, &tcm->tunnel_list, list) {
drivers/thunderbolt/tb.c
2245
while (!list_empty(&tcm->dp_resources)) {
drivers/thunderbolt/tb.c
2248
port = list_first_entry(&tcm->dp_resources,
drivers/thunderbolt/tb.c
2278
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
2315
list_add_tail(&tunnel->list, &tcm->tunnel_list);
drivers/thunderbolt/tb.c
2323
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
2355
list_add_tail(&tunnel->list, &tcm->tunnel_list);
drivers/thunderbolt/tb.c
2372
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
2381
list_for_each_entry_safe(tunnel, n, &tcm->tunnel_list, list) {
drivers/thunderbolt/tb.c
2425
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
2433
if (!tcm->hotplug_active)
drivers/thunderbolt/tb.c
2737
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
2744
if (!tcm->hotplug_active)
drivers/thunderbolt/tb.c
2935
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
2939
cancel_delayed_work(&tcm->remove_work);
drivers/thunderbolt/tb.c
2941
list_for_each_entry_safe(tunnel, n, &tcm->tunnel_list, list) {
drivers/thunderbolt/tb.c
2952
tcm->hotplug_active = false; /* signal tb_handle_hotplug to quit */
drivers/thunderbolt/tb.c
2957
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
2961
for (i = 0; i < ARRAY_SIZE(tcm->groups); i++)
drivers/thunderbolt/tb.c
2962
cancel_delayed_work_sync(&tcm->groups[i].release_work);
drivers/thunderbolt/tb.c
2988
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
3063
tcm->hotplug_active = true;
drivers/thunderbolt/tb.c
3069
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
3075
tcm->hotplug_active = false; /* signal tb_handle_hotplug to quit */
drivers/thunderbolt/tb.c
3115
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
3149
list_for_each_entry_safe(tunnel, n, &tcm->tunnel_list, list) {
drivers/thunderbolt/tb.c
3158
if (!list_empty(&tcm->tunnel_list)) {
drivers/thunderbolt/tb.c
3168
tcm->hotplug_active = true;
drivers/thunderbolt/tb.c
3198
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
3200
tcm->hotplug_active = false;
drivers/thunderbolt/tb.c
3206
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
3208
tcm->hotplug_active = true;
drivers/thunderbolt/tb.c
3227
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
3237
tcm->hotplug_active = false;
drivers/thunderbolt/tb.c
3245
struct tb_cm *tcm = container_of(work, struct tb_cm, remove_work.work);
drivers/thunderbolt/tb.c
3246
struct tb *tb = tcm_to_tb(tcm);
drivers/thunderbolt/tb.c
3258
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
3265
list_for_each_entry_safe(tunnel, n, &tcm->tunnel_list, list)
drivers/thunderbolt/tb.c
3268
tcm->hotplug_active = true;
drivers/thunderbolt/tb.c
3276
queue_delayed_work(tb->wq, &tcm->remove_work, msecs_to_jiffies(50));
drivers/thunderbolt/tb.c
3368
struct tb_cm *tcm;
drivers/thunderbolt/tb.c
3371
tb = tb_domain_alloc(nhi, TB_TIMEOUT, sizeof(*tcm));
drivers/thunderbolt/tb.c
3382
tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
3383
INIT_LIST_HEAD(&tcm->tunnel_list);
drivers/thunderbolt/tb.c
3384
INIT_LIST_HEAD(&tcm->dp_resources);
drivers/thunderbolt/tb.c
3385
INIT_DELAYED_WORK(&tcm->remove_work, tb_remove_work);
drivers/thunderbolt/tb.c
3386
tb_init_bandwidth_groups(tcm);
drivers/thunderbolt/tb.c
494
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
497
list_for_each_entry(tunnel, &tcm->tunnel_list, list) {
drivers/thunderbolt/tb.c
613
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
624
list_for_each_entry(tunnel, &tcm->tunnel_list, list) {
drivers/thunderbolt/tb.c
72
static inline struct tb *tcm_to_tb(struct tb_cm *tcm)
drivers/thunderbolt/tb.c
74
return ((void *)tcm - sizeof(struct tb));
drivers/thunderbolt/tb.c
910
struct tb_cm *tcm = tb_priv(tb);
drivers/thunderbolt/tb.c
982
list_add_tail(&tunnel->list, &tcm->tunnel_list);
drivers/usb/gadget/function/f_tcm.c
2583
struct f_uas *tcm = to_f_uas(f);
drivers/usb/gadget/function/f_tcm.c
2585
kfree(tcm);
drivers/usb/gadget/function/f_tcm.c
2629
DECLARE_USB_FUNCTION(tcm, tcm_alloc_inst, tcm_alloc);
include/net/sch_priv.h
18
struct tcmsg *tcm);
include/net/sch_priv.h
22
struct sk_buff *skb, struct tcmsg *tcm);
net/sched/cls_api.c
2057
struct tcmsg *tcm;
net/sched/cls_api.c
2062
nlh = nlmsg_put(skb, portid, seq, event, sizeof(*tcm), flags);
net/sched/cls_api.c
2065
tcm = nlmsg_data(nlh);
net/sched/cls_api.c
2066
tcm->tcm_family = AF_UNSPEC;
net/sched/cls_api.c
2067
tcm->tcm__pad1 = 0;
net/sched/cls_api.c
2068
tcm->tcm__pad2 = 0;
net/sched/cls_api.c
2070
tcm->tcm_ifindex = qdisc_dev(q)->ifindex;
net/sched/cls_api.c
2071
tcm->tcm_parent = parent;
net/sched/cls_api.c
2073
tcm->tcm_ifindex = TCM_IFINDEX_MAGIC_BLOCK;
net/sched/cls_api.c
2074
tcm->tcm_block_index = block->index;
net/sched/cls_api.c
2076
tcm->tcm_info = TC_H_MAKE(tp->prio, tp->protocol);
net/sched/cls_api.c
2082
tcm->tcm_handle = 0;
net/sched/cls_api.c
2085
if (tp->ops->terse_dump(net, tp, fh, skb, tcm,
net/sched/cls_api.c
2093
tp->ops->dump(net, tp, fh, skb, tcm, rtnl_held) < 0)
net/sched/cls_api.c
2783
struct tcmsg *tcm = nlmsg_data(cb->nlh);
net/sched/cls_api.c
2795
if (TC_H_MAJ(tcm->tcm_info) &&
net/sched/cls_api.c
2796
TC_H_MAJ(tcm->tcm_info) != tp->prio)
net/sched/cls_api.c
2798
if (TC_H_MIN(tcm->tcm_info) &&
net/sched/cls_api.c
2799
TC_H_MIN(tcm->tcm_info) != tp->protocol)
net/sched/cls_api.c
2851
struct tcmsg *tcm = nlmsg_data(cb->nlh);
net/sched/cls_api.c
2858
if (nlmsg_len(cb->nlh) < sizeof(*tcm))
net/sched/cls_api.c
2861
err = nlmsg_parse_deprecated(cb->nlh, sizeof(*tcm), tca, TCA_MAX,
net/sched/cls_api.c
2873
if (tcm->tcm_ifindex == TCM_IFINDEX_MAGIC_BLOCK) {
net/sched/cls_api.c
2874
block = tcf_block_refcnt_get(net, tcm->tcm_block_index);
net/sched/cls_api.c
2889
dev = __dev_get_by_index(net, tcm->tcm_ifindex);
net/sched/cls_api.c
2893
parent = tcm->tcm_parent;
net/sched/cls_api.c
2897
q = qdisc_lookup(dev, TC_H_MAJ(tcm->tcm_parent));
net/sched/cls_api.c
2905
if (TC_H_MIN(tcm->tcm_parent)) {
net/sched/cls_api.c
2906
cl = cops->find(q, tcm->tcm_parent);
net/sched/cls_api.c
2937
if (tcm->tcm_ifindex == TCM_IFINDEX_MAGIC_BLOCK)
net/sched/cls_api.c
2958
struct tcmsg *tcm;
net/sched/cls_api.c
2964
nlh = nlmsg_put(skb, portid, seq, event, sizeof(*tcm), flags);
net/sched/cls_api.c
2967
tcm = nlmsg_data(nlh);
net/sched/cls_api.c
2968
tcm->tcm_family = AF_UNSPEC;
net/sched/cls_api.c
2969
tcm->tcm__pad1 = 0;
net/sched/cls_api.c
2970
tcm->tcm__pad2 = 0;
net/sched/cls_api.c
2971
tcm->tcm_handle = 0;
net/sched/cls_api.c
2972
tcm->tcm_info = 0;
net/sched/cls_api.c
2974
tcm->tcm_ifindex = qdisc_dev(block->q)->ifindex;
net/sched/cls_api.c
2975
tcm->tcm_parent = block->q->handle;
net/sched/cls_api.c
2977
tcm->tcm_ifindex = TCM_IFINDEX_MAGIC_BLOCK;
net/sched/cls_api.c
2978
tcm->tcm_block_index = block->index;
net/sched/cls_api.c
3251
struct tcmsg *tcm = nlmsg_data(cb->nlh);
net/sched/cls_api.c
3257
if (nlmsg_len(cb->nlh) < sizeof(*tcm))
net/sched/cls_api.c
3260
err = nlmsg_parse_deprecated(cb->nlh, sizeof(*tcm), tca, TCA_MAX,
net/sched/cls_api.c
3265
if (tcm->tcm_ifindex == TCM_IFINDEX_MAGIC_BLOCK) {
net/sched/cls_api.c
3266
block = tcf_block_refcnt_get(net, tcm->tcm_block_index);
net/sched/cls_api.c
3274
dev = __dev_get_by_index(net, tcm->tcm_ifindex);
net/sched/cls_api.c
3278
if (!tcm->tcm_parent)
net/sched/cls_api.c
3281
q = qdisc_lookup(dev, TC_H_MAJ(tcm->tcm_parent));
net/sched/cls_api.c
3290
if (TC_H_MIN(tcm->tcm_parent)) {
net/sched/cls_api.c
3291
cl = cops->find(q, tcm->tcm_parent);
net/sched/cls_api.c
3327
if (tcm->tcm_ifindex == TCM_IFINDEX_MAGIC_BLOCK)
net/sched/sch_api.c
1471
struct tcmsg *tcm)
net/sched/sch_api.c
1479
clid = tcm->tcm_parent;
net/sched/sch_api.c
1502
if (tcm->tcm_handle && q->handle != tcm->tcm_handle) {
net/sched/sch_api.c
1507
q = qdisc_lookup(dev, tcm->tcm_handle);
net/sched/sch_api.c
1541
struct tcmsg *tcm = nlmsg_data(n);
net/sched/sch_api.c
1546
err = nlmsg_parse_deprecated(n, sizeof(*tcm), tca, TCA_MAX,
net/sched/sch_api.c
1551
dev = __dev_get_by_index(net, tcm->tcm_ifindex);
net/sched/sch_api.c
1556
err = __tc_get_qdisc(skb, n, extack, dev, tca, tcm);
net/sched/sch_api.c
1585
struct tcmsg *tcm)
net/sched/sch_api.c
1592
clid = tcm->tcm_parent;
net/sched/sch_api.c
1621
if (!q || !tcm->tcm_handle || q->handle != tcm->tcm_handle) {
net/sched/sch_api.c
1622
if (tcm->tcm_handle) {
net/sched/sch_api.c
1627
if (TC_H_MIN(tcm->tcm_handle)) {
net/sched/sch_api.c
1631
q = qdisc_lookup(dev, tcm->tcm_handle);
net/sched/sch_api.c
1634
if (q->parent != tcm->tcm_parent) {
net/sched/sch_api.c
1703
if (!tcm->tcm_handle) {
net/sched/sch_api.c
1707
q = qdisc_lookup(dev, tcm->tcm_handle);
net/sched/sch_api.c
1737
tcm->tcm_parent, tcm->tcm_parent,
net/sched/sch_api.c
1747
dev_queue = p->ops->cl_ops->select_queue(p, tcm);
net/sched/sch_api.c
1754
tcm->tcm_parent, tcm->tcm_handle,
net/sched/sch_api.c
1801
struct tcmsg *tcm;
net/sched/sch_api.c
1804
err = nlmsg_parse_deprecated(n, sizeof(*tcm), tca, TCA_MAX,
net/sched/sch_api.c
1811
tcm = nlmsg_data(n);
net/sched/sch_api.c
1812
dev = __dev_get_by_index(net, tcm->tcm_ifindex);
net/sched/sch_api.c
1817
err = __tc_modify_qdisc(skb, n, extack, dev, tca, tcm);
net/sched/sch_api.c
1946
struct tcmsg *tcm;
net/sched/sch_api.c
1953
nlh = nlmsg_put(skb, portid, seq, event, sizeof(*tcm), flags);
net/sched/sch_api.c
1956
tcm = nlmsg_data(nlh);
net/sched/sch_api.c
1957
tcm->tcm_family = AF_UNSPEC;
net/sched/sch_api.c
1958
tcm->tcm__pad1 = 0;
net/sched/sch_api.c
1959
tcm->tcm__pad2 = 0;
net/sched/sch_api.c
1960
tcm->tcm_ifindex = qdisc_dev(q)->ifindex;
net/sched/sch_api.c
1961
tcm->tcm_parent = q->handle;
net/sched/sch_api.c
1962
tcm->tcm_handle = q->handle;
net/sched/sch_api.c
1963
tcm->tcm_info = 0;
net/sched/sch_api.c
1966
if (cl_ops->dump && cl_ops->dump(q, cl, skb, tcm) < 0)
net/sched/sch_api.c
2164
struct tcmsg *tcm)
net/sched/sch_api.c
2191
portid = tcm->tcm_parent;
net/sched/sch_api.c
2192
clid = tcm->tcm_handle;
net/sched/sch_api.c
2294
struct tcmsg *tcm = nlmsg_data(n);
net/sched/sch_api.c
2299
err = nlmsg_parse_deprecated(n, sizeof(*tcm), tca, TCA_MAX,
net/sched/sch_api.c
2304
dev = __dev_get_by_index(net, tcm->tcm_ifindex);
net/sched/sch_api.c
2309
err = __tc_ctl_tclass(skb, n, extack, dev, tca, tcm);
net/sched/sch_api.c
2332
struct tcmsg *tcm, struct netlink_callback *cb,
net/sched/sch_api.c
2339
(tcm->tcm_parent &&
net/sched/sch_api.c
2340
TC_H_MAJ(tcm->tcm_parent) != q->handle)) {
net/sched/sch_api.c
2361
struct tcmsg *tcm, struct netlink_callback *cb,
net/sched/sch_api.c
2370
if (tc_dump_tclass_qdisc(root, skb, tcm, cb, t_p, s_t) < 0)
net/sched/sch_api.c
2376
if (tcm->tcm_parent) {
net/sched/sch_api.c
2377
q = qdisc_match_from_root(root, TC_H_MAJ(tcm->tcm_parent));
net/sched/sch_api.c
2379
tc_dump_tclass_qdisc(q, skb, tcm, cb, t_p, s_t) < 0)
net/sched/sch_api.c
2384
if (tc_dump_tclass_qdisc(q, skb, tcm, cb, t_p, s_t) < 0)
net/sched/sch_api.c
2392
struct tcmsg *tcm, struct net_device *dev)
net/sched/sch_api.c
2401
skb, tcm, cb, &t, s_t, true) < 0)
net/sched/sch_api.c
2407
skb, tcm, cb, &t, s_t, false) < 0)
net/sched/sch_api.c
2418
struct tcmsg *tcm = nlmsg_data(cb->nlh);
net/sched/sch_api.c
2423
if (nlmsg_len(cb->nlh) < sizeof(*tcm))
net/sched/sch_api.c
2426
dev = dev_get_by_index(net, tcm->tcm_ifindex);
net/sched/sch_api.c
2431
err = __tc_dump_tclass(skb, cb, tcm, dev);
net/sched/sch_api.c
904
struct tcmsg *tcm;
net/sched/sch_api.c
913
nlh = nlmsg_put(skb, portid, seq, event, sizeof(*tcm), flags);
net/sched/sch_api.c
916
tcm = nlmsg_data(nlh);
net/sched/sch_api.c
917
tcm->tcm_family = AF_UNSPEC;
net/sched/sch_api.c
918
tcm->tcm__pad1 = 0;
net/sched/sch_api.c
919
tcm->tcm__pad2 = 0;
net/sched/sch_api.c
920
tcm->tcm_ifindex = qdisc_dev(q)->ifindex;
net/sched/sch_api.c
921
tcm->tcm_parent = clid;
net/sched/sch_api.c
922
tcm->tcm_handle = q->handle;
net/sched/sch_api.c
923
tcm->tcm_info = refcount_read(&q->refcnt);
net/sched/sch_cake.c
3109
struct sk_buff *skb, struct tcmsg *tcm)
net/sched/sch_cake.c
3111
tcm->tcm_handle |= TC_H_MIN(cl);
net/sched/sch_cbs.c
479
struct sk_buff *skb, struct tcmsg *tcm)
net/sched/sch_cbs.c
486
tcm->tcm_handle |= TC_H_MIN(1);
net/sched/sch_cbs.c
487
tcm->tcm_info = q->qdisc->handle;
net/sched/sch_drr.c
242
struct sk_buff *skb, struct tcmsg *tcm)
net/sched/sch_drr.c
247
tcm->tcm_parent = TC_H_ROOT;
net/sched/sch_drr.c
248
tcm->tcm_handle = cl->common.classid;
net/sched/sch_drr.c
249
tcm->tcm_info = cl->qdisc->handle;
net/sched/sch_ets.c
309
struct sk_buff *skb, struct tcmsg *tcm)
net/sched/sch_ets.c
315
tcm->tcm_parent = TC_H_ROOT;
net/sched/sch_ets.c
316
tcm->tcm_handle = ets_class_id(sch, cl);
net/sched/sch_ets.c
317
tcm->tcm_info = cl->qdisc->handle;
net/sched/sch_fq_codel.c
638
struct sk_buff *skb, struct tcmsg *tcm)
net/sched/sch_fq_codel.c
640
tcm->tcm_handle |= TC_H_MIN(cl);
net/sched/sch_hfsc.c
1309
struct tcmsg *tcm)
net/sched/sch_hfsc.c
1314
tcm->tcm_parent = cl->cl_parent ? cl->cl_parent->cl_common.classid :
net/sched/sch_hfsc.c
1316
tcm->tcm_handle = cl->cl_common.classid;
net/sched/sch_hfsc.c
1318
tcm->tcm_info = cl->qdisc->handle;
net/sched/sch_htb.c
1243
struct sk_buff *skb, struct tcmsg *tcm)
net/sched/sch_htb.c
1253
tcm->tcm_parent = cl->parent ? cl->parent->common.classid : TC_H_ROOT;
net/sched/sch_htb.c
1254
tcm->tcm_handle = cl->common.classid;
net/sched/sch_htb.c
1256
tcm->tcm_info = cl->leaf.q->handle;
net/sched/sch_htb.c
1363
htb_select_queue(struct Qdisc *sch, struct tcmsg *tcm)
net/sched/sch_htb.c
1375
.classid = TC_H_MIN(tcm->tcm_parent),
net/sched/sch_mq.c
190
struct tcmsg *tcm)
net/sched/sch_mq.c
192
return mq_queue_get(sch, TC_H_MIN(tcm->tcm_parent));
net/sched/sch_mq.c
241
struct sk_buff *skb, struct tcmsg *tcm)
net/sched/sch_mq.c
245
tcm->tcm_parent = TC_H_ROOT;
net/sched/sch_mq.c
246
tcm->tcm_handle |= TC_H_MIN(cl);
net/sched/sch_mq.c
247
tcm->tcm_info = rtnl_dereference(dev_queue->qdisc_sleeping)->handle;
net/sched/sch_mqprio.c
641
struct sk_buff *skb, struct tcmsg *tcm)
net/sched/sch_mqprio.c
648
tcm->tcm_parent = (tc < 0) ? 0 :
net/sched/sch_mqprio.c
651
tcm->tcm_info = rtnl_dereference(dev_queue->qdisc_sleeping)->handle;
net/sched/sch_mqprio.c
653
tcm->tcm_parent = TC_H_ROOT;
net/sched/sch_mqprio.c
654
tcm->tcm_info = 0;
net/sched/sch_mqprio.c
656
tcm->tcm_handle |= TC_H_MIN(cl);
net/sched/sch_mqprio.c
748
struct tcmsg *tcm)
net/sched/sch_mqprio.c
750
return mqprio_queue_get(sch, TC_H_MIN(tcm->tcm_parent));
net/sched/sch_multiq.c
324
struct sk_buff *skb, struct tcmsg *tcm)
net/sched/sch_multiq.c
328
tcm->tcm_handle |= TC_H_MIN(cl);
net/sched/sch_multiq.c
329
tcm->tcm_info = q->queues[cl - 1]->handle;
net/sched/sch_netem.c
1296
struct sk_buff *skb, struct tcmsg *tcm)
net/sched/sch_netem.c
1303
tcm->tcm_handle |= TC_H_MIN(1);
net/sched/sch_netem.c
1304
tcm->tcm_info = q->qdisc->handle;
net/sched/sch_prio.c
346
struct tcmsg *tcm)
net/sched/sch_prio.c
350
tcm->tcm_handle |= TC_H_MIN(cl);
net/sched/sch_prio.c
351
tcm->tcm_info = q->queues[cl-1]->handle;
net/sched/sch_qfq.c
628
struct sk_buff *skb, struct tcmsg *tcm)
net/sched/sch_qfq.c
634
tcm->tcm_parent = TC_H_ROOT;
net/sched/sch_qfq.c
635
tcm->tcm_handle = cl->common.classid;
net/sched/sch_qfq.c
636
tcm->tcm_info = cl->qdisc->handle;
net/sched/sch_red.c
474
struct sk_buff *skb, struct tcmsg *tcm)
net/sched/sch_red.c
478
tcm->tcm_handle |= TC_H_MIN(1);
net/sched/sch_red.c
479
tcm->tcm_info = q->qdisc->handle;
net/sched/sch_sfb.c
616
struct sk_buff *skb, struct tcmsg *tcm)
net/sched/sch_sfq.c
898
struct sk_buff *skb, struct tcmsg *tcm)
net/sched/sch_sfq.c
900
tcm->tcm_handle |= TC_H_MIN(cl);
net/sched/sch_skbprio.c
243
struct sk_buff *skb, struct tcmsg *tcm)
net/sched/sch_skbprio.c
245
tcm->tcm_handle |= TC_H_MIN(cl);
net/sched/sch_taprio.c
2469
struct sk_buff *skb, struct tcmsg *tcm)
net/sched/sch_taprio.c
2473
tcm->tcm_parent = TC_H_ROOT;
net/sched/sch_taprio.c
2474
tcm->tcm_handle |= TC_H_MIN(cl);
net/sched/sch_taprio.c
2475
tcm->tcm_info = child->handle;
net/sched/sch_taprio.c
2516
struct tcmsg *tcm)
net/sched/sch_taprio.c
2518
return taprio_queue_get(sch, TC_H_MIN(tcm->tcm_parent));
net/sched/sch_tbf.c
553
struct sk_buff *skb, struct tcmsg *tcm)
net/sched/sch_tbf.c
557
tcm->tcm_handle |= TC_H_MIN(1);
net/sched/sch_tbf.c
558
tcm->tcm_info = q->qdisc->handle;