Symbol: to_dev
drivers/dma/ioat/dma.c
177
dev_dbg(to_dev(ioat_chan),
drivers/dma/ioat/dma.c
213
dev_err(to_dev(ioat_chan),
drivers/dma/ioat/dma.c
218
dev_dbg(to_dev(ioat_chan),
drivers/dma/ioat/dma.c
256
dev_dbg(to_dev(ioat_chan),
drivers/dma/ioat/dma.c
317
dev_dbg(to_dev(ioat_chan), "%s: cookie: %d\n", __func__, cookie);
drivers/dma/ioat/dma.c
391
descs->virt = dma_alloc_coherent(to_dev(ioat_chan),
drivers/dma/ioat/dma.c
398
dma_free_coherent(to_dev(ioat_chan),
drivers/dma/ioat/dma.c
420
dma_free_coherent(to_dev(ioat_chan),
drivers/dma/ioat/dma.c
472
dev_dbg(to_dev(ioat_chan), "%s: num_descs: %d (%x:%x:%x)\n",
drivers/dma/ioat/dma.c
480
dev_dbg_ratelimited(to_dev(ioat_chan),
drivers/dma/ioat/dma.c
537
dev_dbg(to_dev(ioat_chan), "%s: phys_complete: %#llx\n", __func__,
drivers/dma/ioat/dma.c
599
dev_dbg(to_dev(ioat_chan), "%s: head: %#x tail: %#x issued: %#x\n",
drivers/dma/ioat/dma.c
657
dev_dbg(to_dev(ioat_chan), "%s: cancel completion timeout\n",
drivers/dma/ioat/dma.c
794
dev_dbg(to_dev(ioat_chan), "%s: error = %x:%x\n",
drivers/dma/ioat/dma.c
80
dev_err(to_dev(ioat_chan), "Err(%d): %s\n",
drivers/dma/ioat/dma.c
836
dev_err(to_dev(ioat_chan), "%s: fatal error (%x:%x)\n",
drivers/dma/ioat/dma.c
838
dev_err(to_dev(ioat_chan), "Errors handled:\n");
drivers/dma/ioat/dma.c
840
dev_err(to_dev(ioat_chan), "Errors not handled:\n");
drivers/dma/ioat/dma.c
892
dev_warn(to_dev(ioat_chan), "Reset channel...\n");
drivers/dma/ioat/dma.c
894
dev_warn(to_dev(ioat_chan), "Restart channel...\n");
drivers/dma/ioat/dma.c
918
dev_err(to_dev(ioat_chan), "%s: Channel halted (%x)\n",
drivers/dma/ioat/dma.c
920
dev_err(to_dev(ioat_chan), "Errors:\n");
drivers/dma/ioat/dma.c
959
dev_err(to_dev(ioat_chan), "CHANSTS: %#Lx CHANERR: %#x\n",
drivers/dma/ioat/dma.c
961
dev_err(to_dev(ioat_chan), "Errors:\n");
drivers/dma/ioat/dma.c
964
dev_dbg(to_dev(ioat_chan), "Active descriptors: %d\n",
drivers/dma/ioat/dma.c
974
dev_warn(to_dev(ioat_chan),
drivers/dma/ioat/dma.h
224
struct device *dev = to_dev(ioat_chan);
drivers/dma/ioat/init.c
626
dev_dbg(to_dev(ioat_chan), "freeing %d idle descriptors\n", descs);
drivers/dma/ioat/init.c
633
dev_err(to_dev(ioat_chan), "Freeing %d in use descriptors!\n",
drivers/dma/ioat/init.c
643
dma_free_coherent(to_dev(ioat_chan), IOAT_CHUNK_SIZE,
drivers/dma/ioat/init.c
747
dev_WARN(to_dev(ioat_chan),
drivers/dma/ioat/prep.c
279
struct device *dev = to_dev(ioat_chan);
drivers/dma/ioat/prep.c
306
struct device *dev = to_dev(ioat_chan);
drivers/dma/ioat/prep.c
358
dev_dbg(to_dev(ioat_chan), "%s\n", __func__);
drivers/dma/ioat/prep.c
480
dev_dbg(to_dev(ioat_chan), "%s\n", __func__);
drivers/dma/ioat/prep.c
507
dev_err(to_dev(ioat_chan),
drivers/dma/ioat/sysfs.c
88
dev_warn(to_dev(ioat_chan),
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1234
struct kfd_topology_device *dev, *to_dev;
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1284
to_dev = kfd_topology_device_by_proximity_domain_no_lock(id_to);
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1285
if (!to_dev)
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1295
to_dev->node_props.io_links_count++;
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1296
list_add_tail(&props2->list, &to_dev->io_link_props);
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
1238
static void kfd_set_iolink_non_coherent(struct kfd_topology_device *to_dev,
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
1243
if (!to_dev->gpu &&
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
1247
if (to_dev->gpu) {
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
1253
KFD_GC_VERSION(to_dev->gpu) == IP_VERSION(9, 4, 0))) {
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
1271
static void kfd_set_recommended_sdma_engines(struct kfd_topology_device *to_dev,
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
1284
bool support_rec_eng = !amdgpu_sriov_vf(adev) && to_dev->gpu &&
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
1291
int dst_socket_id = to_dev->gpu->adev->gmc.xgmi.physical_node_id;
drivers/gpu/drm/amd/amdkfd/kfd_topology.c
1307
num_xgmi_sdma_engines && to_dev->gpu) ? xgmi_sdma_eng_id_mask :
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
1185
const struct net_device *to_dev;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_flex_actions.c
144
agent_parms.to_dev = out_dev;
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
271
mall_entry->mirror.to_dev = act->dev;
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
36
if (!mall_entry->mirror.to_dev) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
41
agent_parms.to_dev = mall_entry->mirror.to_dev;
drivers/net/ethernet/mellanox/mlxsw/spectrum_matchall.c
418
.to_dev = NULL, /* Mirror to CPU. */
drivers/net/ethernet/mellanox/mlxsw/spectrum_qdisc.c
1833
.to_dev = mall_entry->mirror.to_dev,
drivers/net/ethernet/mellanox/mlxsw/spectrum_qdisc.c
2030
mall_entry->mirror.to_dev = act->dev;
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
1017
const struct net_device *to_dev)
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
1023
if (span->span_entry_ops_arr[i]->can_handle(to_dev))
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
1049
err = curr->ops->parms_set(mlxsw_sp, curr->to_dev, &sparms);
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
1071
const struct net_device *to_dev = parms->to_dev;
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
1079
ops = mlxsw_sp_span_entry_ops(mlxsw_sp, to_dev);
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
1086
err = ops->parms_set(mlxsw_sp, to_dev, &sparms);
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
1093
span_entry = mlxsw_sp_span_entry_get(mlxsw_sp, to_dev, ops, sparms);
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
140
const struct net_device *to_dev,
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
169
const struct net_device *to_dev,
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
172
sparmsp->dest_port = netdev_priv(to_dev);
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
412
mlxsw_sp_span_gretap4_route(const struct net_device *to_dev,
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
415
struct ip_tunnel *tun = netdev_priv(to_dev);
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
424
parms = mlxsw_sp_ipip_netdev_parms4(to_dev);
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
450
const struct net_device *to_dev,
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
453
struct ip_tunnel_parm_kern tparm = mlxsw_sp_ipip_netdev_parms4(to_dev);
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
461
if (!(to_dev->flags & IFF_UP) ||
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
471
l3edev = mlxsw_sp_span_gretap4_route(to_dev, &saddr.addr4, &gw.addr4);
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
521
mlxsw_sp_span_gretap6_route(const struct net_device *to_dev,
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
525
struct ip6_tnl *t = netdev_priv(to_dev);
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
555
const struct net_device *to_dev,
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
558
struct __ip6_tnl_parm tparm = mlxsw_sp_ipip_netdev_parms6(to_dev);
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
566
if (!(to_dev->flags & IFF_UP) ||
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
576
l3edev = mlxsw_sp_span_gretap6_route(to_dev, &saddr.addr6, &gw.addr6);
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
633
const struct net_device *to_dev,
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
639
if (!(to_dev->flags & IFF_UP))
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
642
real_dev = mlxsw_sp_span_entry_vlan(to_dev, &vid);
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
701
const struct net_device *to_dev,
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
751
const struct net_device *to_dev,
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
852
const struct net_device *to_dev,
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
881
span_entry->to_dev = to_dev;
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
898
const struct net_device *to_dev)
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
905
if (refcount_read(&curr->ref_count) && curr->to_dev == to_dev)
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
934
const struct net_device *to_dev,
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
942
if (refcount_read(&curr->ref_count) && curr->to_dev == to_dev &&
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
953
const struct net_device *to_dev,
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
959
span_entry = mlxsw_sp_span_entry_find_by_parms(mlxsw_sp, to_dev,
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.c
967
return mlxsw_sp_span_entry_create(mlxsw_sp, to_dev, ops, sparms);
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.h
56
const struct net_device *to_dev;
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.h
71
const struct net_device *to_dev;
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.h
80
bool (*can_handle)(const struct net_device *to_dev);
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.h
82
const struct net_device *to_dev,
drivers/net/ethernet/mellanox/mlxsw/spectrum_span.h
95
const struct net_device *to_dev);
drivers/nvdimm/btt.c
102
dev_err_ratelimited(to_dev(arena),
drivers/nvdimm/btt.c
1258
dev_warn_ratelimited(to_dev(arena),
drivers/nvdimm/btt.c
143
dev_err_ratelimited(to_dev(arena),
drivers/nvdimm/btt.c
161
dev_err_ratelimited(to_dev(arena),
drivers/nvdimm/btt.c
342
dev_err(to_dev(arena),
drivers/nvdimm/btt.c
425
dev_WARN_ONCE(to_dev(arena), !IS_ALIGNED(arena->mapoff, 512),
drivers/nvdimm/btt.c
431
dev_WARN_ONCE(to_dev(arena), size < 512,
drivers/nvdimm/btt.c
469
dev_WARN_ONCE(to_dev(arena), !IS_ALIGNED(arena->logoff, 512),
drivers/nvdimm/btt.c
475
dev_WARN_ONCE(to_dev(arena), size < 512,
drivers/nvdimm/btt.c
569
dev_err_ratelimited(to_dev(arena),
drivers/nvdimm/btt.c
70
dev_WARN_ONCE(to_dev(arena), !IS_ALIGNED(arena->infooff, 512),
drivers/nvdimm/btt.c
711
dev_err(to_dev(arena), "Found an unknown padding scheme\n");
drivers/nvdimm/btt.c
717
dev_dbg(to_dev(arena), "log_index_0 = %d\n", log_index[0]);
drivers/nvdimm/btt.c
718
dev_dbg(to_dev(arena), "log_index_1 = %d\n", log_index[1]);
drivers/nvdimm/btt.c
72
dev_WARN_ONCE(to_dev(arena), !IS_ALIGNED(arena->info2off, 512),
drivers/nvdimm/btt.c
873
dev_info(to_dev(arena), "No existing arenas\n");
drivers/nvdimm/btt.c
876
dev_err(to_dev(arena),
drivers/nvdimm/btt.c
888
dev_err(to_dev(arena),
drivers/nvdimm/pmem.c
101
long cleared = nvdimm_clear_poison(to_dev(pmem), phys, len);
drivers/nvdimm/pmem.c
45
return to_nd_region(to_dev(pmem)->parent);
include/trace/events/xdp.h
253
const struct net_device *to_dev,
include/trace/events/xdp.h
256
TP_ARGS(from_dev, to_dev, sent, drops, err),
include/trace/events/xdp.h
270
__entry->to_ifindex = to_dev->ifindex;
samples/bpf/xdp_sample.bpf.c
207
const struct net_device *to_dev, int sent, int drops, int err)
samples/bpf/xdp_sample.bpf.c
214
idx_out = to_dev->ifindex;
samples/bpf/xdp_sample.bpf.c
238
const struct net_device *to_dev, int sent, int drops, int err)
samples/bpf/xdp_sample.bpf.c
246
idx_out = to_dev->ifindex;
tools/testing/selftests/bpf/progs/core_kern.c
44
*from_dev, const struct net_device *to_dev, int sent, int drops,