_d
#define ECARD_DEV(_d) container_of((_d), struct expansion_card, dev)
#define ECARD_DRV(_d) container_of((_d), struct ecard_driver, drv)
#define LOCOMO_DEV(_d) container_of((_d), struct locomo_dev, dev)
#define LOCOMO_DRV(_d) container_of_const((_d), struct locomo_driver, drv)
#define SA1111_DRV(_d) container_of_const((_d), struct sa1111_driver, drv)
#define dc21285_request_irq(_a, _b, _c, _d, _e) \
WARN_ON(request_irq(_a, _b, _c, _d, _e) < 0)
#define LOCAL_HUB_S(_r, _d) __raw_writeq((_d), LOCAL_HUB_PTR(_r))
#define REMOTE_HUB_S(_n, _r, _d) __raw_writeq((_d), \
#define BAT_SHOW_603(_m, _n, _l, _u, _d) bat_show_603(_m, _n, mfspr(_l), mfspr(_u), _d)
#define dump_node_name(_a, _b, _c, _d, _e) \
_dump_node_name(_a, _b, _c, _d, _e, __func__, __LINE__)
#define dump_node(_a, _b, _c, _d, _e, _f, _g) \
_dump_node(_a, _b, _c, _d, _e, _f, _g, __func__, __LINE__)
#define dump_areas(_a, _b, _c, _d, _e) \
_dump_areas(_a, _b, _c, _d, _e, __func__, __LINE__)
#define LPC32XX_DEFINE_CLK(_idx, _reg, _e, _em, _d, _dm, _b, _bm, _ops) \
.disable = (_d), \
DEFINE_CLK_RPMH_VRM(rf_clk1, _d, "rfclkd1", 1);
DEFINE_CLK_RPMH_VRM(rf_clk2, _d, "rfclkd2", 1);
DEFINE_CLK_RPMH_VRM(rf_clk3, _d, "rfclkd3", 1);
DEFINE_CLK_RPMH_VRM(rf_clk4, _d, "rfclkd4", 1);
_shift, _width, _cf, _v, _m, _d) \
.div = _d, \
struct dma_device *device, *_d;
list_for_each_entry_safe(device, _d, &dma_device_list, global_node) {
struct dma_device *device, *_d;
list_for_each_entry_safe(device, _d, &dma_device_list, global_node) {
struct dma_device *d, *_d;
list_for_each_entry_safe(d, _d, &dma_device_list, global_node) {
struct dma_device *device, *_d;
list_for_each_entry_safe(device, _d, &dma_device_list, global_node) {
struct ep93xx_dma_desc *desc, *_d;
list_for_each_entry_safe(desc, _d, &list, node)
struct pch_dma_desc *desc, *_d;
list_for_each_entry_safe(desc, _d, &list, desc_node)
struct pch_dma_desc *desc, *_d;
list_for_each_entry_safe(desc, _d, &pd_chan->free_list, desc_node) {
struct pch_dma_desc *desc, *_d;
list_for_each_entry_safe(desc, _d, &tmp_list, desc_node)
struct pch_dma_desc *desc, *_d;
list_for_each_entry_safe(desc, _d, &list, desc_node)
struct d40_desc *_d;
list_for_each_entry_safe(d40d, _d, &d40c->client, node) {
list_for_each_entry_safe(d40d, _d,
struct d40_desc *_d;
list_for_each_entry_safe(d, _d, &d40c->client, node) {
u32 _d = (d); \
PUSH_PRINTF(_p, "%08x"f, _d, ##a); \
*_p->cur++ = _d; \
u64 _a = (a), _c = (c), _d = (d), _o = _a >> s, _s = _c << s; \
if (_d) { \
iowrite##t##_native(_d, &_m[_o++]); \
memset_io(&_m[_o], _d, _s); \
nvkm_wo##t((o), _a, _d); \
#define _ARG13(_1, _2, _3, _4, _5, _6, _7, _8, _9, _a, _b, _c, _d) _1; _2; _3; _4; _5; _6; _7; _8; _9; _a; _b; _c; _d;
#define _ARG14(_1, _2, _3, _4, _5, _6, _7, _8, _9, _a, _b, _c, _d, _e) _1; _2; _3; _4; _5; _6; _7; _8; _9; _a; _b; _c; _d; _e;
#define _ARG15(_1, _2, _3, _4, _5, _6, _7, _8, _9, _a, _b, _c, _d, _e, _f) _1; _2; _3; _4; _5; _6; _7; _8; _9; _a; _b; _c; _d; _e; _f;
#define to_intel_th_driver(_d) \
container_of_const((_d), struct intel_th_driver, driver)
#define to_intel_th_driver_or_null(_d) \
((_d) ? to_intel_th_driver(_d) : NULL)
#define to_intel_th_device(_d) \
container_of((_d), struct intel_th_device, dev)
#define to_stm_device(_d) \
container_of((_d), struct stm_device, dev)
#define to_stm_source_device(_d) \
container_of((_d), struct stm_source_device, dev)
#define DESC_DEC(_d, _s, _m) (((_d) >> (_s)) & (_m))
#define DESC_ENC(_d, _v, _s, _m) \
(_d) &= ~((u64)(_m) << (_s)); \
(_d) |= (((u64)(_v) & (_m)) << (_s)); \
typeof(d) _d = (d); \
_d ? _n / _d : zero; \
#define AR_PHY_POWER_TX_RATE(_d) (AR_SM_BASE + 0x1c0 + ((_d) << 2))
struct vring_tx_desc *_d = &pmc->pring_va[i];
*_d = *d;
volatile struct vring_tx_desc *_d =
_d->dma.status = TX_DMA_STATUS_DU;
volatile struct vring_tx_desc *_d;
_d = &ring->va[i].tx.legacy;
*_d = *d;
_d = &ring->va[i].tx.legacy;
*_d = *d;
volatile struct vring_tx_desc *_d =
_d = &ring->va[i].tx.legacy;
*d = *_d;
_d->dma.status = TX_DMA_STATUS_DU;
*d = *_d;
volatile struct vring_rx_desc *_d =
*d = *_d;
volatile struct vring_tx_desc *_d;
_d = &vring->va[lf].tx.legacy;
if (unlikely(!(_d->dma.status & TX_DMA_STATUS_DU)))
_d = &vring->va[vring->swtail].tx.legacy;
*d = *_d;
volatile struct vring_rx_desc *_d = &vring->va[i].rx.legacy;
*_d = *d;
struct vring_rx_desc *_d;
_d = (struct vring_rx_desc *)&ring->va[ring->swhead].rx.legacy;
if (_d->dma.status & RX_DMA_STATUS_DU)
volatile struct vring_rx_desc *_d;
_d = &vring->va[i].rx.legacy;
if (unlikely(!(_d->dma.status & RX_DMA_STATUS_DU))) {
*d = *_d;
struct wil_tx_enhanced_desc *_d;
_d = (struct wil_tx_enhanced_desc *)
*d = *_d;
struct wil_rx_enhanced_desc *_d = (struct wil_rx_enhanced_desc *)
*_d = *d;
struct wil_tx_enhanced_desc *_d =
*d = *_d;
typeof(_dev) _d = (_dev); \
if (test_bit(RTW89_FLAG_PROBE_DONE, _d->flags)) \
lockdep_assert_wiphy(_d->hw->wiphy); \
_d->sar._cfg_name = *(_cfg_data); \
_d->sar.src = _s; \
struct tsi721_tx_desc *desc, *_d;
list_for_each_entry_safe(desc, _d, &list, desc_node)
static void virtio_ccw_release_dev(struct device *_d)
struct virtio_device *dev = dev_to_virtio(_d);
#define FNIC_TRACE(_fn, _hn, _t, _a, _b, _c, _d, _e) \
trace_buf->data[3] = (u64)(unsigned long)_d; \
u32 *_d = dest->q.state;
*_d++ = *_a++ & *_b++;
u32 *_d = dest->q.state;
*_d++ = *_a++ ^ *_b++;
static int udc_debug_show(struct seq_file *m, void *_d)
static ssize_t device_show(struct device *_d,
struct virtio_device *dev = dev_to_virtio(_d);
static ssize_t vendor_show(struct device *_d,
struct virtio_device *dev = dev_to_virtio(_d);
static int virtio_dev_probe(struct device *_d)
struct virtio_device *dev = dev_to_virtio(_d);
static ssize_t status_show(struct device *_d,
struct virtio_device *dev = dev_to_virtio(_d);
static void virtio_dev_remove(struct device *_d)
struct virtio_device *dev = dev_to_virtio(_d);
static ssize_t modalias_show(struct device *_d,
static const struct cpumask *virtio_irq_get_affinity(struct device *_d,
struct virtio_device *dev = dev_to_virtio(_d);
static void virtio_dev_shutdown(struct device *_d)
struct virtio_device *dev = dev_to_virtio(_d);
struct virtio_device *dev = dev_to_virtio(_d);
static ssize_t features_show(struct device *_d,
struct virtio_device *dev = dev_to_virtio(_d);
static void virtio_mmio_release_dev(struct device *_d)
container_of(_d, struct virtio_device, dev);
static void virtio_pci_release_dev(struct device *_d)
struct virtio_device *vdev = dev_to_virtio(_d);
static void virtio_vdpa_release_dev(struct device *_d)
container_of(_d, struct virtio_device, dev);
#define __BPF_DISPATCHER_UPDATE(_d, _new) \
__static_call_update((_d)->sc_key, (_d)->sc_tramp, (_new))
#define __BPF_DISPATCHER_UPDATE(_d, _new)
typeof(d) _d = d; \
*_d = (_a_full << _to_shift); \
(_to_shift != _s || *_d < 0 || _a < 0 || \
(*_d >> _to_shift) != _a); \
_FP_W_TYPE _d; \
doit(_d, _FP_FRAC_WORD_4(_z,0), X##_f0, Y##_f0); \
_FP_FRAC_WORD_4(_z,1), (_c1 & _c2), 0, _d, \
0, _d, _FP_FRAC_WORD_4(_z,0)); \
_FP_FRAC_DECL_2(_d); _FP_FRAC_DECL_2(_e); _FP_FRAC_DECL_2(_f); \
#define GF_POLY_SZ(_d) (sizeof(struct gf_poly)+((_d)+1)*sizeof(unsigned int))
u8 *d, _d;
sizeof(_d), &_d);
#define TEST_BTF_DUMP_DATA(_b, _d, _prefix, _str, _type, _flags, \
(void) btf_dump_data(_b, _d, _ptrtype, _prefix, _flags, \
#define TEST_BTF_DUMP_DATA_C(_b, _d, _prefix, _str, _type, _flags, \
TEST_BTF_DUMP_DATA(_b, _d, _prefix, _str, _type, _flags, \
#define TEST_BTF_DUMP_DATA_OVER(_b, _d, _prefix, _str, _type, _type_sz, \
(void) btf_dump_data(_b, _d, _ptrtype, _prefix, 0, \
#define TEST_BTF_DUMP_VAR(_b, _d, _prefix, _str, _var, _type, _flags, \
(void) btf_dump_data(_b, _d, _var, _prefix, _flags, \
#define VGIC_DEV_IS_V2(_d) ((_d) == KVM_DEV_TYPE_ARM_VGIC_V2)
#define VGIC_DEV_IS_V3(_d) ((_d) == KVM_DEV_TYPE_ARM_VGIC_V3)
type_t _a = __val1, _d = __val1, rm = __val2; \
type_t a = _a, d = _d, ex_a = _a, ex_d = _d; \
ex_v, insn, (uint64_t)_a, (uint64_t)_d, (uint64_t)rm, v); \
(uint64_t)_d, (uint64_t)rm, (uint64_t)a, (uint64_t)d); \
ex_flags, insn, (uint64_t)_a, (uint64_t)_d, (uint64_t)rm, flags);\
#define vfio_pci_config_readb(_d, _o) vfio_pci_config_read(_d, _o, u8)
#define vfio_pci_config_readw(_d, _o) vfio_pci_config_read(_d, _o, u16)
#define vfio_pci_config_readl(_d, _o) vfio_pci_config_read(_d, _o, u32)
#define vfio_pci_config_writeb(_d, _o, _v) vfio_pci_config_write(_d, _o, _v, u8)
#define vfio_pci_config_writew(_d, _o, _v) vfio_pci_config_write(_d, _o, _v, u16)
#define vfio_pci_config_writel(_d, _o, _v) vfio_pci_config_write(_d, _o, _v, u32)