Symbol: mmio_read
arch/x86/coco/tdx/tdx.c
655
if (!mmio_read(size, ve->gpa, &val))
drivers/fpga/dfl-fme-perf.c
640
FME_EVENT_FABRIC(mmio_read, FAB_EVNT_MMIO_RD);
drivers/fpga/dfl-fme-perf.c
649
FME_EVENT_FABRIC_PORT(mmio_read, FAB_EVNT_MMIO_RD);
drivers/gpu/drm/i915/intel_uncore.h
356
return uncore->funcs.mmio_read##s__(uncore, reg, (trace__)); \
drivers/infiniband/hw/efa/efa_com.c
104
if (read_resp->req_id != mmio_read->seq_num) {
drivers/infiniband/hw/efa/efa_com.c
108
mmio_read->seq_num, offset, read_resp->req_id,
drivers/infiniband/hw/efa/efa_com.c
124
spin_unlock(&mmio_read->lock);
drivers/infiniband/hw/efa/efa_com.c
78
struct efa_com_mmio_read *mmio_read = &edev->mmio_read;
drivers/infiniband/hw/efa/efa_com.c
84
read_resp = mmio_read->read_resp;
drivers/infiniband/hw/efa/efa_com.c
86
spin_lock(&mmio_read->lock);
drivers/infiniband/hw/efa/efa_com.c
87
mmio_read->seq_num++;
drivers/infiniband/hw/efa/efa_com.c
897
struct efa_com_mmio_read *mmio_read = &edev->mmio_read;
drivers/infiniband/hw/efa/efa_com.c
90
read_resp->req_id = mmio_read->seq_num + 0x9aL;
drivers/infiniband/hw/efa/efa_com.c
902
addr_high = (mmio_read->read_resp_dma_addr >> 32) & GENMASK(31, 0);
drivers/infiniband/hw/efa/efa_com.c
903
addr_low = mmio_read->read_resp_dma_addr & GENMASK(31, 0);
drivers/infiniband/hw/efa/efa_com.c
911
struct efa_com_mmio_read *mmio_read = &edev->mmio_read;
drivers/infiniband/hw/efa/efa_com.c
913
spin_lock_init(&mmio_read->lock);
drivers/infiniband/hw/efa/efa_com.c
914
mmio_read->read_resp =
drivers/infiniband/hw/efa/efa_com.c
915
dma_alloc_coherent(edev->dmadev, sizeof(*mmio_read->read_resp),
drivers/infiniband/hw/efa/efa_com.c
916
&mmio_read->read_resp_dma_addr, GFP_KERNEL);
drivers/infiniband/hw/efa/efa_com.c
917
if (!mmio_read->read_resp)
drivers/infiniband/hw/efa/efa_com.c
922
mmio_read->read_resp->req_id = 0;
drivers/infiniband/hw/efa/efa_com.c
923
mmio_read->seq_num = 0;
drivers/infiniband/hw/efa/efa_com.c
924
mmio_read->mmio_read_timeout = EFA_REG_READ_TIMEOUT_US;
drivers/infiniband/hw/efa/efa_com.c
93
mmio_read->seq_num);
drivers/infiniband/hw/efa/efa_com.c
931
struct efa_com_mmio_read *mmio_read = &edev->mmio_read;
drivers/infiniband/hw/efa/efa_com.c
933
dma_free_coherent(edev->dmadev, sizeof(*mmio_read->read_resp),
drivers/infiniband/hw/efa/efa_com.c
934
mmio_read->read_resp, mmio_read->read_resp_dma_addr);
drivers/infiniband/hw/efa/efa_com.c
97
exp_time = jiffies + usecs_to_jiffies(mmio_read->mmio_read_timeout);
drivers/infiniband/hw/efa/efa_com.c
99
if (READ_ONCE(read_resp->req_id) == mmio_read->seq_num)
drivers/infiniband/hw/efa/efa_com.h
115
struct efa_com_mmio_read mmio_read;
drivers/infiniband/hw/efa/efa_main.c
235
edev->mmio_read.mmio_read_timeout =
drivers/net/ethernet/amazon/ena/ena_com.c
1913
struct ena_com_mmio_read *mmio_read = &ena_dev->mmio_read;
drivers/net/ethernet/amazon/ena/ena_com.c
1915
spin_lock_init(&mmio_read->lock);
drivers/net/ethernet/amazon/ena/ena_com.c
1916
mmio_read->read_resp = dma_alloc_coherent(ena_dev->dmadev, sizeof(*mmio_read->read_resp),
drivers/net/ethernet/amazon/ena/ena_com.c
1917
&mmio_read->read_resp_dma_addr, GFP_KERNEL);
drivers/net/ethernet/amazon/ena/ena_com.c
1918
if (unlikely(!mmio_read->read_resp))
drivers/net/ethernet/amazon/ena/ena_com.c
1923
mmio_read->read_resp->req_id = 0x0;
drivers/net/ethernet/amazon/ena/ena_com.c
1924
mmio_read->seq_num = 0x0;
drivers/net/ethernet/amazon/ena/ena_com.c
1925
mmio_read->readless_supported = true;
drivers/net/ethernet/amazon/ena/ena_com.c
1936
struct ena_com_mmio_read *mmio_read = &ena_dev->mmio_read;
drivers/net/ethernet/amazon/ena/ena_com.c
1938
mmio_read->readless_supported = readless_supported;
drivers/net/ethernet/amazon/ena/ena_com.c
1943
struct ena_com_mmio_read *mmio_read = &ena_dev->mmio_read;
drivers/net/ethernet/amazon/ena/ena_com.c
1948
dma_free_coherent(ena_dev->dmadev, sizeof(*mmio_read->read_resp), mmio_read->read_resp,
drivers/net/ethernet/amazon/ena/ena_com.c
1949
mmio_read->read_resp_dma_addr);
drivers/net/ethernet/amazon/ena/ena_com.c
1951
mmio_read->read_resp = NULL;
drivers/net/ethernet/amazon/ena/ena_com.c
1956
struct ena_com_mmio_read *mmio_read = &ena_dev->mmio_read;
drivers/net/ethernet/amazon/ena/ena_com.c
1959
addr_low = ENA_DMA_ADDR_TO_UINT32_LOW(mmio_read->read_resp_dma_addr);
drivers/net/ethernet/amazon/ena/ena_com.c
1960
addr_high = ENA_DMA_ADDR_TO_UINT32_HIGH(mmio_read->read_resp_dma_addr);
drivers/net/ethernet/amazon/ena/ena_com.c
796
struct ena_com_mmio_read *mmio_read = &ena_dev->mmio_read;
drivers/net/ethernet/amazon/ena/ena_com.c
798
mmio_read->read_resp;
drivers/net/ethernet/amazon/ena/ena_com.c
801
u32 timeout = mmio_read->reg_read_to;
drivers/net/ethernet/amazon/ena/ena_com.c
809
if (!mmio_read->readless_supported)
drivers/net/ethernet/amazon/ena/ena_com.c
812
spin_lock_irqsave(&mmio_read->lock, flags);
drivers/net/ethernet/amazon/ena/ena_com.c
813
mmio_read->seq_num++;
drivers/net/ethernet/amazon/ena/ena_com.c
815
read_resp->req_id = mmio_read->seq_num + 0xDEAD;
drivers/net/ethernet/amazon/ena/ena_com.c
818
mmio_read_reg |= mmio_read->seq_num &
drivers/net/ethernet/amazon/ena/ena_com.c
824
if (READ_ONCE(read_resp->req_id) == mmio_read->seq_num)
drivers/net/ethernet/amazon/ena/ena_com.c
833
mmio_read->seq_num, offset, read_resp->req_id, read_resp->reg_off);
drivers/net/ethernet/amazon/ena/ena_com.c
845
spin_unlock_irqrestore(&mmio_read->lock, flags);
drivers/net/ethernet/amazon/ena/ena_com.h
368
struct ena_com_mmio_read mmio_read;
drivers/net/ethernet/amazon/ena/ena_netdev.c
3646
adapter->ena_dev->mmio_read.reg_read_to =
drivers/video/fbdev/broadsheetfb.c
252
if (par->board->mmio_read)
drivers/video/fbdev/broadsheetfb.c
253
return par->board->mmio_read(par);
include/video/broadsheetfb.h
72
u16 (*mmio_read)(struct broadsheetfb_par *);
kernel/trace/trace_mmiotrace.c
282
.read = mmio_read,