Symbol: rr
arch/arm/mach-rpc/include/mach/acornfb.h
48
u_int rr, v, p;
arch/arm/mach-rpc/include/mach/acornfb.h
51
rr = 41667 * r;
arch/arm/mach-rpc/include/mach/acornfb.h
53
v = (rr + pixclk / 2) / pixclk;
arch/arm/mach-rpc/include/mach/acornfb.h
58
p = (rr + v / 2) / v;
arch/mips/include/uapi/asm/inst.h
1165
struct m16e_rr rr;
arch/mips/kernel/branch.c
374
if (inst.rr.func == MIPS16e_jr_func) {
arch/mips/kernel/branch.c
376
if (inst.rr.ra)
arch/mips/kernel/branch.c
380
regs->regs[reg16to32[inst.rr.rx]];
arch/mips/kernel/branch.c
382
if (inst.rr.l) {
arch/mips/kernel/branch.c
383
if (inst.rr.nd)
arch/powerpc/kvm/book3s_32_mmu_host.c
139
register int rr = 0;
arch/powerpc/kvm/book3s_32_mmu_host.c
169
if (rr == 16) {
arch/powerpc/kvm/book3s_32_mmu_host.c
172
rr = 0;
arch/powerpc/kvm/book3s_32_mmu_host.c
178
if (!evict && (pteg[rr] & PTE_V)) {
arch/powerpc/kvm/book3s_32_mmu_host.c
179
rr += 2;
arch/powerpc/kvm/book3s_32_mmu_host.c
183
dprintk_mmu("KVM: old PTEG: %p (%d)\n", pteg, rr);
arch/powerpc/kvm/book3s_32_mmu_host.c
209
if (pteg[rr]) {
arch/powerpc/kvm/book3s_32_mmu_host.c
210
pteg[rr] = 0;
arch/powerpc/kvm/book3s_32_mmu_host.c
213
pteg[rr + 1] = pteg1;
arch/powerpc/kvm/book3s_32_mmu_host.c
214
pteg[rr] = pteg0;
arch/powerpc/kvm/book3s_32_mmu_host.c
245
pte->slot = (ulong)&pteg[rr];
arch/x86/mm/numa.c
163
int rr, i;
arch/x86/mm/numa.c
165
rr = first_node(node_online_map);
arch/x86/mm/numa.c
169
numa_set_node(i, rr);
arch/x86/mm/numa.c
170
rr = next_node_in(rr, node_online_map);
block/partitions/acorn.c
75
struct riscix_record *rr;
block/partitions/acorn.c
77
rr = read_part_sector(state, first_sect, &sect);
block/partitions/acorn.c
78
if (!rr)
block/partitions/acorn.c
84
if (rr->magic == RISCIX_MAGIC) {
block/partitions/acorn.c
92
if (rr->part[part].one &&
block/partitions/acorn.c
93
memcmp(rr->part[part].name, "All\0", 4)) {
block/partitions/acorn.c
95
le32_to_cpu(rr->part[part].start),
block/partitions/acorn.c
96
le32_to_cpu(rr->part[part].length));
block/partitions/acorn.c
98
strlcat(state->pp_buf, rr->part[part].name, PAGE_SIZE);
crypto/camellia_generic.c
319
#define ROLDQ(ll, lr, rl, rr, w0, w1, bits) ({ \
crypto/camellia_generic.c
323
rl = (rl << bits) + (rr >> (32 - bits)); \
crypto/camellia_generic.c
324
rr = (rr << bits) + (w0 >> (32 - bits)); \
crypto/camellia_generic.c
327
#define ROLDQo32(ll, lr, rl, rr, w0, w1, bits) ({ \
crypto/camellia_generic.c
331
lr = (rl << (bits - 32)) + (rr >> (64 - bits)); \
crypto/camellia_generic.c
332
rl = (rr << (bits - 32)) + (w0 >> (64 - bits)); \
crypto/camellia_generic.c
333
rr = (w0 << (bits - 32)) + (w1 >> (64 - bits)); \
crypto/camellia_generic.c
811
#define CAMELLIA_FLS(ll, lr, rl, rr, kll, klr, krl, krr, t0, t1, t2, t3) ({ \
crypto/camellia_generic.c
815
t2 |= rr; \
crypto/camellia_generic.c
823
rr ^= rol32(t3, 1); \
crypto/cast5_generic.c
339
if (!(c->rr)) {
crypto/cast5_generic.c
371
if (!(c->rr)) {
crypto/cast5_generic.c
484
c->rr = key_len <= 10 ? 1 : 0;
drivers/acpi/arm64/iort.c
1048
list_add_tail(&rmr_data->rr.list, head);
drivers/acpi/arm64/iort.c
926
rmr_data = container_of(region, struct iommu_iort_rmr_data, rr);
drivers/acpi/arm64/iort.c
965
region = &rmr_data->rr;
drivers/acpi/reboot.c
11
static void acpi_pci_reboot(struct acpi_generic_address *rr, u8 reset_value)
drivers/acpi/reboot.c
21
devfn = PCI_DEVFN((rr->address >> 32) & 0xffff,
drivers/acpi/reboot.c
22
(rr->address >> 16) & 0xffff);
drivers/acpi/reboot.c
26
(rr->address & 0xffff), reset_value);
drivers/acpi/reboot.c
29
static inline void acpi_pci_reboot(struct acpi_generic_address *rr,
drivers/acpi/reboot.c
38
struct acpi_generic_address *rr;
drivers/acpi/reboot.c
44
rr = &acpi_gbl_FADT.reset_register;
drivers/acpi/reboot.c
61
switch (rr->space_id) {
drivers/acpi/reboot.c
63
acpi_pci_reboot(rr, reset_value);
drivers/block/drbd/drbd_main.c
2139
int rr;
drivers/block/drbd/drbd_main.c
2141
rr = drbd_free_peer_reqs(device, &device->active_ee);
drivers/block/drbd/drbd_main.c
2142
if (rr)
drivers/block/drbd/drbd_main.c
2143
drbd_err(device, "%d EEs in active list found!\n", rr);
drivers/block/drbd/drbd_main.c
2145
rr = drbd_free_peer_reqs(device, &device->sync_ee);
drivers/block/drbd/drbd_main.c
2146
if (rr)
drivers/block/drbd/drbd_main.c
2147
drbd_err(device, "%d EEs in sync list found!\n", rr);
drivers/block/drbd/drbd_main.c
2149
rr = drbd_free_peer_reqs(device, &device->read_ee);
drivers/block/drbd/drbd_main.c
2150
if (rr)
drivers/block/drbd/drbd_main.c
2151
drbd_err(device, "%d EEs in read list found!\n", rr);
drivers/block/drbd/drbd_main.c
2153
rr = drbd_free_peer_reqs(device, &device->done_ee);
drivers/block/drbd/drbd_main.c
2154
if (rr)
drivers/block/drbd/drbd_main.c
2155
drbd_err(device, "%d EEs in done list found!\n", rr);
drivers/block/drbd/drbd_receiver.c
659
int rr;
drivers/block/drbd/drbd_receiver.c
665
rr = drbd_recv_short(*sock, tb, 4, MSG_DONTWAIT | MSG_PEEK);
drivers/block/drbd/drbd_receiver.c
667
if (rr > 0 || rr == -EAGAIN) {
drivers/cxl/core/region.c
955
struct cxl_region_ref *rr = cxl_rr_load(port, cxlr_iter);
drivers/cxl/core/region.c
956
struct cxl_decoder *cxld_iter = rr->decoder;
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
1376
uint8_t rr = le16_to_cpu(lvds->usSupportedRefreshRate);
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
1378
if (SUPPORTED_LCD_REFRESHRATE_30Hz & rr)
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
1380
else if (SUPPORTED_LCD_REFRESHRATE_40Hz & rr)
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
1382
else if (SUPPORTED_LCD_REFRESHRATE_48Hz & rr)
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
1384
else if (SUPPORTED_LCD_REFRESHRATE_50Hz & rr)
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
1386
else if (SUPPORTED_LCD_REFRESHRATE_60Hz & rr)
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
1502
uint8_t rr = lvds->sRefreshRateSupport.ucSupportedRefreshRate;
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
1516
if (SUPPORTED_LCD_REFRESHRATE_30Hz & rr)
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
1518
else if (SUPPORTED_LCD_REFRESHRATE_40Hz & rr)
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
1520
else if (SUPPORTED_LCD_REFRESHRATE_48Hz & rr)
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
1522
else if (SUPPORTED_LCD_REFRESHRATE_50Hz & rr)
drivers/gpu/drm/amd/display/dc/bios/bios_parser.c
1524
else if (SUPPORTED_LCD_REFRESHRATE_60Hz & rr)
drivers/gpu/drm/omapdrm/dss/dispc.c
1290
coef_r = FLD_VAL(coefs->rr, 31, 22) | FLD_VAL(coefs->rg, 20, 11) |
drivers/gpu/drm/omapdrm/dss/omapdss.h
144
s16 rr, rg, rb;
drivers/gpu/drm/omapdrm/omap_crtc.c
385
cpr->rr = omap_crtc_s31_32_to_s2_8(ctm->matrix[0]);
drivers/gpu/drm/tidss/tidss_dispc.c
1299
int r = rate / 100, rr = real_rate / 100;
drivers/gpu/drm/tidss/tidss_dispc.c
1301
return (unsigned int)(abs(((rr - r) * 100) / r));
drivers/gpu/drm/tilcdc/tilcdc_crtc.c
197
int r = rate / 100, rr = real_rate / 100;
drivers/gpu/drm/tilcdc/tilcdc_crtc.c
199
return (unsigned int)(abs(((rr - r) * 100) / r));
drivers/gpu/drm/xe/xe_gt_sriov_vf.c
1082
struct vf_runtime_reg *rr;
drivers/gpu/drm/xe/xe_gt_sriov_vf.c
1093
rr = vf_lookup_reg(gt, addr);
drivers/gpu/drm/xe/xe_gt_sriov_vf.c
1094
if (!rr) {
drivers/gpu/drm/xe/xe_gt_sriov_vf.c
1101
xe_gt_sriov_dbg_verbose(gt, "runtime[%#x] = %#x\n", addr, rr->value);
drivers/gpu/drm/xe/xe_gt_sriov_vf.c
1102
return rr->value;
drivers/hwmon/dme1737.c
498
int rr = (ix == 1) ? reg >> 4 : reg;
drivers/hwmon/dme1737.c
500
return (rr & 0x08) ? PWM_RR[rr & 0x07] : 0;
drivers/i3c/master/i3c-master-cdns.c
1038
u32 rr;
drivers/i3c/master/i3c-master-cdns.c
1041
rr = readl(master->regs + DEV_ID_RR0(slot));
drivers/i3c/master/i3c-master-cdns.c
1042
info->dyn_addr = DEV_ID_RR0_GET_DEV_ADDR(rr);
drivers/i3c/master/i3c-master-cdns.c
1043
rr = readl(master->regs + DEV_ID_RR2(slot));
drivers/i3c/master/i3c-master-cdns.c
1044
info->dcr = rr;
drivers/i3c/master/i3c-master-cdns.c
1045
info->bcr = rr >> 8;
drivers/i3c/master/i3c-master-cdns.c
1046
info->pid = rr >> 16;
drivers/i3c/master/i3c-master-cdns.c
891
u32 rr;
drivers/i3c/master/i3c-master-cdns.c
893
rr = prepare_rr0_dev_address(dev->info.dyn_addr ?
drivers/i3c/master/i3c-master-cdns.c
896
writel(DEV_ID_RR0_IS_I3C | rr, master->regs + DEV_ID_RR0(data->id));
drivers/i3c/master/i3c-master-cdns.c
903
u32 rr;
drivers/i3c/master/i3c-master-cdns.c
917
rr = readl(master->regs + DEV_ID_RR0(i));
drivers/i3c/master/i3c-master-cdns.c
918
if (!(rr & DEV_ID_RR0_IS_I3C) ||
drivers/i3c/master/i3c-master-cdns.c
919
DEV_ID_RR0_GET_DEV_ADDR(rr) != dyn_addr)
drivers/infiniband/hw/erdma/erdma_cm.c
336
struct mpa_rr *rr = &cep->mpa.hdr;
drivers/infiniband/hw/erdma/erdma_cm.c
345
rr->params.pd_len = cpu_to_be16(pd_len);
drivers/infiniband/hw/erdma/erdma_cm.c
347
iov[iovec_num].iov_base = rr;
drivers/infiniband/hw/erdma/erdma_cm.c
348
iov[iovec_num].iov_len = sizeof(*rr);
drivers/infiniband/hw/erdma/erdma_cm.c
350
mpa_len = sizeof(*rr);
drivers/infiniband/hw/mlx5/main.c
3465
MLX5_SET(mkc, mkc, rr, 1);
drivers/infiniband/hw/mlx5/mr.c
71
MLX5_SET(mkc, mkc, rr, !!(acc & IB_ACCESS_REMOTE_READ));
drivers/infiniband/hw/mlx5/odp.c
578
MLX5_SET(mkc, mkc, rr, 1);
drivers/infiniband/hw/mlx5/umr.c
452
MLX5_SET(mkc, seg, rr, !!(access_flags & IB_ACCESS_REMOTE_READ));
drivers/infiniband/sw/siw/siw_cm.c
517
struct mpa_rr *rr = &cep->mpa.hdr;
drivers/infiniband/sw/siw/siw_cm.c
526
iov[iovec_num].iov_base = rr;
drivers/infiniband/sw/siw/siw_cm.c
527
iov[iovec_num].iov_len = sizeof(*rr);
drivers/infiniband/sw/siw/siw_cm.c
528
mpa_len = sizeof(*rr);
drivers/infiniband/sw/siw/siw_cm.c
545
rr->params.pd_len = cpu_to_be16(pd_len);
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
4753
rmr = container_of(e, struct iommu_iort_rmr_data, rr);
drivers/iommu/arm/arm-smmu/arm-smmu.c
2097
rmr = container_of(e, struct iommu_iort_rmr_data, rr);
drivers/isdn/mISDN/dsp_cmx.c
1305
int r, rr, t, tt, o_r, o_rr;
drivers/isdn/mISDN/dsp_cmx.c
1363
rr = (r + len) & CMX_BUFF_MASK;
drivers/isdn/mISDN/dsp_cmx.c
1385
while (r != rr && t != tt) {
drivers/isdn/mISDN/dsp_cmx.c
1395
if (r == rr) {
drivers/isdn/mISDN/dsp_cmx.c
1412
while (r != rr && t != tt) {
drivers/isdn/mISDN/dsp_cmx.c
1417
if (r != rr) {
drivers/isdn/mISDN/dsp_cmx.c
1421
memset(d, dsp_silence, (rr - r) & CMX_BUFF_MASK);
drivers/isdn/mISDN/dsp_cmx.c
1429
while (r != rr && t != tt) {
drivers/isdn/mISDN/dsp_cmx.c
1434
while (r != rr) {
drivers/isdn/mISDN/dsp_cmx.c
1457
o_r = (o_rr - rr + r) & CMX_BUFF_MASK;
drivers/isdn/mISDN/dsp_cmx.c
1480
while (r != rr && t != tt) {
drivers/isdn/mISDN/dsp_cmx.c
1494
while (r != rr) {
drivers/isdn/mISDN/dsp_cmx.c
1510
while (r != rr && t != tt) {
drivers/isdn/mISDN/dsp_cmx.c
1522
while (r != rr) {
drivers/isdn/mISDN/dsp_cmx.c
1538
while (r != rr && t != tt) {
drivers/isdn/mISDN/dsp_cmx.c
1549
while (r != rr) {
drivers/isdn/mISDN/dsp_cmx.c
1626
int r, rr;
drivers/isdn/mISDN/dsp_cmx.c
1706
rr = (r + length) & CMX_BUFF_MASK;
drivers/isdn/mISDN/dsp_cmx.c
1708
while (r != rr) {
drivers/isdn/mISDN/dsp_cmx.c
1732
rr = (r + length) & CMX_BUFF_MASK;
drivers/isdn/mISDN/dsp_cmx.c
1734
while (r != rr) {
drivers/isdn/mISDN/dsp_cmx.c
1778
rr = (r + delay - (dsp_poll >> 1))
drivers/isdn/mISDN/dsp_cmx.c
1781
while (r != rr) {
drivers/isdn/mISDN/dsp_cmx.c
1809
rr = (r + delay - (dsp_poll >> 1))
drivers/isdn/mISDN/dsp_cmx.c
1812
while (r != rr) {
drivers/md/dm-ebs-target.c
124
int r = 0, rr;
drivers/md/dm-ebs-target.c
129
rr = __ebs_rw_bvec(ec, op, &bv, &iter);
drivers/md/dm-ebs-target.c
130
if (rr)
drivers/md/dm-ebs-target.c
131
r = rr;
drivers/media/i2c/tda7432.c
276
u8 lf, lr, rf, rr;
drivers/media/i2c/tda7432.c
282
rr = rf = -t->balance->val;
drivers/media/i2c/tda7432.c
286
rr = rf = TDA7432_ATTEN_0DB;
drivers/media/i2c/tda7432.c
290
rr = rf = TDA7432_ATTEN_0DB;
drivers/media/i2c/tda7432.c
297
rr |= TDA7432_MUTE;
drivers/media/i2c/tda7432.c
303
tda7432_write(sd, TDA7432_RR, rr);
drivers/mtd/nand/raw/nand_hynix.c
290
struct hynix_read_retry *rr = NULL;
drivers/mtd/nand/raw/nand_hynix.c
314
rr = kzalloc(sizeof(*rr) + (nregs * nmodes), GFP_KERNEL);
drivers/mtd/nand/raw/nand_hynix.c
315
if (!rr) {
drivers/mtd/nand/raw/nand_hynix.c
322
u8 *val = rr->values + (i * nregs);
drivers/mtd/nand/raw/nand_hynix.c
336
rr->nregs = nregs;
drivers/mtd/nand/raw/nand_hynix.c
337
rr->regs = hynix_1xnm_mlc_read_retry_regs;
drivers/mtd/nand/raw/nand_hynix.c
338
hynix->read_retry = rr;
drivers/mtd/nand/raw/nand_hynix.c
346
kfree(rr);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7054
req->rr = cpu_to_le16(bp->grp_info[grp_idx].rx_fw_ring_id);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
69
MLX5_SET(mkc, mkc, rr, 1);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_send.c
1147
MLX5_SET(mkc, mkc, rr, 1);
drivers/net/ethernet/spacemit/k1_emac.c
462
struct emac_desc_ring *rr = &priv->rx_ring;
drivers/net/ethernet/spacemit/k1_emac.c
467
kfree(rr->rx_desc_buf);
drivers/net/ethernet/spacemit/k1_emac.c
468
rr->rx_desc_buf = NULL;
drivers/net/ethernet/spacemit/k1_emac.c
470
dma_free_coherent(dev, rr->total_size, rr->desc_addr,
drivers/net/ethernet/spacemit/k1_emac.c
471
rr->desc_dma_addr);
drivers/net/ethernet/spacemit/k1_emac.c
472
rr->desc_addr = NULL;
drivers/net/wireless/ath/ath10k/wmi.c
4054
const struct phyerr_radar_report *rr,
drivers/net/wireless/ath/ath10k/wmi.c
4065
reg0 = __le32_to_cpu(rr->reg0);
drivers/net/wireless/ath/ath10k/wmi.c
4066
reg1 = __le32_to_cpu(rr->reg1);
drivers/net/wireless/ath/ath10k/wmi.c
4214
const struct phyerr_radar_report *rr;
drivers/net/wireless/ath/ath10k/wmi.c
4246
if (i + sizeof(*tlv) + sizeof(*rr) > buf_len) {
drivers/net/wireless/ath/ath10k/wmi.c
4252
rr = (struct phyerr_radar_report *)tlv_buf;
drivers/net/wireless/ath/ath10k/wmi.c
4253
ath10k_dfs_radar_report(ar, phyerr, rr, tsf);
drivers/net/wireless/ath/carl9170/tx.c
941
txc->s.rr[i - 1] = phy_set;
drivers/net/wireless/ath/carl9170/wlan.h
260
struct ar9170_tx_hw_phy_control rr[CARL9170_TX_MAX_RETRY_RATES];
drivers/net/wireless/ath/carl9170/wlan.h
324
__le32 rr[CARL9170_TX_MAX_RETRY_RATES];
drivers/net/wireless/broadcom/brcm80211/brcmsmac/ampdu.c
501
bool rr;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/ampdu.c
543
rr = true;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/ampdu.c
546
rr = false;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/ampdu.c
554
if (rr) {
drivers/net/wireless/mediatek/mt76/mmio.c
99
.rr = mt76_mmio_rr,
drivers/net/wireless/mediatek/mt76/mt76.h
1170
#define __mt76_rr(dev, ...) (dev)->bus->rr((dev), __VA_ARGS__)
drivers/net/wireless/mediatek/mt76/mt76.h
1179
#define mt76_rr(dev, ...) (dev)->mt76.bus->rr(&((dev)->mt76), __VA_ARGS__)
drivers/net/wireless/mediatek/mt76/mt76.h
94
u32 (*rr)(struct mt76_dev *dev, u32 offset);
drivers/net/wireless/mediatek/mt76/mt7603/init.c
405
return dev->bus_ops->rr(mdev, addr);
drivers/net/wireless/mediatek/mt76/mt7603/init.c
508
bus_ops->rr = mt7603_rr;
drivers/net/wireless/mediatek/mt76/mt7615/mmio.c
147
return dev->bus_ops->rr(mdev, addr);
drivers/net/wireless/mediatek/mt76/mt7615/mmio.c
219
bus_ops->rr = mt7615_rr;
drivers/net/wireless/mediatek/mt76/mt7615/sdio.c
93
.rr = mt76s_rr,
drivers/net/wireless/mediatek/mt76/mt7615/usb.c
129
.rr = mt7663u_rr,
drivers/net/wireless/mediatek/mt76/mt76x02_mac.h
158
switch (dev->bus->rr(dev, MAC_CSR0)) {
drivers/net/wireless/mediatek/mt76/mt7915/mmio.c
432
dev->bus_ops->rr(&dev->mt76, l1_remap);
drivers/net/wireless/mediatek/mt76/mt7915/mmio.c
450
dev->bus_ops->rr(&dev->mt76, MT_HIF_REMAP_L2);
drivers/net/wireless/mediatek/mt76/mt7915/mmio.c
462
dev->bus_ops->rr(&dev->mt76, MT_HIF_REMAP_L2_MT7916 + ofs);
drivers/net/wireless/mediatek/mt76/mt7915/mmio.c
541
return dev->bus_ops->rr(mdev, addr);
drivers/net/wireless/mediatek/mt76/mt7915/mmio.c
544
val = dev->bus_ops->rr(mdev, __mt7915_reg_remap_addr(dev, offset));
drivers/net/wireless/mediatek/mt76/mt7915/mmio.c
781
bus_ops->rr = mt7915_rr;
drivers/net/wireless/mediatek/mt76/mt7921/pci.c
149
return dev->bus_ops->rr(mdev, addr);
drivers/net/wireless/mediatek/mt76/mt7921/pci.c
342
bus_ops->rr = mt7921_rr;
drivers/net/wireless/mediatek/mt76/mt7921/sdio.c
109
.rr = mt76s_rr,
drivers/net/wireless/mediatek/mt76/mt7921/usb.c
171
.rr = mt792xu_rr,
drivers/net/wireless/mediatek/mt76/mt7925/pci.c
100
dev->bus_ops->rr(&dev->mt76, MT_HIF_REMAP_L1);
drivers/net/wireless/mediatek/mt76/mt7925/pci.c
194
return dev->bus_ops->rr(mdev, addr);
drivers/net/wireless/mediatek/mt76/mt7925/pci.c
376
bus_ops->rr = mt7925_rr;
drivers/net/wireless/mediatek/mt76/mt7925/pci.c
76
dev->backup_l1 = dev->bus_ops->rr(&dev->mt76, MT_HIF_REMAP_L1);
drivers/net/wireless/mediatek/mt76/mt7925/pci.c
83
dev->bus_ops->rr(&dev->mt76, MT_HIF_REMAP_L1);
drivers/net/wireless/mediatek/mt76/mt7925/pci.c
92
dev->backup_l2 = dev->bus_ops->rr(&dev->mt76, MT_HIF_REMAP_L1);
drivers/net/wireless/mediatek/mt76/mt7925/usb.c
158
.rr = mt792xu_rr,
drivers/net/wireless/mediatek/mt76/mt7996/mmio.c
275
dev->bus_ops->rr(&dev->mt76, MT_HIF_REMAP_L1);
drivers/net/wireless/mediatek/mt76/mt7996/mmio.c
298
dev->bus_ops->rr(&dev->mt76, MT_HIF_REMAP_L2);
drivers/net/wireless/mediatek/mt76/mt7996/mmio.c
312
dev->bus_ops->rr(&dev->mt76, MT_HIF_REMAP_CBTOP);
drivers/net/wireless/mediatek/mt76/mt7996/mmio.c
386
return dev->bus_ops->rr(mdev, addr);
drivers/net/wireless/mediatek/mt76/mt7996/mmio.c
389
val = dev->bus_ops->rr(mdev, __mt7996_reg_remap_addr(dev, offset));
drivers/net/wireless/mediatek/mt76/mt7996/mmio.c
667
bus_ops->rr = mt7996_rr;
drivers/net/wireless/mediatek/mt76/usb.c
1123
.rr = mt76u_rr,
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1627
struct read_regs_int *rr = &intr->read_regs;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1628
struct usb_int_regs *regs = (struct usb_int_regs *)rr->buffer;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1633
if (rr->length < struct_size(regs, regs, count)) {
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1636
rr->length, struct_size(regs, regs, count));
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1640
if (rr->length > sizeof(rr->buffer)) {
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1643
rr->length, sizeof(rr->buffer));
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1668
struct read_regs_int *rr = &intr->read_regs;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1669
struct usb_int_regs *regs = (struct usb_int_regs *)rr->buffer;
drivers/s390/cio/chsc.c
1322
} *rr;
drivers/s390/cio/chsc.c
1326
rr = page;
drivers/s390/cio/chsc.c
1327
rr->request.length = 0x0020;
drivers/s390/cio/chsc.c
1328
rr->request.code = 0x0033;
drivers/s390/cio/chsc.c
1329
rr->op = op;
drivers/s390/cio/chsc.c
1330
rr->ctrl = ctrl;
drivers/s390/cio/chsc.c
1331
rc = chsc(rr);
drivers/s390/cio/chsc.c
1334
rc = (rr->response.code == 0x0001) ? 0 : -EIO;
drivers/s390/cio/chsc.c
1336
*clock_delta = rr->clock_delta;
drivers/s390/cio/chsc.c
1347
} *rr;
drivers/s390/cio/chsc.c
1351
rr = page;
drivers/s390/cio/chsc.c
1352
rr->request.length = 0x0010;
drivers/s390/cio/chsc.c
1353
rr->request.code = 0x0038;
drivers/s390/cio/chsc.c
1354
rc = chsc(rr);
drivers/s390/cio/chsc.c
1357
memcpy(result, &rr->data, size);
drivers/s390/cio/chsc.c
1358
return (rr->response.code == 0x0001) ? 0 : -EIO;
drivers/s390/cio/chsc.c
1368
} *rr;
drivers/s390/cio/chsc.c
1372
rr = page;
drivers/s390/cio/chsc.c
1373
rr->request.length = 0x0010;
drivers/s390/cio/chsc.c
1374
rr->request.code = 0x003e;
drivers/s390/cio/chsc.c
1375
rc = chsc(rr);
drivers/s390/cio/chsc.c
1378
memcpy(result, &rr->data, size);
drivers/s390/cio/chsc.c
1379
return (rr->response.code == 0x0001) ? 0 : -EIO;
drivers/s390/net/qeth_l2_main.c
556
struct chsc_pnso_area *rr;
drivers/s390/net/qeth_l2_main.c
562
rr = (struct chsc_pnso_area *)get_zeroed_page(GFP_KERNEL);
drivers/s390/net/qeth_l2_main.c
563
if (rr == NULL)
drivers/s390/net/qeth_l2_main.c
568
rc = ccw_device_pnso(ddev, rr, oc, rr->naihdr.resume_token,
drivers/s390/net/qeth_l2_main.c
575
size = rr->naihdr.naids;
drivers/s390/net/qeth_l2_main.c
581
elems = (rr->response.length - sizeof(struct chsc_header) -
drivers/s390/net/qeth_l2_main.c
584
if (!isfirstblock && (rr->naihdr.instance != prev_instance)) {
drivers/s390/net/qeth_l2_main.c
591
prev_instance = rr->naihdr.instance;
drivers/s390/net/qeth_l2_main.c
593
(*cb)(priv, &rr->entries[i]);
drivers/s390/net/qeth_l2_main.c
596
(rr->naihdr.resume_token.t1 || rr->naihdr.resume_token.t2)));
drivers/s390/net/qeth_l2_main.c
599
QETH_CARD_TEXT_(card, 2, "PNrp%04x", rr->response.code);
drivers/s390/net/qeth_l2_main.c
601
free_page((unsigned long)rr);
drivers/scsi/lpfc/lpfc_hw.h
2857
uint32_t rr:1;
drivers/scsi/lpfc/lpfc_hw.h
2869
uint32_t rr:1;
drivers/scsi/lpfc/lpfc_init.c
198
if (mb->un.varRdRev.rr == 0) {
drivers/soc/fsl/qbman/bman.c
164
union bm_mc_result *rr;
drivers/soc/fsl/qbman/bman.c
433
mc->rr = portal->addr.ce + BM_CL_RR0;
drivers/soc/fsl/qbman/bman.c
469
union bm_mc_result *rr = mc->rr + mc->rridx;
drivers/soc/fsl/qbman/bman.c
475
dpaa_invalidate_touch_ro(rr);
drivers/soc/fsl/qbman/bman.c
484
union bm_mc_result *rr = mc->rr + mc->rridx;
drivers/soc/fsl/qbman/bman.c
492
if (!rr->verb) {
drivers/soc/fsl/qbman/bman.c
493
dpaa_invalidate_touch_ro(rr);
drivers/soc/fsl/qbman/bman.c
501
return rr;
drivers/soc/fsl/qbman/qman.c
1455
struct qman_cgrs rr, c;
drivers/soc/fsl/qbman/qman.c
1472
qman_cgrs_and(&rr, (struct qman_cgrs *)&mcr->querycongestion.state,
drivers/soc/fsl/qbman/qman.c
1475
qman_cgrs_xor(&c, &rr, &p->cgrs[1]);
drivers/soc/fsl/qbman/qman.c
1477
qman_cgrs_cp(&p->cgrs[1], &rr);
drivers/soc/fsl/qbman/qman.c
1481
cgr->cb(p, cgr, qman_cgrs_get(&rr, cgr->cgrid));
drivers/soc/fsl/qbman/qman.c
328
union qm_mc_result *rr;
drivers/soc/fsl/qbman/qman.c
860
mc->rr = portal->addr.ce + QM_CL_RR0;
drivers/soc/fsl/qbman/qman.c
868
rr0 = mc->rr->verb;
drivers/soc/fsl/qbman/qman.c
869
rr1 = (mc->rr+1)->verb;
drivers/soc/fsl/qbman/qman.c
907
union qm_mc_result *rr = mc->rr + mc->rridx;
drivers/soc/fsl/qbman/qman.c
913
dpaa_invalidate_touch_ro(rr);
drivers/soc/fsl/qbman/qman.c
922
union qm_mc_result *rr = mc->rr + mc->rridx;
drivers/soc/fsl/qbman/qman.c
930
if (!rr->verb) {
drivers/soc/fsl/qbman/qman.c
931
dpaa_invalidate_touch_ro(rr);
drivers/soc/fsl/qbman/qman.c
939
return rr;
drivers/vfio/pci/mlx5/cmd.c
336
MLX5_SET(mkc, mkc, rr, 1);
drivers/video/fbdev/cyber2000fb.c
691
u_int rr, t_mult, t_pll_ps;
drivers/video/fbdev/cyber2000fb.c
697
rr = ref_ps * t_div1;
drivers/video/fbdev/cyber2000fb.c
698
t_mult = (rr + pll_ps / 2) / pll_ps;
drivers/video/fbdev/cyber2000fb.c
710
t_pll_ps = (rr + t_mult / 2) / t_mult;
drivers/video/fbdev/hpfb.c
126
static void topcat_blit(int x0, int y0, int x1, int y1, int w, int h, int rr)
drivers/video/fbdev/hpfb.c
128
if (rr >= 0) {
drivers/video/fbdev/hpfb.c
133
if (rr >= 0) {
drivers/video/fbdev/hpfb.c
135
out_8(fb_regs + WMRR, rr);
drivers/video/fbdev/omap2/omapfb/dss/dispc.c
1062
coef_r = FLD_VAL(coefs->rr, 31, 22) | FLD_VAL(coefs->rg, 20, 11) |
drivers/video/fbdev/omap2/omapfb/dss/manager-sysfs.c
364
info.cpr_coefs.rr,
drivers/video/fbdev/omap2/omapfb/dss/manager-sysfs.c
387
&coefs.rr, &coefs.rg, &coefs.rb,
drivers/video/fbdev/omap2/omapfb/dss/manager-sysfs.c
392
arr = (s16[]){ coefs.rr, coefs.rg, coefs.rb,
fs/gfs2/recovery.c
100
b = (where < rr->rr_where);
fs/gfs2/recovery.c
109
struct gfs2_revoke_replay *rr;
fs/gfs2/recovery.c
112
rr = list_first_entry(head, struct gfs2_revoke_replay, rr_list);
fs/gfs2/recovery.c
113
list_del(&rr->rr_list);
fs/gfs2/recovery.c
114
kfree(rr);
fs/gfs2/recovery.c
58
struct gfs2_revoke_replay *rr = NULL, *iter;
fs/gfs2/recovery.c
62
rr = iter;
fs/gfs2/recovery.c
67
if (rr) {
fs/gfs2/recovery.c
68
rr->rr_where = where;
fs/gfs2/recovery.c
72
rr = kmalloc_obj(struct gfs2_revoke_replay, GFP_NOFS);
fs/gfs2/recovery.c
73
if (!rr)
fs/gfs2/recovery.c
76
rr->rr_blkno = blkno;
fs/gfs2/recovery.c
77
rr->rr_where = where;
fs/gfs2/recovery.c
78
list_add(&rr->rr_list, head);
fs/gfs2/recovery.c
85
struct gfs2_revoke_replay *rr = NULL, *iter;
fs/gfs2/recovery.c
90
rr = iter;
fs/gfs2/recovery.c
95
if (!rr)
fs/gfs2/recovery.c
98
wrap = (rr->rr_where < jd->jd_replay_tail);
fs/isofs/rock.c
202
struct rock_ridge *rr;
fs/isofs/rock.c
219
rr = (struct rock_ridge *)rs.chr;
fs/isofs/rock.c
225
if (rr->len < 3)
fs/isofs/rock.c
230
rs.chr += rr->len;
fs/isofs/rock.c
231
rs.len -= rr->len;
fs/isofs/rock.c
241
if ((rr->u.RR.flags[0] & RR_NM) == 0)
fs/isofs/rock.c
245
if (check_sp(rr, inode))
fs/isofs/rock.c
249
rs.cont_extent = isonum_733(rr->u.CE.extent);
fs/isofs/rock.c
250
rs.cont_offset = isonum_733(rr->u.CE.offset);
fs/isofs/rock.c
251
rs.cont_size = isonum_733(rr->u.CE.size);
fs/isofs/rock.c
256
if (rr->len < 5)
fs/isofs/rock.c
265
if (rr->u.NM.flags & 6)
fs/isofs/rock.c
268
if (rr->u.NM.flags & ~1) {
fs/isofs/rock.c
270
rr->u.NM.flags);
fs/isofs/rock.c
273
len = rr->len - 5;
fs/isofs/rock.c
278
p = memchr(rr->u.NM.name, '\0', len);
fs/isofs/rock.c
280
len = p - rr->u.NM.name;
fs/isofs/rock.c
281
memcpy(retname + retnamlen, rr->u.NM.name, len);
fs/isofs/rock.c
316
struct rock_ridge *rr;
fs/isofs/rock.c
335
rr = (struct rock_ridge *)rs.chr;
fs/isofs/rock.c
341
if (rr->len < 3)
fs/isofs/rock.c
346
rs.chr += rr->len;
fs/isofs/rock.c
347
rs.len -= rr->len;
fs/isofs/rock.c
358
if ((rr->u.RR.flags[0] &
fs/isofs/rock.c
364
if (check_sp(rr, inode))
fs/isofs/rock.c
368
rs.cont_extent = isonum_733(rr->u.CE.extent);
fs/isofs/rock.c
369
rs.cont_offset = isonum_733(rr->u.CE.offset);
fs/isofs/rock.c
370
rs.cont_size = isonum_733(rr->u.CE.size);
fs/isofs/rock.c
374
if (rr->u.ER.len_id + offsetof(struct rock_ridge, u.ER.data) > rr->len)
fs/isofs/rock.c
380
for (p = 0; p < rr->u.ER.len_id; p++)
fs/isofs/rock.c
381
printk(KERN_CONT "%c", rr->u.ER.data[p]);
fs/isofs/rock.c
386
inode->i_mode = isonum_733(rr->u.PX.mode);
fs/isofs/rock.c
387
set_nlink(inode, isonum_733(rr->u.PX.n_links));
fs/isofs/rock.c
388
i_uid_write(inode, isonum_733(rr->u.PX.uid));
fs/isofs/rock.c
389
i_gid_write(inode, isonum_733(rr->u.PX.gid));
fs/isofs/rock.c
394
high = isonum_733(rr->u.PN.dev_high);
fs/isofs/rock.c
395
low = isonum_733(rr->u.PN.dev_low);
fs/isofs/rock.c
418
flags = rr->u.TF.flags & TF_LONG_FORM ? ISO_DATE_LONG_FORM : 0;
fs/isofs/rock.c
419
size = rr->u.TF.flags & TF_LONG_FORM ? 17 : 7;
fs/isofs/rock.c
420
slen = rr->len - 5;
fs/isofs/rock.c
428
if ((rr->u.TF.flags & TF_CREATE) && size <= slen) {
fs/isofs/rock.c
43
static int check_sp(struct rock_ridge *rr, struct inode *inode)
fs/isofs/rock.c
430
iso_date(rr->u.TF.data + size * cnt++, flags));
fs/isofs/rock.c
433
if ((rr->u.TF.flags & TF_MODIFY) && size <= slen) {
fs/isofs/rock.c
435
iso_date(rr->u.TF.data + size * cnt++, flags));
fs/isofs/rock.c
438
if ((rr->u.TF.flags & TF_ACCESS) && size <= slen) {
fs/isofs/rock.c
440
iso_date(rr->u.TF.data + size * cnt++, flags));
fs/isofs/rock.c
443
if ((rr->u.TF.flags & TF_ATTRIBUTES) && size <= slen) {
fs/isofs/rock.c
445
iso_date(rr->u.TF.data + size * cnt++, flags));
fs/isofs/rock.c
45
if (rr->u.SP.magic[0] != 0xbe)
fs/isofs/rock.c
455
slen = rr->len - 5;
fs/isofs/rock.c
456
slp = &rr->u.SL.link;
fs/isofs/rock.c
47
if (rr->u.SP.magic[1] != 0xef)
fs/isofs/rock.c
485
if (((rr->u.SL.
fs/isofs/rock.c
49
ISOFS_SB(inode->i_sb)->s_rock_offset = rr->u.SP.skip;
fs/isofs/rock.c
517
reloc_block = isonum_733(rr->u.CL.location);
fs/isofs/rock.c
549
algo = isonum_721(rr->u.ZF.algorithm);
fs/isofs/rock.c
552
isonum_711(&rr->u.ZF.parms[1]);
fs/isofs/rock.c
571
isonum_711(&rr->u.ZF.parms[0]);
fs/isofs/rock.c
573
isonum_711(&rr->u.ZF.parms[1]);
fs/isofs/rock.c
575
isonum_733(rr->u.ZF.
fs/isofs/rock.c
582
rr->u.ZF.algorithm[0],
fs/isofs/rock.c
583
rr->u.ZF.algorithm[1]);
fs/isofs/rock.c
605
static char *get_symlink_chunk(char *rpnt, struct rock_ridge *rr, char *plimit)
fs/isofs/rock.c
611
slen = rr->len - 5;
fs/isofs/rock.c
612
slp = &rr->u.SL.link;
fs/isofs/rock.c
652
if ((!rootflag) && (rr->u.SL.flags & 1) &&
fs/isofs/rock.c
708
struct rock_ridge *rr;
fs/isofs/rock.c
741
rr = (struct rock_ridge *)rs.chr;
fs/isofs/rock.c
742
if (rr->len < 3)
fs/isofs/rock.c
747
rs.chr += rr->len;
fs/isofs/rock.c
748
rs.len -= rr->len;
fs/isofs/rock.c
754
if ((rr->u.RR.flags[0] & RR_SL) == 0)
fs/isofs/rock.c
758
if (check_sp(rr, inode))
fs/isofs/rock.c
762
rpnt = get_symlink_chunk(rpnt, rr,
fs/isofs/rock.c
769
rs.cont_extent = isonum_733(rr->u.CE.extent);
fs/isofs/rock.c
770
rs.cont_offset = isonum_733(rr->u.CE.offset);
fs/isofs/rock.c
771
rs.cont_size = isonum_733(rr->u.CE.size);
fs/jffs2/scan.c
389
struct jffs2_raw_xref *rr, uint32_t ofs,
fs/jffs2/scan.c
396
crc = crc32(0, rr, sizeof(*rr) - 4);
fs/jffs2/scan.c
397
if (crc != je32_to_cpu(rr->node_crc)) {
fs/jffs2/scan.c
399
ofs, je32_to_cpu(rr->node_crc), crc);
fs/jffs2/scan.c
400
if ((err = jffs2_scan_dirty_space(c, jeb, PAD(je32_to_cpu(rr->totlen)))))
fs/jffs2/scan.c
405
if (PAD(sizeof(struct jffs2_raw_xref)) != je32_to_cpu(rr->totlen)) {
fs/jffs2/scan.c
407
ofs, je32_to_cpu(rr->totlen),
fs/jffs2/scan.c
409
if ((err = jffs2_scan_dirty_space(c, jeb, je32_to_cpu(rr->totlen))))
fs/jffs2/scan.c
427
ref->ino = je32_to_cpu(rr->ino);
fs/jffs2/scan.c
428
ref->xid = je32_to_cpu(rr->xid);
fs/jffs2/scan.c
429
ref->xseqno = je32_to_cpu(rr->xseqno);
fs/jffs2/scan.c
435
jffs2_link_node_ref(c, jeb, ofs | REF_PRISTINE, PAD(je32_to_cpu(rr->totlen)), (void *)ref);
fs/jffs2/scan.c
438
jffs2_sum_add_xref_mem(s, rr, ofs - jeb->offset);
fs/jffs2/summary.c
176
int jffs2_sum_add_xref_mem(struct jffs2_summary *s, struct jffs2_raw_xref *rr, uint32_t ofs)
fs/jffs2/summary.c
184
temp->nodetype = rr->nodetype;
fs/jffs2/summary.h
188
int jffs2_sum_add_xref_mem(struct jffs2_summary *s, struct jffs2_raw_xref *rr, uint32_t ofs);
fs/jffs2/xattr.c
448
struct jffs2_raw_xref rr;
fs/jffs2/xattr.c
459
rc = jffs2_flash_read(c, offset, sizeof(rr), &readlen, (char *)&rr);
fs/jffs2/xattr.c
460
if (rc || sizeof(rr) != readlen) {
fs/jffs2/xattr.c
462
rc, sizeof(rr), readlen, offset);
fs/jffs2/xattr.c
466
crc = crc32(0, &rr, sizeof(rr) - 4);
fs/jffs2/xattr.c
467
if (crc != je32_to_cpu(rr.node_crc)) {
fs/jffs2/xattr.c
469
offset, je32_to_cpu(rr.node_crc), crc);
fs/jffs2/xattr.c
472
if (je16_to_cpu(rr.magic) != JFFS2_MAGIC_BITMASK
fs/jffs2/xattr.c
473
|| je16_to_cpu(rr.nodetype) != JFFS2_NODETYPE_XREF
fs/jffs2/xattr.c
474
|| je32_to_cpu(rr.totlen) != PAD(sizeof(rr))) {
fs/jffs2/xattr.c
477
offset, je16_to_cpu(rr.magic), JFFS2_MAGIC_BITMASK,
fs/jffs2/xattr.c
478
je16_to_cpu(rr.nodetype), JFFS2_NODETYPE_XREF,
fs/jffs2/xattr.c
479
je32_to_cpu(rr.totlen), PAD(sizeof(rr)));
fs/jffs2/xattr.c
482
ref->ino = je32_to_cpu(rr.ino);
fs/jffs2/xattr.c
483
ref->xid = je32_to_cpu(rr.xid);
fs/jffs2/xattr.c
484
ref->xseqno = je32_to_cpu(rr.xseqno);
fs/jffs2/xattr.c
509
struct jffs2_raw_xref rr;
fs/jffs2/xattr.c
514
rr.magic = cpu_to_je16(JFFS2_MAGIC_BITMASK);
fs/jffs2/xattr.c
515
rr.nodetype = cpu_to_je16(JFFS2_NODETYPE_XREF);
fs/jffs2/xattr.c
516
rr.totlen = cpu_to_je32(PAD(sizeof(rr)));
fs/jffs2/xattr.c
517
rr.hdr_crc = cpu_to_je32(crc32(0, &rr, sizeof(struct jffs2_unknown_node) - 4));
fs/jffs2/xattr.c
522
rr.ino = cpu_to_je32(ref->ino);
fs/jffs2/xattr.c
523
rr.xid = cpu_to_je32(ref->xid);
fs/jffs2/xattr.c
525
rr.ino = cpu_to_je32(ref->ic->ino);
fs/jffs2/xattr.c
526
rr.xid = cpu_to_je32(ref->xd->xid);
fs/jffs2/xattr.c
528
rr.xseqno = cpu_to_je32(xseqno);
fs/jffs2/xattr.c
529
rr.node_crc = cpu_to_je32(crc32(0, &rr, sizeof(rr) - 4));
fs/jffs2/xattr.c
531
ret = jffs2_flash_write(c, phys_ofs, sizeof(rr), &length, (char *)&rr);
fs/jffs2/xattr.c
532
if (ret || sizeof(rr) != length) {
fs/jffs2/xattr.c
534
ret, sizeof(rr), length, phys_ofs);
fs/jffs2/xattr.c
537
jffs2_add_physical_node_ref(c, phys_ofs | REF_OBSOLETE, PAD(sizeof(rr)), NULL);
fs/jffs2/xattr.c
543
jffs2_add_physical_node_ref(c, phys_ofs | REF_PRISTINE, PAD(sizeof(rr)), (void *)ref);
fs/ocfs2/dlm/dlmmaster.c
1894
int rr, err = 0;
fs/ocfs2/dlm/dlmmaster.c
1942
rr = kref_read(&mle->mle_refs);
fs/ocfs2/dlm/dlmmaster.c
1944
if (extra_ref && rr < 3)
fs/ocfs2/dlm/dlmmaster.c
1946
else if (!extra_ref && rr < 2)
fs/ocfs2/dlm/dlmmaster.c
1949
if (extra_ref && rr < 2)
fs/ocfs2/dlm/dlmmaster.c
1951
else if (!extra_ref && rr < 1)
fs/ocfs2/dlm/dlmmaster.c
1958
assert->node_idx, rr, extra_ref, mle->inuse);
fs/resctrl/ctrlmondata.c
554
void mon_event_read(struct rmid_read *rr, struct rdt_resource *r,
fs/resctrl/ctrlmondata.c
566
rr->rgrp = rdtgrp;
fs/resctrl/ctrlmondata.c
567
rr->evt = evt;
fs/resctrl/ctrlmondata.c
568
rr->r = r;
fs/resctrl/ctrlmondata.c
569
rr->hdr = hdr;
fs/resctrl/ctrlmondata.c
570
rr->first = first;
fs/resctrl/ctrlmondata.c
573
rr->is_mbm_cntr = true;
fs/resctrl/ctrlmondata.c
575
rr->arch_mon_ctx = resctrl_arch_mon_ctx_alloc(r, evt->evtid);
fs/resctrl/ctrlmondata.c
576
if (IS_ERR(rr->arch_mon_ctx)) {
fs/resctrl/ctrlmondata.c
577
rr->err = -EINVAL;
fs/resctrl/ctrlmondata.c
583
mon_event_count(rr);
fs/resctrl/ctrlmondata.c
596
smp_call_function_any(cpumask, mon_event_count, rr, 1);
fs/resctrl/ctrlmondata.c
598
smp_call_on_cpu(cpu, smp_mon_event_count, rr, false);
fs/resctrl/ctrlmondata.c
601
if (rr->arch_mon_ctx)
fs/resctrl/ctrlmondata.c
602
resctrl_arch_mon_ctx_free(r, evt->evtid, rr->arch_mon_ctx);
fs/resctrl/ctrlmondata.c
681
struct rmid_read rr = {0};
fs/resctrl/ctrlmondata.c
726
rr.ci = ci;
fs/resctrl/ctrlmondata.c
727
mon_event_read(&rr, r, NULL, rdtgrp,
fs/resctrl/ctrlmondata.c
744
mon_event_read(&rr, r, hdr, rdtgrp, &hdr->cpu_mask, evt, false);
fs/resctrl/ctrlmondata.c
753
if (rr.err == -EIO)
fs/resctrl/ctrlmondata.c
755
else if (rr.err == -EINVAL)
fs/resctrl/ctrlmondata.c
757
else if (rr.err == -ENOENT)
fs/resctrl/ctrlmondata.c
760
print_event_value(m, evt->binary_bits, rr.val);
fs/resctrl/ctrlmondata.c
762
seq_printf(m, "%llu\n", rr.val);
fs/resctrl/internal.h
388
void mon_event_read(struct rmid_read *rr, struct rdt_resource *r,
fs/resctrl/monitor.c
418
static int __l3_mon_event_count(struct rdtgroup *rdtgrp, struct rmid_read *rr)
fs/resctrl/monitor.c
428
if (!domain_header_is_valid(rr->hdr, RESCTRL_MON_DOMAIN, RDT_RESOURCE_L3)) {
fs/resctrl/monitor.c
429
rr->err = -EIO;
fs/resctrl/monitor.c
432
d = container_of(rr->hdr, struct rdt_l3_mon_domain, hdr);
fs/resctrl/monitor.c
434
if (rr->is_mbm_cntr) {
fs/resctrl/monitor.c
435
cntr_id = mbm_cntr_get(rr->r, d, rdtgrp, rr->evt->evtid);
fs/resctrl/monitor.c
437
rr->err = -ENOENT;
fs/resctrl/monitor.c
442
if (rr->first) {
fs/resctrl/monitor.c
443
if (rr->is_mbm_cntr)
fs/resctrl/monitor.c
444
resctrl_arch_reset_cntr(rr->r, d, closid, rmid, cntr_id, rr->evt->evtid);
fs/resctrl/monitor.c
446
resctrl_arch_reset_rmid(rr->r, d, closid, rmid, rr->evt->evtid);
fs/resctrl/monitor.c
447
m = get_mbm_state(d, closid, rmid, rr->evt->evtid);
fs/resctrl/monitor.c
456
if (rr->is_mbm_cntr)
fs/resctrl/monitor.c
457
rr->err = resctrl_arch_cntr_read(rr->r, d, closid, rmid, cntr_id,
fs/resctrl/monitor.c
458
rr->evt->evtid, &tval);
fs/resctrl/monitor.c
460
rr->err = resctrl_arch_rmid_read(rr->r, rr->hdr, closid, rmid,
fs/resctrl/monitor.c
461
rr->evt->evtid, rr->evt->arch_priv,
fs/resctrl/monitor.c
462
&tval, rr->arch_mon_ctx);
fs/resctrl/monitor.c
463
if (rr->err)
fs/resctrl/monitor.c
464
return rr->err;
fs/resctrl/monitor.c
466
rr->val += tval;
fs/resctrl/monitor.c
471
static int __l3_mon_event_count_sum(struct rdtgroup *rdtgrp, struct rmid_read *rr)
fs/resctrl/monitor.c
485
if (rr->is_mbm_cntr) {
fs/resctrl/monitor.c
487
rr->err = -EINVAL;
fs/resctrl/monitor.c
492
if (!cpumask_test_cpu(cpu, &rr->ci->shared_cpu_map))
fs/resctrl/monitor.c
503
list_for_each_entry(d, &rr->r->mon_domains, hdr.list) {
fs/resctrl/monitor.c
504
if (d->ci_id != rr->ci->id)
fs/resctrl/monitor.c
506
err = resctrl_arch_rmid_read(rr->r, &d->hdr, closid, rmid,
fs/resctrl/monitor.c
507
rr->evt->evtid, rr->evt->arch_priv,
fs/resctrl/monitor.c
508
&tval, rr->arch_mon_ctx);
fs/resctrl/monitor.c
510
rr->val += tval;
fs/resctrl/monitor.c
516
rr->err = ret;
fs/resctrl/monitor.c
521
static int __mon_event_count(struct rdtgroup *rdtgrp, struct rmid_read *rr)
fs/resctrl/monitor.c
523
switch (rr->r->rid) {
fs/resctrl/monitor.c
525
WARN_ON_ONCE(rr->evt->any_cpu);
fs/resctrl/monitor.c
526
if (rr->hdr)
fs/resctrl/monitor.c
527
return __l3_mon_event_count(rdtgrp, rr);
fs/resctrl/monitor.c
529
return __l3_mon_event_count_sum(rdtgrp, rr);
fs/resctrl/monitor.c
533
rr->err = resctrl_arch_rmid_read(rr->r, rr->hdr, rdtgrp->closid,
fs/resctrl/monitor.c
534
rdtgrp->mon.rmid, rr->evt->evtid,
fs/resctrl/monitor.c
535
rr->evt->arch_priv,
fs/resctrl/monitor.c
536
&tval, rr->arch_mon_ctx);
fs/resctrl/monitor.c
537
if (rr->err)
fs/resctrl/monitor.c
538
return rr->err;
fs/resctrl/monitor.c
540
rr->val += tval;
fs/resctrl/monitor.c
545
rr->err = -EINVAL;
fs/resctrl/monitor.c
562
static void mbm_bw_count(struct rdtgroup *rdtgrp, struct rmid_read *rr)
fs/resctrl/monitor.c
570
if (!domain_header_is_valid(rr->hdr, RESCTRL_MON_DOMAIN, RDT_RESOURCE_L3))
fs/resctrl/monitor.c
572
d = container_of(rr->hdr, struct rdt_l3_mon_domain, hdr);
fs/resctrl/monitor.c
573
m = get_mbm_state(d, closid, rmid, rr->evt->evtid);
fs/resctrl/monitor.c
577
cur_bytes = rr->val;
fs/resctrl/monitor.c
593
struct rmid_read *rr = info;
fs/resctrl/monitor.c
597
rdtgrp = rr->rgrp;
fs/resctrl/monitor.c
599
ret = __mon_event_count(rdtgrp, rr);
fs/resctrl/monitor.c
610
if (__mon_event_count(entry, rr) == 0)
fs/resctrl/monitor.c
621
rr->err = 0;
fs/resctrl/monitor.c
743
struct rmid_read rr = {0};
fs/resctrl/monitor.c
745
rr.r = r;
fs/resctrl/monitor.c
746
rr.hdr = &d->hdr;
fs/resctrl/monitor.c
747
rr.evt = &mon_event_all[evtid];
fs/resctrl/monitor.c
749
rr.is_mbm_cntr = true;
fs/resctrl/monitor.c
751
rr.arch_mon_ctx = resctrl_arch_mon_ctx_alloc(rr.r, evtid);
fs/resctrl/monitor.c
752
if (IS_ERR(rr.arch_mon_ctx)) {
fs/resctrl/monitor.c
754
PTR_ERR(rr.arch_mon_ctx));
fs/resctrl/monitor.c
759
__mon_event_count(rdtgrp, &rr);
fs/resctrl/monitor.c
766
mbm_bw_count(rdtgrp, &rr);
fs/resctrl/monitor.c
768
if (rr.arch_mon_ctx)
fs/resctrl/monitor.c
769
resctrl_arch_mon_ctx_free(rr.r, evtid, rr.arch_mon_ctx);
fs/resctrl/rdtgroup.c
3296
struct rmid_read rr = {0};
fs/resctrl/rdtgroup.c
3324
mon_event_read(&rr, r, hdr, prgrp, &hdr->cpu_mask, mevt, true);
fs/xfs/libxfs/xfs_refcount.c
2028
struct xfs_refcount_recovery *rr;
fs/xfs/libxfs/xfs_refcount.c
2036
rr = kmalloc_obj(struct xfs_refcount_recovery,
fs/xfs/libxfs/xfs_refcount.c
2038
INIT_LIST_HEAD(&rr->rr_list);
fs/xfs/libxfs/xfs_refcount.c
2039
xfs_refcount_btrec_to_irec(rec, &rr->rr_rrec);
fs/xfs/libxfs/xfs_refcount.c
2041
if (xfs_refcount_check_btrec(cur, &rr->rr_rrec) != NULL ||
fs/xfs/libxfs/xfs_refcount.c
2043
rr->rr_rrec.rc_domain != XFS_REFC_DOMAIN_COW)) {
fs/xfs/libxfs/xfs_refcount.c
2045
kfree(rr);
fs/xfs/libxfs/xfs_refcount.c
2049
list_add_tail(&rr->rr_list, debris);
fs/xfs/libxfs/xfs_refcount.c
2063
struct xfs_refcount_recovery *rr, *n;
fs/xfs/libxfs/xfs_refcount.c
2126
list_for_each_entry_safe(rr, n, &debris, rr_list) {
fs/xfs/libxfs/xfs_refcount.c
2133
fsb = xfs_gbno_to_fsb(xg, rr->rr_rrec.rc_startblock);
fs/xfs/libxfs/xfs_refcount.c
2135
rr->rr_rrec.rc_blockcount);
fs/xfs/libxfs/xfs_refcount.c
2139
rr->rr_rrec.rc_blockcount, NULL,
fs/xfs/libxfs/xfs_refcount.c
2149
list_del(&rr->rr_list);
fs/xfs/libxfs/xfs_refcount.c
2150
kfree(rr);
fs/xfs/libxfs/xfs_refcount.c
2158
list_for_each_entry_safe(rr, n, &debris, rr_list) {
fs/xfs/libxfs/xfs_refcount.c
2159
list_del(&rr->rr_list);
fs/xfs/libxfs/xfs_refcount.c
2160
kfree(rr);
fs/xfs/scrub/refcount_repair.c
163
struct xrep_refc *rr,
fs/xfs/scrub/refcount_repair.c
174
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/refcount_repair.c
182
error = xrep_refc_check_ext(rr->sc, &irec);
fs/xfs/scrub/refcount_repair.c
188
return xfarray_append(rr->refcount_records, &irec);
fs/xfs/scrub/refcount_repair.c
194
struct xrep_refc *rr,
fs/xfs/scrub/refcount_repair.c
198
return xrep_refc_stash(rr, XFS_REFC_DOMAIN_COW, agbno, len, 1);
fs/xfs/scrub/refcount_repair.c
229
struct xrep_refc *rr,
fs/xfs/scrub/refcount_repair.c
233
struct xfs_btree_cur *cur = rr->sc->sa.rmap_cur;
fs/xfs/scrub/refcount_repair.c
247
if (xchk_should_terminate(rr->sc, &error))
fs/xfs/scrub/refcount_repair.c
265
error = xrep_refc_stash_cow(rr, rmap->rm_startblock,
fs/xfs/scrub/refcount_repair.c
271
rr->btblocks += rmap->rm_blockcount;
fs/xfs/scrub/refcount_repair.c
272
error = xagb_bitmap_set(&rr->old_refcountbt_blocks,
fs/xfs/scrub/refcount_repair.c
323
struct xrep_refc *rr)
fs/xfs/scrub/refcount_repair.c
331
error = xfarray_sort(rr->refcount_records, xrep_refc_extent_cmp,
fs/xfs/scrub/refcount_repair.c
336
foreach_xfarray_idx(rr->refcount_records, cur) {
fs/xfs/scrub/refcount_repair.c
337
if (xchk_should_terminate(rr->sc, &error))
fs/xfs/scrub/refcount_repair.c
340
error = xfarray_load(rr->refcount_records, cur, &irec);
fs/xfs/scrub/refcount_repair.c
369
struct xrep_refc *rr,
fs/xfs/scrub/refcount_repair.c
375
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/refcount_repair.c
380
error = rcbag_add(rcstack, rr->sc->tp, rmap);
fs/xfs/scrub/refcount_repair.c
384
error = xrep_refc_walk_rmaps(rr, rmap, have);
fs/xfs/scrub/refcount_repair.c
403
struct xrep_refc *rr)
fs/xfs/scrub/refcount_repair.c
405
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/refcount_repair.c
435
error = xrep_refc_walk_rmaps(rr, &rmap, &have);
fs/xfs/scrub/refcount_repair.c
441
error = xrep_refc_push_rmaps_at(rr, rcstack, sbno, &rmap,
fs/xfs/scrub/refcount_repair.c
462
error = xrep_refc_walk_rmaps(rr, &rmap, &have);
fs/xfs/scrub/refcount_repair.c
466
error = xrep_refc_push_rmaps_at(rr, rcstack,
fs/xfs/scrub/refcount_repair.c
476
error = xrep_refc_stash(rr,
fs/xfs/scrub/refcount_repair.c
520
struct xrep_refc *rr = priv;
fs/xfs/scrub/refcount_repair.c
526
error = xfarray_load(rr->refcount_records, rr->array_cur++,
fs/xfs/scrub/refcount_repair.c
545
struct xrep_refc *rr = priv;
fs/xfs/scrub/refcount_repair.c
547
return xrep_newbt_claim_block(cur, &rr->new_btree, ptr);
fs/xfs/scrub/refcount_repair.c
553
struct xrep_refc *rr)
fs/xfs/scrub/refcount_repair.c
555
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/refcount_repair.c
582
struct xrep_refc *rr)
fs/xfs/scrub/refcount_repair.c
584
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/refcount_repair.c
589
error = xrep_refc_sort_records(rr);
fs/xfs/scrub/refcount_repair.c
599
xrep_newbt_init_ag(&rr->new_btree, sc, &XFS_RMAP_OINFO_REFC,
fs/xfs/scrub/refcount_repair.c
602
rr->new_btree.bload.get_records = xrep_refc_get_records;
fs/xfs/scrub/refcount_repair.c
603
rr->new_btree.bload.claim_block = xrep_refc_claim_block;
fs/xfs/scrub/refcount_repair.c
607
xfs_btree_stage_afakeroot(refc_cur, &rr->new_btree.afake);
fs/xfs/scrub/refcount_repair.c
609
&rr->new_btree.bload,
fs/xfs/scrub/refcount_repair.c
610
xfarray_length(rr->refcount_records));
fs/xfs/scrub/refcount_repair.c
619
error = xrep_newbt_alloc_blocks(&rr->new_btree,
fs/xfs/scrub/refcount_repair.c
620
rr->new_btree.bload.nr_blocks);
fs/xfs/scrub/refcount_repair.c
630
pag->pagf_repair_refcount_level = rr->new_btree.bload.btree_height;
fs/xfs/scrub/refcount_repair.c
633
rr->array_cur = XFARRAY_CURSOR_INIT;
fs/xfs/scrub/refcount_repair.c
634
error = xfs_btree_bload(refc_cur, &rr->new_btree.bload, rr);
fs/xfs/scrub/refcount_repair.c
646
error = xrep_refc_reset_counters(rr);
fs/xfs/scrub/refcount_repair.c
651
error = xrep_newbt_commit(&rr->new_btree);
fs/xfs/scrub/refcount_repair.c
662
xrep_newbt_cancel(&rr->new_btree);
fs/xfs/scrub/refcount_repair.c
672
struct xrep_refc *rr)
fs/xfs/scrub/refcount_repair.c
674
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/refcount_repair.c
679
error = xrep_reap_agblocks(sc, &rr->old_refcountbt_blocks,
fs/xfs/scrub/refcount_repair.c
699
struct xrep_refc *rr;
fs/xfs/scrub/refcount_repair.c
707
rr = kzalloc_obj(struct xrep_refc, XCHK_GFP_FLAGS);
fs/xfs/scrub/refcount_repair.c
708
if (!rr)
fs/xfs/scrub/refcount_repair.c
710
rr->sc = sc;
fs/xfs/scrub/refcount_repair.c
715
&rr->refcount_records);
fs/xfs/scrub/refcount_repair.c
720
xagb_bitmap_init(&rr->old_refcountbt_blocks);
fs/xfs/scrub/refcount_repair.c
721
error = xrep_refc_find_refcounts(rr);
fs/xfs/scrub/refcount_repair.c
726
error = xrep_refc_build_new_tree(rr);
fs/xfs/scrub/refcount_repair.c
731
error = xrep_refc_remove_old_tree(rr);
fs/xfs/scrub/refcount_repair.c
736
xagb_bitmap_destroy(&rr->old_refcountbt_blocks);
fs/xfs/scrub/refcount_repair.c
737
xfarray_destroy(rr->refcount_records);
fs/xfs/scrub/refcount_repair.c
739
kfree(rr);
fs/xfs/scrub/rmap_repair.c
1035
struct xrep_rmap *rr,
fs/xfs/scrub/rmap_repair.c
1043
.agno = pag_agno(rr->sc->sa.pag),
fs/xfs/scrub/rmap_repair.c
1045
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/rmap_repair.c
1060
nr_blocks = rr->new_btree.bload.nr_blocks;
fs/xfs/scrub/rmap_repair.c
1068
error = xrep_newbt_alloc_blocks(&rr->new_btree,
fs/xfs/scrub/rmap_repair.c
1073
*blocks_reserved = rr->new_btree.bload.nr_blocks;
fs/xfs/scrub/rmap_repair.c
1097
rr->freesp_btblocks = xagb_bitmap_hweight(freesp_blocks) - 2;
fs/xfs/scrub/rmap_repair.c
1100
list_for_each_entry_safe(resv, n, &rr->new_btree.resv_list, list) {
fs/xfs/scrub/rmap_repair.c
1120
&rr->new_btree.bload, rr->nr_records + freesp_records);
fs/xfs/scrub/rmap_repair.c
1125
*done = nr_blocks >= rr->new_btree.bload.nr_blocks;
fs/xfs/scrub/rmap_repair.c
1135
struct xrep_rmap *rr,
fs/xfs/scrub/rmap_repair.c
1145
&rr->new_btree.bload, rr->nr_records);
fs/xfs/scrub/rmap_repair.c
1150
if (xchk_should_terminate(rr->sc, &error))
fs/xfs/scrub/rmap_repair.c
1162
error = xrep_rmap_try_reserve(rr, rmap_cur, &freesp_blocks,
fs/xfs/scrub/rmap_repair.c
1169
xrep_ag_btcur_init(rr->sc, &rr->sc->sa);
fs/xfs/scrub/rmap_repair.c
1170
error = xrep_rmap_stash_bitmap(rr, &freesp_blocks, &XFS_RMAP_OINFO_AG);
fs/xfs/scrub/rmap_repair.c
1171
xchk_ag_btcur_free(&rr->sc->sa);
fs/xfs/scrub/rmap_repair.c
1183
struct xrep_rmap *rr)
fs/xfs/scrub/rmap_repair.c
1185
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/rmap_repair.c
1194
rmap_btblocks = rr->new_btree.afake.af_blocks - 1;
fs/xfs/scrub/rmap_repair.c
1195
agf->agf_btreeblks = cpu_to_be32(rr->freesp_btblocks + rmap_btblocks);
fs/xfs/scrub/rmap_repair.c
1224
struct xrep_rmap *rr = priv;
fs/xfs/scrub/rmap_repair.c
1232
error = xfs_btree_increment(rr->mcur, 0, &stat);
fs/xfs/scrub/rmap_repair.c
1238
error = xfs_rmap_get_rec(rr->mcur, &cur->bc_rec.r, &stat);
fs/xfs/scrub/rmap_repair.c
1258
struct xrep_rmap *rr = priv;
fs/xfs/scrub/rmap_repair.c
1260
return xrep_newbt_claim_block(cur, &rr->new_btree, ptr);
fs/xfs/scrub/rmap_repair.c
1333
struct xrep_rmap *rr)
fs/xfs/scrub/rmap_repair.c
1335
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/rmap_repair.c
1346
rr->old_rmapbt_fsbcount = be32_to_cpu(agf->agf_rmap_blocks);
fs/xfs/scrub/rmap_repair.c
1356
xrep_newbt_init_ag(&rr->new_btree, sc, &XFS_RMAP_OINFO_SKIP_UPDATE,
fs/xfs/scrub/rmap_repair.c
1359
rr->new_btree.bload.get_records = xrep_rmap_get_records;
fs/xfs/scrub/rmap_repair.c
1360
rr->new_btree.bload.claim_block = xrep_rmap_claim_block;
fs/xfs/scrub/rmap_repair.c
1361
rr->new_btree.alloc_vextent = xrep_rmap_alloc_vextent;
fs/xfs/scrub/rmap_repair.c
1363
xfs_btree_stage_afakeroot(rmap_cur, &rr->new_btree.afake);
fs/xfs/scrub/rmap_repair.c
1369
error = xrep_rmap_reserve_space(rr, rmap_cur);
fs/xfs/scrub/rmap_repair.c
1377
rr->mcur = xfs_rmapbt_mem_cursor(rr->sc->sa.pag, NULL,
fs/xfs/scrub/rmap_repair.c
1378
&rr->rmap_btree);
fs/xfs/scrub/rmap_repair.c
1380
error = xrep_rmap_count_records(rr->mcur, &rr->nr_records);
fs/xfs/scrub/rmap_repair.c
1390
pag->pagf_repair_rmap_level = rr->new_btree.bload.btree_height;
fs/xfs/scrub/rmap_repair.c
1396
error = xfs_btree_goto_left_edge(rr->mcur);
fs/xfs/scrub/rmap_repair.c
1401
error = xfs_btree_bload(rmap_cur, &rr->new_btree.bload, rr);
fs/xfs/scrub/rmap_repair.c
1411
xfs_btree_del_cursor(rr->mcur, 0);
fs/xfs/scrub/rmap_repair.c
1412
rr->mcur = NULL;
fs/xfs/scrub/rmap_repair.c
1418
xchk_iscan_abort(&rr->iscan);
fs/xfs/scrub/rmap_repair.c
1426
rr->new_btree.oinfo = XFS_RMAP_OINFO_AG;
fs/xfs/scrub/rmap_repair.c
1429
error = xrep_rmap_reset_counters(rr);
fs/xfs/scrub/rmap_repair.c
1434
error = xrep_newbt_commit(&rr->new_btree);
fs/xfs/scrub/rmap_repair.c
1443
xfs_btree_del_cursor(rr->mcur, error);
fs/xfs/scrub/rmap_repair.c
1447
xrep_newbt_cancel(&rr->new_btree);
fs/xfs/scrub/rmap_repair.c
1501
struct xrep_rmap *rr)
fs/xfs/scrub/rmap_repair.c
1506
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/rmap_repair.c
1516
mcur = xfs_rmapbt_mem_cursor(rr->sc->sa.pag, NULL, &rr->rmap_btree);
fs/xfs/scrub/rmap_repair.c
1606
struct xrep_rmap *rr;
fs/xfs/scrub/rmap_repair.c
1612
rr = container_of(nb, struct xrep_rmap, rhook.rmap_hook.nb);
fs/xfs/scrub/rmap_repair.c
1613
mp = rr->sc->mp;
fs/xfs/scrub/rmap_repair.c
1615
if (!xrep_rmapbt_want_live_update(&rr->iscan, &p->oinfo))
fs/xfs/scrub/rmap_repair.c
1618
trace_xrep_rmap_live_update(pag_group(rr->sc->sa.pag), action, p);
fs/xfs/scrub/rmap_repair.c
1622
mutex_lock(&rr->lock);
fs/xfs/scrub/rmap_repair.c
1623
mcur = xfs_rmapbt_mem_cursor(rr->sc->sa.pag, tp, &rr->rmap_btree);
fs/xfs/scrub/rmap_repair.c
1630
error = xfbtree_trans_commit(&rr->rmap_btree, tp);
fs/xfs/scrub/rmap_repair.c
1635
mutex_unlock(&rr->lock);
fs/xfs/scrub/rmap_repair.c
1639
xfbtree_trans_cancel(&rr->rmap_btree, tp);
fs/xfs/scrub/rmap_repair.c
1641
mutex_unlock(&rr->lock);
fs/xfs/scrub/rmap_repair.c
1642
xchk_iscan_abort(&rr->iscan);
fs/xfs/scrub/rmap_repair.c
1650
struct xrep_rmap *rr)
fs/xfs/scrub/rmap_repair.c
1652
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/rmap_repair.c
1655
mutex_init(&rr->lock);
fs/xfs/scrub/rmap_repair.c
1658
error = xfs_rmapbt_mem_init(sc->mp, &rr->rmap_btree, sc->xmbtp,
fs/xfs/scrub/rmap_repair.c
166
struct xrep_rmap *rr;
fs/xfs/scrub/rmap_repair.c
1664
xchk_iscan_start(sc, 30000, 100, &rr->iscan);
fs/xfs/scrub/rmap_repair.c
1673
xfs_rmap_hook_setup(&rr->rhook, xrep_rmapbt_live_update);
fs/xfs/scrub/rmap_repair.c
1674
error = xfs_rmap_hook_add(pag_group(sc->sa.pag), &rr->rhook);
fs/xfs/scrub/rmap_repair.c
1680
xchk_iscan_teardown(&rr->iscan);
fs/xfs/scrub/rmap_repair.c
1681
xfbtree_destroy(&rr->rmap_btree);
fs/xfs/scrub/rmap_repair.c
1683
mutex_destroy(&rr->lock);
fs/xfs/scrub/rmap_repair.c
1690
struct xrep_rmap *rr)
fs/xfs/scrub/rmap_repair.c
1692
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/rmap_repair.c
1694
xchk_iscan_abort(&rr->iscan);
fs/xfs/scrub/rmap_repair.c
1695
xfs_rmap_hook_del(pag_group(sc->sa.pag), &rr->rhook);
fs/xfs/scrub/rmap_repair.c
1696
xchk_iscan_teardown(&rr->iscan);
fs/xfs/scrub/rmap_repair.c
1697
xfbtree_destroy(&rr->rmap_btree);
fs/xfs/scrub/rmap_repair.c
1698
mutex_destroy(&rr->lock);
fs/xfs/scrub/rmap_repair.c
1706
struct xrep_rmap *rr = sc->buf;
fs/xfs/scrub/rmap_repair.c
1709
error = xrep_rmap_setup_scan(rr);
fs/xfs/scrub/rmap_repair.c
1717
error = xrep_rmap_find_rmaps(rr);
fs/xfs/scrub/rmap_repair.c
1722
error = xrep_rmap_build_new_tree(rr);
fs/xfs/scrub/rmap_repair.c
1727
error = xrep_rmap_remove_old_tree(rr);
fs/xfs/scrub/rmap_repair.c
1732
xrep_rmap_teardown(rr);
fs/xfs/scrub/rmap_repair.c
175
rr = kzalloc_obj(struct xrep_rmap, XCHK_GFP_FLAGS);
fs/xfs/scrub/rmap_repair.c
176
if (!rr)
fs/xfs/scrub/rmap_repair.c
179
rr->sc = sc;
fs/xfs/scrub/rmap_repair.c
180
sc->buf = rr;
fs/xfs/scrub/rmap_repair.c
210
struct xrep_rmap *rr,
fs/xfs/scrub/rmap_repair.c
224
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/rmap_repair.c
231
if (xchk_iscan_aborted(&rr->iscan))
fs/xfs/scrub/rmap_repair.c
236
mutex_lock(&rr->lock);
fs/xfs/scrub/rmap_repair.c
237
mcur = xfs_rmapbt_mem_cursor(sc->sa.pag, sc->tp, &rr->rmap_btree);
fs/xfs/scrub/rmap_repair.c
243
error = xfbtree_trans_commit(&rr->rmap_btree, sc->tp);
fs/xfs/scrub/rmap_repair.c
247
mutex_unlock(&rr->lock);
fs/xfs/scrub/rmap_repair.c
251
xfbtree_trans_cancel(&rr->rmap_btree, sc->tp);
fs/xfs/scrub/rmap_repair.c
253
xchk_iscan_abort(&rr->iscan);
fs/xfs/scrub/rmap_repair.c
254
mutex_unlock(&rr->lock);
fs/xfs/scrub/rmap_repair.c
259
struct xrep_rmap *rr;
fs/xfs/scrub/rmap_repair.c
271
struct xrep_rmap *rr = rsr->rr;
fs/xfs/scrub/rmap_repair.c
273
return xrep_rmap_stash(rr, start, len, rsr->owner, 0, rsr->rmap_flags);
fs/xfs/scrub/rmap_repair.c
282
struct xrep_rmap *rr,
fs/xfs/scrub/rmap_repair.c
287
.rr = rr,
fs/xfs/scrub/rmap_repair.c
313
struct xrep_rmap *rr;
fs/xfs/scrub/rmap_repair.c
327
return xrep_rmap_stash(rf->rr, rf->accum.rm_startblock,
fs/xfs/scrub/rmap_repair.c
340
struct xfs_mount *mp = rf->rr->sc->mp;
fs/xfs/scrub/rmap_repair.c
347
pag_agno(rf->rr->sc->sa.pag))
fs/xfs/scrub/rmap_repair.c
394
if (XFS_FSB_TO_AGNO(cur->bc_mp, fsbno) != pag_agno(rf->rr->sc->sa.pag))
fs/xfs/scrub/rmap_repair.c
411
struct xrep_rmap *rr = rf->rr;
fs/xfs/scrub/rmap_repair.c
424
error = xrep_rmap_stash_bitmap(rr, &rf->bmbt_blocks, &oinfo);
fs/xfs/scrub/rmap_repair.c
446
struct xrep_rmap *rr = rf->rr;
fs/xfs/scrub/rmap_repair.c
453
cur = xfs_bmbt_init_cursor(rr->sc->mp, rr->sc->tp, ip, rf->whichfork);
fs/xfs/scrub/rmap_repair.c
507
struct xfs_scrub *sc = rf->rr->sc;
fs/xfs/scrub/rmap_repair.c
568
struct xrep_rmap *rr,
fs/xfs/scrub/rmap_repair.c
574
.rr = rr,
fs/xfs/scrub/rmap_repair.c
636
struct xrep_rmap *rr,
fs/xfs/scrub/rmap_repair.c
643
error = xrep_rmap_scan_ifork(rr, ip, XFS_DATA_FORK);
fs/xfs/scrub/rmap_repair.c
648
error = xrep_rmap_scan_ifork(rr, ip, XFS_ATTR_FORK);
fs/xfs/scrub/rmap_repair.c
654
xchk_iscan_mark_visited(&rr->iscan, ip);
fs/xfs/scrub/rmap_repair.c
663
struct xrep_rmap *rr;
fs/xfs/scrub/rmap_repair.c
729
struct xrep_rmap *rr)
fs/xfs/scrub/rmap_repair.c
732
.rr = rr,
fs/xfs/scrub/rmap_repair.c
734
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/rmap_repair.c
770
error = xrep_rmap_stash_bitmap(rr, &ri.inobt_blocks,
fs/xfs/scrub/rmap_repair.c
774
error = xrep_rmap_stash_bitmap(rr, &ri.ichunk_blocks,
fs/xfs/scrub/rmap_repair.c
805
struct xrep_rmap *rr)
fs/xfs/scrub/rmap_repair.c
817
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/rmap_repair.c
838
error = xrep_rmap_stash_bitmap(rr, &cow_blocks, &XFS_RMAP_OINFO_COW);
fs/xfs/scrub/rmap_repair.c
841
error = xrep_rmap_stash_bitmap(rr, &refcountbt_blocks,
fs/xfs/scrub/rmap_repair.c
853
struct xrep_rmap *rr)
fs/xfs/scrub/rmap_repair.c
855
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/rmap_repair.c
858
return xrep_rmap_stash(rr, XFS_SB_BLOCK(sc->mp),
fs/xfs/scrub/rmap_repair.c
866
struct xrep_rmap *rr)
fs/xfs/scrub/rmap_repair.c
868
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/rmap_repair.c
873
return xrep_rmap_stash(rr,
fs/xfs/scrub/rmap_repair.c
885
struct xrep_rmap *rr = priv;
fs/xfs/scrub/rmap_repair.c
888
error = xrep_rmap_check_mapping(rr->sc, rec);
fs/xfs/scrub/rmap_repair.c
892
rr->nr_records++;
fs/xfs/scrub/rmap_repair.c
904
struct xrep_rmap *rr)
fs/xfs/scrub/rmap_repair.c
906
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/rmap_repair.c
915
error = xrep_rmap_find_inode_rmaps(rr);
fs/xfs/scrub/rmap_repair.c
919
error = xrep_rmap_find_refcount_rmaps(rr);
fs/xfs/scrub/rmap_repair.c
923
error = xrep_rmap_find_agheader_rmaps(rr);
fs/xfs/scrub/rmap_repair.c
927
error = xrep_rmap_find_log_rmaps(rr);
fs/xfs/scrub/rmap_repair.c
954
while ((error = xchk_iscan_iter(&rr->iscan, &ip)) == 1) {
fs/xfs/scrub/rmap_repair.c
955
error = xrep_rmap_scan_inode(rr, ip);
fs/xfs/scrub/rmap_repair.c
963
xchk_iscan_iter_finish(&rr->iscan);
fs/xfs/scrub/rmap_repair.c
983
if (xchk_iscan_aborted(&rr->iscan))
fs/xfs/scrub/rmap_repair.c
993
mcur = xfs_rmapbt_mem_cursor(rr->sc->sa.pag, NULL, &rr->rmap_btree);
fs/xfs/scrub/rmap_repair.c
997
rr->nr_records = 0;
fs/xfs/scrub/rmap_repair.c
998
error = xfs_rmap_query_all(mcur, xrep_rmap_check_record, rr);
fs/xfs/scrub/rtrefcount_repair.c
160
struct xrep_rtrefc *rr,
fs/xfs/scrub/rtrefcount_repair.c
174
if (xchk_should_terminate(rr->sc, &error))
fs/xfs/scrub/rtrefcount_repair.c
179
error = xrep_rtrefc_check_ext(rr->sc, &irec);
fs/xfs/scrub/rtrefcount_repair.c
183
trace_xrep_refc_found(rtg_group(rr->sc->sr.rtg), &irec);
fs/xfs/scrub/rtrefcount_repair.c
185
return xfarray_append(rr->refcount_records, &irec);
fs/xfs/scrub/rtrefcount_repair.c
191
struct xrep_rtrefc *rr,
fs/xfs/scrub/rtrefcount_repair.c
195
return xrep_rtrefc_stash(rr, XFS_REFC_DOMAIN_COW, bno, len, 1);
fs/xfs/scrub/rtrefcount_repair.c
217
struct xrep_rtrefc *rr,
fs/xfs/scrub/rtrefcount_repair.c
221
struct xfs_btree_cur *cur = rr->sc->sr.rmap_cur;
fs/xfs/scrub/rtrefcount_repair.c
235
if (xchk_should_terminate(rr->sc, &error))
fs/xfs/scrub/rtrefcount_repair.c
253
error = xrep_rtrefc_stash_cow(rr, rmap->rm_startblock,
fs/xfs/scrub/rtrefcount_repair.c
311
struct xrep_rtrefc *rr)
fs/xfs/scrub/rtrefcount_repair.c
319
error = xfarray_sort(rr->refcount_records, xrep_rtrefc_extent_cmp,
fs/xfs/scrub/rtrefcount_repair.c
324
foreach_xfarray_idx(rr->refcount_records, cur) {
fs/xfs/scrub/rtrefcount_repair.c
325
if (xchk_should_terminate(rr->sc, &error))
fs/xfs/scrub/rtrefcount_repair.c
328
error = xfarray_load(rr->refcount_records, cur, &irec);
fs/xfs/scrub/rtrefcount_repair.c
356
struct xrep_rtrefc *rr = priv;
fs/xfs/scrub/rtrefcount_repair.c
359
if (xchk_should_terminate(rr->sc, &error))
fs/xfs/scrub/rtrefcount_repair.c
363
if (rec->rm_owner != rr->sc->ip->i_ino)
fs/xfs/scrub/rtrefcount_repair.c
366
error = xrep_check_ino_btree_mapping(rr->sc, rec);
fs/xfs/scrub/rtrefcount_repair.c
370
return xfsb_bitmap_set(&rr->old_rtrefcountbt_blocks,
fs/xfs/scrub/rtrefcount_repair.c
383
struct xrep_rtrefc *rr,
fs/xfs/scrub/rtrefcount_repair.c
389
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/rtrefcount_repair.c
394
error = rcbag_add(rcstack, rr->sc->tp, rmap);
fs/xfs/scrub/rtrefcount_repair.c
398
error = xrep_rtrefc_walk_rmaps(rr, rmap, have);
fs/xfs/scrub/rtrefcount_repair.c
417
struct xrep_rtrefc *rr,
fs/xfs/scrub/rtrefcount_repair.c
420
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/rtrefcount_repair.c
427
error = xfs_rmap_query_all(sc->sa.rmap_cur, xrep_rtrefc_walk_rmap, rr);
fs/xfs/scrub/rtrefcount_repair.c
435
struct xrep_rtrefc *rr)
fs/xfs/scrub/rtrefcount_repair.c
437
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/rtrefcount_repair.c
449
error = xrep_rtrefc_scan_ag(rr, pag);
fs/xfs/scrub/rtrefcount_repair.c
477
error = xrep_rtrefc_walk_rmaps(rr, &rmap, &have);
fs/xfs/scrub/rtrefcount_repair.c
483
error = xrep_rtrefc_push_rmaps_at(rr, rcstack, sbno, &rmap,
fs/xfs/scrub/rtrefcount_repair.c
504
error = xrep_rtrefc_walk_rmaps(rr, &rmap, &have);
fs/xfs/scrub/rtrefcount_repair.c
508
error = xrep_rtrefc_push_rmaps_at(rr, rcstack,
fs/xfs/scrub/rtrefcount_repair.c
518
error = xrep_rtrefc_stash(rr,
fs/xfs/scrub/rtrefcount_repair.c
561
struct xrep_rtrefc *rr = priv;
fs/xfs/scrub/rtrefcount_repair.c
567
error = xfarray_load(rr->refcount_records, rr->array_cur++,
fs/xfs/scrub/rtrefcount_repair.c
586
struct xrep_rtrefc *rr = priv;
fs/xfs/scrub/rtrefcount_repair.c
588
return xrep_newbt_claim_block(cur, &rr->new_btree, ptr);
fs/xfs/scrub/rtrefcount_repair.c
610
struct xrep_rtrefc *rr)
fs/xfs/scrub/rtrefcount_repair.c
612
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/rtrefcount_repair.c
617
error = xrep_rtrefc_sort_records(rr);
fs/xfs/scrub/rtrefcount_repair.c
627
error = xrep_newbt_init_metadir_inode(&rr->new_btree, sc);
fs/xfs/scrub/rtrefcount_repair.c
631
rr->new_btree.bload.get_records = xrep_rtrefc_get_records;
fs/xfs/scrub/rtrefcount_repair.c
632
rr->new_btree.bload.claim_block = xrep_rtrefc_claim_block;
fs/xfs/scrub/rtrefcount_repair.c
633
rr->new_btree.bload.iroot_size = xrep_rtrefc_iroot_size;
fs/xfs/scrub/rtrefcount_repair.c
636
xfs_btree_stage_ifakeroot(refc_cur, &rr->new_btree.ifake);
fs/xfs/scrub/rtrefcount_repair.c
639
error = xfs_btree_bload_compute_geometry(refc_cur, &rr->new_btree.bload,
fs/xfs/scrub/rtrefcount_repair.c
640
xfarray_length(rr->refcount_records));
fs/xfs/scrub/rtrefcount_repair.c
656
rr->new_btree.bload.nr_blocks, 0, true);
fs/xfs/scrub/rtrefcount_repair.c
661
error = xrep_newbt_alloc_blocks(&rr->new_btree,
fs/xfs/scrub/rtrefcount_repair.c
662
rr->new_btree.bload.nr_blocks);
fs/xfs/scrub/rtrefcount_repair.c
667
rr->new_btree.ifake.if_fork->if_format = XFS_DINODE_FMT_META_BTREE;
fs/xfs/scrub/rtrefcount_repair.c
668
rr->array_cur = XFARRAY_CURSOR_INIT;
fs/xfs/scrub/rtrefcount_repair.c
669
error = xfs_btree_bload(refc_cur, &rr->new_btree.bload, rr);
fs/xfs/scrub/rtrefcount_repair.c
679
xrep_inode_set_nblocks(rr->sc, rr->new_btree.ifake.if_blocks);
fs/xfs/scrub/rtrefcount_repair.c
683
error = xrep_newbt_commit(&rr->new_btree);
fs/xfs/scrub/rtrefcount_repair.c
690
xrep_newbt_cancel(&rr->new_btree);
fs/xfs/scrub/rtrefcount_repair.c
699
struct xrep_rtrefc *rr;
fs/xfs/scrub/rtrefcount_repair.c
712
rr = kzalloc_obj(struct xrep_rtrefc, XCHK_GFP_FLAGS);
fs/xfs/scrub/rtrefcount_repair.c
713
if (!rr)
fs/xfs/scrub/rtrefcount_repair.c
715
rr->sc = sc;
fs/xfs/scrub/rtrefcount_repair.c
720
&rr->refcount_records);
fs/xfs/scrub/rtrefcount_repair.c
725
xfsb_bitmap_init(&rr->old_rtrefcountbt_blocks);
fs/xfs/scrub/rtrefcount_repair.c
726
error = xrep_rtrefc_find_refcounts(rr);
fs/xfs/scrub/rtrefcount_repair.c
733
error = xrep_rtrefc_build_new_tree(rr);
fs/xfs/scrub/rtrefcount_repair.c
741
error = xrep_reap_metadir_fsblocks(rr->sc,
fs/xfs/scrub/rtrefcount_repair.c
742
&rr->old_rtrefcountbt_blocks);
fs/xfs/scrub/rtrefcount_repair.c
747
xfsb_bitmap_destroy(&rr->old_rtrefcountbt_blocks);
fs/xfs/scrub/rtrefcount_repair.c
748
xfarray_destroy(rr->refcount_records);
fs/xfs/scrub/rtrefcount_repair.c
750
kfree(rr);
fs/xfs/scrub/rtrmap_repair.c
105
struct xrep_rtrmap *rr;
fs/xfs/scrub/rtrmap_repair.c
114
rr = kzalloc_obj(struct xrep_rtrmap, XCHK_GFP_FLAGS);
fs/xfs/scrub/rtrmap_repair.c
115
if (!rr)
fs/xfs/scrub/rtrmap_repair.c
118
rr->sc = sc;
fs/xfs/scrub/rtrmap_repair.c
119
sc->buf = rr;
fs/xfs/scrub/rtrmap_repair.c
140
struct xrep_rtrmap *rr,
fs/xfs/scrub/rtrmap_repair.c
154
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/rtrmap_repair.c
161
if (xchk_iscan_aborted(&rr->iscan))
fs/xfs/scrub/rtrmap_repair.c
167
mutex_lock(&rr->lock);
fs/xfs/scrub/rtrmap_repair.c
168
mcur = xfs_rtrmapbt_mem_cursor(sc->sr.rtg, sc->tp, &rr->rtrmap_btree);
fs/xfs/scrub/rtrmap_repair.c
174
error = xfbtree_trans_commit(&rr->rtrmap_btree, sc->tp);
fs/xfs/scrub/rtrmap_repair.c
178
mutex_unlock(&rr->lock);
fs/xfs/scrub/rtrmap_repair.c
182
xfbtree_trans_cancel(&rr->rtrmap_btree, sc->tp);
fs/xfs/scrub/rtrmap_repair.c
184
xchk_iscan_abort(&rr->iscan);
fs/xfs/scrub/rtrmap_repair.c
185
mutex_unlock(&rr->lock);
fs/xfs/scrub/rtrmap_repair.c
199
struct xrep_rtrmap *rr;
fs/xfs/scrub/rtrmap_repair.c
210
return xrep_rtrmap_stash(rf->rr, rf->accum.rm_startblock,
fs/xfs/scrub/rtrmap_repair.c
224
struct xfs_mount *mp = rf->rr->sc->mp;
fs/xfs/scrub/rtrmap_repair.c
230
rtg_rgno(rf->rr->sc->sr.rtg))
fs/xfs/scrub/rtrmap_repair.c
269
struct xrep_rtrmap *rr = rf->rr;
fs/xfs/scrub/rtrmap_repair.c
285
cur = xfs_bmbt_init_cursor(rr->sc->mp, rr->sc->tp, ip, XFS_DATA_FORK);
fs/xfs/scrub/rtrmap_repair.c
323
struct xrep_rtrmap *rr,
fs/xfs/scrub/rtrmap_repair.c
328
.rr = rr,
fs/xfs/scrub/rtrmap_repair.c
356
struct xrep_rtrmap *rr,
fs/xfs/scrub/rtrmap_repair.c
363
if (rr->sc->ip == ip)
fs/xfs/scrub/rtrmap_repair.c
370
error = xrep_rtrmap_scan_dfork(rr, ip);
fs/xfs/scrub/rtrmap_repair.c
375
xchk_iscan_mark_visited(&rr->iscan, ip);
fs/xfs/scrub/rtrmap_repair.c
388
struct xrep_rtrmap *rr = priv;
fs/xfs/scrub/rtrmap_repair.c
391
if (xchk_should_terminate(rr->sc, &error))
fs/xfs/scrub/rtrmap_repair.c
395
if (rec->rm_owner != rr->sc->ip->i_ino)
fs/xfs/scrub/rtrmap_repair.c
398
error = xrep_check_ino_btree_mapping(rr->sc, rec);
fs/xfs/scrub/rtrmap_repair.c
402
return xfsb_bitmap_set(&rr->old_rtrmapbt_blocks,
fs/xfs/scrub/rtrmap_repair.c
410
struct xrep_rtrmap *rr,
fs/xfs/scrub/rtrmap_repair.c
413
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/rtrmap_repair.c
420
error = xfs_rmap_query_all(sc->sa.rmap_cur, xrep_rtrmap_walk_rmap, rr);
fs/xfs/scrub/rtrmap_repair.c
426
struct xrep_rtrmap *rr;
fs/xfs/scrub/rtrmap_repair.c
437
struct xrep_rtrmap *rr = rsr->rr;
fs/xfs/scrub/rtrmap_repair.c
440
return xrep_rtrmap_stash(rr, rgbno, len, rsr->owner, 0, 0);
fs/xfs/scrub/rtrmap_repair.c
449
struct xrep_rtrmap *rr,
fs/xfs/scrub/rtrmap_repair.c
454
.rr = rr,
fs/xfs/scrub/rtrmap_repair.c
484
struct xrep_rtrmap *rr)
fs/xfs/scrub/rtrmap_repair.c
495
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/rtrmap_repair.c
510
error = xrep_rtrmap_stash_bitmap(rr, &cow_blocks, &XFS_RMAP_OINFO_COW);
fs/xfs/scrub/rtrmap_repair.c
526
struct xrep_rtrmap *rr = priv;
fs/xfs/scrub/rtrmap_repair.c
529
error = xrep_rtrmap_check_mapping(rr->sc, rec);
fs/xfs/scrub/rtrmap_repair.c
533
rr->nr_records++;
fs/xfs/scrub/rtrmap_repair.c
540
struct xrep_rtrmap *rr)
fs/xfs/scrub/rtrmap_repair.c
542
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/rtrmap_repair.c
549
if (xfs_has_rtsb(sc->mp) && rtg_rgno(rr->sc->sr.rtg) == 0) {
fs/xfs/scrub/rtrmap_repair.c
550
error = xrep_rtrmap_stash(rr, 0, sc->mp->m_sb.sb_rextsize,
fs/xfs/scrub/rtrmap_repair.c
558
error = xrep_rtrmap_find_refcount_rmaps(rr);
fs/xfs/scrub/rtrmap_repair.c
582
while ((error = xchk_iscan_iter(&rr->iscan, &ip)) == 1) {
fs/xfs/scrub/rtrmap_repair.c
583
error = xrep_rtrmap_scan_inode(rr, ip);
fs/xfs/scrub/rtrmap_repair.c
591
xchk_iscan_iter_finish(&rr->iscan);
fs/xfs/scrub/rtrmap_repair.c
611
if (xchk_iscan_aborted(&rr->iscan))
fs/xfs/scrub/rtrmap_repair.c
616
error = xrep_rtrmap_scan_ag(rr, pag);
fs/xfs/scrub/rtrmap_repair.c
630
mcur = xfs_rtrmapbt_mem_cursor(rr->sc->sr.rtg, NULL, &rr->rtrmap_btree);
fs/xfs/scrub/rtrmap_repair.c
631
rr->nr_records = 0;
fs/xfs/scrub/rtrmap_repair.c
632
error = xfs_rmap_query_all(mcur, xrep_rtrmap_check_record, rr);
fs/xfs/scrub/rtrmap_repair.c
649
struct xrep_rtrmap *rr = priv;
fs/xfs/scrub/rtrmap_repair.c
657
error = xfs_btree_increment(rr->mcur, 0, &stat);
fs/xfs/scrub/rtrmap_repair.c
663
error = xfs_rmap_get_rec(rr->mcur, &cur->bc_rec.r, &stat);
fs/xfs/scrub/rtrmap_repair.c
683
struct xrep_rtrmap *rr = priv;
fs/xfs/scrub/rtrmap_repair.c
685
return xrep_newbt_claim_block(cur, &rr->new_btree, ptr);
fs/xfs/scrub/rtrmap_repair.c
707
struct xrep_rtrmap *rr)
fs/xfs/scrub/rtrmap_repair.c
709
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/rtrmap_repair.c
720
error = xrep_newbt_init_metadir_inode(&rr->new_btree, sc);
fs/xfs/scrub/rtrmap_repair.c
724
rr->new_btree.bload.get_records = xrep_rtrmap_get_records;
fs/xfs/scrub/rtrmap_repair.c
725
rr->new_btree.bload.claim_block = xrep_rtrmap_claim_block;
fs/xfs/scrub/rtrmap_repair.c
726
rr->new_btree.bload.iroot_size = xrep_rtrmap_iroot_size;
fs/xfs/scrub/rtrmap_repair.c
729
xfs_btree_stage_ifakeroot(rmap_cur, &rr->new_btree.ifake);
fs/xfs/scrub/rtrmap_repair.c
733
&rr->new_btree.bload, rr->nr_records);
fs/xfs/scrub/rtrmap_repair.c
749
rr->new_btree.bload.nr_blocks, 0, true);
fs/xfs/scrub/rtrmap_repair.c
754
error = xrep_newbt_alloc_blocks(&rr->new_btree,
fs/xfs/scrub/rtrmap_repair.c
755
rr->new_btree.bload.nr_blocks);
fs/xfs/scrub/rtrmap_repair.c
763
rr->mcur = xfs_rtrmapbt_mem_cursor(sc->sr.rtg, NULL, &rr->rtrmap_btree);
fs/xfs/scrub/rtrmap_repair.c
764
error = xfs_btree_goto_left_edge(rr->mcur);
fs/xfs/scrub/rtrmap_repair.c
769
rr->new_btree.ifake.if_fork->if_format = XFS_DINODE_FMT_META_BTREE;
fs/xfs/scrub/rtrmap_repair.c
770
error = xfs_btree_bload(rmap_cur, &rr->new_btree.bload, rr);
fs/xfs/scrub/rtrmap_repair.c
780
xrep_inode_set_nblocks(rr->sc, rr->new_btree.ifake.if_blocks);
fs/xfs/scrub/rtrmap_repair.c
782
xfs_btree_del_cursor(rr->mcur, 0);
fs/xfs/scrub/rtrmap_repair.c
783
rr->mcur = NULL;
fs/xfs/scrub/rtrmap_repair.c
789
xchk_iscan_abort(&rr->iscan);
fs/xfs/scrub/rtrmap_repair.c
792
error = xrep_newbt_commit(&rr->new_btree);
fs/xfs/scrub/rtrmap_repair.c
799
xfs_btree_del_cursor(rr->mcur, error);
fs/xfs/scrub/rtrmap_repair.c
802
xrep_newbt_cancel(&rr->new_btree);
fs/xfs/scrub/rtrmap_repair.c
840
struct xrep_rtrmap *rr;
fs/xfs/scrub/rtrmap_repair.c
846
rr = container_of(nb, struct xrep_rtrmap, rhook.rmap_hook.nb);
fs/xfs/scrub/rtrmap_repair.c
847
mp = rr->sc->mp;
fs/xfs/scrub/rtrmap_repair.c
849
if (!xrep_rtrmapbt_want_live_update(&rr->iscan, &p->oinfo))
fs/xfs/scrub/rtrmap_repair.c
852
trace_xrep_rmap_live_update(rtg_group(rr->sc->sr.rtg), action, p);
fs/xfs/scrub/rtrmap_repair.c
856
mutex_lock(&rr->lock);
fs/xfs/scrub/rtrmap_repair.c
857
mcur = xfs_rtrmapbt_mem_cursor(rr->sc->sr.rtg, tp, &rr->rtrmap_btree);
fs/xfs/scrub/rtrmap_repair.c
864
error = xfbtree_trans_commit(&rr->rtrmap_btree, tp);
fs/xfs/scrub/rtrmap_repair.c
869
mutex_unlock(&rr->lock);
fs/xfs/scrub/rtrmap_repair.c
873
xfbtree_trans_cancel(&rr->rtrmap_btree, tp);
fs/xfs/scrub/rtrmap_repair.c
875
xchk_iscan_abort(&rr->iscan);
fs/xfs/scrub/rtrmap_repair.c
876
mutex_unlock(&rr->lock);
fs/xfs/scrub/rtrmap_repair.c
884
struct xrep_rtrmap *rr)
fs/xfs/scrub/rtrmap_repair.c
886
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/rtrmap_repair.c
889
mutex_init(&rr->lock);
fs/xfs/scrub/rtrmap_repair.c
890
xfsb_bitmap_init(&rr->old_rtrmapbt_blocks);
fs/xfs/scrub/rtrmap_repair.c
893
error = xfs_rtrmapbt_mem_init(sc->mp, &rr->rtrmap_btree, sc->xmbtp,
fs/xfs/scrub/rtrmap_repair.c
899
xchk_iscan_start(sc, 30000, 100, &rr->iscan);
fs/xfs/scrub/rtrmap_repair.c
908
xfs_rmap_hook_setup(&rr->rhook, xrep_rtrmapbt_live_update);
fs/xfs/scrub/rtrmap_repair.c
909
error = xfs_rmap_hook_add(rtg_group(sc->sr.rtg), &rr->rhook);
fs/xfs/scrub/rtrmap_repair.c
915
xchk_iscan_teardown(&rr->iscan);
fs/xfs/scrub/rtrmap_repair.c
916
xfbtree_destroy(&rr->rtrmap_btree);
fs/xfs/scrub/rtrmap_repair.c
918
xfsb_bitmap_destroy(&rr->old_rtrmapbt_blocks);
fs/xfs/scrub/rtrmap_repair.c
919
mutex_destroy(&rr->lock);
fs/xfs/scrub/rtrmap_repair.c
926
struct xrep_rtrmap *rr)
fs/xfs/scrub/rtrmap_repair.c
928
struct xfs_scrub *sc = rr->sc;
fs/xfs/scrub/rtrmap_repair.c
930
xchk_iscan_abort(&rr->iscan);
fs/xfs/scrub/rtrmap_repair.c
931
xfs_rmap_hook_del(rtg_group(sc->sr.rtg), &rr->rhook);
fs/xfs/scrub/rtrmap_repair.c
932
xchk_iscan_teardown(&rr->iscan);
fs/xfs/scrub/rtrmap_repair.c
933
xfbtree_destroy(&rr->rtrmap_btree);
fs/xfs/scrub/rtrmap_repair.c
934
xfsb_bitmap_destroy(&rr->old_rtrmapbt_blocks);
fs/xfs/scrub/rtrmap_repair.c
935
mutex_destroy(&rr->lock);
fs/xfs/scrub/rtrmap_repair.c
943
struct xrep_rtrmap *rr = sc->buf;
fs/xfs/scrub/rtrmap_repair.c
951
error = xrep_rtrmap_setup_scan(rr);
fs/xfs/scrub/rtrmap_repair.c
956
error = xrep_rtrmap_find_rmaps(rr);
fs/xfs/scrub/rtrmap_repair.c
963
error = xrep_rtrmap_build_new_tree(rr);
fs/xfs/scrub/rtrmap_repair.c
971
error = xrep_reap_metadir_fsblocks(rr->sc, &rr->old_rtrmapbt_blocks);
fs/xfs/scrub/rtrmap_repair.c
976
xrep_rtrmap_teardown(rr);
include/crypto/cast5.h
16
int rr; /* rr ? rounds = 12 : rounds = 16; (rfc 2144) */
include/linux/bnge/hsi.h
8526
__le16 rr;
include/linux/bnxt/hsi.h
7789
__le16 rr;
include/linux/iommu.h
313
struct iommu_resv_region rr;
include/linux/mlx5/mlx5_ifc.h
4492
u8 rr[0x1];
include/net/inet_sock.h
47
unsigned char rr;
include/uapi/linux/batadv_packet.h
390
__u8 rr[BATADV_RR_LEN][ETH_ALEN];
include/video/omapfb_dss.h
289
s16 rr, rg, rb;
io_uring/rsrc.c
382
struct io_uring_rsrc_register rr;
io_uring/rsrc.c
385
if (size != sizeof(rr))
io_uring/rsrc.c
388
memset(&rr, 0, sizeof(rr));
io_uring/rsrc.c
389
if (copy_from_user(&rr, arg, size))
io_uring/rsrc.c
391
if (!rr.nr || rr.resv2)
io_uring/rsrc.c
393
if (rr.flags & ~IORING_RSRC_REGISTER_SPARSE)
io_uring/rsrc.c
398
if (rr.flags & IORING_RSRC_REGISTER_SPARSE && rr.data)
io_uring/rsrc.c
400
return io_sqe_files_register(ctx, u64_to_user_ptr(rr.data),
io_uring/rsrc.c
401
rr.nr, u64_to_user_ptr(rr.tags));
io_uring/rsrc.c
403
if (rr.flags & IORING_RSRC_REGISTER_SPARSE && rr.data)
io_uring/rsrc.c
405
return io_sqe_buffers_register(ctx, u64_to_user_ptr(rr.data),
io_uring/rsrc.c
406
rr.nr, u64_to_user_ptr(rr.tags));
net/batman-adv/routing.c
385
ether_addr_copy(icmp_packet_rr->rr[icmp_packet_rr->rr_cur],
net/bluetooth/smp.c
107
u8 rr[16]; /* Remote OOB ra/rb value */
net/bluetooth/smp.c
1471
memcpy(r, smp->rr, 16);
net/bluetooth/smp.c
2754
smp->rr, 0, cfm.confirm_val);
net/bluetooth/smp.c
691
memcpy(smp->rr, oob_data->rand256, 16);
net/bluetooth/smp.c
694
SMP_DBG("OOB Remote Random: %16phN", smp->rr);
net/ipv4/cipso_ipv4.c
2006
if (opt->opt.srr || opt->opt.rr || opt->opt.ts || opt->opt.router_alert) {
net/ipv4/cipso_ipv4.c
2018
if (opt->opt.rr > opt->opt.cipso)
net/ipv4/cipso_ipv4.c
2019
opt->opt.rr -= cipso_len;
net/ipv4/ip_options.c
220
opt->rr = 0;
net/ipv4/ip_options.c
312
if (opt->rr) {
net/ipv4/ip_options.c
337
opt->rr = optptr - iph;
net/ipv4/ip_options.c
491
unsigned char *optptr = opt->__data + opt->rr - sizeof(struct iphdr);
net/ipv4/ip_options.c
546
optptr = (unsigned char *)raw + opt->rr;
net/ipv4/ip_options.c
57
ip_rt_get_source(iph + opt->rr + iph[opt->rr + 2] - 5, skb, rt);
net/ipv4/ip_options.c
92
if (sopt->rr) {
net/ipv4/ip_options.c
93
optlen = sptr[sopt->rr+1];
net/ipv4/ip_options.c
94
soffset = sptr[sopt->rr+2];
net/ipv4/ip_options.c
95
dopt->rr = dopt->optlen + sizeof(struct iphdr);
net/ipv4/ip_options.c
96
memcpy(dptr, sptr+sopt->rr, optlen);
net/netfilter/nft_exthdr.c
122
if (!opt->rr)
net/netfilter/nft_exthdr.c
124
*offset = opt->rr;
net/wireless/reg.c
1623
const struct ieee80211_reg_rule *rr;
net/wireless/reg.c
1626
rr = &regd->reg_rules[i];
net/wireless/reg.c
1627
fr = &rr->freq_range;
net/wireless/reg.c
1640
return rr;
tools/perf/builtin-lock.c
423
static void insert_to(struct rb_root *rr, struct lock_stat *st,
tools/perf/builtin-lock.c
426
struct rb_node **rb = &rr->rb_node;
tools/perf/builtin-lock.c
441
rb_insert_color(&st->rb, rr);
tools/perf/builtin-lock.c
453
static inline struct lock_stat *pop_from(struct rb_root *rr)
tools/perf/builtin-lock.c
455
struct rb_node *node = rr->rb_node;
tools/perf/builtin-lock.c
463
rb_erase(node, rr);