arch/arm/vfp/vfp.h
105
u64 rh, rl;
arch/arm/vfp/vfp.h
106
mul64to128(&rh, &rl, n, m);
arch/arm/vfp/vfp.h
107
return rh | (rl != 0);
arch/arm/vfp/vfp.h
73
u64 rh, rma, rmb, rl;
arch/arm/vfp/vfp.h
77
rl = (u64)nl * ml;
arch/arm/vfp/vfp.h
90
rl += rma;
arch/arm/vfp/vfp.h
91
rh += (rl < rma);
arch/arm/vfp/vfp.h
93
*resl = rl;
arch/arm64/include/asm/atomic_ll_sc.h
323
[rl] "=&r" (r.low), [rh] "=&r" (r.high), \
arch/mips/include/asm/octeon/cvmx-pciercx-defs.h
197
__BITFIELD_FIELD(uint32_t rl:1,
arch/riscv/include/asm/bitops.h
313
__op_bit_ord(and, __NOT, nr, addr, .rl);
arch/riscv/net/bpf_jit.h
276
static inline u32 rv_amo_insn(u8 funct5, u8 aq, u8 rl, u8 rs2, u8 rs1,
arch/riscv/net/bpf_jit.h
279
u8 funct7 = (funct5 << 2) | (aq << 1) | rl;
arch/riscv/net/bpf_jit.h
559
static inline u32 rv_amoadd_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
561
return rv_amo_insn(0, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
564
static inline u32 rv_amoand_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
566
return rv_amo_insn(0xc, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
569
static inline u32 rv_amoor_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
571
return rv_amo_insn(0x8, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
574
static inline u32 rv_amoxor_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
576
return rv_amo_insn(0x4, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
579
static inline u32 rv_amoswap_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
581
return rv_amo_insn(0x1, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
584
static inline u32 rv_lr_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
586
return rv_amo_insn(0x2, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
589
static inline u32 rv_sc_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
591
return rv_amo_insn(0x3, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
751
static inline u32 rvzacas_amocas_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
753
return rv_amo_insn(0x5, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
756
static inline u32 rvzacas_amocas_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
758
return rv_amo_insn(0x5, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
893
static inline u32 rv_amoadd_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
895
return rv_amo_insn(0, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
898
static inline u32 rv_amoand_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
900
return rv_amo_insn(0xc, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
903
static inline u32 rv_amoor_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
905
return rv_amo_insn(0x8, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
908
static inline u32 rv_amoxor_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
910
return rv_amo_insn(0x4, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
913
static inline u32 rv_amoswap_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
915
return rv_amo_insn(0x1, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
918
static inline u32 rv_lr_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
920
return rv_amo_insn(0x2, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
923
static inline u32 rv_sc_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
925
return rv_amo_insn(0x3, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/s390/kvm/guestdbg.c
471
s32 rl = *((s32 *) (opcode + 1));
arch/s390/kvm/guestdbg.c
474
*addr += (u64)(s64) rl * 2;
block/blk-cgroup.h
498
#define blk_queue_for_each_rl(rl, q) \
block/blk-cgroup.h
499
for ((rl) = &(q)->root_rl; (rl); (rl) = NULL)
block/sed-opal.c
1472
bool rl, bool wl)
block/sed-opal.c
1494
add_token_u8(&err, dev, rl);
crypto/camellia_generic.c
319
#define ROLDQ(ll, lr, rl, rr, w0, w1, bits) ({ \
crypto/camellia_generic.c
322
lr = (lr << bits) + (rl >> (32 - bits)); \
crypto/camellia_generic.c
323
rl = (rl << bits) + (rr >> (32 - bits)); \
crypto/camellia_generic.c
327
#define ROLDQo32(ll, lr, rl, rr, w0, w1, bits) ({ \
crypto/camellia_generic.c
330
ll = (lr << (bits - 32)) + (rl >> (64 - bits)); \
crypto/camellia_generic.c
331
lr = (rl << (bits - 32)) + (rr >> (64 - bits)); \
crypto/camellia_generic.c
332
rl = (rr << (bits - 32)) + (w0 >> (64 - bits)); \
crypto/camellia_generic.c
811
#define CAMELLIA_FLS(ll, lr, rl, rr, kll, klr, krl, krr, t0, t1, t2, t3) ({ \
crypto/camellia_generic.c
816
rl ^= t2; \
crypto/camellia_generic.c
820
t3 &= rl; \
crypto/crypto_user.c
106
struct crypto_report_larval rl;
crypto/crypto_user.c
108
memset(&rl, 0, sizeof(rl));
crypto/crypto_user.c
109
strscpy(rl.type, "larval", sizeof(rl.type));
crypto/crypto_user.c
110
if (nla_put(skb, CRYPTOCFGA_REPORT_LARVAL, sizeof(rl), &rl))
drivers/block/drbd/drbd_main.c
1089
unsigned long rl;
drivers/block/drbd/drbd_main.c
1122
rl = tmp - c->bit_offset;
drivers/block/drbd/drbd_main.c
1125
if (rl == 0) {
drivers/block/drbd/drbd_main.c
1138
if (rl == 0) {
drivers/block/drbd/drbd_main.c
1144
bits = vli_encode_bits(&bs, rl);
drivers/block/drbd/drbd_main.c
1153
plain_bits += rl;
drivers/block/drbd/drbd_receiver.c
4508
u64 rl;
drivers/block/drbd/drbd_receiver.c
4522
for (have = bits; have > 0; s += rl, toggle = !toggle) {
drivers/block/drbd/drbd_receiver.c
4523
bits = vli_decode_bits(&rl, look_ahead);
drivers/block/drbd/drbd_receiver.c
4528
e = s + rl -1;
drivers/crypto/intel/qat/qat_common/adf_rl.c
1053
struct adf_rl *rl;
drivers/crypto/intel/qat/qat_common/adf_rl.c
1068
rl = kzalloc_obj(*rl);
drivers/crypto/intel/qat/qat_common/adf_rl.c
1069
if (!rl) {
drivers/crypto/intel/qat/qat_common/adf_rl.c
1074
mutex_init(&rl->rl_lock);
drivers/crypto/intel/qat/qat_common/adf_rl.c
1075
rl->device_data = &accel_dev->hw_device->rl_data;
drivers/crypto/intel/qat/qat_common/adf_rl.c
1076
rl->accel_dev = accel_dev;
drivers/crypto/intel/qat/qat_common/adf_rl.c
1077
init_rwsem(&rl->user_input.lock);
drivers/crypto/intel/qat/qat_common/adf_rl.c
1078
accel_dev->rate_limiting = rl;
drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c
1157
if (!is_psp_fw_valid(psp->rl))
drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c
1163
memcpy(psp->fw_pri_buf, psp->rl.start_addr, psp->rl.size_bytes);
drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c
1168
cmd->cmd.cmd_load_ip_fw.fw_size = psp->rl.size_bytes;
drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c
3591
psp->rl.fw_version = le32_to_cpu(desc->fw_version);
drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c
3592
psp->rl.feature_version = le32_to_cpu(desc->fw_version);
drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c
3593
psp->rl.size_bytes = le32_to_cpu(desc->size_bytes);
drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c
3594
psp->rl.start_addr = ucode_start_addr;
drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c
3745
adev->psp.rl.size_bytes = le32_to_cpu(sos_hdr_v1_3->rl.size_bytes);
drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c
3746
adev->psp.rl.start_addr = ucode_array_start_addr +
drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c
3747
le32_to_cpu(sos_hdr_v1_3->rl.offset_bytes);
drivers/gpu/drm/amd/amdgpu/amdgpu_psp.h
382
struct psp_bin_desc rl;
drivers/gpu/drm/amd/amdgpu/amdgpu_ucode.h
106
struct psp_fw_legacy_bin_desc rl;
drivers/hwmon/asc7621.c
799
#define PREAD(name, n, pri, rm, rl, m, s, r) \
drivers/hwmon/asc7621.c
801
.priority = pri, .msb[0] = rm, .lsb[0] = rl, .mask[0] = m, \
drivers/hwmon/asc7621.c
804
#define PWRITE(name, n, pri, rm, rl, m, s, r) \
drivers/hwmon/asc7621.c
806
.priority = pri, .msb[0] = rm, .lsb[0] = rl, .mask[0] = m, \
drivers/hwmon/asc7621.c
813
#define PWRITEM(name, n, pri, rm, rl, m, s, r) \
drivers/hwmon/asc7621.c
815
.priority = pri, .msb = rm, .lsb = rl, .mask = m, .shift = s,}
drivers/infiniband/hw/mlx5/mlx5_ib.h
537
struct mlx5_rate_limit rl;
drivers/infiniband/hw/mlx5/qp.c
3845
struct mlx5_rate_limit old_rl = ibqp->rl;
drivers/infiniband/hw/mlx5/qp.c
3870
new_rl = raw_qp_param->rl;
drivers/infiniband/hw/mlx5/qp.c
3908
ibqp->rl = new_rl;
drivers/infiniband/hw/mlx5/qp.c
4383
raw_qp_param.rl.rate = attr->rate_limit;
drivers/infiniband/hw/mlx5/qp.c
4388
raw_qp_param.rl.max_burst_sz =
drivers/infiniband/hw/mlx5/qp.c
4399
raw_qp_param.rl.typical_pkt_sz =
drivers/infiniband/hw/mlx5/qp.c
72
struct mlx5_rate_limit rl;
drivers/net/ethernet/hisilicon/hns3/hns3_ethtool.c
1440
u32 rl;
drivers/net/ethernet/hisilicon/hns3/hns3_ethtool.c
1455
rl = hns3_rl_round_down(cmd->rx_coalesce_usecs_high);
drivers/net/ethernet/hisilicon/hns3/hns3_ethtool.c
1456
if (rl != cmd->rx_coalesce_usecs_high) {
drivers/net/ethernet/hisilicon/hns3/hns3_ethtool.c
1459
cmd->rx_coalesce_usecs_high, rl);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2870
struct virtchnl2_rss_lut *rl __free(kfree) = NULL;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2884
buf_size = struct_size(rl, lut, rss_data->rss_lut_size);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2885
rl = kzalloc(buf_size, GFP_KERNEL);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2886
if (!rl)
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2889
rl->vport_id = cpu_to_le32(vport_id);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2892
xn_params.send_buf.iov_base = rl;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2903
rl->lut_entries = cpu_to_le16(rss_data->rss_lut_size);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2905
rl->lut[i] = rxhash_ena ?
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
466
void mlx5e_mqprio_rl_free(struct mlx5e_mqprio_rl *rl)
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
468
kvfree(rl);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
471
int mlx5e_mqprio_rl_init(struct mlx5e_mqprio_rl *rl, struct mlx5_core_dev *mdev, u8 num_tc,
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
484
rl->mdev = mdev;
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
485
rl->num_tc = num_tc;
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
486
rl->leaves_id = kvcalloc(num_tc, sizeof(*rl->leaves_id), GFP_KERNEL);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
487
if (!rl->leaves_id)
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
490
err = mlx5_qos_create_root_node(mdev, &rl->root_id);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
494
qos_dbg(mdev, "Root created, id %#x\n", rl->root_id);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
500
err = mlx5_qos_create_leaf_node(mdev, rl->root_id, 0, max_average_bw,
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
501
&rl->leaves_id[tc]);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
506
tc, rl->leaves_id[tc], max_average_bw);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
512
mlx5_qos_destroy_node(mdev, rl->leaves_id[tc]);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
513
mlx5_qos_destroy_node(mdev, rl->root_id);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
515
kvfree(rl->leaves_id);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
519
void mlx5e_mqprio_rl_cleanup(struct mlx5e_mqprio_rl *rl)
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
523
for (tc = 0; tc < rl->num_tc; tc++)
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
524
mlx5_qos_destroy_node(rl->mdev, rl->leaves_id[tc]);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
525
mlx5_qos_destroy_node(rl->mdev, rl->root_id);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
526
kvfree(rl->leaves_id);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
529
int mlx5e_mqprio_rl_get_node_hw_id(struct mlx5e_mqprio_rl *rl, int tc, u32 *hw_id)
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
531
if (tc >= rl->num_tc)
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
534
*hw_id = rl->leaves_id[tc];
drivers/net/ethernet/mellanox/mlx5/core/en/qos.h
46
void mlx5e_mqprio_rl_free(struct mlx5e_mqprio_rl *rl);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.h
47
int mlx5e_mqprio_rl_init(struct mlx5e_mqprio_rl *rl, struct mlx5_core_dev *mdev, u8 num_tc,
drivers/net/ethernet/mellanox/mlx5/core/en/qos.h
49
void mlx5e_mqprio_rl_cleanup(struct mlx5e_mqprio_rl *rl);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.h
50
int mlx5e_mqprio_rl_get_node_hw_id(struct mlx5e_mqprio_rl *rl, int tc, u32 *hw_id);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1982
struct mlx5_rate_limit rl = {0};
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1989
rl.rate = sq->rate_limit;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1990
mlx5_rl_remove_rate(mdev, &rl);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
2492
struct mlx5_rate_limit rl = {0};
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
2501
rl.rate = sq->rate_limit;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
2503
mlx5_rl_remove_rate(mdev, &rl);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
2509
rl.rate = rate;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
2510
err = mlx5_rl_add_rate(mdev, &rl_index, &rl);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
2528
mlx5_rl_remove_rate(mdev, &rl);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3796
struct mlx5e_mqprio_rl *rl)
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3803
if (rl)
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3804
mlx5e_mqprio_rl_get_node_hw_id(rl, tc, &hw_id);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3811
struct mlx5e_mqprio_rl *rl)
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3821
mlx5e_mqprio_rl_update_params(params, rl);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3925
struct mlx5e_mqprio_rl *rl;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3931
rl = mlx5e_mqprio_rl_alloc();
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3932
if (!rl)
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3935
err = mlx5e_mqprio_rl_init(rl, mdev, num_tc, max_rate);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3937
mlx5e_mqprio_rl_free(rl);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3941
return rl;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3948
struct mlx5e_mqprio_rl *rl;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3955
rl = mlx5e_mqprio_rl_create(priv->mdev, mqprio->qopt.num_tc, mqprio->max_rate);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3956
if (IS_ERR(rl))
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3957
return PTR_ERR(rl);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3960
mlx5e_params_mqprio_channel_set(&new_params, mqprio, rl);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3965
if (rl) {
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3966
mlx5e_mqprio_rl_cleanup(rl);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3967
mlx5e_mqprio_rl_free(rl);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3976
priv->mqprio_rl = rl;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
6126
struct mlx5e_mqprio_rl *rl;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
6132
rl = mlx5e_mqprio_rl_create(priv->mdev, params->mqprio.num_tc,
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
6134
if (IS_ERR(rl))
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
6135
rl = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
6136
priv->mqprio_rl = rl;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
6137
mlx5e_mqprio_rl_update_params(params, rl);
drivers/net/ethernet/mellanox/mlx5/core/rl.c
377
struct mlx5_rate_limit *rl)
drivers/net/ethernet/mellanox/mlx5/core/rl.c
381
MLX5_SET(set_pp_rate_limit_context, rl_raw, rate_limit, rl->rate);
drivers/net/ethernet/mellanox/mlx5/core/rl.c
383
rl->max_burst_sz);
drivers/net/ethernet/mellanox/mlx5/core/rl.c
385
rl->typical_pkt_sz);
drivers/net/ethernet/mellanox/mlx5/core/rl.c
394
void mlx5_rl_remove_rate(struct mlx5_core_dev *dev, struct mlx5_rate_limit *rl)
drivers/net/ethernet/mellanox/mlx5/core/rl.c
401
if (rl->rate == 0)
drivers/net/ethernet/mellanox/mlx5/core/rl.c
404
MLX5_SET(set_pp_rate_limit_context, rl_raw, rate_limit, rl->rate);
drivers/net/ethernet/mellanox/mlx5/core/rl.c
406
rl->max_burst_sz);
drivers/net/ethernet/mellanox/mlx5/core/rl.c
408
rl->typical_pkt_sz);
drivers/net/ethernet/mellanox/mlx5/core/rl.c
416
rl->rate, rl->max_burst_sz, rl->typical_pkt_sz);
drivers/net/ethernet/microchip/enc28j60.c
292
int rl, rh;
drivers/net/ethernet/microchip/enc28j60.c
295
rl = spi_read_op(priv, ENC28J60_READ_CTRL_REG, address);
drivers/net/ethernet/microchip/enc28j60.c
298
return (rh << 8) | rl;
drivers/net/ethernet/qlogic/qed/qed_init_fw_funcs.c
194
#define PQ_INFO_ELEMENT(vp, pf, tc, port, rl_valid, rl) \
drivers/net/ethernet/qlogic/qed/qed_init_fw_funcs.c
196
((rl_valid ? 1 : 0) << 22) | (((rl) & 255) << 24) | \
drivers/net/ethernet/qlogic/qed/qed_init_fw_funcs.c
197
(((rl) >> 8) << 9))
drivers/nvme/host/trace.c
109
u16 rl = get_unaligned_le16(cdw10 + 12);
drivers/nvme/host/trace.c
113
slba, mndw, rl, atype);
drivers/nvme/target/trace.c
42
u16 rl = get_unaligned_le16(cdw10 + 12);
drivers/nvme/target/trace.c
46
slba, mndw, rl, atype);
drivers/s390/scsi/zfcp_reqlist.h
107
zfcp_reqlist_find(struct zfcp_reqlist *rl, u64 req_id)
drivers/s390/scsi/zfcp_reqlist.h
112
spin_lock_irqsave(&rl->lock, flags);
drivers/s390/scsi/zfcp_reqlist.h
113
req = _zfcp_reqlist_find(rl, req_id);
drivers/s390/scsi/zfcp_reqlist.h
114
spin_unlock_irqrestore(&rl->lock, flags);
drivers/s390/scsi/zfcp_reqlist.h
132
zfcp_reqlist_find_rm(struct zfcp_reqlist *rl, u64 req_id)
drivers/s390/scsi/zfcp_reqlist.h
137
spin_lock_irqsave(&rl->lock, flags);
drivers/s390/scsi/zfcp_reqlist.h
138
req = _zfcp_reqlist_find(rl, req_id);
drivers/s390/scsi/zfcp_reqlist.h
141
spin_unlock_irqrestore(&rl->lock, flags);
drivers/s390/scsi/zfcp_reqlist.h
156
static inline void zfcp_reqlist_add(struct zfcp_reqlist *rl,
drivers/s390/scsi/zfcp_reqlist.h
164
spin_lock_irqsave(&rl->lock, flags);
drivers/s390/scsi/zfcp_reqlist.h
165
list_add_tail(&req->list, &rl->buckets[i]);
drivers/s390/scsi/zfcp_reqlist.h
166
spin_unlock_irqrestore(&rl->lock, flags);
drivers/s390/scsi/zfcp_reqlist.h
174
static inline void zfcp_reqlist_move(struct zfcp_reqlist *rl,
drivers/s390/scsi/zfcp_reqlist.h
180
spin_lock_irqsave(&rl->lock, flags);
drivers/s390/scsi/zfcp_reqlist.h
182
list_splice_init(&rl->buckets[i], list);
drivers/s390/scsi/zfcp_reqlist.h
183
spin_unlock_irqrestore(&rl->lock, flags);
drivers/s390/scsi/zfcp_reqlist.h
200
zfcp_reqlist_apply_for_all(struct zfcp_reqlist *rl,
drivers/s390/scsi/zfcp_reqlist.h
207
spin_lock_irqsave(&rl->lock, flags);
drivers/s390/scsi/zfcp_reqlist.h
209
list_for_each_entry(req, &rl->buckets[i], list)
drivers/s390/scsi/zfcp_reqlist.h
211
spin_unlock_irqrestore(&rl->lock, flags);
drivers/s390/scsi/zfcp_reqlist.h
43
struct zfcp_reqlist *rl;
drivers/s390/scsi/zfcp_reqlist.h
45
rl = kzalloc_obj(struct zfcp_reqlist);
drivers/s390/scsi/zfcp_reqlist.h
46
if (!rl)
drivers/s390/scsi/zfcp_reqlist.h
49
spin_lock_init(&rl->lock);
drivers/s390/scsi/zfcp_reqlist.h
52
INIT_LIST_HEAD(&rl->buckets[i]);
drivers/s390/scsi/zfcp_reqlist.h
54
return rl;
drivers/s390/scsi/zfcp_reqlist.h
63
static inline int zfcp_reqlist_isempty(struct zfcp_reqlist *rl)
drivers/s390/scsi/zfcp_reqlist.h
68
if (!list_empty(&rl->buckets[i]))
drivers/s390/scsi/zfcp_reqlist.h
77
static inline void zfcp_reqlist_free(struct zfcp_reqlist *rl)
drivers/s390/scsi/zfcp_reqlist.h
80
BUG_ON(!zfcp_reqlist_isempty(rl));
drivers/s390/scsi/zfcp_reqlist.h
82
kfree(rl);
drivers/s390/scsi/zfcp_reqlist.h
86
_zfcp_reqlist_find(struct zfcp_reqlist *rl, u64 req_id)
drivers/s390/scsi/zfcp_reqlist.h
92
list_for_each_entry(req, &rl->buckets[i], list)
drivers/slimbus/messaging.c
152
unsigned long ms = txn->rl + HZ;
drivers/slimbus/messaging.c
261
txn->rl += msg->num_bytes;
drivers/slimbus/messaging.c
268
txn->rl++;
drivers/slimbus/qcom-ngd-ctrl.c
1058
txn.rl = txn.msg->num_bytes + 4;
drivers/slimbus/qcom-ngd-ctrl.c
1070
txn.rl = txn.msg->num_bytes + 4;
drivers/slimbus/qcom-ngd-ctrl.c
1105
txn.rl = 11;
drivers/slimbus/qcom-ngd-ctrl.c
1270
txn.rl = 8;
drivers/slimbus/qcom-ngd-ctrl.c
809
txn->rl > SLIM_MSGQ_BUF_LEN) {
drivers/slimbus/qcom-ngd-ctrl.c
814
pbuf = qcom_slim_ngd_tx_msg_get(ctrl, txn->rl, &tx_sent);
drivers/slimbus/qcom-ngd-ctrl.c
859
txn->rl = txn->msg->num_bytes + 4;
drivers/slimbus/qcom-ngd-ctrl.c
863
txn->rl--;
drivers/slimbus/qcom-ngd-ctrl.c
867
*pbuf = SLIM_MSG_ASM_FIRST_WORD(txn->rl, txn->mt, txn->mc, 0,
drivers/slimbus/qcom-ngd-ctrl.c
871
*pbuf = SLIM_MSG_ASM_FIRST_WORD(txn->rl, txn->mt, txn->mc, 1,
drivers/slimbus/qcom-ngd-ctrl.c
888
ret = qcom_slim_ngd_tx_msg_post(ctrl, pbuf, txn->rl);
drivers/slimbus/sched.c
105
txn.rl = 3;
drivers/slimbus/sched.c
97
txn.rl = 4;
drivers/slimbus/slimbus.h
128
u8 rl;
drivers/slimbus/slimbus.h
140
#define DEFINE_SLIM_LDEST_TXN(name, mc, rl, la, msg) \
drivers/slimbus/slimbus.h
141
struct slim_msg_txn name = { rl, 0, mc, SLIM_MSG_DEST_LOGICALADDR, 0,\
drivers/slimbus/slimbus.h
144
#define DEFINE_SLIM_BCAST_TXN(name, mc, rl, la, msg) \
drivers/slimbus/slimbus.h
145
struct slim_msg_txn name = { rl, 0, mc, SLIM_MSG_DEST_BROADCAST, 0,\
drivers/slimbus/slimbus.h
148
#define DEFINE_SLIM_EDEST_TXN(name, mc, rl, la, msg) \
drivers/slimbus/slimbus.h
149
struct slim_msg_txn name = { rl, 0, mc, SLIM_MSG_DEST_ENUMADDR, 0,\
drivers/usb/isp1760/isp1760-hcd.c
822
u32 rl = RL_COUNTER;
drivers/usb/isp1760/isp1760-hcd.c
863
rl = 0;
drivers/usb/isp1760/isp1760-hcd.c
874
ptd->dw2 |= TO_DW2_RL(rl);
drivers/video/fbdev/metronomefb.c
276
unsigned char rl;
drivers/video/fbdev/metronomefb.c
289
rl = mem[wfm_idx++];
drivers/video/fbdev/metronomefb.c
290
for (i = 0; i <= rl; i++)
fs/binfmt_flat.c
341
static void old_reloc(unsigned long rl)
fs/binfmt_flat.c
348
r.value = rl;
fs/dlm/lock.c
5530
struct rcom_lock *rl = (struct rcom_lock *) rc->rc_buf;
fs/dlm/lock.c
5533
lkb->lkb_ownpid = le32_to_cpu(rl->rl_ownpid);
fs/dlm/lock.c
5534
lkb->lkb_remid = le32_to_cpu(rl->rl_lkid);
fs/dlm/lock.c
5535
lkb->lkb_exflags = le32_to_cpu(rl->rl_exflags);
fs/dlm/lock.c
5536
dlm_set_dflags_val(lkb, le32_to_cpu(rl->rl_flags));
fs/dlm/lock.c
5538
lkb->lkb_lvbseq = le32_to_cpu(rl->rl_lvbseq);
fs/dlm/lock.c
5539
lkb->lkb_rqmode = rl->rl_rqmode;
fs/dlm/lock.c
5540
lkb->lkb_grmode = rl->rl_grmode;
fs/dlm/lock.c
5543
lkb->lkb_bastfn = (rl->rl_asts & DLM_CB_BAST) ? &fake_bastfn : NULL;
fs/dlm/lock.c
5544
lkb->lkb_astfn = (rl->rl_asts & DLM_CB_CAST) ? &fake_astfn : NULL;
fs/dlm/lock.c
5554
memcpy(lkb->lkb_lvbptr, rl->rl_lvb, lvblen);
fs/dlm/lock.c
5561
if (rl->rl_status == DLM_LKSTS_CONVERT && middle_conversion(lkb)) {
fs/dlm/lock.c
5582
struct rcom_lock *rl = (struct rcom_lock *) rc->rc_buf;
fs/dlm/lock.c
5590
*rl_remid = rl->rl_remid;
fs/dlm/lock.c
5592
if (rl->rl_parent_lkid) {
fs/dlm/lock.c
5597
remid = le32_to_cpu(rl->rl_lkid);
fs/dlm/lock.c
5607
error = find_rsb(ls, rl->rl_name, le16_to_cpu(rl->rl_namelen),
fs/dlm/lock.c
5638
add_lkb(r, lkb, rl->rl_status);
fs/dlm/lock.c
5666
struct rcom_lock *rl = (struct rcom_lock *) rc->rc_buf;
fs/dlm/lock.c
5672
lkid = le32_to_cpu(rl->rl_lkid);
fs/dlm/lock.c
5673
remid = le32_to_cpu(rl->rl_remid);
fs/dlm/lock.c
5674
result = le32_to_cpu(rl->rl_result);
fs/dlm/rcom.c
418
struct rcom_lock *rl)
fs/dlm/rcom.c
420
memset(rl, 0, sizeof(*rl));
fs/dlm/rcom.c
422
rl->rl_ownpid = cpu_to_le32(lkb->lkb_ownpid);
fs/dlm/rcom.c
423
rl->rl_lkid = cpu_to_le32(lkb->lkb_id);
fs/dlm/rcom.c
424
rl->rl_exflags = cpu_to_le32(lkb->lkb_exflags);
fs/dlm/rcom.c
425
rl->rl_flags = cpu_to_le32(dlm_dflags_val(lkb));
fs/dlm/rcom.c
426
rl->rl_lvbseq = cpu_to_le32(lkb->lkb_lvbseq);
fs/dlm/rcom.c
427
rl->rl_rqmode = lkb->lkb_rqmode;
fs/dlm/rcom.c
428
rl->rl_grmode = lkb->lkb_grmode;
fs/dlm/rcom.c
429
rl->rl_status = lkb->lkb_status;
fs/dlm/rcom.c
430
rl->rl_wait_type = cpu_to_le16(lkb->lkb_wait_type);
fs/dlm/rcom.c
433
rl->rl_asts |= DLM_CB_BAST;
fs/dlm/rcom.c
435
rl->rl_asts |= DLM_CB_CAST;
fs/dlm/rcom.c
437
rl->rl_namelen = cpu_to_le16(r->res_length);
fs/dlm/rcom.c
438
memcpy(rl->rl_name, r->res_name, r->res_length);
fs/dlm/rcom.c
444
memcpy(rl->rl_lvb, lkb->lkb_lvbptr, r->res_ls->ls_lvblen);
fs/dlm/rcom.c
452
struct rcom_lock *rl;
fs/dlm/rcom.c
463
rl = (struct rcom_lock *) rc->rc_buf;
fs/dlm/rcom.c
464
pack_rcom_lock(r, lkb, rl);
fs/dlm/rcom.c
477
struct rcom_lock *rl;
fs/dlm/rcom.c
490
rl = (struct rcom_lock *)rc->rc_buf;
fs/dlm/rcom.c
492
rl->rl_remid = rl_remid;
fs/dlm/rcom.c
493
rl->rl_result = rl_result;
fs/ntfs/attrib.c
1768
struct runlist_element *rl;
fs/ntfs/attrib.c
184
rl = ntfs_mapping_pairs_decompress(ni->vol, a, &ni->runlist, &new_rl_count);
fs/ntfs/attrib.c
1847
rl = ntfs_cluster_alloc(vol, 0,
fs/ntfs/attrib.c
185
if (IS_ERR(rl))
fs/ntfs/attrib.c
1850
if (IS_ERR(rl)) {
fs/ntfs/attrib.c
1851
err = PTR_ERR(rl);
fs/ntfs/attrib.c
1858
rl = NULL;
fs/ntfs/attrib.c
186
err = PTR_ERR(rl);
fs/ntfs/attrib.c
1864
mp_size = ntfs_get_size_for_mapping_pairs(vol, rl, 0, -1, -1);
fs/ntfs/attrib.c
188
ni->runlist.rl = rl;
fs/ntfs/attrib.c
1944
arec_size - mp_ofs, rl, 0, -1, NULL, NULL, NULL);
fs/ntfs/attrib.c
1952
ni->runlist.rl = rl;
fs/ntfs/attrib.c
1953
if (rl) {
fs/ntfs/attrib.c
1954
for (ni->runlist.count = 1; rl->length != 0; rl++)
fs/ntfs/attrib.c
2055
if (rl) {
fs/ntfs/attrib.c
2056
if (ntfs_cluster_free_from_rl(vol, rl) < 0) {
fs/ntfs/attrib.c
2061
kvfree(rl);
fs/ntfs/attrib.c
2071
ni->runlist.rl = NULL;
fs/ntfs/attrib.c
2844
ni->runlist.rl = NULL;
fs/ntfs/attrib.c
299
if (likely(ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn) <=
fs/ntfs/attrib.c
3075
if (NInoNonResident(ni) && ni->runlist.rl)
fs/ntfs/attrib.c
3076
kvfree(ni->runlist.rl);
fs/ntfs/attrib.c
308
struct runlist_element *rl = ni->runlist.rl;
fs/ntfs/attrib.c
3107
if (NInoFullyMapped(ni) && ni->runlist.rl)
fs/ntfs/attrib.c
312
if (!rl) {
fs/ntfs/attrib.c
3125
struct runlist_element *rl;
fs/ntfs/attrib.c
3128
if (ntfs_rl_vcn_to_lcn(ni->runlist.rl, next_vcn) == LCN_RL_NOT_MAPPED)
fs/ntfs/attrib.c
3140
rl = ntfs_mapping_pairs_decompress(ni->vol, a, &ni->runlist,
fs/ntfs/attrib.c
3142
if (IS_ERR(rl)) {
fs/ntfs/attrib.c
3143
err = PTR_ERR(rl);
fs/ntfs/attrib.c
3146
ni->runlist.rl = rl;
fs/ntfs/attrib.c
316
rl = ni->runlist.rl;
fs/ntfs/attrib.c
321
while (rl->length && rl[1].vcn <= vcn)
fs/ntfs/attrib.c
322
rl++;
fs/ntfs/attrib.c
323
*lcn = ntfs_rl_vcn_to_lcn(rl, vcn);
fs/ntfs/attrib.c
328
rl = ni->runlist.rl;
fs/ntfs/attrib.c
333
return rl;
fs/ntfs/attrib.c
3411
sparse = ntfs_rl_sparse(ni->runlist.rl);
fs/ntfs/attrib.c
3492
new_compr_size = ntfs_rl_get_compressed_size(ni->vol, ni->runlist.rl);
fs/ntfs/attrib.c
3545
if (!ni || !ni->runlist.rl)
fs/ntfs/attrib.c
3571
start_rl = ni->runlist.rl;
fs/ntfs/attrib.c
3597
first_lcn = ntfs_rl_vcn_to_lcn(ni->runlist.rl, stop_vcn);
fs/ntfs/attrib.c
376
if (!ni->runlist.rl) {
fs/ntfs/attrib.c
386
lcn = ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn);
fs/ntfs/attrib.c
3987
kvfree(ni->runlist.rl);
fs/ntfs/attrib.c
3988
ni->runlist.rl = NULL;
fs/ntfs/attrib.c
400
if (unlikely(ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn) !=
fs/ntfs/attrib.c
4094
kvfree(ni->runlist.rl);
fs/ntfs/attrib.c
4095
ni->runlist.rl = NULL;
fs/ntfs/attrib.c
4186
struct runlist_element *rl, *rln;
fs/ntfs/attrib.c
4254
while (ni->runlist.rl[last].length)
fs/ntfs/attrib.c
4257
rl = ntfs_rl_realloc(ni->runlist.rl, last + 1,
fs/ntfs/attrib.c
4259
if (IS_ERR(rl)) {
fs/ntfs/attrib.c
4266
rl[last].vcn = ntfs_bytes_to_cluster(vol,
fs/ntfs/attrib.c
4268
rl[last].length = ni->itype.compressed.block_clusters -
fs/ntfs/attrib.c
4269
(rl[last].vcn &
fs/ntfs/attrib.c
4271
rl[last].lcn = LCN_HOLE;
fs/ntfs/attrib.c
4276
rl[last].vcn = first_free_vcn;
fs/ntfs/attrib.c
4277
rl[last].lcn = LCN_ENOENT;
fs/ntfs/attrib.c
4278
rl[last].length = 0;
fs/ntfs/attrib.c
4280
ni->runlist.rl = rl;
fs/ntfs/attrib.c
4283
rl = kmalloc(sizeof(struct runlist_element) * 2, GFP_NOFS);
fs/ntfs/attrib.c
4284
if (!rl) {
fs/ntfs/attrib.c
4289
rl[0].vcn = ntfs_bytes_to_cluster(vol, ni->allocated_size);
fs/ntfs/attrib.c
4290
rl[0].lcn = LCN_HOLE;
fs/ntfs/attrib.c
4291
rl[0].length = first_free_vcn -
fs/ntfs/attrib.c
4293
rl[1].vcn = first_free_vcn;
fs/ntfs/attrib.c
4294
rl[1].lcn = LCN_ENOENT;
fs/ntfs/attrib.c
4295
rl[1].length = 0;
fs/ntfs/attrib.c
4306
if (ni->runlist.rl->length) {
fs/ntfs/attrib.c
4308
for (rl = ni->runlist.rl; (rl + 1)->length; rl++)
fs/ntfs/attrib.c
4314
while (rl->lcn < 0 && rl != ni->runlist.rl)
fs/ntfs/attrib.c
4315
rl--;
fs/ntfs/attrib.c
4319
if (rl->lcn >= 0)
fs/ntfs/attrib.c
4320
lcn_seek_from = rl->lcn + rl->length;
fs/ntfs/attrib.c
4323
rl = ntfs_cluster_alloc(vol,
fs/ntfs/attrib.c
4328
if (IS_ERR(rl)) {
fs/ntfs/attrib.c
4332
return PTR_ERR(rl);
fs/ntfs/attrib.c
4338
rln = ntfs_runlists_merge(&ni->runlist, rl, 0, &new_rl_count);
fs/ntfs/attrib.c
4342
ntfs_cluster_free_from_rl(vol, rl);
fs/ntfs/attrib.c
4343
kvfree(rl);
fs/ntfs/attrib.c
4346
ni->runlist.rl = rln;
fs/ntfs/attrib.c
438
if (runlist->count - 1 > rh && runlist->rl[rh].vcn <= vcn) {
fs/ntfs/attrib.c
439
if (vcn < runlist->rl[rh].vcn + runlist->rl[rh].length)
fs/ntfs/attrib.c
440
return &runlist->rl[rh];
fs/ntfs/attrib.c
4407
kvfree(ni->runlist.rl);
fs/ntfs/attrib.c
4408
ni->runlist.rl = NULL;
fs/ntfs/attrib.c
446
run = &runlist->rl[0];
fs/ntfs/attrib.c
457
run = &runlist->rl[runlist->count - 2];
fs/ntfs/attrib.c
469
run = &runlist->rl[idx];
fs/ntfs/attrib.c
4823
struct runlist_element *rl, *rlc;
fs/ntfs/attrib.c
4851
rl = ntfs_attr_find_vcn_nolock(ni, vcn, ctx);
fs/ntfs/attrib.c
4852
if (IS_ERR(rl)) {
fs/ntfs/attrib.c
4854
err = PTR_ERR(rl);
fs/ntfs/attrib.c
4858
lcn = ntfs_rl_vcn_to_lcn(rl, vcn);
fs/ntfs/attrib.c
4859
clu_count = min(max_clu_count, rl->length - (vcn - rl->vcn));
fs/ntfs/attrib.c
4878
rlc = rl;
fs/ntfs/attrib.c
4898
rlc = rl;
fs/ntfs/attrib.c
4921
rl = ntfs_runlists_merge(&ni->runlist, rlc, 0, &new_rl_count);
fs/ntfs/attrib.c
4922
if (IS_ERR(rl)) {
fs/ntfs/attrib.c
4924
err = PTR_ERR(rl);
fs/ntfs/attrib.c
4930
ni->runlist.rl = rl;
fs/ntfs/attrib.c
5144
struct runlist_element *hole_rl, *rl;
fs/ntfs/attrib.c
5171
rl = ntfs_rl_find_vcn_nolock(ni->runlist.rl, start_vcn);
fs/ntfs/attrib.c
5172
if (!rl) {
fs/ntfs/attrib.c
5178
rl = ntfs_rl_insert_range(ni->runlist.rl, (int)ni->runlist.count,
fs/ntfs/attrib.c
5180
if (IS_ERR(rl)) {
fs/ntfs/attrib.c
5183
return PTR_ERR(rl);
fs/ntfs/attrib.c
5185
ni->runlist.rl = rl;
fs/ntfs/attrib.c
520
struct runlist_element *rl;
fs/ntfs/attrib.c
5222
struct runlist_element *punch_rl, *rl;
fs/ntfs/attrib.c
5244
for (rl = ni->runlist.rl, dst_cnt = 0; rl && rl->length; rl++)
fs/ntfs/attrib.c
5246
rl = ntfs_rl_find_vcn_nolock(ni->runlist.rl, start_vcn);
fs/ntfs/attrib.c
5247
if (!rl) {
fs/ntfs/attrib.c
5252
rl = ntfs_rl_collapse_range(ni->runlist.rl, dst_cnt + 1,
fs/ntfs/attrib.c
5254
if (IS_ERR(rl)) {
fs/ntfs/attrib.c
5256
return PTR_ERR(rl);
fs/ntfs/attrib.c
5258
ni->runlist.rl = rl;
fs/ntfs/attrib.c
526
if (!ni->runlist.rl) {
fs/ntfs/attrib.c
5317
struct runlist_element *punch_rl, *rl;
fs/ntfs/attrib.c
5338
for (rl = ni->runlist.rl, dst_cnt = 0; rl && rl->length; rl++)
fs/ntfs/attrib.c
5340
rl = ntfs_rl_find_vcn_nolock(ni->runlist.rl, start_vcn);
fs/ntfs/attrib.c
5341
if (!rl) {
fs/ntfs/attrib.c
5346
rl = ntfs_rl_punch_hole(ni->runlist.rl, dst_cnt + 1,
fs/ntfs/attrib.c
5348
if (IS_ERR(rl)) {
fs/ntfs/attrib.c
5350
return PTR_ERR(rl);
fs/ntfs/attrib.c
5352
ni->runlist.rl = rl;
fs/ntfs/attrib.c
536
rl = ni->runlist.rl;
fs/ntfs/attrib.c
537
if (likely(rl && vcn >= rl[0].vcn)) {
fs/ntfs/attrib.c
538
rl = __ntfs_attr_find_vcn_nolock(&ni->runlist, vcn);
fs/ntfs/attrib.c
5380
struct runlist_element *rl;
fs/ntfs/attrib.c
539
if (IS_ERR(rl))
fs/ntfs/attrib.c
540
err = PTR_ERR(rl);
fs/ntfs/attrib.c
541
else if (rl->lcn >= LCN_HOLE)
fs/ntfs/attrib.c
542
return rl;
fs/ntfs/attrib.c
543
else if (rl->lcn <= LCN_ENOENT)
fs/ntfs/attrib.c
5457
rl = ntfs_attr_find_vcn_nolock(ni, vcn, NULL);
fs/ntfs/attrib.c
5459
if (IS_ERR(rl)) {
fs/ntfs/attrib.c
5460
err = PTR_ERR(rl);
fs/ntfs/attrib.c
5464
if (rl->lcn > 0) {
fs/ntfs/attrib.c
5465
vcn += rl->length - (vcn - rl->vcn);
fs/ntfs/attrib.c
5466
} else if (rl->lcn == LCN_DELALLOC || rl->lcn == LCN_HOLE) {
fs/ntfs/attrib.c
5467
try_alloc_cnt = min(rl->length - (vcn - rl->vcn),
fs/ntfs/attrib.c
5470
if (rl->lcn == LCN_DELALLOC) {
fs/ntfs/attrib.c
85
struct runlist_element *rl;
fs/ntfs/compress.c
1291
struct runlist_element *rlc, *rl;
fs/ntfs/compress.c
1393
rl = ntfs_runlists_merge(&ni->runlist, rlc, 0, &new_rl_count);
fs/ntfs/compress.c
1394
if (IS_ERR(rl)) {
fs/ntfs/compress.c
1397
err = PTR_ERR(rl);
fs/ntfs/compress.c
1405
ni->runlist.rl = rl;
fs/ntfs/compress.c
472
struct runlist_element *rl;
fs/ntfs/compress.c
600
rl = NULL;
fs/ntfs/compress.c
605
if (!rl) {
fs/ntfs/compress.c
608
rl = ni->runlist.rl;
fs/ntfs/compress.c
610
if (likely(rl != NULL)) {
fs/ntfs/compress.c
612
while (rl->length && rl[1].vcn <= vcn)
fs/ntfs/compress.c
613
rl++;
fs/ntfs/compress.c
614
lcn = ntfs_rl_vcn_to_lcn(rl, vcn);
fs/ntfs/compress.c
663
if (rl)
fs/ntfs/debug.c
131
void ntfs_debug_dump_runlist(const struct runlist_element *rl)
fs/ntfs/debug.c
141
if (!rl) {
fs/ntfs/debug.c
147
s64 lcn = (rl + i)->lcn;
fs/ntfs/debug.c
155
(long long)(rl + i)->vcn, lcn_str[index],
fs/ntfs/debug.c
156
(long long)(rl + i)->length,
fs/ntfs/debug.c
157
(rl + i)->length ? "" :
fs/ntfs/debug.c
161
(long long)(rl + i)->vcn,
fs/ntfs/debug.c
162
(long long)(rl + i)->lcn,
fs/ntfs/debug.c
163
(long long)(rl + i)->length,
fs/ntfs/debug.c
164
(rl + i)->length ? "" :
fs/ntfs/debug.c
166
if (!(rl + i)->length)
fs/ntfs/debug.h
33
void ntfs_debug_dump_runlist(const struct runlist_element *rl);
fs/ntfs/debug.h
43
#define ntfs_debug_dump_runlist(rl) \
fs/ntfs/debug.h
46
(void)rl; \
fs/ntfs/ea.c
640
mp_size = ntfs_get_size_for_mapping_pairs(ni->vol, ni->runlist.rl, 0, -1, -1);
fs/ntfs/file.c
100
while (rc >= 0 && rl[rc].lcn == LCN_HOLE && vcn_ds <= rl[rc].vcn) {
fs/ntfs/file.c
101
vcn_tr = rl[rc].vcn;
fs/ntfs/file.c
108
kvfree(ni->runlist.rl);
fs/ntfs/file.c
109
ni->runlist.rl = NULL;
fs/ntfs/file.c
81
struct runlist_element *rl;
fs/ntfs/file.c
98
rl = ni->runlist.rl;
fs/ntfs/inode.c
1714
struct runlist_element *rl, u8 *al_start, const s64 size,
fs/ntfs/inode.c
1725
if (!vol || !rl || !al || size <= 0 || initialized_size < 0 ||
fs/ntfs/inode.c
1735
while (rl->length) {
fs/ntfs/inode.c
1736
lcn = ntfs_rl_vcn_to_lcn(rl, rl->vcn);
fs/ntfs/inode.c
1738
(unsigned long long)rl->vcn,
fs/ntfs/inode.c
1747
rl_byte_len = ntfs_cluster_to_bytes(vol, rl->length);
fs/ntfs/inode.c
1766
rl++;
fs/ntfs/inode.c
1952
struct runlist_element *rl;
fs/ntfs/inode.c
1962
rl = ntfs_mapping_pairs_decompress(vol, a, NULL, &new_rl_count);
fs/ntfs/inode.c
1963
if (IS_ERR(rl)) {
fs/ntfs/inode.c
1964
err = PTR_ERR(rl);
fs/ntfs/inode.c
1971
err = load_attribute_list_mount(vol, rl, ni->attr_list, ni->attr_list_size,
fs/ntfs/inode.c
1973
kvfree(rl);
fs/ntfs/inode.c
2070
ni->runlist.rl = nrl;
fs/ntfs/inode.c
2193
if (NInoNonResident(ni) && ni->runlist.rl) {
fs/ntfs/inode.c
2194
kvfree(ni->runlist.rl);
fs/ntfs/inode.c
2195
ni->runlist.rl = NULL;
fs/ntfs/inode.c
2660
struct runlist_element *rl;
fs/ntfs/inode.c
2668
rl = mft_ni->runlist.rl;
fs/ntfs/inode.c
2669
if (!rl) {
fs/ntfs/inode.c
2675
while (rl->length && rl[1].vcn <= vcn)
fs/ntfs/inode.c
2676
rl++;
fs/ntfs/inode.c
2677
ni->mft_lcn[0] = ntfs_rl_vcn_to_lcn(rl, vcn);
fs/ntfs/inode.c
2681
(rl->length - (vcn - rl->vcn)) <= 1) {
fs/ntfs/inode.c
2682
rl++;
fs/ntfs/inode.c
2683
ni->mft_lcn[1] = ntfs_rl_vcn_to_lcn(rl, vcn + 1);
fs/ntfs/inode.c
340
err = ntfs_cluster_free_from_rl(ni->vol, ni->runlist.rl);
fs/ntfs/inode.c
349
struct runlist_element *rl;
fs/ntfs/inode.c
352
rl = ntfs_mapping_pairs_decompress(ni->vol, actx->attr, NULL,
fs/ntfs/inode.c
354
if (IS_ERR(rl)) {
fs/ntfs/inode.c
355
err = PTR_ERR(rl);
fs/ntfs/inode.c
361
err = ntfs_cluster_free_from_rl(ni->vol, rl);
fs/ntfs/inode.c
365
kvfree(rl);
fs/ntfs/inode.c
3658
struct runlist_element *rl;
fs/ntfs/inode.c
3665
rl = ntfs_attr_vcn_to_rl(ni, vcn, &lcn);
fs/ntfs/inode.c
3666
if (IS_ERR(rl)) {
fs/ntfs/inode.c
3667
ret = PTR_ERR(rl);
fs/ntfs/inode.c
3672
rl_length = rl->length - (vcn - rl->vcn);
fs/ntfs/iomap.c
197
struct runlist_element *rl;
fs/ntfs/iomap.c
206
rl = ntfs_attr_vcn_to_rl(ni, vcn, &lcn);
fs/ntfs/iomap.c
207
if (IS_ERR(rl)) {
fs/ntfs/iomap.c
209
return PTR_ERR(rl);
fs/ntfs/iomap.c
239
rl_length = ntfs_cluster_to_bytes(vol, rl->length - (vcn - rl->vcn));
fs/ntfs/iomap.c
241
if (rl_length == 0 && rl->lcn > LCN_DELALLOC) {
fs/ntfs/iomap.c
244
rl->vcn, rl->length, rl->lcn);
fs/ntfs/iomap.c
308
struct runlist_element *rl;
fs/ntfs/iomap.c
317
rl = __ntfs_attr_find_vcn_nolock(&ni->runlist, vcn);
fs/ntfs/iomap.c
318
if (IS_ERR(rl)) {
fs/ntfs/iomap.c
322
lcn = ntfs_rl_vcn_to_lcn(rl, vcn);
fs/ntfs/iomap.c
385
struct runlist_element *rl, *rlc;
fs/ntfs/iomap.c
396
rl = ni->runlist.rl;
fs/ntfs/iomap.c
397
if (!rl) {
fs/ntfs/iomap.c
405
rl = ni->runlist.rl;
fs/ntfs/iomap.c
412
rl = __ntfs_attr_find_vcn_nolock(&ni->runlist, vcn);
fs/ntfs/iomap.c
413
if (IS_ERR(rl)) {
fs/ntfs/iomap.c
418
lcn = ntfs_rl_vcn_to_lcn(rl, vcn);
fs/ntfs/iomap.c
423
rl = ni->runlist.rl;
fs/ntfs/iomap.c
428
max_clu_count = min(max_clu_count, rl->length - (vcn - rl->vcn));
fs/ntfs/iomap.c
432
rl->vcn, rl->length);
fs/ntfs/iomap.c
475
rl = ntfs_runlists_merge(&ni->runlist, rlc, 0,
fs/ntfs/iomap.c
477
if (IS_ERR(rl)) {
fs/ntfs/iomap.c
482
return PTR_ERR(rl);
fs/ntfs/iomap.c
485
ni->runlist.rl = rl;
fs/ntfs/lcnalloc.c
1000
start_sector = ALIGN(NTFS_CLU_TO_B(vol, rl->lcn + rl_off),
fs/ntfs/lcnalloc.c
1003
rl->lcn + rl_off + to_discard),
fs/ntfs/lcnalloc.c
1015
++rl;
fs/ntfs/lcnalloc.c
195
struct runlist_element *rl = NULL;
fs/ntfs/lcnalloc.c
37
const struct runlist_element *rl)
fs/ntfs/lcnalloc.c
378
if ((rlpos + 2) * sizeof(*rl) > rlsize) {
fs/ntfs/lcnalloc.c
382
if (!rl)
fs/ntfs/lcnalloc.c
391
memcpy(rl2, rl, rlsize);
fs/ntfs/lcnalloc.c
392
kvfree(rl);
fs/ntfs/lcnalloc.c
393
rl = rl2;
fs/ntfs/lcnalloc.c
416
rl[rlpos - 1].lcn,
fs/ntfs/lcnalloc.c
417
rl[rlpos - 1].length);
fs/ntfs/lcnalloc.c
418
rl[rlpos - 1].length = ++prev_run_len;
fs/ntfs/lcnalloc.c
420
rl[rlpos - 1].lcn,
fs/ntfs/lcnalloc.c
421
rl[rlpos - 1].length,
fs/ntfs/lcnalloc.c
426
rl[rlpos - 1].lcn, rl[rlpos - 1].length);
fs/ntfs/lcnalloc.c
427
rl[rlpos].vcn = rl[rlpos - 1].vcn +
fs/ntfs/lcnalloc.c
431
rl[rlpos].vcn = start_vcn;
fs/ntfs/lcnalloc.c
433
rl[rlpos].lcn = prev_lcn = lcn + bmp_pos;
fs/ntfs/lcnalloc.c
434
rl[rlpos].length = prev_run_len = 1;
fs/ntfs/lcnalloc.c
44
if (!rl)
fs/ntfs/lcnalloc.c
50
for (; rl->length; rl++) {
fs/ntfs/lcnalloc.c
53
if (rl->lcn < 0)
fs/ntfs/lcnalloc.c
55
err = ntfs_bitmap_clear_run(lcnbmp_vi, rl->lcn, rl->length);
fs/ntfs/lcnalloc.c
576
tc = rl[rlpos - 1].lcn +
fs/ntfs/lcnalloc.c
577
rl[rlpos - 1].length;
fs/ntfs/lcnalloc.c
59
nr_freed += rl->length;
fs/ntfs/lcnalloc.c
612
tc = rl[rlpos - 1].lcn +
fs/ntfs/lcnalloc.c
613
rl[rlpos - 1].length;
fs/ntfs/lcnalloc.c
646
tc = rl[rlpos - 1].lcn +
fs/ntfs/lcnalloc.c
647
rl[rlpos - 1].length;
fs/ntfs/lcnalloc.c
719
if (likely(rl)) {
fs/ntfs/lcnalloc.c
720
rl[rlpos].vcn = rl[rlpos - 1].vcn + rl[rlpos - 1].length;
fs/ntfs/lcnalloc.c
721
rl[rlpos].lcn = is_extension ? LCN_ENOENT : LCN_RL_NOT_MAPPED;
fs/ntfs/lcnalloc.c
722
rl[rlpos].length = 0;
fs/ntfs/lcnalloc.c
735
if (!rl) {
fs/ntfs/lcnalloc.c
740
ntfs_release_dirty_clusters(vol, rl->length);
fs/ntfs/lcnalloc.c
748
if (rl) {
fs/ntfs/lcnalloc.c
753
rl[0].lcn, count - clusters);
fs/ntfs/lcnalloc.c
756
err2 = ntfs_cluster_free_from_rl_nolock(vol, rl);
fs/ntfs/lcnalloc.c
764
kvfree(rl);
fs/ntfs/lcnalloc.c
774
return err < 0 ? ERR_PTR(err) : rl;
fs/ntfs/lcnalloc.c
843
struct runlist_element *rl;
fs/ntfs/lcnalloc.c
872
rl = ntfs_attr_find_vcn_nolock(ni, start_vcn, ctx);
fs/ntfs/lcnalloc.c
873
if (IS_ERR(rl)) {
fs/ntfs/lcnalloc.c
874
err = PTR_ERR(rl);
fs/ntfs/lcnalloc.c
889
if (unlikely(rl->lcn < LCN_HOLE)) {
fs/ntfs/lcnalloc.c
896
delta = start_vcn - rl->vcn;
fs/ntfs/lcnalloc.c
899
to_free = rl->length - delta;
fs/ntfs/lcnalloc.c
903
if (likely(rl->lcn >= 0)) {
fs/ntfs/lcnalloc.c
905
err = ntfs_bitmap_set_bits_in_run(lcnbmp_vi, rl->lcn + delta,
fs/ntfs/lcnalloc.c
918
++rl;
fs/ntfs/lcnalloc.c
928
for (; rl->length && count != 0; ++rl) {
fs/ntfs/lcnalloc.c
929
if (unlikely(rl->lcn < LCN_HOLE)) {
fs/ntfs/lcnalloc.c
933
vcn = rl->vcn;
fs/ntfs/lcnalloc.c
934
rl = ntfs_attr_find_vcn_nolock(ni, vcn, ctx);
fs/ntfs/lcnalloc.c
935
if (IS_ERR(rl)) {
fs/ntfs/lcnalloc.c
936
err = PTR_ERR(rl);
fs/ntfs/lcnalloc.c
942
if (unlikely(rl->lcn < LCN_HOLE)) {
fs/ntfs/lcnalloc.c
946
rl->lcn);
fs/ntfs/lcnalloc.c
952
to_free = rl->length;
fs/ntfs/lcnalloc.c
956
if (likely(rl->lcn >= 0)) {
fs/ntfs/lcnalloc.c
958
err = ntfs_bitmap_set_bits_in_run(lcnbmp_vi, rl->lcn,
fs/ntfs/lcnalloc.c
987
rl = ntfs_attr_find_vcn_nolock(ni, start_vcn, ctx);
fs/ntfs/lcnalloc.c
988
if (IS_ERR(rl))
fs/ntfs/lcnalloc.c
990
rl_off = start_vcn - rl->vcn;
fs/ntfs/lcnalloc.c
991
while (rl->length && total_discarded < total_freed) {
fs/ntfs/lcnalloc.c
992
s64 to_discard = rl->length - rl_off;
fs/ntfs/lcnalloc.c
996
if (rl->lcn >= 0) {
fs/ntfs/lcnalloc.h
102
const struct runlist_element *rl);
fs/ntfs/lcnalloc.h
121
const struct runlist_element *rl)
fs/ntfs/lcnalloc.h
128
ret = ntfs_cluster_free_from_rl_nolock(vol, rl);
fs/ntfs/logfile.c
647
struct runlist_element *rl;
fs/ntfs/logfile.c
674
rl = log_ni->runlist.rl;
fs/ntfs/logfile.c
675
if (unlikely(!rl || vcn < rl->vcn || !rl->length)) {
fs/ntfs/logfile.c
682
rl = log_ni->runlist.rl;
fs/ntfs/logfile.c
685
while (rl->length && vcn >= rl[1].vcn)
fs/ntfs/logfile.c
686
rl++;
fs/ntfs/logfile.c
709
lcn = rl->lcn;
fs/ntfs/logfile.c
711
vcn = rl->vcn;
fs/ntfs/logfile.c
716
if (unlikely(!rl->length || lcn < LCN_HOLE))
fs/ntfs/logfile.c
722
len = rl->length;
fs/ntfs/logfile.c
723
if (rl[1].vcn > end_vcn)
fs/ntfs/logfile.c
724
len = end_vcn - rl->vcn;
fs/ntfs/logfile.c
754
} while ((++rl)->vcn < end_vcn);
fs/ntfs/mft.c
1138
struct runlist_element *rl, *rl2 = NULL;
fs/ntfs/mft.c
1164
rl = ntfs_attr_find_vcn_nolock(mftbmp_ni,
fs/ntfs/mft.c
1166
if (IS_ERR(rl) || unlikely(!rl->length || rl->lcn < 0)) {
fs/ntfs/mft.c
1170
if (!IS_ERR(rl))
fs/ntfs/mft.c
1173
ret = PTR_ERR(rl);
fs/ntfs/mft.c
1176
lcn = rl->lcn + rl->length;
fs/ntfs/mft.c
1206
rl->length++;
fs/ntfs/mft.c
1207
rl[1].vcn++;
fs/ntfs/mft.c
1216
rl2 = ntfs_cluster_alloc(vol, rl[1].vcn, 1, lcn, DATA_ZONE,
fs/ntfs/mft.c
1224
rl = ntfs_runlists_merge(&mftbmp_ni->runlist, rl2, 0, &new_rl_count);
fs/ntfs/mft.c
1225
if (IS_ERR(rl)) {
fs/ntfs/mft.c
1234
return PTR_ERR(rl);
fs/ntfs/mft.c
1236
mftbmp_ni->runlist.rl = rl;
fs/ntfs/mft.c
1241
for (; rl[1].length; rl++)
fs/ntfs/mft.c
1261
mftbmp_ni->name_len, CASE_SENSITIVE, rl[1].vcn, NULL,
fs/ntfs/mft.c
1273
for (rl2 = rl; rl2 > mftbmp_ni->runlist.rl; rl2--) {
fs/ntfs/mft.c
1312
a->data.non_resident.highest_vcn = cpu_to_le64(rl[1].vcn - 1);
fs/ntfs/mft.c
1352
mftbmp_ni->name_len, CASE_SENSITIVE, rl[1].vcn, NULL,
fs/ntfs/mft.c
1370
a->data.non_resident.highest_vcn = cpu_to_le64(rl[1].vcn - 2);
fs/ntfs/mft.c
1374
rl->length--;
fs/ntfs/mft.c
1375
rl[1].vcn--;
fs/ntfs/mft.c
1377
lcn = rl->lcn;
fs/ntfs/mft.c
1379
rl->lcn = rl[1].lcn;
fs/ntfs/mft.c
1380
rl->length = 0;
fs/ntfs/mft.c
1571
struct runlist_element *rl, *rl2;
fs/ntfs/mft.c
1591
rl = ntfs_attr_find_vcn_nolock(mft_ni,
fs/ntfs/mft.c
1593
if (IS_ERR(rl) || unlikely(!rl->length || rl->lcn < 0)) {
fs/ntfs/mft.c
1597
if (!IS_ERR(rl))
fs/ntfs/mft.c
1600
ret = PTR_ERR(rl);
fs/ntfs/mft.c
1603
lcn = rl->lcn + rl->length;
fs/ntfs/mft.c
1630
old_last_vcn = rl[1].vcn;
fs/ntfs/mft.c
1661
rl = ntfs_runlists_merge(&mft_ni->runlist, rl2, 0, &new_rl_count);
fs/ntfs/mft.c
1662
if (IS_ERR(rl)) {
fs/ntfs/mft.c
1671
return PTR_ERR(rl);
fs/ntfs/mft.c
1673
mft_ni->runlist.rl = rl;
fs/ntfs/mft.c
1677
for (; rl[1].length; rl++)
fs/ntfs/mft.c
1696
CASE_SENSITIVE, rl[1].vcn, NULL, 0, ctx);
fs/ntfs/mft.c
1708
for (rl2 = rl; rl2 > mft_ni->runlist.rl; rl2--) {
fs/ntfs/mft.c
1749
a->data.non_resident.highest_vcn = cpu_to_le64(rl[1].vcn - 1);
fs/ntfs/mft.c
1789
CASE_SENSITIVE, rl[1].vcn, NULL, 0, ctx)) {
fs/ntfs/mft.c
2717
struct runlist_element *rl;
fs/ntfs/mft.c
2792
rl = ntfs_attr_vcn_to_rl(ni, vcn_off, &lcn);
fs/ntfs/mft.c
2794
if (IS_ERR(rl) || lcn < 0) {
fs/ntfs/mft.c
2823
rl->length - (vcn_off - rl->vcn) == 1 ||
fs/ntfs/runlist.c
1000
return rl[i].lcn + (vcn - rl[i].vcn);
fs/ntfs/runlist.c
1001
return rl[i].lcn;
fs/ntfs/runlist.c
1008
if (likely(rl[i].lcn < 0))
fs/ntfs/runlist.c
1009
return rl[i].lcn;
fs/ntfs/runlist.c
1027
struct runlist_element *ntfs_rl_find_vcn_nolock(struct runlist_element *rl, const s64 vcn)
fs/ntfs/runlist.c
1029
if (unlikely(!rl || vcn < rl[0].vcn))
fs/ntfs/runlist.c
1031
while (likely(rl->length)) {
fs/ntfs/runlist.c
1032
if (unlikely(vcn < rl[1].vcn)) {
fs/ntfs/runlist.c
1033
if (likely(rl->lcn >= LCN_HOLE))
fs/ntfs/runlist.c
1034
return rl;
fs/ntfs/runlist.c
1037
rl++;
fs/ntfs/runlist.c
1039
if (likely(rl->lcn == LCN_ENOENT))
fs/ntfs/runlist.c
1040
return rl;
fs/ntfs/runlist.c
1099
const struct runlist_element *rl, const s64 first_vcn,
fs/ntfs/runlist.c
1112
if (!rl) {
fs/ntfs/runlist.c
1120
while (rl->length && first_vcn >= rl[1].vcn)
fs/ntfs/runlist.c
1121
rl++;
fs/ntfs/runlist.c
1122
if (unlikely((!rl->length && first_vcn > rl->vcn) ||
fs/ntfs/runlist.c
1123
first_vcn < rl->vcn))
fs/ntfs/runlist.c
1129
if (first_vcn > rl->vcn) {
fs/ntfs/runlist.c
1130
s64 delta, length = rl->length;
fs/ntfs/runlist.c
1133
if (unlikely(length < 0 || rl->lcn < LCN_HOLE))
fs/ntfs/runlist.c
1139
if (unlikely(last_vcn >= 0 && rl[1].vcn > last_vcn)) {
fs/ntfs/runlist.c
1142
if (unlikely(rl[1].vcn > s1))
fs/ntfs/runlist.c
1143
length = s1 - rl->vcn;
fs/ntfs/runlist.c
1146
delta = first_vcn - rl->vcn;
fs/ntfs/runlist.c
1156
if (likely(rl->lcn >= 0 || vol->major_ver < 3)) {
fs/ntfs/runlist.c
1157
prev_lcn = rl->lcn;
fs/ntfs/runlist.c
1158
if (likely(rl->lcn >= 0))
fs/ntfs/runlist.c
116
static inline struct runlist_element *ntfs_rl_realloc_nofail(struct runlist_element *rl,
fs/ntfs/runlist.c
1164
rl++;
fs/ntfs/runlist.c
1167
for (; rl->length && !the_end; rl++) {
fs/ntfs/runlist.c
1168
s64 length = rl->length;
fs/ntfs/runlist.c
1170
if (unlikely(length < 0 || rl->lcn < LCN_HOLE))
fs/ntfs/runlist.c
1176
if (unlikely(last_vcn >= 0 && rl[1].vcn > last_vcn)) {
fs/ntfs/runlist.c
1179
if (unlikely(rl[1].vcn > s1))
fs/ntfs/runlist.c
1180
length = s1 - rl->vcn;
fs/ntfs/runlist.c
1192
if (likely(rl->lcn >= 0 || vol->major_ver < 3)) {
fs/ntfs/runlist.c
1194
rls += ntfs_get_nr_significant_bytes(rl->lcn -
fs/ntfs/runlist.c
1196
prev_lcn = rl->lcn;
fs/ntfs/runlist.c
1204
if (rl->lcn == LCN_RL_NOT_MAPPED)
fs/ntfs/runlist.c
121
old_size = old_size * sizeof(*rl);
fs/ntfs/runlist.c
122
new_size = new_size * sizeof(*rl);
fs/ntfs/runlist.c
124
return rl;
fs/ntfs/runlist.c
127
if (likely(rl != NULL)) {
fs/ntfs/runlist.c
130
memcpy(new_rl, rl, old_size);
fs/ntfs/runlist.c
1303
const int dst_len, const struct runlist_element *rl,
fs/ntfs/runlist.c
131
kvfree(rl);
fs/ntfs/runlist.c
1319
if (!rl) {
fs/ntfs/runlist.c
1328
while (rl->length && first_vcn >= rl[1].vcn)
fs/ntfs/runlist.c
1329
rl++;
fs/ntfs/runlist.c
1330
if (unlikely((!rl->length && first_vcn > rl->vcn) ||
fs/ntfs/runlist.c
1331
first_vcn < rl->vcn))
fs/ntfs/runlist.c
1340
if (first_vcn > rl->vcn) {
fs/ntfs/runlist.c
1341
s64 delta, length = rl->length;
fs/ntfs/runlist.c
1344
if (unlikely(length < 0 || rl->lcn < LCN_HOLE))
fs/ntfs/runlist.c
1350
if (unlikely(last_vcn >= 0 && rl[1].vcn > last_vcn)) {
fs/ntfs/runlist.c
1353
if (unlikely(rl[1].vcn > s1))
fs/ntfs/runlist.c
1354
length = s1 - rl->vcn;
fs/ntfs/runlist.c
1357
delta = first_vcn - rl->vcn;
fs/ntfs/runlist.c
1369
if (likely(rl->lcn >= 0 || vol->major_ver < 3)) {
fs/ntfs/runlist.c
1370
prev_lcn = rl->lcn;
fs/ntfs/runlist.c
1371
if (likely(rl->lcn >= 0))
fs/ntfs/runlist.c
1388
rl++;
fs/ntfs/runlist.c
1391
for (; rl->length && !the_end; rl++) {
fs/ntfs/runlist.c
1392
s64 length = rl->length;
fs/ntfs/runlist.c
1394
if (unlikely(length < 0 || rl->lcn < LCN_HOLE))
fs/ntfs/runlist.c
1400
if (unlikely(last_vcn >= 0 && rl[1].vcn > last_vcn)) {
fs/ntfs/runlist.c
1403
if (unlikely(rl[1].vcn > s1))
fs/ntfs/runlist.c
1404
length = s1 - rl->vcn;
fs/ntfs/runlist.c
1418
if (likely(rl->lcn >= 0 || vol->major_ver < 3)) {
fs/ntfs/runlist.c
1421
len_len, dst_max, rl->lcn - prev_lcn);
fs/ntfs/runlist.c
1424
prev_lcn = rl->lcn;
fs/ntfs/runlist.c
1426
if (rl->lcn == LCN_DELALLOC)
fs/ntfs/runlist.c
1427
de_cnt += rl->length;
fs/ntfs/runlist.c
1445
*stop_vcn = rl->vcn;
fs/ntfs/runlist.c
1447
*stop_rl = (struct runlist_element *)rl;
fs/ntfs/runlist.c
1452
if (rl->lcn == LCN_RL_NOT_MAPPED)
fs/ntfs/runlist.c
1486
struct runlist_element *rl;
fs/ntfs/runlist.c
1494
rl = runlist->rl;
fs/ntfs/runlist.c
1495
if (new_length < rl->vcn)
fs/ntfs/runlist.c
1499
while (likely(rl->length && new_length >= rl[1].vcn))
fs/ntfs/runlist.c
1500
rl++;
fs/ntfs/runlist.c
1505
if (rl->length) {
fs/ntfs/runlist.c
1511
trl = rl + 1;
fs/ntfs/runlist.c
1514
old_size = trl - runlist->rl + 1;
fs/ntfs/runlist.c
1516
rl->length = new_length - rl->vcn;
fs/ntfs/runlist.c
1522
if (rl->length) {
fs/ntfs/runlist.c
1523
rl++;
fs/ntfs/runlist.c
1524
if (!rl->length)
fs/ntfs/runlist.c
1526
rl->vcn = new_length;
fs/ntfs/runlist.c
1527
rl->length = 0;
fs/ntfs/runlist.c
1529
rl->lcn = LCN_ENOENT;
fs/ntfs/runlist.c
1530
runlist->count = rl - runlist->rl + 1;
fs/ntfs/runlist.c
1533
int new_size = rl - runlist->rl + 1;
fs/ntfs/runlist.c
1535
rl = ntfs_rl_realloc(runlist->rl, old_size, new_size);
fs/ntfs/runlist.c
1536
if (IS_ERR(rl))
fs/ntfs/runlist.c
1540
runlist->rl = rl;
fs/ntfs/runlist.c
1542
} else if (likely(/* !rl->length && */ new_length > rl->vcn)) {
fs/ntfs/runlist.c
1549
if ((rl > runlist->rl) && ((rl - 1)->lcn == LCN_HOLE))
fs/ntfs/runlist.c
1550
(rl - 1)->length = new_length - (rl - 1)->vcn;
fs/ntfs/runlist.c
1553
old_size = rl - runlist->rl + 1;
fs/ntfs/runlist.c
1555
rl = ntfs_rl_realloc(runlist->rl, old_size,
fs/ntfs/runlist.c
1557
if (IS_ERR(rl)) {
fs/ntfs/runlist.c
1559
return PTR_ERR(rl);
fs/ntfs/runlist.c
1561
runlist->rl = rl;
fs/ntfs/runlist.c
1566
rl += old_size - 1;
fs/ntfs/runlist.c
1568
rl->lcn = LCN_HOLE;
fs/ntfs/runlist.c
1569
rl->length = new_length - rl->vcn;
fs/ntfs/runlist.c
1571
rl++;
fs/ntfs/runlist.c
1572
rl->length = 0;
fs/ntfs/runlist.c
1575
rl->vcn = new_length;
fs/ntfs/runlist.c
1576
rl->lcn = LCN_ENOENT;
fs/ntfs/runlist.c
1579
rl->lcn = LCN_ENOENT;
fs/ntfs/runlist.c
1591
int ntfs_rl_sparse(struct runlist_element *rl)
fs/ntfs/runlist.c
1595
if (!rl)
fs/ntfs/runlist.c
1598
for (rlc = rl; rlc->length; rlc++)
fs/ntfs/runlist.c
1616
s64 ntfs_rl_get_compressed_size(struct ntfs_volume *vol, struct runlist_element *rl)
fs/ntfs/runlist.c
1621
if (!rl)
fs/ntfs/runlist.c
1624
for (rlc = rl; rlc->length; rlc++) {
fs/ntfs/runlist.c
1747
struct runlist_element *rl, *rl_3rd;
fs/ntfs/runlist.c
1758
rl = &new_rl[new_1st_cnt + new_2nd_cnt];
fs/ntfs/runlist.c
1761
rl = &new_rl[new_1st_cnt + new_2nd_cnt - 1];
fs/ntfs/runlist.c
1764
rl_3rd->vcn = rl->vcn + rl->length;
fs/ntfs/runlist.c
1767
rl = rl_3rd;
fs/ntfs/runlist.c
527
struct runlist_element *drl = d_runlist->rl, *rl;
fs/ntfs/runlist.c
589
rl = __ntfs_attr_find_vcn_nolock(d_runlist, srl[sstart].vcn);
fs/ntfs/runlist.c
590
if (IS_ERR(rl))
fs/ntfs/runlist.c
593
di = (int)(rl - d_runlist->rl);
fs/ntfs/runlist.c
72
struct runlist_element *ntfs_rl_realloc(struct runlist_element *rl,
fs/ntfs/runlist.c
736
struct runlist_element *rl, *new_rl; /* The output runlist. */
fs/ntfs/runlist.c
765
rl = kvzalloc(rlsize = PAGE_SIZE, GFP_NOFS);
fs/ntfs/runlist.c
766
if (unlikely(!rl))
fs/ntfs/runlist.c
77
old_size = old_size * sizeof(*rl);
fs/ntfs/runlist.c
770
rl->vcn = 0;
fs/ntfs/runlist.c
771
rl->lcn = LCN_RL_NOT_MAPPED;
fs/ntfs/runlist.c
772
rl->length = vcn;
fs/ntfs/runlist.c
78
new_size = new_size * sizeof(*rl);
fs/ntfs/runlist.c
781
if (((rlpos + 3) * sizeof(*rl)) > rlsize) {
fs/ntfs/runlist.c
786
kvfree(rl);
fs/ntfs/runlist.c
789
memcpy(rl2, rl, rlsize);
fs/ntfs/runlist.c
790
kvfree(rl);
fs/ntfs/runlist.c
791
rl = rl2;
fs/ntfs/runlist.c
795
rl[rlpos].vcn = vcn;
fs/ntfs/runlist.c
80
return rl;
fs/ntfs/runlist.c
825
rl[rlpos].length = deltaxcn;
fs/ntfs/runlist.c
834
rl[rlpos].lcn = LCN_HOLE;
fs/ntfs/runlist.c
86
if (likely(rl != NULL)) {
fs/ntfs/runlist.c
868
if ((lcn != -1) && !rl[rlpos].length) {
fs/ntfs/runlist.c
876
rl[rlpos].lcn = lcn;
fs/ntfs/runlist.c
879
if (rl[rlpos].length)
fs/ntfs/runlist.c
89
memcpy(new_rl, rl, old_size);
fs/ntfs/runlist.c
90
kvfree(rl);
fs/ntfs/runlist.c
917
rl[rlpos].vcn = vcn;
fs/ntfs/runlist.c
918
vcn += rl[rlpos].length = max_cluster -
fs/ntfs/runlist.c
920
rl[rlpos].lcn = LCN_RL_NOT_MAPPED;
fs/ntfs/runlist.c
929
rl[rlpos].lcn = LCN_ENOENT;
fs/ntfs/runlist.c
931
rl[rlpos].lcn = LCN_RL_NOT_MAPPED;
fs/ntfs/runlist.c
934
rl[rlpos].vcn = vcn;
fs/ntfs/runlist.c
935
rl[rlpos].length = (s64)0;
fs/ntfs/runlist.c
937
if (!old_runlist || !old_runlist->rl) {
fs/ntfs/runlist.c
940
ntfs_debug_dump_runlist(rl);
fs/ntfs/runlist.c
941
return rl;
fs/ntfs/runlist.c
944
new_rl = ntfs_runlists_merge(old_runlist, rl, rlpos + 1, new_rl_count);
fs/ntfs/runlist.c
947
kvfree(rl);
fs/ntfs/runlist.c
953
kvfree(rl);
fs/ntfs/runlist.c
981
s64 ntfs_rl_vcn_to_lcn(const struct runlist_element *rl, const s64 vcn)
fs/ntfs/runlist.c
990
if (unlikely(!rl))
fs/ntfs/runlist.c
994
if (unlikely(vcn < rl[0].vcn))
fs/ntfs/runlist.c
997
for (i = 0; likely(rl[i].length); i++) {
fs/ntfs/runlist.c
998
if (vcn < rl[i+1].vcn) {
fs/ntfs/runlist.c
999
if (likely(rl[i].lcn >= 0))
fs/ntfs/runlist.h
44
struct runlist_element *rl;
fs/ntfs/runlist.h
50
static inline void ntfs_init_runlist(struct runlist *rl)
fs/ntfs/runlist.h
52
rl->rl = NULL;
fs/ntfs/runlist.h
53
init_rwsem(&rl->lock);
fs/ntfs/runlist.h
54
rl->count = 0;
fs/ntfs/runlist.h
55
rl->rl_hint = -1;
fs/ntfs/runlist.h
74
s64 ntfs_rl_vcn_to_lcn(const struct runlist_element *rl, const s64 vcn);
fs/ntfs/runlist.h
75
struct runlist_element *ntfs_rl_find_vcn_nolock(struct runlist_element *rl, const s64 vcn);
fs/ntfs/runlist.h
77
const struct runlist_element *rl, const s64 first_vcn,
fs/ntfs/runlist.h
80
const int dst_len, const struct runlist_element *rl,
fs/ntfs/runlist.h
85
int ntfs_rl_sparse(struct runlist_element *rl);
fs/ntfs/runlist.h
86
s64 ntfs_rl_get_compressed_size(struct ntfs_volume *vol, struct runlist_element *rl);
fs/ntfs/runlist.h
95
struct runlist_element *ntfs_rl_realloc(struct runlist_element *rl, int old_size,
fs/ntfs/super.c
1005
rl = mirr_ni->runlist.rl;
fs/ntfs/super.c
1009
if (rl2[i].vcn != rl[i].vcn || rl2[i].lcn != rl[i].lcn ||
fs/ntfs/super.c
1010
rl2[i].length != rl[i].length) {
fs/ntfs/super.c
908
struct runlist_element *rl, rl2[2];
fs/ocfs2/refcounttree.c
1242
struct ocfs2_refcount_list *rl = &rb->rf_records;
fs/ocfs2/refcounttree.c
1243
struct ocfs2_refcount_rec *rec = &rl->rl_recs[index];
fs/ocfs2/refcounttree.c
1258
if (index != le16_to_cpu(rl->rl_used) - 1) {
fs/ocfs2/refcounttree.c
1260
(le16_to_cpu(rl->rl_used) - index - 1) *
fs/ocfs2/refcounttree.c
1262
memset(&rl->rl_recs[le16_to_cpu(rl->rl_used) - 1],
fs/ocfs2/refcounttree.c
1266
le16_add_cpu(&rl->rl_used, -1);
fs/ocfs2/refcounttree.c
1405
static int ocfs2_find_refcount_split_pos(struct ocfs2_refcount_list *rl,
fs/ocfs2/refcounttree.c
1408
int num_used = le16_to_cpu(rl->rl_used);
fs/ocfs2/refcounttree.c
1414
&rl->rl_recs[middle - delta - 1],
fs/ocfs2/refcounttree.c
1415
&rl->rl_recs[middle - delta])) {
fs/ocfs2/refcounttree.c
1426
&rl->rl_recs[middle + delta],
fs/ocfs2/refcounttree.c
1427
&rl->rl_recs[middle + delta + 1])) {
fs/ocfs2/refcounttree.c
1436
*split_pos = ocfs2_get_ref_rec_low_cpos(&rl->rl_recs[*split_index]);
fs/ocfs2/refcounttree.c
1448
struct ocfs2_refcount_list *rl = &rb->rf_records;
fs/ocfs2/refcounttree.c
1455
le16_to_cpu(rl->rl_count), le16_to_cpu(rl->rl_used));
fs/ocfs2/refcounttree.c
1469
sort(&rl->rl_recs, le16_to_cpu(rl->rl_used),
fs/ocfs2/refcounttree.c
1473
ret = ocfs2_find_refcount_split_pos(rl, &cpos, &split_index);
fs/ocfs2/refcounttree.c
1482
num_moved = le16_to_cpu(rl->rl_used) - split_index;
fs/ocfs2/refcounttree.c
1483
memcpy(new_rl->rl_recs, &rl->rl_recs[split_index],
fs/ocfs2/refcounttree.c
1487
memset(&rl->rl_recs[split_index], 0,
fs/ocfs2/refcounttree.c
1491
le16_add_cpu(&rl->rl_used, -num_moved);
fs/ocfs2/refcounttree.c
1494
sort(&rl->rl_recs, le16_to_cpu(rl->rl_used),
fs/qnx4/inode.c
161
int rd, rl;
fs/qnx4/inode.c
168
rl = le32_to_cpu(s->RootDir.di_first_xtnt.xtnt_size);
fs/qnx4/inode.c
169
for (j = 0; j < rl; j++) {
include/linux/blk-mq.h
251
static inline int rq_list_empty(const struct rq_list *rl)
include/linux/blk-mq.h
253
return rl->head == NULL;
include/linux/blk-mq.h
256
static inline void rq_list_init(struct rq_list *rl)
include/linux/blk-mq.h
258
rl->head = NULL;
include/linux/blk-mq.h
259
rl->tail = NULL;
include/linux/blk-mq.h
265
if (rl->tail)
include/linux/blk-mq.h
266
rl->tail->rq_next = rq;
include/linux/blk-mq.h
268
rl->head = rq;
include/linux/blk-mq.h
269
rl->tail = rq;
include/linux/blk-mq.h
274
rq->rq_next = rl->head;
include/linux/blk-mq.h
275
rl->head = rq;
include/linux/blk-mq.h
276
if (!rl->tail)
include/linux/blk-mq.h
277
rl->tail = rq;
include/linux/blk-mq.h
280
static inline struct request *rq_list_pop(struct rq_list *rl)
include/linux/blk-mq.h
282
struct request *rq = rl->head;
include/linux/blk-mq.h
285
rl->head = rl->head->rq_next;
include/linux/blk-mq.h
286
if (!rl->head)
include/linux/blk-mq.h
287
rl->tail = NULL;
include/linux/blk-mq.h
294
static inline struct request *rq_list_peek(struct rq_list *rl)
include/linux/blk-mq.h
296
return rl->head;
include/linux/blk-mq.h
300
for (pos = rq_list_peek((rl)); (pos); pos = pos->rq_next)
include/linux/blk-mq.h
302
#define rq_list_for_each_safe(rl, pos, nxt) \
include/linux/blk-mq.h
303
for (pos = rq_list_peek((rl)), nxt = pos->rq_next; \
include/linux/jump_label_ratelimit.h
41
jump_label_rate_limit(struct static_key_deferred *key, unsigned long rl);
include/linux/jump_label_ratelimit.h
45
#define DEFINE_STATIC_KEY_DEFERRED_TRUE(name, rl) \
include/linux/jump_label_ratelimit.h
48
.timeout = (rl), \
include/linux/jump_label_ratelimit.h
54
#define DEFINE_STATIC_KEY_DEFERRED_FALSE(name, rl) \
include/linux/jump_label_ratelimit.h
57
.timeout = (rl), \
include/linux/jump_label_ratelimit.h
73
#define DEFINE_STATIC_KEY_DEFERRED_TRUE(name, rl) \
include/linux/jump_label_ratelimit.h
75
#define DEFINE_STATIC_KEY_DEFERRED_FALSE(name, rl) \
include/linux/jump_label_ratelimit.h
91
unsigned long rl)
include/linux/math64.h
215
} rl, rm, rn, rh, a0, b0;
include/linux/math64.h
221
rl.ll = mul_u32_u32(a0.l.low, b0.l.low);
include/linux/math64.h
231
rl.l.high = c = (u64)rl.l.high + rm.l.low + rn.l.low;
include/linux/math64.h
240
return rl.ll;
include/linux/math64.h
242
return (rl.ll >> shift) | (rh.ll << (64 - shift));
include/linux/math64.h
279
} u, rl, rh;
include/linux/math64.h
282
rl.ll = mul_u32_u32(u.l.low, mul);
include/linux/math64.h
283
rh.ll = mul_u32_u32(u.l.high, mul) + rl.l.high;
include/linux/math64.h
286
rl.l.high = do_div(rh.ll, divisor);
include/linux/math64.h
289
do_div(rl.ll, divisor);
include/linux/math64.h
291
rl.l.high = rh.l.low;
include/linux/math64.h
292
return rl.ll;
include/linux/mlx5/driver.h
1091
struct mlx5_rate_limit *rl);
include/linux/mlx5/driver.h
1092
void mlx5_rl_remove_rate(struct mlx5_core_dev *dev, struct mlx5_rate_limit *rl);
include/math-emu/op-2.h
147
#define __FP_FRAC_ADD_2(rh, rl, xh, xl, yh, yl) \
include/math-emu/op-2.h
148
(rh = xh + yh + ((rl = xl + yl) < xl))
include/math-emu/op-2.h
151
#define __FP_FRAC_SUB_2(rh, rl, xh, xl, yh, yl) \
include/math-emu/op-2.h
152
(rh = xh - yh - ((rl = xl - yl) > xl))
io_uring/io_uring.c
2076
unsigned int cq_entries, struct io_rings_layout *rl)
io_uring/io_uring.c
2091
rl->sq_array_offset = SIZE_MAX;
io_uring/io_uring.c
2097
rl->sq_size = array_size(sqe_size, sq_entries);
io_uring/io_uring.c
2098
if (rl->sq_size == SIZE_MAX)
io_uring/io_uring.c
2116
rl->sq_array_offset = off;
io_uring/io_uring.c
2124
rl->rings_size = off;
io_uring/io_uring.c
2718
struct io_rings_layout *rl = &config->layout;
io_uring/io_uring.c
2728
rd.size = PAGE_ALIGN(rl->rings_size);
io_uring/io_uring.c
2739
ctx->sq_array = (u32 *)((char *)rings + rl->sq_array_offset);
io_uring/io_uring.c
2742
rd.size = PAGE_ALIGN(rl->sq_size);
io_uring/register.c
505
struct io_rings_layout *rl = &config.layout;
io_uring/register.c
526
rd.size = PAGE_ALIGN(rl->rings_size);
io_uring/register.c
556
rd.size = PAGE_ALIGN(rl->sq_size);
io_uring/register.c
639
ctx->sq_array = (u32 *)((char *)n.rings + rl->sq_array_offset);
kernel/jump_label.c
367
unsigned long rl)
kernel/jump_label.c
370
key->timeout = rl;
sound/soc/codecs/rt1305.c
1058
regmap_read(rt1305->regmap, RT1305_PR_BASE + 0x56, &rl);
sound/soc/codecs/rt1305.c
1059
rhl = (rh << 16) | rl;
sound/soc/codecs/rt1305.c
1062
pr_debug("Left_rhl = 0x%x rh=0x%x rl=0x%x\n", rhl, rh, rl);
sound/soc/codecs/rt1305.c
1075
regmap_read(rt1305->regmap, RT1305_PR_BASE + 0x56, &rl);
sound/soc/codecs/rt1305.c
1076
rhl = (rh << 16) | rl;
sound/soc/codecs/rt1305.c
1079
pr_debug("Right_rhl = 0x%x rh=0x%x rl=0x%x\n", rhl, rh, rl);
sound/soc/codecs/rt1305.c
993
unsigned int rh, rl, rhl, r0ohm;
sound/usb/usx2y/usbus428ctldefs.h
68
rl;
tools/perf/bench/epoll-ctl.c
319
struct rlimit rl, prevrl;
tools/perf/bench/epoll-ctl.c
358
rl.rlim_cur = rl.rlim_max = nfds * nthreads * 2 + 50;
tools/perf/bench/epoll-ctl.c
360
(uint64_t)prevrl.rlim_max, (uint64_t)rl.rlim_max);
tools/perf/bench/epoll-ctl.c
361
if (setrlimit(RLIMIT_NOFILE, &rl) < 0)
tools/perf/bench/epoll-wait.c
439
struct rlimit rl, prevrl;
tools/perf/bench/epoll-wait.c
483
rl.rlim_cur = rl.rlim_max = nfds * nthreads * 2 + 50;
tools/perf/bench/epoll-wait.c
485
(uint64_t)prevrl.rlim_max, (uint64_t)rl.rlim_max);
tools/perf/bench/epoll-wait.c
486
if (setrlimit(RLIMIT_NOFILE, &rl) < 0)
tools/perf/util/print-events.c
212
static struct rb_node *mep_new(struct rblist *rl __maybe_unused, const void *entry)
tools/perf/util/print-events.c
223
static void mep_delete(struct rblist *rl __maybe_unused,
tools/testing/selftests/kvm/lib/kvm_util.c
446
struct rlimit rl;
tools/testing/selftests/kvm/lib/kvm_util.c
452
TEST_ASSERT(!getrlimit(RLIMIT_NOFILE, &rl), "getrlimit() failed!");
tools/testing/selftests/kvm/lib/kvm_util.c
454
if (rl.rlim_cur < nr_fds_wanted) {
tools/testing/selftests/kvm/lib/kvm_util.c
455
rl.rlim_cur = nr_fds_wanted;
tools/testing/selftests/kvm/lib/kvm_util.c
456
if (rl.rlim_max < nr_fds_wanted) {
tools/testing/selftests/kvm/lib/kvm_util.c
457
int old_rlim_max = rl.rlim_max;
tools/testing/selftests/kvm/lib/kvm_util.c
459
rl.rlim_max = nr_fds_wanted;
tools/testing/selftests/kvm/lib/kvm_util.c
460
__TEST_REQUIRE(setrlimit(RLIMIT_NOFILE, &rl) >= 0,
tools/testing/selftests/kvm/lib/kvm_util.c
464
TEST_ASSERT(!setrlimit(RLIMIT_NOFILE, &rl), "setrlimit() failed!");
tools/testing/selftests/landlock/fs_test.c
4746
struct rlimit rl = {
tools/testing/selftests/landlock/fs_test.c
4751
ASSERT_EQ(0, setrlimit(RLIMIT_CORE, &rl));