Symbol: rl
arch/arm/vfp/vfp.h
105
u64 rh, rl;
arch/arm/vfp/vfp.h
106
mul64to128(&rh, &rl, n, m);
arch/arm/vfp/vfp.h
107
return rh | (rl != 0);
arch/arm/vfp/vfp.h
73
u64 rh, rma, rmb, rl;
arch/arm/vfp/vfp.h
77
rl = (u64)nl * ml;
arch/arm/vfp/vfp.h
90
rl += rma;
arch/arm/vfp/vfp.h
91
rh += (rl < rma);
arch/arm/vfp/vfp.h
93
*resl = rl;
arch/arm64/include/asm/atomic_ll_sc.h
323
[rl] "=&r" (r.low), [rh] "=&r" (r.high), \
arch/mips/include/asm/octeon/cvmx-pciercx-defs.h
197
__BITFIELD_FIELD(uint32_t rl:1,
arch/riscv/include/asm/bitops.h
313
__op_bit_ord(and, __NOT, nr, addr, .rl);
arch/riscv/net/bpf_jit.h
282
static inline u32 rv_amo_insn(u8 funct5, u8 aq, u8 rl, u8 rs2, u8 rs1,
arch/riscv/net/bpf_jit.h
285
u8 funct7 = (funct5 << 2) | (aq << 1) | rl;
arch/riscv/net/bpf_jit.h
565
static inline u32 rv_amoadd_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
567
return rv_amo_insn(0, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
570
static inline u32 rv_amoand_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
572
return rv_amo_insn(0xc, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
575
static inline u32 rv_amoor_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
577
return rv_amo_insn(0x8, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
580
static inline u32 rv_amoxor_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
582
return rv_amo_insn(0x4, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
585
static inline u32 rv_amoswap_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
587
return rv_amo_insn(0x1, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
590
static inline u32 rv_lr_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
592
return rv_amo_insn(0x2, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
595
static inline u32 rv_sc_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
597
return rv_amo_insn(0x3, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
757
static inline u32 rvzacas_amocas_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
759
return rv_amo_insn(0x5, aq, rl, rs2, rs1, 2, rd, 0x2f);
arch/riscv/net/bpf_jit.h
762
static inline u32 rvzacas_amocas_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
764
return rv_amo_insn(0x5, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
899
static inline u32 rv_amoadd_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
901
return rv_amo_insn(0, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
904
static inline u32 rv_amoand_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
906
return rv_amo_insn(0xc, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
909
static inline u32 rv_amoor_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
911
return rv_amo_insn(0x8, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
914
static inline u32 rv_amoxor_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
916
return rv_amo_insn(0x4, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
919
static inline u32 rv_amoswap_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
921
return rv_amo_insn(0x1, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
924
static inline u32 rv_lr_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
926
return rv_amo_insn(0x2, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/riscv/net/bpf_jit.h
929
static inline u32 rv_sc_d(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl)
arch/riscv/net/bpf_jit.h
931
return rv_amo_insn(0x3, aq, rl, rs2, rs1, 3, rd, 0x2f);
arch/s390/kvm/guestdbg.c
471
s32 rl = *((s32 *) (opcode + 1));
arch/s390/kvm/guestdbg.c
474
*addr += (u64)(s64) rl * 2;
block/blk-cgroup.h
498
#define blk_queue_for_each_rl(rl, q) \
block/blk-cgroup.h
499
for ((rl) = &(q)->root_rl; (rl); (rl) = NULL)
block/sed-opal.c
1468
bool rl, bool wl)
block/sed-opal.c
1490
add_token_u8(&err, dev, rl);
crypto/camellia_generic.c
319
#define ROLDQ(ll, lr, rl, rr, w0, w1, bits) ({ \
crypto/camellia_generic.c
322
lr = (lr << bits) + (rl >> (32 - bits)); \
crypto/camellia_generic.c
323
rl = (rl << bits) + (rr >> (32 - bits)); \
crypto/camellia_generic.c
327
#define ROLDQo32(ll, lr, rl, rr, w0, w1, bits) ({ \
crypto/camellia_generic.c
330
ll = (lr << (bits - 32)) + (rl >> (64 - bits)); \
crypto/camellia_generic.c
331
lr = (rl << (bits - 32)) + (rr >> (64 - bits)); \
crypto/camellia_generic.c
332
rl = (rr << (bits - 32)) + (w0 >> (64 - bits)); \
crypto/camellia_generic.c
811
#define CAMELLIA_FLS(ll, lr, rl, rr, kll, klr, krl, krr, t0, t1, t2, t3) ({ \
crypto/camellia_generic.c
816
rl ^= t2; \
crypto/camellia_generic.c
820
t3 &= rl; \
crypto/crypto_user.c
106
struct crypto_report_larval rl;
crypto/crypto_user.c
108
memset(&rl, 0, sizeof(rl));
crypto/crypto_user.c
109
strscpy(rl.type, "larval", sizeof(rl.type));
crypto/crypto_user.c
110
if (nla_put(skb, CRYPTOCFGA_REPORT_LARVAL, sizeof(rl), &rl))
drivers/block/drbd/drbd_main.c
1089
unsigned long rl;
drivers/block/drbd/drbd_main.c
1122
rl = tmp - c->bit_offset;
drivers/block/drbd/drbd_main.c
1125
if (rl == 0) {
drivers/block/drbd/drbd_main.c
1138
if (rl == 0) {
drivers/block/drbd/drbd_main.c
1144
bits = vli_encode_bits(&bs, rl);
drivers/block/drbd/drbd_main.c
1153
plain_bits += rl;
drivers/block/drbd/drbd_receiver.c
4508
u64 rl;
drivers/block/drbd/drbd_receiver.c
4522
for (have = bits; have > 0; s += rl, toggle = !toggle) {
drivers/block/drbd/drbd_receiver.c
4523
bits = vli_decode_bits(&rl, look_ahead);
drivers/block/drbd/drbd_receiver.c
4528
e = s + rl -1;
drivers/crypto/intel/qat/qat_common/adf_rl.c
1053
struct adf_rl *rl;
drivers/crypto/intel/qat/qat_common/adf_rl.c
1068
rl = kzalloc_obj(*rl);
drivers/crypto/intel/qat/qat_common/adf_rl.c
1069
if (!rl) {
drivers/crypto/intel/qat/qat_common/adf_rl.c
1074
mutex_init(&rl->rl_lock);
drivers/crypto/intel/qat/qat_common/adf_rl.c
1075
rl->device_data = &accel_dev->hw_device->rl_data;
drivers/crypto/intel/qat/qat_common/adf_rl.c
1076
rl->accel_dev = accel_dev;
drivers/crypto/intel/qat/qat_common/adf_rl.c
1077
init_rwsem(&rl->user_input.lock);
drivers/crypto/intel/qat/qat_common/adf_rl.c
1078
accel_dev->rate_limiting = rl;
drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c
1157
if (!is_psp_fw_valid(psp->rl))
drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c
1163
memcpy(psp->fw_pri_buf, psp->rl.start_addr, psp->rl.size_bytes);
drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c
1168
cmd->cmd.cmd_load_ip_fw.fw_size = psp->rl.size_bytes;
drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c
3584
psp->rl.fw_version = le32_to_cpu(desc->fw_version);
drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c
3585
psp->rl.feature_version = le32_to_cpu(desc->fw_version);
drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c
3586
psp->rl.size_bytes = le32_to_cpu(desc->size_bytes);
drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c
3587
psp->rl.start_addr = ucode_start_addr;
drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c
3738
adev->psp.rl.size_bytes = le32_to_cpu(sos_hdr_v1_3->rl.size_bytes);
drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c
3739
adev->psp.rl.start_addr = ucode_array_start_addr +
drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c
3740
le32_to_cpu(sos_hdr_v1_3->rl.offset_bytes);
drivers/gpu/drm/amd/amdgpu/amdgpu_psp.h
382
struct psp_bin_desc rl;
drivers/gpu/drm/amd/amdgpu/amdgpu_ucode.h
106
struct psp_fw_legacy_bin_desc rl;
drivers/hwmon/asc7621.c
799
#define PREAD(name, n, pri, rm, rl, m, s, r) \
drivers/hwmon/asc7621.c
801
.priority = pri, .msb[0] = rm, .lsb[0] = rl, .mask[0] = m, \
drivers/hwmon/asc7621.c
804
#define PWRITE(name, n, pri, rm, rl, m, s, r) \
drivers/hwmon/asc7621.c
806
.priority = pri, .msb[0] = rm, .lsb[0] = rl, .mask[0] = m, \
drivers/hwmon/asc7621.c
813
#define PWRITEM(name, n, pri, rm, rl, m, s, r) \
drivers/hwmon/asc7621.c
815
.priority = pri, .msb = rm, .lsb = rl, .mask = m, .shift = s,}
drivers/infiniband/hw/mlx5/mlx5_ib.h
538
struct mlx5_rate_limit rl;
drivers/infiniband/hw/mlx5/qp.c
3845
struct mlx5_rate_limit old_rl = ibqp->rl;
drivers/infiniband/hw/mlx5/qp.c
3870
new_rl = raw_qp_param->rl;
drivers/infiniband/hw/mlx5/qp.c
3908
ibqp->rl = new_rl;
drivers/infiniband/hw/mlx5/qp.c
4383
raw_qp_param.rl.rate = attr->rate_limit;
drivers/infiniband/hw/mlx5/qp.c
4388
raw_qp_param.rl.max_burst_sz =
drivers/infiniband/hw/mlx5/qp.c
4399
raw_qp_param.rl.typical_pkt_sz =
drivers/infiniband/hw/mlx5/qp.c
72
struct mlx5_rate_limit rl;
drivers/net/ethernet/hisilicon/hns3/hns3_ethtool.c
1440
u32 rl;
drivers/net/ethernet/hisilicon/hns3/hns3_ethtool.c
1455
rl = hns3_rl_round_down(cmd->rx_coalesce_usecs_high);
drivers/net/ethernet/hisilicon/hns3/hns3_ethtool.c
1456
if (rl != cmd->rx_coalesce_usecs_high) {
drivers/net/ethernet/hisilicon/hns3/hns3_ethtool.c
1459
cmd->rx_coalesce_usecs_high, rl);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2868
struct virtchnl2_rss_lut *rl __free(kfree) = NULL;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2882
buf_size = struct_size(rl, lut, rss_data->rss_lut_size);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2883
rl = kzalloc(buf_size, GFP_KERNEL);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2884
if (!rl)
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2887
rl->vport_id = cpu_to_le32(vport_id);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2890
xn_params.send_buf.iov_base = rl;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2901
rl->lut_entries = cpu_to_le16(rss_data->rss_lut_size);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
2903
rl->lut[i] = rxhash_ena ?
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
466
void mlx5e_mqprio_rl_free(struct mlx5e_mqprio_rl *rl)
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
468
kvfree(rl);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
471
int mlx5e_mqprio_rl_init(struct mlx5e_mqprio_rl *rl, struct mlx5_core_dev *mdev, u8 num_tc,
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
484
rl->mdev = mdev;
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
485
rl->num_tc = num_tc;
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
486
rl->leaves_id = kvcalloc(num_tc, sizeof(*rl->leaves_id), GFP_KERNEL);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
487
if (!rl->leaves_id)
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
490
err = mlx5_qos_create_root_node(mdev, &rl->root_id);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
494
qos_dbg(mdev, "Root created, id %#x\n", rl->root_id);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
500
err = mlx5_qos_create_leaf_node(mdev, rl->root_id, 0, max_average_bw,
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
501
&rl->leaves_id[tc]);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
506
tc, rl->leaves_id[tc], max_average_bw);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
512
mlx5_qos_destroy_node(mdev, rl->leaves_id[tc]);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
513
mlx5_qos_destroy_node(mdev, rl->root_id);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
515
kvfree(rl->leaves_id);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
519
void mlx5e_mqprio_rl_cleanup(struct mlx5e_mqprio_rl *rl)
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
523
for (tc = 0; tc < rl->num_tc; tc++)
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
524
mlx5_qos_destroy_node(rl->mdev, rl->leaves_id[tc]);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
525
mlx5_qos_destroy_node(rl->mdev, rl->root_id);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
526
kvfree(rl->leaves_id);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
529
int mlx5e_mqprio_rl_get_node_hw_id(struct mlx5e_mqprio_rl *rl, int tc, u32 *hw_id)
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
531
if (tc >= rl->num_tc)
drivers/net/ethernet/mellanox/mlx5/core/en/qos.c
534
*hw_id = rl->leaves_id[tc];
drivers/net/ethernet/mellanox/mlx5/core/en/qos.h
46
void mlx5e_mqprio_rl_free(struct mlx5e_mqprio_rl *rl);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.h
47
int mlx5e_mqprio_rl_init(struct mlx5e_mqprio_rl *rl, struct mlx5_core_dev *mdev, u8 num_tc,
drivers/net/ethernet/mellanox/mlx5/core/en/qos.h
49
void mlx5e_mqprio_rl_cleanup(struct mlx5e_mqprio_rl *rl);
drivers/net/ethernet/mellanox/mlx5/core/en/qos.h
50
int mlx5e_mqprio_rl_get_node_hw_id(struct mlx5e_mqprio_rl *rl, int tc, u32 *hw_id);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1924
struct mlx5_rate_limit rl = {0};
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1931
rl.rate = sq->rate_limit;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
1932
mlx5_rl_remove_rate(mdev, &rl);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
2434
struct mlx5_rate_limit rl = {0};
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
2443
rl.rate = sq->rate_limit;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
2445
mlx5_rl_remove_rate(mdev, &rl);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
2451
rl.rate = rate;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
2452
err = mlx5_rl_add_rate(mdev, &rl_index, &rl);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
2470
mlx5_rl_remove_rate(mdev, &rl);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3733
struct mlx5e_mqprio_rl *rl)
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3740
if (rl)
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3741
mlx5e_mqprio_rl_get_node_hw_id(rl, tc, &hw_id);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3748
struct mlx5e_mqprio_rl *rl)
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3758
mlx5e_mqprio_rl_update_params(params, rl);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3862
struct mlx5e_mqprio_rl *rl;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3868
rl = mlx5e_mqprio_rl_alloc();
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3869
if (!rl)
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3872
err = mlx5e_mqprio_rl_init(rl, mdev, num_tc, max_rate);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3874
mlx5e_mqprio_rl_free(rl);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3878
return rl;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3885
struct mlx5e_mqprio_rl *rl;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3892
rl = mlx5e_mqprio_rl_create(priv->mdev, mqprio->qopt.num_tc, mqprio->max_rate);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3893
if (IS_ERR(rl))
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3894
return PTR_ERR(rl);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3897
mlx5e_params_mqprio_channel_set(&new_params, mqprio, rl);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3902
if (rl) {
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3903
mlx5e_mqprio_rl_cleanup(rl);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3904
mlx5e_mqprio_rl_free(rl);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
3913
priv->mqprio_rl = rl;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
6017
struct mlx5e_mqprio_rl *rl;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
6023
rl = mlx5e_mqprio_rl_create(priv->mdev, params->mqprio.num_tc,
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
6025
if (IS_ERR(rl))
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
6026
rl = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
6027
priv->mqprio_rl = rl;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
6028
mlx5e_mqprio_rl_update_params(params, rl);
drivers/net/ethernet/mellanox/mlx5/core/rl.c
377
struct mlx5_rate_limit *rl)
drivers/net/ethernet/mellanox/mlx5/core/rl.c
381
MLX5_SET(set_pp_rate_limit_context, rl_raw, rate_limit, rl->rate);
drivers/net/ethernet/mellanox/mlx5/core/rl.c
383
rl->max_burst_sz);
drivers/net/ethernet/mellanox/mlx5/core/rl.c
385
rl->typical_pkt_sz);
drivers/net/ethernet/mellanox/mlx5/core/rl.c
394
void mlx5_rl_remove_rate(struct mlx5_core_dev *dev, struct mlx5_rate_limit *rl)
drivers/net/ethernet/mellanox/mlx5/core/rl.c
401
if (rl->rate == 0)
drivers/net/ethernet/mellanox/mlx5/core/rl.c
404
MLX5_SET(set_pp_rate_limit_context, rl_raw, rate_limit, rl->rate);
drivers/net/ethernet/mellanox/mlx5/core/rl.c
406
rl->max_burst_sz);
drivers/net/ethernet/mellanox/mlx5/core/rl.c
408
rl->typical_pkt_sz);
drivers/net/ethernet/mellanox/mlx5/core/rl.c
416
rl->rate, rl->max_burst_sz, rl->typical_pkt_sz);
drivers/net/ethernet/microchip/enc28j60.c
292
int rl, rh;
drivers/net/ethernet/microchip/enc28j60.c
295
rl = spi_read_op(priv, ENC28J60_READ_CTRL_REG, address);
drivers/net/ethernet/microchip/enc28j60.c
298
return (rh << 8) | rl;
drivers/net/ethernet/qlogic/qed/qed_init_fw_funcs.c
194
#define PQ_INFO_ELEMENT(vp, pf, tc, port, rl_valid, rl) \
drivers/net/ethernet/qlogic/qed/qed_init_fw_funcs.c
196
((rl_valid ? 1 : 0) << 22) | (((rl) & 255) << 24) | \
drivers/net/ethernet/qlogic/qed/qed_init_fw_funcs.c
197
(((rl) >> 8) << 9))
drivers/nvme/host/trace.c
109
u16 rl = get_unaligned_le16(cdw10 + 12);
drivers/nvme/host/trace.c
113
slba, mndw, rl, atype);
drivers/nvme/target/trace.c
42
u16 rl = get_unaligned_le16(cdw10 + 12);
drivers/nvme/target/trace.c
46
slba, mndw, rl, atype);
drivers/s390/scsi/zfcp_reqlist.h
107
zfcp_reqlist_find(struct zfcp_reqlist *rl, u64 req_id)
drivers/s390/scsi/zfcp_reqlist.h
112
spin_lock_irqsave(&rl->lock, flags);
drivers/s390/scsi/zfcp_reqlist.h
113
req = _zfcp_reqlist_find(rl, req_id);
drivers/s390/scsi/zfcp_reqlist.h
114
spin_unlock_irqrestore(&rl->lock, flags);
drivers/s390/scsi/zfcp_reqlist.h
132
zfcp_reqlist_find_rm(struct zfcp_reqlist *rl, u64 req_id)
drivers/s390/scsi/zfcp_reqlist.h
137
spin_lock_irqsave(&rl->lock, flags);
drivers/s390/scsi/zfcp_reqlist.h
138
req = _zfcp_reqlist_find(rl, req_id);
drivers/s390/scsi/zfcp_reqlist.h
141
spin_unlock_irqrestore(&rl->lock, flags);
drivers/s390/scsi/zfcp_reqlist.h
156
static inline void zfcp_reqlist_add(struct zfcp_reqlist *rl,
drivers/s390/scsi/zfcp_reqlist.h
164
spin_lock_irqsave(&rl->lock, flags);
drivers/s390/scsi/zfcp_reqlist.h
165
list_add_tail(&req->list, &rl->buckets[i]);
drivers/s390/scsi/zfcp_reqlist.h
166
spin_unlock_irqrestore(&rl->lock, flags);
drivers/s390/scsi/zfcp_reqlist.h
174
static inline void zfcp_reqlist_move(struct zfcp_reqlist *rl,
drivers/s390/scsi/zfcp_reqlist.h
180
spin_lock_irqsave(&rl->lock, flags);
drivers/s390/scsi/zfcp_reqlist.h
182
list_splice_init(&rl->buckets[i], list);
drivers/s390/scsi/zfcp_reqlist.h
183
spin_unlock_irqrestore(&rl->lock, flags);
drivers/s390/scsi/zfcp_reqlist.h
200
zfcp_reqlist_apply_for_all(struct zfcp_reqlist *rl,
drivers/s390/scsi/zfcp_reqlist.h
207
spin_lock_irqsave(&rl->lock, flags);
drivers/s390/scsi/zfcp_reqlist.h
209
list_for_each_entry(req, &rl->buckets[i], list)
drivers/s390/scsi/zfcp_reqlist.h
211
spin_unlock_irqrestore(&rl->lock, flags);
drivers/s390/scsi/zfcp_reqlist.h
43
struct zfcp_reqlist *rl;
drivers/s390/scsi/zfcp_reqlist.h
45
rl = kzalloc_obj(struct zfcp_reqlist);
drivers/s390/scsi/zfcp_reqlist.h
46
if (!rl)
drivers/s390/scsi/zfcp_reqlist.h
49
spin_lock_init(&rl->lock);
drivers/s390/scsi/zfcp_reqlist.h
52
INIT_LIST_HEAD(&rl->buckets[i]);
drivers/s390/scsi/zfcp_reqlist.h
54
return rl;
drivers/s390/scsi/zfcp_reqlist.h
63
static inline int zfcp_reqlist_isempty(struct zfcp_reqlist *rl)
drivers/s390/scsi/zfcp_reqlist.h
68
if (!list_empty(&rl->buckets[i]))
drivers/s390/scsi/zfcp_reqlist.h
77
static inline void zfcp_reqlist_free(struct zfcp_reqlist *rl)
drivers/s390/scsi/zfcp_reqlist.h
80
BUG_ON(!zfcp_reqlist_isempty(rl));
drivers/s390/scsi/zfcp_reqlist.h
82
kfree(rl);
drivers/s390/scsi/zfcp_reqlist.h
86
_zfcp_reqlist_find(struct zfcp_reqlist *rl, u64 req_id)
drivers/s390/scsi/zfcp_reqlist.h
92
list_for_each_entry(req, &rl->buckets[i], list)
drivers/slimbus/messaging.c
152
unsigned long ms = txn->rl + HZ;
drivers/slimbus/messaging.c
261
txn->rl += msg->num_bytes;
drivers/slimbus/messaging.c
268
txn->rl++;
drivers/slimbus/qcom-ngd-ctrl.c
1058
txn.rl = txn.msg->num_bytes + 4;
drivers/slimbus/qcom-ngd-ctrl.c
1070
txn.rl = txn.msg->num_bytes + 4;
drivers/slimbus/qcom-ngd-ctrl.c
1105
txn.rl = 11;
drivers/slimbus/qcom-ngd-ctrl.c
1270
txn.rl = 8;
drivers/slimbus/qcom-ngd-ctrl.c
809
txn->rl > SLIM_MSGQ_BUF_LEN) {
drivers/slimbus/qcom-ngd-ctrl.c
814
pbuf = qcom_slim_ngd_tx_msg_get(ctrl, txn->rl, &tx_sent);
drivers/slimbus/qcom-ngd-ctrl.c
859
txn->rl = txn->msg->num_bytes + 4;
drivers/slimbus/qcom-ngd-ctrl.c
863
txn->rl--;
drivers/slimbus/qcom-ngd-ctrl.c
867
*pbuf = SLIM_MSG_ASM_FIRST_WORD(txn->rl, txn->mt, txn->mc, 0,
drivers/slimbus/qcom-ngd-ctrl.c
871
*pbuf = SLIM_MSG_ASM_FIRST_WORD(txn->rl, txn->mt, txn->mc, 1,
drivers/slimbus/qcom-ngd-ctrl.c
888
ret = qcom_slim_ngd_tx_msg_post(ctrl, pbuf, txn->rl);
drivers/slimbus/sched.c
105
txn.rl = 3;
drivers/slimbus/sched.c
97
txn.rl = 4;
drivers/slimbus/slimbus.h
128
u8 rl;
drivers/slimbus/slimbus.h
140
#define DEFINE_SLIM_LDEST_TXN(name, mc, rl, la, msg) \
drivers/slimbus/slimbus.h
141
struct slim_msg_txn name = { rl, 0, mc, SLIM_MSG_DEST_LOGICALADDR, 0,\
drivers/slimbus/slimbus.h
144
#define DEFINE_SLIM_BCAST_TXN(name, mc, rl, la, msg) \
drivers/slimbus/slimbus.h
145
struct slim_msg_txn name = { rl, 0, mc, SLIM_MSG_DEST_BROADCAST, 0,\
drivers/slimbus/slimbus.h
148
#define DEFINE_SLIM_EDEST_TXN(name, mc, rl, la, msg) \
drivers/slimbus/slimbus.h
149
struct slim_msg_txn name = { rl, 0, mc, SLIM_MSG_DEST_ENUMADDR, 0,\
drivers/usb/isp1760/isp1760-hcd.c
822
u32 rl = RL_COUNTER;
drivers/usb/isp1760/isp1760-hcd.c
863
rl = 0;
drivers/usb/isp1760/isp1760-hcd.c
874
ptd->dw2 |= TO_DW2_RL(rl);
drivers/video/fbdev/metronomefb.c
276
unsigned char rl;
drivers/video/fbdev/metronomefb.c
289
rl = mem[wfm_idx++];
drivers/video/fbdev/metronomefb.c
290
for (i = 0; i <= rl; i++)
fs/binfmt_flat.c
341
static void old_reloc(unsigned long rl)
fs/binfmt_flat.c
348
r.value = rl;
fs/dlm/lock.c
5530
struct rcom_lock *rl = (struct rcom_lock *) rc->rc_buf;
fs/dlm/lock.c
5533
lkb->lkb_ownpid = le32_to_cpu(rl->rl_ownpid);
fs/dlm/lock.c
5534
lkb->lkb_remid = le32_to_cpu(rl->rl_lkid);
fs/dlm/lock.c
5535
lkb->lkb_exflags = le32_to_cpu(rl->rl_exflags);
fs/dlm/lock.c
5536
dlm_set_dflags_val(lkb, le32_to_cpu(rl->rl_flags));
fs/dlm/lock.c
5538
lkb->lkb_lvbseq = le32_to_cpu(rl->rl_lvbseq);
fs/dlm/lock.c
5539
lkb->lkb_rqmode = rl->rl_rqmode;
fs/dlm/lock.c
5540
lkb->lkb_grmode = rl->rl_grmode;
fs/dlm/lock.c
5543
lkb->lkb_bastfn = (rl->rl_asts & DLM_CB_BAST) ? &fake_bastfn : NULL;
fs/dlm/lock.c
5544
lkb->lkb_astfn = (rl->rl_asts & DLM_CB_CAST) ? &fake_astfn : NULL;
fs/dlm/lock.c
5554
memcpy(lkb->lkb_lvbptr, rl->rl_lvb, lvblen);
fs/dlm/lock.c
5561
if (rl->rl_status == DLM_LKSTS_CONVERT && middle_conversion(lkb)) {
fs/dlm/lock.c
5582
struct rcom_lock *rl = (struct rcom_lock *) rc->rc_buf;
fs/dlm/lock.c
5590
*rl_remid = rl->rl_remid;
fs/dlm/lock.c
5592
if (rl->rl_parent_lkid) {
fs/dlm/lock.c
5597
remid = le32_to_cpu(rl->rl_lkid);
fs/dlm/lock.c
5607
error = find_rsb(ls, rl->rl_name, le16_to_cpu(rl->rl_namelen),
fs/dlm/lock.c
5638
add_lkb(r, lkb, rl->rl_status);
fs/dlm/lock.c
5666
struct rcom_lock *rl = (struct rcom_lock *) rc->rc_buf;
fs/dlm/lock.c
5672
lkid = le32_to_cpu(rl->rl_lkid);
fs/dlm/lock.c
5673
remid = le32_to_cpu(rl->rl_remid);
fs/dlm/lock.c
5674
result = le32_to_cpu(rl->rl_result);
fs/dlm/rcom.c
418
struct rcom_lock *rl)
fs/dlm/rcom.c
420
memset(rl, 0, sizeof(*rl));
fs/dlm/rcom.c
422
rl->rl_ownpid = cpu_to_le32(lkb->lkb_ownpid);
fs/dlm/rcom.c
423
rl->rl_lkid = cpu_to_le32(lkb->lkb_id);
fs/dlm/rcom.c
424
rl->rl_exflags = cpu_to_le32(lkb->lkb_exflags);
fs/dlm/rcom.c
425
rl->rl_flags = cpu_to_le32(dlm_dflags_val(lkb));
fs/dlm/rcom.c
426
rl->rl_lvbseq = cpu_to_le32(lkb->lkb_lvbseq);
fs/dlm/rcom.c
427
rl->rl_rqmode = lkb->lkb_rqmode;
fs/dlm/rcom.c
428
rl->rl_grmode = lkb->lkb_grmode;
fs/dlm/rcom.c
429
rl->rl_status = lkb->lkb_status;
fs/dlm/rcom.c
430
rl->rl_wait_type = cpu_to_le16(lkb->lkb_wait_type);
fs/dlm/rcom.c
433
rl->rl_asts |= DLM_CB_BAST;
fs/dlm/rcom.c
435
rl->rl_asts |= DLM_CB_CAST;
fs/dlm/rcom.c
437
rl->rl_namelen = cpu_to_le16(r->res_length);
fs/dlm/rcom.c
438
memcpy(rl->rl_name, r->res_name, r->res_length);
fs/dlm/rcom.c
444
memcpy(rl->rl_lvb, lkb->lkb_lvbptr, r->res_ls->ls_lvblen);
fs/dlm/rcom.c
452
struct rcom_lock *rl;
fs/dlm/rcom.c
463
rl = (struct rcom_lock *) rc->rc_buf;
fs/dlm/rcom.c
464
pack_rcom_lock(r, lkb, rl);
fs/dlm/rcom.c
477
struct rcom_lock *rl;
fs/dlm/rcom.c
490
rl = (struct rcom_lock *)rc->rc_buf;
fs/dlm/rcom.c
492
rl->rl_remid = rl_remid;
fs/dlm/rcom.c
493
rl->rl_result = rl_result;
fs/ocfs2/refcounttree.c
1243
struct ocfs2_refcount_list *rl = &rb->rf_records;
fs/ocfs2/refcounttree.c
1244
struct ocfs2_refcount_rec *rec = &rl->rl_recs[index];
fs/ocfs2/refcounttree.c
1259
if (index != le16_to_cpu(rl->rl_used) - 1) {
fs/ocfs2/refcounttree.c
1261
(le16_to_cpu(rl->rl_used) - index - 1) *
fs/ocfs2/refcounttree.c
1263
memset(&rl->rl_recs[le16_to_cpu(rl->rl_used) - 1],
fs/ocfs2/refcounttree.c
1267
le16_add_cpu(&rl->rl_used, -1);
fs/ocfs2/refcounttree.c
1406
static int ocfs2_find_refcount_split_pos(struct ocfs2_refcount_list *rl,
fs/ocfs2/refcounttree.c
1409
int num_used = le16_to_cpu(rl->rl_used);
fs/ocfs2/refcounttree.c
1415
&rl->rl_recs[middle - delta - 1],
fs/ocfs2/refcounttree.c
1416
&rl->rl_recs[middle - delta])) {
fs/ocfs2/refcounttree.c
1427
&rl->rl_recs[middle + delta],
fs/ocfs2/refcounttree.c
1428
&rl->rl_recs[middle + delta + 1])) {
fs/ocfs2/refcounttree.c
1437
*split_pos = ocfs2_get_ref_rec_low_cpos(&rl->rl_recs[*split_index]);
fs/ocfs2/refcounttree.c
1449
struct ocfs2_refcount_list *rl = &rb->rf_records;
fs/ocfs2/refcounttree.c
1456
le16_to_cpu(rl->rl_count), le16_to_cpu(rl->rl_used));
fs/ocfs2/refcounttree.c
1470
sort(&rl->rl_recs, le16_to_cpu(rl->rl_used),
fs/ocfs2/refcounttree.c
1474
ret = ocfs2_find_refcount_split_pos(rl, &cpos, &split_index);
fs/ocfs2/refcounttree.c
1483
num_moved = le16_to_cpu(rl->rl_used) - split_index;
fs/ocfs2/refcounttree.c
1484
memcpy(new_rl->rl_recs, &rl->rl_recs[split_index],
fs/ocfs2/refcounttree.c
1488
memset(&rl->rl_recs[split_index], 0,
fs/ocfs2/refcounttree.c
1492
le16_add_cpu(&rl->rl_used, -num_moved);
fs/ocfs2/refcounttree.c
1495
sort(&rl->rl_recs, le16_to_cpu(rl->rl_used),
fs/qnx4/inode.c
161
int rd, rl;
fs/qnx4/inode.c
168
rl = le32_to_cpu(s->RootDir.di_first_xtnt.xtnt_size);
fs/qnx4/inode.c
169
for (j = 0; j < rl; j++) {
include/linux/blk-mq.h
251
static inline int rq_list_empty(const struct rq_list *rl)
include/linux/blk-mq.h
253
return rl->head == NULL;
include/linux/blk-mq.h
256
static inline void rq_list_init(struct rq_list *rl)
include/linux/blk-mq.h
258
rl->head = NULL;
include/linux/blk-mq.h
259
rl->tail = NULL;
include/linux/blk-mq.h
265
if (rl->tail)
include/linux/blk-mq.h
266
rl->tail->rq_next = rq;
include/linux/blk-mq.h
268
rl->head = rq;
include/linux/blk-mq.h
269
rl->tail = rq;
include/linux/blk-mq.h
274
rq->rq_next = rl->head;
include/linux/blk-mq.h
275
rl->head = rq;
include/linux/blk-mq.h
276
if (!rl->tail)
include/linux/blk-mq.h
277
rl->tail = rq;
include/linux/blk-mq.h
280
static inline struct request *rq_list_pop(struct rq_list *rl)
include/linux/blk-mq.h
282
struct request *rq = rl->head;
include/linux/blk-mq.h
285
rl->head = rl->head->rq_next;
include/linux/blk-mq.h
286
if (!rl->head)
include/linux/blk-mq.h
287
rl->tail = NULL;
include/linux/blk-mq.h
294
static inline struct request *rq_list_peek(struct rq_list *rl)
include/linux/blk-mq.h
296
return rl->head;
include/linux/blk-mq.h
300
for (pos = rq_list_peek((rl)); (pos); pos = pos->rq_next)
include/linux/blk-mq.h
302
#define rq_list_for_each_safe(rl, pos, nxt) \
include/linux/blk-mq.h
303
for (pos = rq_list_peek((rl)), nxt = pos->rq_next; \
include/linux/jump_label_ratelimit.h
41
jump_label_rate_limit(struct static_key_deferred *key, unsigned long rl);
include/linux/jump_label_ratelimit.h
45
#define DEFINE_STATIC_KEY_DEFERRED_TRUE(name, rl) \
include/linux/jump_label_ratelimit.h
48
.timeout = (rl), \
include/linux/jump_label_ratelimit.h
54
#define DEFINE_STATIC_KEY_DEFERRED_FALSE(name, rl) \
include/linux/jump_label_ratelimit.h
57
.timeout = (rl), \
include/linux/jump_label_ratelimit.h
73
#define DEFINE_STATIC_KEY_DEFERRED_TRUE(name, rl) \
include/linux/jump_label_ratelimit.h
75
#define DEFINE_STATIC_KEY_DEFERRED_FALSE(name, rl) \
include/linux/jump_label_ratelimit.h
91
unsigned long rl)
include/linux/math64.h
215
} rl, rm, rn, rh, a0, b0;
include/linux/math64.h
221
rl.ll = mul_u32_u32(a0.l.low, b0.l.low);
include/linux/math64.h
231
rl.l.high = c = (u64)rl.l.high + rm.l.low + rn.l.low;
include/linux/math64.h
240
return rl.ll;
include/linux/math64.h
242
return (rl.ll >> shift) | (rh.ll << (64 - shift));
include/linux/math64.h
279
} u, rl, rh;
include/linux/math64.h
282
rl.ll = mul_u32_u32(u.l.low, mul);
include/linux/math64.h
283
rh.ll = mul_u32_u32(u.l.high, mul) + rl.l.high;
include/linux/math64.h
286
rl.l.high = do_div(rh.ll, divisor);
include/linux/math64.h
289
do_div(rl.ll, divisor);
include/linux/math64.h
291
rl.l.high = rh.l.low;
include/linux/math64.h
292
return rl.ll;
include/linux/mlx5/driver.h
1091
struct mlx5_rate_limit *rl);
include/linux/mlx5/driver.h
1092
void mlx5_rl_remove_rate(struct mlx5_core_dev *dev, struct mlx5_rate_limit *rl);
include/math-emu/op-2.h
147
#define __FP_FRAC_ADD_2(rh, rl, xh, xl, yh, yl) \
include/math-emu/op-2.h
148
(rh = xh + yh + ((rl = xl + yl) < xl))
include/math-emu/op-2.h
151
#define __FP_FRAC_SUB_2(rh, rl, xh, xl, yh, yl) \
include/math-emu/op-2.h
152
(rh = xh - yh - ((rl = xl - yl) > xl))
io_uring/io_uring.c
2074
unsigned int cq_entries, struct io_rings_layout *rl)
io_uring/io_uring.c
2089
rl->sq_array_offset = SIZE_MAX;
io_uring/io_uring.c
2095
rl->sq_size = array_size(sqe_size, sq_entries);
io_uring/io_uring.c
2096
if (rl->sq_size == SIZE_MAX)
io_uring/io_uring.c
2114
rl->sq_array_offset = off;
io_uring/io_uring.c
2122
rl->rings_size = off;
io_uring/io_uring.c
2691
struct io_rings_layout *rl = &config->layout;
io_uring/io_uring.c
2701
rd.size = PAGE_ALIGN(rl->rings_size);
io_uring/io_uring.c
2712
ctx->sq_array = (u32 *)((char *)rings + rl->sq_array_offset);
io_uring/io_uring.c
2715
rd.size = PAGE_ALIGN(rl->sq_size);
io_uring/register.c
505
struct io_rings_layout *rl = &config.layout;
io_uring/register.c
526
rd.size = PAGE_ALIGN(rl->rings_size);
io_uring/register.c
556
rd.size = PAGE_ALIGN(rl->sq_size);
io_uring/register.c
639
ctx->sq_array = (u32 *)((char *)n.rings + rl->sq_array_offset);
kernel/jump_label.c
367
unsigned long rl)
kernel/jump_label.c
370
key->timeout = rl;
sound/soc/codecs/rt1305.c
1058
regmap_read(rt1305->regmap, RT1305_PR_BASE + 0x56, &rl);
sound/soc/codecs/rt1305.c
1059
rhl = (rh << 16) | rl;
sound/soc/codecs/rt1305.c
1062
pr_debug("Left_rhl = 0x%x rh=0x%x rl=0x%x\n", rhl, rh, rl);
sound/soc/codecs/rt1305.c
1075
regmap_read(rt1305->regmap, RT1305_PR_BASE + 0x56, &rl);
sound/soc/codecs/rt1305.c
1076
rhl = (rh << 16) | rl;
sound/soc/codecs/rt1305.c
1079
pr_debug("Right_rhl = 0x%x rh=0x%x rl=0x%x\n", rhl, rh, rl);
sound/soc/codecs/rt1305.c
993
unsigned int rh, rl, rhl, r0ohm;
sound/usb/usx2y/usbus428ctldefs.h
68
rl;
tools/perf/bench/epoll-ctl.c
319
struct rlimit rl, prevrl;
tools/perf/bench/epoll-ctl.c
358
rl.rlim_cur = rl.rlim_max = nfds * nthreads * 2 + 50;
tools/perf/bench/epoll-ctl.c
360
(uint64_t)prevrl.rlim_max, (uint64_t)rl.rlim_max);
tools/perf/bench/epoll-ctl.c
361
if (setrlimit(RLIMIT_NOFILE, &rl) < 0)
tools/perf/bench/epoll-wait.c
439
struct rlimit rl, prevrl;
tools/perf/bench/epoll-wait.c
483
rl.rlim_cur = rl.rlim_max = nfds * nthreads * 2 + 50;
tools/perf/bench/epoll-wait.c
485
(uint64_t)prevrl.rlim_max, (uint64_t)rl.rlim_max);
tools/perf/bench/epoll-wait.c
486
if (setrlimit(RLIMIT_NOFILE, &rl) < 0)
tools/perf/util/print-events.c
212
static struct rb_node *mep_new(struct rblist *rl __maybe_unused, const void *entry)
tools/perf/util/print-events.c
223
static void mep_delete(struct rblist *rl __maybe_unused,
tools/testing/selftests/kvm/lib/kvm_util.c
446
struct rlimit rl;
tools/testing/selftests/kvm/lib/kvm_util.c
452
TEST_ASSERT(!getrlimit(RLIMIT_NOFILE, &rl), "getrlimit() failed!");
tools/testing/selftests/kvm/lib/kvm_util.c
454
if (rl.rlim_cur < nr_fds_wanted) {
tools/testing/selftests/kvm/lib/kvm_util.c
455
rl.rlim_cur = nr_fds_wanted;
tools/testing/selftests/kvm/lib/kvm_util.c
456
if (rl.rlim_max < nr_fds_wanted) {
tools/testing/selftests/kvm/lib/kvm_util.c
457
int old_rlim_max = rl.rlim_max;
tools/testing/selftests/kvm/lib/kvm_util.c
459
rl.rlim_max = nr_fds_wanted;
tools/testing/selftests/kvm/lib/kvm_util.c
460
__TEST_REQUIRE(setrlimit(RLIMIT_NOFILE, &rl) >= 0,
tools/testing/selftests/kvm/lib/kvm_util.c
464
TEST_ASSERT(!setrlimit(RLIMIT_NOFILE, &rl), "setrlimit() failed!");