Symbol: inc
arch/arm/include/asm/assembler.h
470
.macro usraccoff, instr, reg, ptr, inc, off, cond, abort, t=TUSER()
arch/arm/include/asm/assembler.h
486
.macro usracc, instr, reg, ptr, inc, cond, rept, abort
arch/arm/include/asm/assembler.h
510
.macro usracc, instr, reg, ptr, inc, cond, rept, abort, t=TUSER()
arch/arm/include/asm/assembler.h
530
.macro strusr, reg, ptr, inc, cond=al, rept=1, abort=9001f
arch/arm/include/asm/assembler.h
534
.macro ldrusr, reg, ptr, inc, cond=al, rept=1, abort=9001f
arch/mips/cavium-octeon/csrc-octeon.c
146
u64 cur, end, inc;
arch/mips/cavium-octeon/csrc-octeon.c
150
inc = us * octeon_udelay_factor;
arch/mips/cavium-octeon/csrc-octeon.c
151
end = cur + inc;
arch/mips/cavium-octeon/csrc-octeon.c
160
u64 cur, end, inc;
arch/mips/cavium-octeon/csrc-octeon.c
164
inc = ((ns * octeon_ndelay_factor) >> 16);
arch/mips/cavium-octeon/csrc-octeon.c
165
end = cur + inc;
arch/mips/cavium-octeon/setup.c
931
u64 inc = addr - *mem;
arch/mips/cavium-octeon/setup.c
932
memblock_add(*mem, inc);
arch/mips/cavium-octeon/setup.c
933
*mem += inc;
arch/mips/cavium-octeon/setup.c
934
*size -= inc;
arch/mips/include/asm/octeon/cvmx-mio-defs.h
810
uint64_t inc:1;
arch/mips/include/asm/octeon/cvmx-mio-defs.h
818
uint64_t inc:1;
arch/mips/include/asm/octeon/cvmx-pko-defs.h
2169
uint64_t inc:8;
arch/mips/include/asm/octeon/cvmx-pko-defs.h
2173
uint64_t inc:8;
arch/mips/include/asm/octeon/cvmx-spxx-defs.h
151
uint64_t inc:1;
arch/mips/include/asm/octeon/cvmx-spxx-defs.h
165
uint64_t inc:1;
arch/mips/kernel/signal.c
100
for (i = 0; i < NUM_FPU_REGS; i += inc) {
arch/mips/kernel/signal.c
78
int inc = test_thread_flag(TIF_32BIT_FPREGS) ? 2 : 1;
arch/mips/kernel/signal.c
80
for (i = 0; i < NUM_FPU_REGS; i += inc) {
arch/mips/kernel/signal.c
97
int inc = test_thread_flag(TIF_32BIT_FPREGS) ? 2 : 1;
arch/powerpc/include/asm/kvm_host.h
504
u8 inc;
arch/powerpc/platforms/powernv/pci-ioda.c
1189
unsigned long start, end, inc;
arch/powerpc/platforms/powernv/pci-ioda.c
1199
inc = (0x1ull << shift);
arch/powerpc/platforms/powernv/pci-ioda.c
1204
start += inc;
arch/powerpc/xmon/xmon.c
2394
int cmd, inc, i, nslash;
arch/powerpc/xmon/xmon.c
2436
inc = size;
arch/powerpc/xmon/xmon.c
2445
inc = size;
arch/powerpc/xmon/xmon.c
2450
inc = 0;
arch/powerpc/xmon/xmon.c
2467
inc = size;
arch/powerpc/xmon/xmon.c
2537
adrs += inc;
arch/s390/mm/cmm.c
129
long inc = 0;
arch/s390/mm/cmm.c
132
inc = min(256L, nr);
arch/s390/mm/cmm.c
133
nr -= inc;
arch/s390/mm/cmm.c
134
inc = __cmm_free_pages(inc, counter, list);
arch/s390/mm/cmm.c
135
if (inc)
arch/s390/mm/cmm.c
139
return nr + inc;
arch/x86/include/asm/asm.h
40
#define _ASM_INC __ASM_SIZE(inc)
arch/x86/include/asm/atomic64_32.h
196
__alternative_atomic64(inc, inc_return,
arch/x86/include/asm/cmpxchg.h
241
#define __xadd(ptr, inc, lock) __xchg_op((ptr), (inc), xadd, lock)
arch/x86/include/asm/cmpxchg.h
242
#define xadd(ptr, inc) __xadd((ptr), (inc), LOCK_PREFIX)
arch/x86/include/asm/xor.h
105
: [inc] XOR_CONSTANT_CONSTRAINT (256UL)
arch/x86/include/asm/xor.h
140
: [inc] XOR_CONSTANT_CONSTRAINT (256UL)
arch/x86/include/asm/xor.h
200
: [inc] XOR_CONSTANT_CONSTRAINT (256UL)
arch/x86/include/asm/xor.h
238
: [inc] XOR_CONSTANT_CONSTRAINT (256UL)
arch/x86/include/asm/xor.h
306
: [inc] XOR_CONSTANT_CONSTRAINT (256UL)
arch/x86/include/asm/xor.h
347
: [inc] XOR_CONSTANT_CONSTRAINT (256UL)
arch/x86/include/asm/xor.h
423
: [inc] XOR_CONSTANT_CONSTRAINT (256UL)
arch/x86/include/asm/xor.h
467
: [inc] XOR_CONSTANT_CONSTRAINT (256UL)
arch/x86/kvm/emulate.c
498
static void masked_increment(ulong *reg, ulong mask, int inc)
arch/x86/kvm/emulate.c
500
assign_masked(reg, *reg + inc, mask);
arch/x86/kvm/emulate.c
504
register_address_increment(struct x86_emulate_ctxt *ctxt, int reg, int inc)
arch/x86/kvm/emulate.c
508
assign_register(preg, *preg + inc, ctxt->ad_bytes);
arch/x86/kvm/emulate.c
511
static void rsp_increment(struct x86_emulate_ctxt *ctxt, int inc)
arch/x86/kvm/emulate.c
513
masked_increment(reg_rmw(ctxt, VCPU_REGS_RSP), stack_mask(ctxt), inc);
arch/x86/kvm/emulate.c
978
EM_ASM_1(inc);
arch/xtensa/kernel/signal.c
104
inc = 3;
arch/xtensa/kernel/signal.c
109
sp = regs->areg[((base + inc) * 4 + 1) % XCHAL_NUM_AREGS];
arch/xtensa/kernel/signal.c
116
base += inc;
arch/xtensa/kernel/signal.c
87
int inc = 0;
arch/xtensa/kernel/signal.c
92
inc = 1;
arch/xtensa/kernel/signal.c
98
inc = 2;
block/blk-crypto-internal.h
30
unsigned int inc);
block/blk-crypto.c
137
unsigned int inc)
block/blk-crypto.c
141
for (i = 0; inc && i < BLK_CRYPTO_DUN_ARRAY_SIZE; i++) {
block/blk-crypto.c
142
dun[i] += inc;
block/blk-crypto.c
147
if (dun[i] < inc)
block/blk-crypto.c
148
inc = 1;
block/blk-crypto.c
150
inc = 0;
drivers/acpi/device_pm.c
876
goto inc;
drivers/acpi/device_pm.c
888
inc:
drivers/atm/iphase.c
466
int idealSlot =0, testSlot, toBeAssigned, inc;
drivers/atm/iphase.c
530
inc = 0;
drivers/atm/iphase.c
538
inc++;
drivers/atm/iphase.c
539
testSlot = idealSlot - inc;
drivers/atm/iphase.c
549
testSlot = idealSlot + inc;
drivers/atm/iphase.c
559
TstSchedTbl,cbrVC,inc);)
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
74
static inline u32 modulo_inc(u32 index, u32 length, u32 inc)
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
76
if (WARN_ON(inc > length))
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
77
inc = length;
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
79
index += inc;
drivers/crypto/marvell/octeontx2/otx2_cptvf_reqmgr.c
62
static inline u32 modulo_inc(u32 index, u32 length, u32 inc)
drivers/crypto/marvell/octeontx2/otx2_cptvf_reqmgr.c
64
if (WARN_ON(inc > length))
drivers/crypto/marvell/octeontx2/otx2_cptvf_reqmgr.c
65
inc = length;
drivers/crypto/marvell/octeontx2/otx2_cptvf_reqmgr.c
67
index += inc;
drivers/dma-buf/sw_sync.c
209
static void sync_timeline_signal(struct sync_timeline *obj, unsigned int inc)
drivers/dma-buf/sw_sync.c
218
obj->value += inc;
drivers/dma/bestcomm/ata.c
101
struct bcom_ata_inc *inc;
drivers/dma/bestcomm/ata.c
103
inc = (struct bcom_ata_inc *) bcom_task_inc(tsk->tasknum);
drivers/dma/bestcomm/ata.c
105
inc->incr_bytes = -(s16)sizeof(u32);
drivers/dma/bestcomm/ata.c
106
inc->incr_src = 0;
drivers/dma/bestcomm/ata.c
107
inc->incr_dst = sizeof(u32);
drivers/dma/bestcomm/ata.c
115
struct bcom_ata_inc *inc;
drivers/dma/bestcomm/ata.c
117
inc = (struct bcom_ata_inc *) bcom_task_inc(tsk->tasknum);
drivers/dma/bestcomm/ata.c
119
inc->incr_bytes = -(s16)sizeof(u32);
drivers/dma/bestcomm/ata.c
120
inc->incr_src = sizeof(u32);
drivers/dma/bestcomm/ata.c
121
inc->incr_dst = 0;
drivers/dma/bestcomm/ata.c
58
struct bcom_ata_inc *inc;
drivers/dma/bestcomm/ata.c
72
inc = (struct bcom_ata_inc *) bcom_task_inc(tsk->tasknum);
drivers/dma/bestcomm/bestcomm.c
141
u32 *desc, *var, *inc;
drivers/dma/bestcomm/bestcomm.c
183
inc = bcom_task_inc(task);
drivers/dma/bestcomm/bestcomm.c
187
memset_io(inc, 0x00, BCOM_INC_SIZE);
drivers/dma/bestcomm/bestcomm.c
195
memcpy_toio(inc, inc_src, hdr->inc_size * sizeof(u32));
drivers/dma/bestcomm/fec.c
111
struct bcom_fec_rx_inc *inc;
drivers/dma/bestcomm/fec.c
118
inc = (struct bcom_fec_rx_inc *) bcom_task_inc(tsk->tasknum);
drivers/dma/bestcomm/fec.c
131
inc->incr_bytes = -(s16)sizeof(u32); /* These should be in the */
drivers/dma/bestcomm/fec.c
132
inc->incr_dst = sizeof(u32); /* task image, but we stick */
drivers/dma/bestcomm/fec.c
133
inc->incr_dst_ma= sizeof(u8); /* to the official ones */
drivers/dma/bestcomm/fec.c
212
struct bcom_fec_tx_inc *inc;
drivers/dma/bestcomm/fec.c
219
inc = (struct bcom_fec_tx_inc *) bcom_task_inc(tsk->tasknum);
drivers/dma/bestcomm/fec.c
232
inc->incr_bytes = -(s16)sizeof(u32); /* These should be in the */
drivers/dma/bestcomm/fec.c
233
inc->incr_src = sizeof(u32); /* task image, but we stick */
drivers/dma/bestcomm/fec.c
234
inc->incr_src_ma= sizeof(u8); /* to the official ones */
drivers/dma/bestcomm/gen_bd.c
118
struct bcom_gen_bd_rx_inc *inc;
drivers/dma/bestcomm/gen_bd.c
125
inc = (struct bcom_gen_bd_rx_inc *) bcom_task_inc(tsk->tasknum);
drivers/dma/bestcomm/gen_bd.c
138
inc->incr_bytes = -(s16)sizeof(u32);
drivers/dma/bestcomm/gen_bd.c
139
inc->incr_dst = sizeof(u32);
drivers/dma/bestcomm/gen_bd.c
202
struct bcom_gen_bd_tx_inc *inc;
drivers/dma/bestcomm/gen_bd.c
209
inc = (struct bcom_gen_bd_tx_inc *) bcom_task_inc(tsk->tasknum);
drivers/dma/bestcomm/gen_bd.c
221
inc->incr_bytes = -(s16)sizeof(u32);
drivers/dma/bestcomm/gen_bd.c
222
inc->incr_src = sizeof(u32);
drivers/dma/bestcomm/gen_bd.c
223
inc->incr_src_ma = sizeof(u8);
drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager.c
111
int inc = cu_inc * NUM_XCC(mm->dev->xcc_mask);
drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager.c
201
i += inc;
drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_v12_1.c
100
i += inc;
drivers/gpu/drm/amd/amdkfd/kfd_mqd_manager_v12_1.c
55
int inc = NUM_XCC(mm->dev->xcc_mask);
drivers/gpu/drm/drm_damage_helper.c
136
uint32_t inc = 1;
drivers/gpu/drm/drm_damage_helper.c
139
inc = 2;
drivers/gpu/drm/drm_damage_helper.c
149
convert_clip_rect_to_rect(clips, rects, num_clips, inc);
drivers/gpu/drm/i915/display/intel_gmbus.c
664
int i = 0, inc, try = 0;
drivers/gpu/drm/i915/display/intel_gmbus.c
676
for (; i < num; i += inc) {
drivers/gpu/drm/i915/display/intel_gmbus.c
677
inc = 1;
drivers/gpu/drm/i915/display/intel_gmbus.c
681
inc = 2; /* an index transmission is two msgs */
drivers/gpu/drm/i915/display/skl_watermark.c
3155
static void increase_wm_latency(struct intel_display *display, int inc)
drivers/gpu/drm/i915/display/skl_watermark.c
3160
wm[0] += inc;
drivers/gpu/drm/i915/display/skl_watermark.c
3166
wm[level] += inc;
drivers/gpu/drm/nouveau/nouveau_exec.c
295
u32 inc = req->wait_count;
drivers/gpu/drm/nouveau/nouveau_exec.c
310
if (inc) {
drivers/gpu/drm/nouveau/nouveau_exec.c
313
args->in_sync.count = inc;
drivers/gpu/drm/nouveau/nouveau_exec.c
314
*s = u_memcpya(ins, inc, sizeof(**s));
drivers/gpu/drm/nouveau/nouveau_sched.h
89
u32 inc, u64 ins,
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1740
u32 inc = req->wait_count;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1758
if (inc) {
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1761
args->in_sync.count = inc;
drivers/gpu/drm/nouveau/nouveau_uvmm.c
1762
*s = u_memcpya(ins, inc, sizeof(**s));
drivers/gpu/drm/omapdrm/dss/dispc.c
1018
enum omap_plane_id plane, s32 inc)
drivers/gpu/drm/omapdrm/dss/dispc.c
1020
dispc_write_reg(dispc, DISPC_OVL_PIXEL_INC(plane), inc);
drivers/gpu/drm/omapdrm/dss/dispc.c
1024
enum omap_plane_id plane, s32 inc)
drivers/gpu/drm/omapdrm/dss/dispc.c
1026
dispc_write_reg(dispc, DISPC_OVL_ROW_INC(plane), inc);
drivers/gpu/drm/omapdrm/dss/dispc.h
106
const struct dispc_coef *dispc_ovl_get_scale_coef(int inc, int five_taps);
drivers/gpu/drm/omapdrm/dss/dispc_coefs.c
276
const struct dispc_coef *dispc_ovl_get_scale_coef(int inc, int five_taps)
drivers/gpu/drm/omapdrm/dss/dispc_coefs.c
307
inc /= 128;
drivers/gpu/drm/omapdrm/dss/dispc_coefs.c
309
if (inc >= coefs[i].Mmin && inc <= coefs[i].Mmax)
drivers/gpu/drm/omapdrm/omap_drv.c
144
int c, i, n, inc;
drivers/gpu/drm/omapdrm/omap_drv.c
179
for (i = 0, inc = 0; i < n; i++) {
drivers/gpu/drm/omapdrm/omap_drv.c
182
states[i]->normalized_zpos = i + inc;
drivers/gpu/drm/omapdrm/omap_drv.c
188
inc++;
drivers/gpu/drm/qxl/qxl_display.c
439
int inc = 1, ret;
drivers/gpu/drm/qxl/qxl_display.c
457
inc = 2; /* skip source rects */
drivers/gpu/drm/qxl/qxl_display.c
461
clips, num_clips, inc, 0);
drivers/gpu/drm/qxl/qxl_draw.c
131
unsigned int num_clips, int inc,
drivers/gpu/drm/qxl/qxl_draw.c
170
for (i = 1, clips_ptr = clips + inc;
drivers/gpu/drm/qxl/qxl_draw.c
171
i < num_clips; i++, clips_ptr += inc) {
drivers/gpu/drm/qxl/qxl_draw.c
245
for (i = 0; i < num_clips; i++, clips_ptr += inc) {
drivers/gpu/drm/qxl/qxl_drv.h
396
unsigned int num_clips, int inc,
drivers/gpu/drm/radeon/radeon_legacy_crtc.c
135
inc = (fp_horz_stretch & RADEON_HORZ_AUTO_RATIO_INC) ? 1 : 0;
drivers/gpu/drm/radeon/radeon_legacy_crtc.c
136
scale = ((xres + inc) * RADEON_HORZ_STRETCH_RATIO_MAX)
drivers/gpu/drm/radeon/radeon_legacy_crtc.c
147
inc = (fp_vert_stretch & RADEON_VERT_AUTO_RATIO_INC) ? 1 : 0;
drivers/gpu/drm/radeon/radeon_legacy_crtc.c
148
scale = ((yres + inc) * RADEON_VERT_STRETCH_RATIO_MAX)
drivers/gpu/drm/radeon/radeon_legacy_crtc.c
62
u32 scale, inc, crtc_more_cntl;
drivers/gpu/drm/tidss/tidss_scale_coefs.c
153
int inc;
drivers/gpu/drm/tidss/tidss_scale_coefs.c
188
inc = firinc / 0x40000;
drivers/gpu/drm/tidss/tidss_scale_coefs.c
190
if (inc >= coefs[i].mmin && inc <= coefs[i].mmax) {
drivers/gpu/drm/tidss/tidss_scale_coefs.c
199
__func__, firinc, inc);
drivers/gpu/drm/vmwgfx/vmwgfx_kms.h
458
unsigned num_clips, int inc,
drivers/gpu/drm/vmwgfx/vmwgfx_kms.h
488
unsigned num_clips, int inc,
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1100
unsigned num_clips, int inc,
drivers/gpu/drm/vmwgfx/vmwgfx_scrn.c
1136
dest_x, dest_y, num_clips, inc,
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
729
unsigned num_clips, int inc,
drivers/gpu/drm/vmwgfx/vmwgfx_stdu.c
762
dest_x, dest_y, num_clips, inc,
drivers/i2c/busses/i2c-img-scb.c
1152
unsigned int i, data, prescale, inc, int_bitrate, filt;
drivers/i2c/busses/i2c-img-scb.c
1200
inc = (256 * 16 * bitrate_khz) / clk_khz;
drivers/i2c/busses/i2c-img-scb.c
1220
inc++;
drivers/i2c/busses/i2c-img-scb.c
1227
data = filt | ((inc & SCB_INC_MASK) << SCB_INC_SHIFT) | (prescale - 1);
drivers/i2c/busses/i2c-img-scb.c
1231
clk_period = (256 * 1000000) / (clk_khz * inc);
drivers/i2c/busses/i2c-octeon-core.c
826
int tclk, thp_base, inc, thp_idx, mdiv_idx, ndiv_idx, foscl, diff;
drivers/i2c/busses/i2c-octeon-core.c
865
for (inc = 0; inc <= 1; inc++) {
drivers/i2c/busses/i2c-octeon-core.c
866
thp_idx = thp_base + inc;
drivers/iio/adc/pac1934.c
596
s64 curr_energy, inc;
drivers/iio/adc/pac1934.c
663
inc = (reg_data->vpower_acc[cnt] >> samp_shift);
drivers/iio/adc/pac1934.c
666
curr_energy += inc;
drivers/iio/light/bh1750.c
201
for (i = chip_info->mtreg_min; i <= chip_info->mtreg_max; i += chip_info->inc)
drivers/iio/light/bh1750.c
63
int inc;
drivers/infiniband/core/multicast.c
228
static void adjust_membership(struct mcast_group *group, u8 join_state, int inc)
drivers/infiniband/core/multicast.c
234
group->members[i] += inc;
drivers/infiniband/hw/cxgb4/qp.c
909
static int ring_kernel_sq_db(struct c4iw_qp *qhp, u16 inc)
drivers/infiniband/hw/cxgb4/qp.c
916
t4_ring_sq_db(&qhp->wq, inc, NULL);
drivers/infiniband/hw/cxgb4/qp.c
919
qhp->wq.sq.wq_pidx_inc += inc;
drivers/infiniband/hw/cxgb4/qp.c
926
static int ring_kernel_rq_db(struct c4iw_qp *qhp, u16 inc)
drivers/infiniband/hw/cxgb4/qp.c
933
t4_ring_rq_db(&qhp->wq, inc, NULL);
drivers/infiniband/hw/cxgb4/qp.c
936
qhp->wq.rq.wq_pidx_inc += inc;
drivers/infiniband/hw/cxgb4/t4.h
582
static inline void t4_ring_srq_db(struct t4_srq *srq, u16 inc, u8 len16,
drivers/infiniband/hw/cxgb4/t4.h
587
if (inc == 1 && srq->bar2_qid == 0 && wqe) {
drivers/infiniband/hw/cxgb4/t4.h
594
writel(PIDX_T5_V(inc) | QID_V(srq->bar2_qid),
drivers/infiniband/hw/cxgb4/t4.h
601
static inline void t4_ring_sq_db(struct t4_wq *wq, u16 inc, union t4_wr *wqe)
drivers/infiniband/hw/cxgb4/t4.h
607
if (inc == 1 && wq->sq.bar2_qid == 0 && wqe) {
drivers/infiniband/hw/cxgb4/t4.h
614
writel(PIDX_T5_V(inc) | QID_V(wq->sq.bar2_qid),
drivers/infiniband/hw/cxgb4/t4.h
622
writel(QID_V(wq->sq.qid) | PIDX_V(inc), wq->db);
drivers/infiniband/hw/cxgb4/t4.h
625
static inline void t4_ring_rq_db(struct t4_wq *wq, u16 inc,
drivers/infiniband/hw/cxgb4/t4.h
632
if (inc == 1 && wq->rq.bar2_qid == 0 && wqe) {
drivers/infiniband/hw/cxgb4/t4.h
639
writel(PIDX_T5_V(inc) | QID_V(wq->rq.bar2_qid),
drivers/infiniband/hw/cxgb4/t4.h
647
writel(QID_V(wq->rq.qid) | PIDX_V(inc), wq->db);
drivers/infiniband/hw/mlx4/mcg.c
474
static void adjust_membership(struct mcast_group *group, u8 join_state, int inc)
drivers/infiniband/hw/mlx4/mcg.c
480
group->members[i] += inc;
drivers/infiniband/hw/mthca/mthca_memfree.c
324
int inc = MTHCA_TABLE_CHUNK_SIZE / table->obj_size;
drivers/infiniband/hw/mthca/mthca_memfree.c
327
for (i = start; i <= end; i += inc) {
drivers/infiniband/hw/mthca/mthca_memfree.c
337
i -= inc;
drivers/iommu/iommu-debug-pagealloc.c
112
unsigned long iova, size_t size, bool inc)
drivers/iommu/iommu-debug-pagealloc.c
126
if (inc)
drivers/iommu/iommufd/io_pagetable.h
310
bool inc, struct pfn_reader_user *user);
drivers/iommu/iommufd/pages.c
1002
rc = __account_locked_vm(pages->source_mm, npages, inc,
drivers/iommu/iommufd/pages.c
1012
bool inc, struct pfn_reader_user *user)
drivers/iommu/iommufd/pages.c
1020
if (inc)
drivers/iommu/iommufd/pages.c
1026
rc = update_mm_locked_vm(pages, npages, inc, user);
drivers/iommu/iommufd/pages.c
1033
if (inc)
drivers/iommu/iommufd/pages.c
1061
bool inc;
drivers/iommu/iommufd/pages.c
1070
inc = false;
drivers/iommu/iommufd/pages.c
1075
inc = true;
drivers/iommu/iommufd/pages.c
1077
return iopt_pages_update_pinned(pages, npages, inc, user);
drivers/iommu/iommufd/pages.c
984
bool inc, struct pfn_reader_user *user)
drivers/md/dm-cache-metadata.c
297
vt.inc = NULL;
drivers/md/dm-era-target.c
1034
int r, inc;
drivers/md/dm-era-target.c
1061
&sb_validator, &clone, &inc);
drivers/md/dm-era-target.c
1067
BUG_ON(!inc);
drivers/md/dm-era-target.c
419
vt->inc = ws_inc;
drivers/md/dm-era-target.c
430
vt.inc = NULL;
drivers/md/dm-thin-metadata.c
1334
int r, inc;
drivers/md/dm-thin-metadata.c
1355
&sb_validator, &copy, &inc);
drivers/md/dm-thin-metadata.c
1359
BUG_ON(!inc);
drivers/md/dm-thin-metadata.c
489
pmd->info.value_type.inc = data_block_inc;
drivers/md/dm-thin-metadata.c
500
pmd->tl_info.value_type.inc = subtree_inc;
drivers/md/dm-thin-metadata.c
508
pmd->bl_info.value_type.inc = data_block_inc;
drivers/md/dm-thin-metadata.c
516
pmd->details_info.value_type.inc = NULL;
drivers/md/md-bitmap.c
1462
sector_t offset, int inc)
drivers/md/md-bitmap.c
1466
bitmap->bp[page].count += inc;
drivers/md/persistent-data/dm-array.c
126
if (vt->inc)
drivers/md/persistent-data/dm-array.c
127
on_entries(info, ab, vt->inc);
drivers/md/persistent-data/dm-array.c
186
if (vt->inc)
drivers/md/persistent-data/dm-array.c
187
vt->inc(vt->context, value, delta);
drivers/md/persistent-data/dm-array.c
283
int inc;
drivers/md/persistent-data/dm-array.c
285
&array_validator, block, &inc);
drivers/md/persistent-data/dm-array.c
290
if (inc)
drivers/md/persistent-data/dm-array.c
651
bvt->inc = block_inc;
drivers/md/persistent-data/dm-array.c
723
if (vt->inc)
drivers/md/persistent-data/dm-array.c
724
vt->inc(vt->context, element_at(info, ab, i), 1);
drivers/md/persistent-data/dm-array.c
836
if (vt->inc)
drivers/md/persistent-data/dm-array.c
837
vt->inc(vt->context, value, 1);
drivers/md/persistent-data/dm-bitset.c
22
.inc = NULL,
drivers/md/persistent-data/dm-btree-remove.c
165
int r, inc;
drivers/md/persistent-data/dm-btree-remove.c
172
&result->block, &inc);
drivers/md/persistent-data/dm-btree-remove.c
178
if (inc)
drivers/md/persistent-data/dm-btree-spine.c
102
int r, inc;
drivers/md/persistent-data/dm-btree-spine.c
105
result, &inc);
drivers/md/persistent-data/dm-btree-spine.c
106
if (!r && inc)
drivers/md/persistent-data/dm-btree-spine.c
260
vt->inc = le64_inc;
drivers/md/persistent-data/dm-btree.c
711
int r, inc;
drivers/md/persistent-data/dm-btree.c
718
result, &inc);
drivers/md/persistent-data/dm-btree.c
724
if (inc)
drivers/md/persistent-data/dm-btree.c
82
else if (vt->inc)
drivers/md/persistent-data/dm-btree.c
83
vt->inc(vt->context, value_ptr(n, 0), nr_entries);
drivers/md/persistent-data/dm-btree.h
62
void (*inc)(void *context, const void *value, unsigned int count);
drivers/md/persistent-data/dm-space-map-common.c
1013
int r, inc;
drivers/md/persistent-data/dm-space-map-common.c
1016
r = dm_tm_shadow_block(ll->tm, ll->bitmap_root, &index_validator, &b, &inc);
drivers/md/persistent-data/dm-space-map-common.c
209
ll->bitmap_info.value_type.inc = NULL;
drivers/md/persistent-data/dm-space-map-common.c
216
ll->ref_count_info.value_type.inc = NULL;
drivers/md/persistent-data/dm-space-map-common.c
431
int inc;
drivers/md/persistent-data/dm-space-map-common.c
439
&dm_sm_bitmap_validator, &nb, &inc);
drivers/md/persistent-data/dm-space-map-common.c
611
int r, inc;
drivers/md/persistent-data/dm-space-map-common.c
614
&dm_sm_bitmap_validator, &ic->bitmap_block, &inc);
drivers/md/raid5-cache.c
256
static sector_t r5l_ring_add(struct r5l_log *log, sector_t start, sector_t inc)
drivers/md/raid5-cache.c
258
start += inc;
drivers/media/dvb-frontends/stv090x.c
2248
s32 timeout, inc, steps_max, srate, car_max;
drivers/media/dvb-frontends/stv090x.c
2259
inc = srate;
drivers/media/dvb-frontends/stv090x.c
2260
inc /= state->internal->mclk / 1000;
drivers/media/dvb-frontends/stv090x.c
2261
inc *= 256;
drivers/media/dvb-frontends/stv090x.c
2262
inc *= 256;
drivers/media/dvb-frontends/stv090x.c
2263
inc /= 1000;
drivers/media/dvb-frontends/stv090x.c
2268
inc *= 3; /* freq step = 3% of srate */
drivers/media/dvb-frontends/stv090x.c
2273
inc *= 4;
drivers/media/dvb-frontends/stv090x.c
2279
inc *= 3;
drivers/media/dvb-frontends/stv090x.c
2283
inc /= 100;
drivers/media/dvb-frontends/stv090x.c
2284
if ((inc > car_max) || (inc < 0))
drivers/media/dvb-frontends/stv090x.c
2285
inc = car_max / 2; /* increment <= 1/8 Mclk */
drivers/media/dvb-frontends/stv090x.c
2294
steps_max = (car_max / inc) + 1; /* min steps = 3 */
drivers/media/dvb-frontends/stv090x.c
2297
inc = car_max / steps_max;
drivers/media/dvb-frontends/stv090x.c
2299
*freq_inc = inc;
drivers/media/dvb-frontends/stv090x.c
2337
static int stv090x_search_car_loop(struct stv090x_state *state, s32 inc, s32 timeout, int zigzag, s32 steps_max)
drivers/media/dvb-frontends/stv090x.c
2353
offst_freq = -car_max + inc;
drivers/media/dvb-frontends/stv090x.c
2372
offst_freq = -offst_freq - 2 * inc;
drivers/media/dvb-frontends/stv090x.c
2376
offst_freq += 2 * inc;
drivers/media/dvb-frontends/stv090x.c
2386
((offst_freq - inc) < car_max) &&
drivers/media/dvb-frontends/stv090x.c
2387
((offst_freq + inc) > -car_max) &&
drivers/media/dvb-frontends/stv090x.c
2407
s32 inc, timeout_step, trials, steps_max;
drivers/media/dvb-frontends/stv090x.c
2410
stv090x_get_loop_params(state, &inc, &timeout_step, &steps_max);
drivers/media/dvb-frontends/stv090x.c
2455
lock = stv090x_search_car_loop(state, inc, timeout_step, zigzag, steps_max);
drivers/media/i2c/ov2680.c
370
u8 sensor_ctrl_0a, inc, fmt1, fmt2;
drivers/media/i2c/ov2680.c
375
inc = 0x31;
drivers/media/i2c/ov2680.c
380
inc = 0x11;
drivers/media/i2c/ov2680.c
404
cci_write(sensor->regmap, OV2680_REG_X_INC, inc, &ret);
drivers/media/i2c/ov2680.c
405
cci_write(sensor->regmap, OV2680_REG_Y_INC, inc, &ret);
drivers/media/platform/amphion/vpu_malone.c
77
#define MALONE_VERSION(maj, min, inc) \
drivers/media/platform/amphion/vpu_malone.c
78
(FIELD_PREP(0xF0000, maj) | FIELD_PREP(0xFF00, min) | FIELD_PREP(0xFF, inc))
drivers/media/platform/st/sti/bdisp/bdisp-debug.c
255
u32 inc;
drivers/media/platform/st/sti/bdisp/bdisp-debug.c
262
inc = val & 0xFFFF;
drivers/media/platform/st/sti/bdisp/bdisp-debug.c
263
seq_printf(s, "H: %d(6.10) / scale~%dx0.1 - ", inc, 1024 * 10 / inc);
drivers/media/platform/st/sti/bdisp/bdisp-debug.c
265
inc = val >> 16;
drivers/media/platform/st/sti/bdisp/bdisp-debug.c
266
seq_printf(s, "V: %d(6.10) / scale~%dx0.1", inc, 1024 * 10 / inc);
drivers/media/platform/st/sti/bdisp/bdisp-hw.c
551
static dma_addr_t bdisp_hw_get_hf_addr(u16 inc)
drivers/media/platform/st/sti/bdisp/bdisp-hw.c
556
if ((bdisp_h_filter[i].min < inc) &&
drivers/media/platform/st/sti/bdisp/bdisp-hw.c
557
(inc <= bdisp_h_filter[i].max))
drivers/media/platform/st/sti/bdisp/bdisp-hw.c
572
static dma_addr_t bdisp_hw_get_vf_addr(u16 inc)
drivers/media/platform/st/sti/bdisp/bdisp-hw.c
577
if ((bdisp_v_filter[i].min < inc) &&
drivers/media/platform/st/sti/bdisp/bdisp-hw.c
578
(inc <= bdisp_v_filter[i].max))
drivers/media/platform/st/sti/bdisp/bdisp-hw.c
595
static int bdisp_hw_get_inc(u32 from, u32 to, u16 *inc)
drivers/media/platform/st/sti/bdisp/bdisp-hw.c
603
*inc = 1 << 10;
drivers/media/platform/st/sti/bdisp/bdisp-hw.c
612
*inc = (u16)tmp;
drivers/media/usb/dvb-usb-v2/anysee.c
188
int ret = 0, inc, i = 0;
drivers/media/usb/dvb-usb-v2/anysee.c
208
inc = 2;
drivers/media/usb/dvb-usb-v2/anysee.c
220
inc = 1;
drivers/media/usb/dvb-usb-v2/anysee.c
225
i += inc;
drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c
237
int inc;
drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c
252
inc = 2;
drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c
259
inc = 1;
drivers/media/usb/ttusb-budget/dvb-ttusb-budget.c
270
i += inc;
drivers/misc/bcm-vk/bcm_vk_msg.c
75
static u32 msgq_inc(const struct bcm_vk_sync_qinfo *qinfo, u32 idx, u32 inc)
drivers/misc/bcm-vk/bcm_vk_msg.c
77
return ((idx + inc) & qinfo->q_mask);
drivers/net/bonding/bond_main.c
843
static int bond_set_promiscuity(struct bonding *bond, int inc)
drivers/net/bonding/bond_main.c
852
err = dev_set_promiscuity(curr_active->dev, inc);
drivers/net/bonding/bond_main.c
857
err = dev_set_promiscuity(slave->dev, inc);
drivers/net/bonding/bond_main.c
866
static int bond_set_allmulti(struct bonding *bond, int inc)
drivers/net/bonding/bond_main.c
875
err = dev_set_allmulti(curr_active->dev, inc);
drivers/net/bonding/bond_main.c
880
err = dev_set_allmulti(slave->dev, inc);
drivers/net/can/dev/rx-offload.c
27
if (offload->inc)
drivers/net/can/dev/rx-offload.c
36
if (offload->inc)
drivers/net/can/dev/rx-offload.c
383
offload->inc = true;
drivers/net/can/dev/rx-offload.c
386
offload->inc = false;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
10523
static void bnx2x_prev_unload_undi_inc(struct bnx2x *bp, u8 inc)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
10534
rcq = BNX2X_PREV_UNDI_RCQ(tmp_reg) + inc;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
10535
bd = BNX2X_PREV_UNDI_BD(tmp_reg) + inc;
drivers/net/ethernet/chelsio/cxgb3/t3_hw.c
2854
unsigned int inc;
drivers/net/ethernet/chelsio/cxgb3/t3_hw.c
2856
inc = max(((mtu - 40) * alpha[w]) / avg_pkts[w],
drivers/net/ethernet/chelsio/cxgb3/t3_hw.c
2860
(w << 16) | (beta[w] << 13) | inc);
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
5871
unsigned int inc;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
5873
inc = max(((mtu - 40) * alpha[w]) / avg_pkts[w],
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
5877
(w << 16) | (beta[w] << 13) | inc);
drivers/net/ethernet/freescale/fec_ptp.c
309
int inc;
drivers/net/ethernet/freescale/fec_ptp.c
311
inc = 1000000000 / fep->cycle_speed;
drivers/net/ethernet/freescale/fec_ptp.c
317
writel(inc << FEC_T_INC_OFFSET, fep->hwp + FEC_ATIME_INC);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1466
static void ice_update_per_q_fltr(struct ice_vsi *vsi, u32 q_index, bool inc)
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1477
if (inc)
drivers/net/ethernet/marvell/octeon_ep/octep_ctrl_mbox.c
59
static u32 octep_ctrl_mbox_circq_inc(u32 index, u32 inc, u32 sz)
drivers/net/ethernet/marvell/octeon_ep/octep_ctrl_mbox.c
61
return (index + inc) % sz;
drivers/net/ethernet/mellanox/mlx4/icm.c
384
int inc = MLX4_TABLE_CHUNK_SIZE / table->obj_size;
drivers/net/ethernet/mellanox/mlx4/icm.c
388
for (i = start; i <= end; i += inc) {
drivers/net/ethernet/mellanox/mlx4/icm.c
398
i -= inc;
drivers/net/ethernet/renesas/ravb_main.c
1857
u64 inc;
drivers/net/ethernet/renesas/ravb_main.c
1869
inc = div64_ul(1000000000ULL << 20, rate);
drivers/net/ethernet/renesas/ravb_main.c
1871
if (inc < GTI_TIV_MIN || inc > GTI_TIV_MAX) {
drivers/net/ethernet/renesas/ravb_main.c
1873
inc, GTI_TIV_MIN, GTI_TIV_MAX);
drivers/net/ethernet/renesas/ravb_main.c
1876
priv->gti_tiv = inc;
drivers/net/hyperv/netvsc_drv.c
70
int inc;
drivers/net/hyperv/netvsc_drv.c
76
inc = (net->flags & IFF_PROMISC) ? 1 : -1;
drivers/net/hyperv/netvsc_drv.c
77
dev_set_promiscuity(vf_netdev, inc);
drivers/net/hyperv/netvsc_drv.c
81
inc = (net->flags & IFF_ALLMULTI) ? 1 : -1;
drivers/net/hyperv/netvsc_drv.c
82
dev_set_allmulti(vf_netdev, inc);
drivers/net/phy/nxp-c45-tja11xx.c
456
bool inc;
drivers/net/phy/nxp-c45-tja11xx.c
459
inc = ppb >= 0;
drivers/net/phy/nxp-c45-tja11xx.c
469
if (inc)
drivers/net/team/team_core.c
1731
int inc;
drivers/net/team/team_core.c
1737
inc = dev->flags & IFF_PROMISC ? 1 : -1;
drivers/net/team/team_core.c
1738
dev_set_promiscuity(port->dev, inc);
drivers/net/team/team_core.c
1741
inc = dev->flags & IFF_ALLMULTI ? 1 : -1;
drivers/net/team/team_core.c
1742
dev_set_allmulti(port->dev, inc);
drivers/net/wireless/ath/ath9k/ar5008_phy.c
293
static const int inc[4] = { 0, 100, 0, 0 };
drivers/net/wireless/ath/ath9k/ar5008_phy.c
311
cur_bin += inc[i];
drivers/pcmcia/cistpl.c
167
u_int inc = 1, card_offset, flags;
drivers/pcmcia/cistpl.c
179
inc++;
drivers/pcmcia/cistpl.c
193
for ( ; len > 0; len--, buf++, sys += inc) {
drivers/pcmcia/cistpl.c
248
u_int inc = 1, card_offset, flags;
drivers/pcmcia/cistpl.c
253
inc++;
drivers/pcmcia/cistpl.c
267
for ( ; len > 0; len--, buf++, sys += inc) {
drivers/pinctrl/pinctrl-rockchip.c
4048
int inc;
drivers/pinctrl/pinctrl-rockchip.c
4084
inc = (iom->type & (IOMUX_WIDTH_4BIT |
drivers/pinctrl/pinctrl-rockchip.c
4088
pmu_offs += inc;
drivers/pinctrl/pinctrl-rockchip.c
4090
grf_offs += inc;
drivers/pinctrl/pinctrl-rockchip.c
4098
inc = 8;
drivers/pinctrl/pinctrl-rockchip.c
4100
inc = 4;
drivers/pinctrl/pinctrl-rockchip.c
4103
drv_pmu_offs += inc;
drivers/pinctrl/pinctrl-rockchip.c
4105
drv_grf_offs += inc;
drivers/platform/x86/lenovo/thinkpad_acpi.c
6563
int cmos_cmd, inc;
drivers/platform/x86/lenovo/thinkpad_acpi.c
6576
inc = (value > current_value) ? 1 : -1;
drivers/platform/x86/lenovo/thinkpad_acpi.c
6578
for (i = current_value; i != value; i += inc)
drivers/s390/cio/qdio.h
316
#define add_buf(bufnr, inc) QDIO_BUFNR((bufnr) + (inc))
drivers/scsi/aha152x.c
285
CMD_INC_RESID(struct scsi_cmnd *cmd, int inc)
drivers/scsi/aha152x.c
287
scsi_set_resid(cmd, scsi_get_resid(cmd) + inc);
drivers/scsi/aic94xx/aic94xx_sds.c
1306
u32 inc;
drivers/scsi/aic94xx/aic94xx_sds.c
1328
inc = asd_ha->hw_prof.flash.wide ? 2 : 1;
drivers/scsi/aic94xx/aic94xx_sds.c
1333
dev_id = asd_read_reg_byte(asd_ha, reg + inc);
drivers/scsi/aic94xx/aic94xx_sds.c
1334
sec_prot = asd_read_reg_byte(asd_ha, reg + inc + inc);
drivers/scsi/aic94xx/aic94xx_sds.c
1398
dev_id = asd_read_reg_byte(asd_ha, reg + inc);
drivers/scsi/aic94xx/aic94xx_sds.c
1399
sec_prot = asd_read_reg_byte(asd_ha, reg + inc + inc);
drivers/spi/spi.c
124
u64 inc;
drivers/spi/spi.c
130
inc = u64_stats_read(field);
drivers/spi/spi.c
132
val += inc;
drivers/staging/media/av7110/av7110_hw.c
1007
rc = LoadBitmap(av7110, w, lpb, inc, data);
drivers/staging/media/av7110/av7110_hw.c
1013
data += lpb * inc;
drivers/staging/media/av7110/av7110_hw.c
1016
rc = LoadBitmap(av7110, w, brest / bpl, inc, data);
drivers/staging/media/av7110/av7110_hw.c
839
u16 dx, u16 dy, int inc, u8 __user *data)
drivers/staging/media/av7110/av7110_hw.c
872
if (copy_from_user(av7110->bmpbuf + 1024 + i * dx, data + i * inc, dx)) {
drivers/staging/media/av7110/av7110_hw.c
974
int x1, int y1, int inc, u8 __user *data)
drivers/staging/media/av7110/av7110_hw.c
982
if (inc <= 0)
drivers/staging/media/av7110/av7110_hw.c
983
inc = w;
drivers/staging/rtl8723bs/core/rtw_mlme_ext.c
297
for (ch = o->min_chan; ch <= o->max_chan; ch += o->inc) {
drivers/staging/rtl8723bs/core/rtw_mlme_ext.c
301
if ((padapter->registrypriv.ht_enable == 0) && (o->inc == 8))
drivers/staging/rtl8723bs/include/rtw_mlme_ext.h
360
u8 inc;
drivers/thermal/k3_j72xx_bandgap.c
132
int inc, i, seg;
drivers/thermal/k3_j72xx_bandgap.c
167
inc = 1;
drivers/thermal/k3_j72xx_bandgap.c
169
derived_table[i] = derived_table[i - 1] + inc * 100;
drivers/tty/mips_ejtag_fdc.c
314
u8 inc[4];
drivers/tty/mips_ejtag_fdc.c
341
inc[buf_len] = i;
drivers/tty/mips_ejtag_fdc.c
344
count -= inc[word.bytes - 1];
drivers/tty/mips_ejtag_fdc.c
345
s += inc[word.bytes - 1];
drivers/ufs/core/ufshcd.c
2333
ktime_t now, inc, lat;
drivers/ufs/core/ufshcd.c
2336
inc = ktime_sub(now, m->busy_start_ts[dir]);
drivers/ufs/core/ufshcd.c
2337
m->total_busy[dir] = ktime_add(m->total_busy[dir], inc);
drivers/usb/dwc2/hcd.h
639
static inline u16 dwc2_frame_num_inc(u16 frame, u16 inc)
drivers/usb/dwc2/hcd.h
641
return (frame + inc) & HFNUM_MAX_FRNUM;
drivers/usb/dwc2/hcd_ddma.c
204
u16 i, j, inc;
drivers/usb/dwc2/hcd_ddma.c
223
inc = dwc2_frame_incr_val(qh);
drivers/usb/dwc2/hcd_ddma.c
235
j = (j + inc) & (FRLISTEN_64_SIZE - 1);
drivers/usb/dwc2/hcd_ddma.c
254
inc = (8 + qh->host_interval - 1) / qh->host_interval;
drivers/usb/dwc2/hcd_ddma.c
255
for (i = 0; i < inc; i++) {
drivers/usb/dwc2/hcd_ddma.c
31
static u16 dwc2_desclist_idx_inc(u16 idx, u16 inc, u8 speed)
drivers/usb/dwc2/hcd_ddma.c
33
return (idx + inc) &
drivers/usb/dwc2/hcd_ddma.c
38
static u16 dwc2_desclist_idx_dec(u16 idx, u16 inc, u8 speed)
drivers/usb/dwc2/hcd_ddma.c
40
return (idx - inc) &
drivers/usb/dwc2/hcd_ddma.c
555
u16 idx, inc, n_desc = 0, ntd_max = 0;
drivers/usb/dwc2/hcd_ddma.c
560
inc = qh->host_interval;
drivers/usb/dwc2/hcd_ddma.c
563
next_idx = dwc2_desclist_idx_inc(qh->td_last, inc, qh->dev_speed);
drivers/usb/dwc2/hcd_ddma.c
573
if (inc < 32) {
drivers/usb/dwc2/hcd_ddma.c
576
qh->td_last = dwc2_desclist_idx_inc(cur_idx, inc,
drivers/usb/dwc2/hcd_ddma.c
603
idx = dwc2_desclist_idx_inc(idx, inc, qh->dev_speed);
drivers/usb/dwc2/hcd_ddma.c
615
idx = dwc2_desclist_idx_dec(qh->td_last, inc, qh->dev_speed);
drivers/usb/dwc2/hcd_ddma.c
640
idx = dwc2_desclist_idx_dec(idx, inc * ((qh->ntd + 1) / 2),
drivers/usb/dwc2/hcd_ddma.c
648
idx = dwc2_desclist_idx_dec(qh->td_last, inc, qh->dev_speed);
drivers/usb/misc/sisusbvga/sisusbvga.c
1556
u32 inc, int testn)
drivers/usb/misc/sisusbvga/sisusbvga.c
1565
j += inc;
drivers/usb/misc/sisusbvga/sisusbvga.c
1573
j += inc;
drivers/usb/misc/sisusbvga/sisusbvga.c
1584
u32 inc;
drivers/usb/misc/sisusbvga/sisusbvga.c
1589
inc = 1 << (rtype[idx][2] + rtype[idx][1] + rtype[idx][0] +
drivers/usb/misc/sisusbvga/sisusbvga.c
1591
ret |= sisusb_check_rbc(sisusb, &i2ret, inc, 2);
drivers/usb/misc/sisusbvga/sisusbvga.c
1596
inc = 1 << (rtype[idx][2] + bw / 64 + 2);
drivers/usb/misc/sisusbvga/sisusbvga.c
1597
ret |= sisusb_check_rbc(sisusb, &i2ret, inc, 4);
drivers/usb/misc/sisusbvga/sisusbvga.c
1601
inc = 1 << (10 + bw / 64);
drivers/usb/misc/sisusbvga/sisusbvga.c
1602
ret |= sisusb_check_rbc(sisusb, &i2ret, inc, 2);
drivers/vfio/pci/mlx5/cmd.c
107
if (inc) {
drivers/vfio/pci/mlx5/cmd.c
142
if (inc)
drivers/vfio/pci/mlx5/cmd.c
740
struct mlx5_vhca_data_buffer *buf, bool inc,
drivers/vfio/pci/mlx5/cmd.c
759
migf->state == MLX5_MIGF_STATE_PRE_COPY_ERROR) && !track && !inc)
drivers/vfio/pci/mlx5/cmd.c
767
inc = false;
drivers/vfio/pci/mlx5/cmd.c
775
MLX5_SET(save_vhca_state_in, in, incremental, inc);
drivers/vfio/pci/mlx5/cmd.c
94
bool inc = query_flags & MLX5VF_QUERY_INC;
drivers/vfio/pci/mlx5/cmd.h
209
struct mlx5_vhca_data_buffer *buf, bool inc,
drivers/video/fbdev/omap2/omapfb/dss/dispc.c
798
static void dispc_ovl_set_pix_inc(enum omap_plane plane, s32 inc)
drivers/video/fbdev/omap2/omapfb/dss/dispc.c
800
dispc_write_reg(DISPC_OVL_PIXEL_INC(plane), inc);
drivers/video/fbdev/omap2/omapfb/dss/dispc.c
803
static void dispc_ovl_set_row_inc(enum omap_plane plane, s32 inc)
drivers/video/fbdev/omap2/omapfb/dss/dispc.c
805
dispc_write_reg(DISPC_OVL_ROW_INC(plane), inc);
drivers/video/fbdev/omap2/omapfb/dss/dispc.h
103
const struct dispc_coef *dispc_ovl_get_scale_coef(int inc, int five_taps);
drivers/video/fbdev/omap2/omapfb/dss/dispc_coefs.c
278
const struct dispc_coef *dispc_ovl_get_scale_coef(int inc, int five_taps)
drivers/video/fbdev/omap2/omapfb/dss/dispc_coefs.c
309
inc /= 128;
drivers/video/fbdev/omap2/omapfb/dss/dispc_coefs.c
311
if (inc >= coefs[i].Mmin && inc <= coefs[i].Mmax)
fs/btrfs/extent-tree.c
2552
bool full_backref, bool inc)
fs/btrfs/extent-tree.c
2580
if (inc)
fs/btrfs/extent-tree.c
2611
if (inc)
fs/btrfs/extent-tree.c
2624
if (inc)
fs/ceph/mdsmap.c
180
s32 mds, inc, state;
fs/ceph/mdsmap.c
209
ceph_decode_32_safe(p, end, inc, bad);
fs/ceph/mdsmap.c
240
mds, inc, ceph_pr_addr(&addr),
fs/ceph/quota.c
13
void ceph_adjust_quota_realms_count(struct inode *inode, bool inc)
fs/ceph/quota.c
16
if (inc)
fs/ceph/super.h
1416
extern void ceph_adjust_quota_realms_count(struct inode *inode, bool inc);
fs/f2fs/data.c
3524
static inline void account_writeback(struct inode *inode, bool inc)
fs/f2fs/data.c
3530
if (inc)
fs/f2fs/f2fs.h
3382
static inline void f2fs_i_links_write(struct inode *inode, bool inc)
fs/f2fs/f2fs.h
3384
if (inc)
fs/f2fs/f2fs.h
3796
int f2fs_pin_file_control(struct inode *inode, bool inc);
fs/f2fs/file.c
3488
int f2fs_pin_file_control(struct inode *inode, bool inc)
fs/f2fs/file.c
3504
if (inc)
fs/ocfs2/dir.c
2659
goto inc;
fs/ocfs2/dir.c
2675
inc:
fs/ocfs2/dir.c
2705
goto inc;
fs/ocfs2/dir.c
2719
inc:
fs/proc/stat.c
62
unsigned int inc;
fs/proc/stat.c
64
inc = min_t(unsigned int, gap, ARRAY_SIZE(zeros) / 2);
fs/proc/stat.c
65
seq_write(p, zeros, 2 * inc);
fs/proc/stat.c
66
gap -= inc;
fs/udf/directory.c
500
int inc)
fs/udf/directory.c
517
if (inc)
fs/udf/directory.c
522
struct long_ad *udf_get_filelongad(uint8_t *ptr, int maxoffset, uint32_t *offset, int inc)
fs/udf/directory.c
539
if (inc)
fs/udf/inode.c
2066
struct kernel_lb_addr *eloc, uint32_t elen, int inc)
fs/udf/inode.c
2089
udf_write_aext(inode, epos, eloc, elen, inc);
fs/udf/inode.c
2100
epos->offset + (inc ? 0 : adsize));
fs/udf/inode.c
2115
struct kernel_lb_addr *eloc, uint32_t elen, int inc)
fs/udf/inode.c
2142
return __udf_add_aext(inode, epos, eloc, elen, inc);
fs/udf/inode.c
2146
struct kernel_lb_addr *eloc, uint32_t elen, int inc)
fs/udf/inode.c
2193
if (inc)
fs/udf/inode.c
2208
int inc)
fs/udf/inode.c
2216
etype, inc);
fs/udf/inode.c
2246
int inc)
fs/udf/inode.c
2279
sad = udf_get_fileshortad(ptr, alen, &epos->offset, inc);
fs/udf/inode.c
2289
lad = udf_get_filelongad(ptr, alen, &epos->offset, inc);
fs/udf/udfdecl.h
167
struct kernel_lb_addr *eloc, uint32_t elen, int inc);
fs/udf/udfdecl.h
175
int8_t *etype, int inc);
fs/udf/udfdecl.h
178
int8_t *etype, int inc);
fs/xfs/libxfs/xfs_alloc_btree.c
46
int inc)
fs/xfs/libxfs/xfs_alloc_btree.c
56
be32_add_cpu(&agf->agf_bno_level, inc);
fs/xfs/libxfs/xfs_alloc_btree.c
57
pag->pagf_bno_level += inc;
fs/xfs/libxfs/xfs_alloc_btree.c
60
be32_add_cpu(&agf->agf_cnt_level, inc);
fs/xfs/libxfs/xfs_alloc_btree.c
61
pag->pagf_cnt_level += inc;
fs/xfs/libxfs/xfs_btree.c
3296
int inc)
fs/xfs/libxfs/xfs_btree.c
3301
cur->bc_ag.afake->af_levels += inc;
fs/xfs/libxfs/xfs_btree.c
3303
cur->bc_ops->set_root(cur, ptr, inc);
fs/xfs/libxfs/xfs_btree_mem.c
28
int inc)
fs/xfs/libxfs/xfs_btree_mem.c
33
cur->bc_mem.xfbtree->nlevels += inc;
fs/xfs/libxfs/xfs_btree_mem.h
51
const union xfs_btree_ptr *ptr, int inc);
fs/xfs/libxfs/xfs_ialloc_btree.c
56
int inc) /* level change */
fs/xfs/libxfs/xfs_ialloc_btree.c
62
be32_add_cpu(&agi->agi_level, inc);
fs/xfs/libxfs/xfs_ialloc_btree.c
70
int inc) /* level change */
fs/xfs/libxfs/xfs_ialloc_btree.c
76
be32_add_cpu(&agi->agi_free_level, inc);
fs/xfs/libxfs/xfs_refcount_btree.c
40
int inc)
fs/xfs/libxfs/xfs_refcount_btree.c
49
be32_add_cpu(&agf->agf_refcount_level, inc);
fs/xfs/libxfs/xfs_refcount_btree.c
50
pag->pagf_refcount_level += inc;
fs/xfs/libxfs/xfs_rmap_btree.c
67
int inc)
fs/xfs/libxfs/xfs_rmap_btree.c
76
be32_add_cpu(&agf->agf_rmap_level, inc);
fs/xfs/libxfs/xfs_rmap_btree.c
77
pag->pagf_rmap_level += inc;
fs/xfs/xfs_stats.h
187
#define XFS_STATS_ADD(mp, v, inc) \
fs/xfs/xfs_stats.h
189
per_cpu_ptr(xfsstats.xs_stats, current_cpu())->s.v += (inc); \
fs/xfs/xfs_stats.h
190
per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v += (inc); \
fs/xfs/xfs_stats.h
205
#define XFS_STATS_ADD_OFF(mp, off, inc) \
fs/xfs/xfs_stats.h
207
per_cpu_ptr(xfsstats.xs_stats, current_cpu())->a[off] += (inc); \
fs/xfs/xfs_stats.h
208
per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->a[off] += (inc); \
include/linux/can/rx-offload.h
31
bool inc;
include/linux/ceph/ceph_features.h
29
#define DEFINE_CEPH_FEATURE_RETIRED(bit, inc, name, unused, unadvertised)
include/linux/ceph/osd_client.h
297
__le32 inc;
include/linux/dmaengine.h
1336
static inline size_t dmaengine_get_icg(bool inc, bool sgl, size_t icg,
include/linux/dmaengine.h
1339
if (inc) {
include/linux/mm.h
3006
int account_locked_vm(struct mm_struct *mm, unsigned long pages, bool inc);
include/linux/mm.h
3007
int __account_locked_vm(struct mm_struct *mm, unsigned long pages, bool inc,
include/linux/netdevice.h
5098
int netif_set_promiscuity(struct net_device *dev, int inc);
include/linux/netdevice.h
5099
int dev_set_promiscuity(struct net_device *dev, int inc);
include/linux/netdevice.h
5100
int netif_set_allmulti(struct net_device *dev, int inc, bool notify);
include/linux/netdevice.h
5101
int dev_set_allmulti(struct net_device *dev, int inc);
include/net/ip_vs.h
1477
int ip_vs_app_inc_get(struct ip_vs_app *inc);
include/net/ip_vs.h
1478
void ip_vs_app_inc_put(struct ip_vs_app *inc);
include/net/ip_vs.h
521
int (*register_app)(struct netns_ipvs *ipvs, struct ip_vs_app *inc);
include/net/ip_vs.h
523
void (*unregister_app)(struct netns_ipvs *ipvs, struct ip_vs_app *inc);
kernel/bpf/helpers.c
1553
bool inc = false;
kernel/bpf/helpers.c
1582
inc = true;
kernel/bpf/helpers.c
1601
if (inc)
kernel/events/core.c
13196
bool inc = false;
kernel/events/core.c
13202
inc = true;
kernel/events/core.c
13219
inc = true;
kernel/events/core.c
13222
inc = true;
kernel/events/core.c
13224
inc = true;
kernel/events/core.c
13232
if (inc) {
kernel/locking/lock_events.h
46
static inline void __lockevent_add(enum lock_events event, int inc)
kernel/locking/lock_events.h
48
raw_cpu_add(lockevents[event], inc);
kernel/sched/fair.c
165
static inline void update_load_add(struct load_weight *lw, unsigned long inc)
kernel/sched/fair.c
167
lw->weight += inc;
kernel/trace/ftrace.c
1767
bool inc)
kernel/trace/ftrace.c
1819
if (inc) {
kernel/trace/ftrace.c
1956
static void ftrace_hash_rec_update_modify(struct ftrace_ops *ops, bool inc)
kernel/trace/ftrace.c
1960
__ftrace_hash_rec_update(ops, inc);
kernel/trace/ftrace.c
1974
__ftrace_hash_rec_update(op, inc);
kernel/trace/ring_buffer.c
7106
bool inc)
kernel/trace/ring_buffer.c
7116
if (inc && cpu_buffer->mapped == UINT_MAX)
kernel/trace/ring_buffer.c
7119
if (WARN_ON(!inc && cpu_buffer->user_mapped == 0))
kernel/trace/ring_buffer.c
7125
if (inc) {
kernel/trace/ring_buffer_benchmark.c
114
int inc;
kernel/trace/ring_buffer_benchmark.c
127
for (i = 0; i < commit && !test_error ; i += inc) {
kernel/trace/ring_buffer_benchmark.c
134
inc = -1;
kernel/trace/ring_buffer_benchmark.c
141
inc = event->array[0] + 4;
kernel/trace/ring_buffer_benchmark.c
144
inc = 8;
kernel/trace/ring_buffer_benchmark.c
157
inc = event->array[0] + 4;
kernel/trace/ring_buffer_benchmark.c
166
inc = ((event->type_len + 1) * 4);
kernel/trace/ring_buffer_benchmark.c
171
if (inc <= 0) {
kernel/trace/trace_events_user.c
1719
goto inc;
kernel/trace/trace_events_user.c
1734
goto inc;
kernel/trace/trace_events_user.c
1752
inc:
kernel/trace/trace_events_user.c
748
goto inc;
kernel/trace/trace_events_user.c
756
inc:
lib/crypto/arm/blake2b.h
14
const u8 *data, size_t nblocks, u32 inc);
lib/crypto/arm/blake2b.h
17
const u8 *data, size_t nblocks, u32 inc)
lib/crypto/arm/blake2b.h
20
blake2b_compress_generic(ctx, data, nblocks, inc);
lib/crypto/arm/blake2b.h
28
blake2b_compress_neon(ctx, data, blocks, inc);
lib/crypto/arm/blake2s.h
5
const u8 *data, size_t nblocks, u32 inc);
lib/crypto/blake2b.c
35
static inline void blake2b_increment_counter(struct blake2b_ctx *ctx, u32 inc)
lib/crypto/blake2b.c
37
ctx->t[0] += inc;
lib/crypto/blake2b.c
38
ctx->t[1] += (ctx->t[0] < inc);
lib/crypto/blake2b.c
43
const u8 *data, size_t nblocks, u32 inc)
lib/crypto/blake2b.c
50
(nblocks > 1 && inc != BLAKE2B_BLOCK_SIZE));
lib/crypto/blake2b.c
53
blake2b_increment_counter(ctx, inc);
lib/crypto/blake2s.c
33
static inline void blake2s_increment_counter(struct blake2s_ctx *ctx, u32 inc)
lib/crypto/blake2s.c
35
ctx->t[0] += inc;
lib/crypto/blake2s.c
36
ctx->t[1] += (ctx->t[0] < inc);
lib/crypto/blake2s.c
41
const u8 *data, size_t nblocks, u32 inc)
lib/crypto/blake2s.c
48
(nblocks > 1 && inc != BLAKE2S_BLOCK_SIZE));
lib/crypto/blake2s.c
51
blake2s_increment_counter(ctx, inc);
lib/crypto/x86/blake2s.h
15
const u8 *data, size_t nblocks, u32 inc);
lib/crypto/x86/blake2s.h
17
const u8 *data, size_t nblocks, u32 inc);
lib/crypto/x86/blake2s.h
23
const u8 *data, size_t nblocks, u32 inc)
lib/crypto/x86/blake2s.h
29
blake2s_compress_generic(ctx, data, nblocks, inc);
lib/crypto/x86/blake2s.h
39
blake2s_compress_avx512(ctx, data, blocks, inc);
lib/crypto/x86/blake2s.h
41
blake2s_compress_ssse3(ctx, data, blocks, inc);
lib/tests/memcpy_kunit.c
385
static int next_step(int idx, int start, int end, int inc)
lib/tests/memcpy_kunit.c
387
start += inc;
lib/tests/memcpy_kunit.c
388
end -= inc;
lib/tests/memcpy_kunit.c
390
if (idx < start || idx + inc > end)
lib/tests/memcpy_kunit.c
391
inc = 1;
lib/tests/memcpy_kunit.c
392
return idx + inc;
mm/util.c
506
int __account_locked_vm(struct mm_struct *mm, unsigned long pages, bool inc,
mm/util.c
515
if (inc) {
mm/util.c
529
(void *)_RET_IP_, (inc) ? '+' : '-', pages << PAGE_SHIFT,
mm/util.c
549
int account_locked_vm(struct mm_struct *mm, unsigned long pages, bool inc)
mm/util.c
557
ret = __account_locked_vm(mm, pages, inc, current,
net/core/dev.c
9567
static int __dev_set_promiscuity(struct net_device *dev, int inc, bool notify)
net/core/dev.c
9576
promiscuity = dev->promiscuity + inc;
net/core/dev.c
9582
if (unlikely(inc > 0)) {
net/core/dev.c
9623
int netif_set_promiscuity(struct net_device *dev, int inc)
net/core/dev.c
9628
err = __dev_set_promiscuity(dev, inc, true);
net/core/dev.c
9636
int netif_set_allmulti(struct net_device *dev, int inc, bool notify)
net/core/dev.c
9643
allmulti = dev->allmulti + inc;
net/core/dev.c
9649
if (unlikely(inc > 0)) {
net/core/dev.c
9785
int inc = (flags & IFF_PROMISC) ? 1 : -1;
net/core/dev.c
9790
if (__dev_set_promiscuity(dev, inc, false) >= 0)
net/core/dev.c
9800
int inc = (flags & IFF_ALLMULTI) ? 1 : -1;
net/core/dev.c
9803
netif_set_allmulti(dev, inc, false);
net/core/dev_api.c
282
int dev_set_promiscuity(struct net_device *dev, int inc)
net/core/dev_api.c
287
ret = netif_set_promiscuity(dev, inc);
net/core/dev_api.c
308
int dev_set_allmulti(struct net_device *dev, int inc)
net/core/dev_api.c
313
ret = netif_set_allmulti(dev, inc, true);
net/dsa/conduit.c
342
static void dsa_conduit_set_promiscuity(struct net_device *dev, int inc)
net/dsa/conduit.c
351
dev_set_promiscuity(dev, inc);
net/ipv4/tcp_veno.c
173
if (veno->inc &&
net/ipv4/tcp_veno.c
176
veno->inc = 0;
net/ipv4/tcp_veno.c
178
veno->inc = 1;
net/ipv4/tcp_veno.c
32
u32 inc; /* decide whether to increase cwnd */
net/ipv4/tcp_veno.c
68
veno->inc = 1;
net/ipv6/ndisc.c
804
bool inc;
net/ipv6/ndisc.c
846
inc = ipv6_addr_is_multicast(daddr);
net/ipv6/ndisc.c
908
inc &&
net/ipv6/ndisc.c
937
if (inc)
net/ipv6/ndisc.c
947
!inc || lladdr || !dev->addr_len);
net/ipv6/ndisc.c
955
true, (ifp != NULL && inc), inc);
net/mac80211/iface.c
2076
u64 mask, start, addr, val, inc;
net/mac80211/iface.c
2168
inc = 1ULL<<__ffs64(mask);
net/mac80211/iface.c
2181
val += inc;
net/netfilter/ipvs/ip_vs_app.c
104
ret = pp->register_app(ipvs, inc);
net/netfilter/ipvs/ip_vs_app.c
108
list_add(&inc->a_list, &app->incs_list);
net/netfilter/ipvs/ip_vs_app.c
110
pp->name, inc->name, ntohs(inc->port));
net/netfilter/ipvs/ip_vs_app.c
115
ip_vs_app_inc_destroy(inc);
net/netfilter/ipvs/ip_vs_app.c
124
ip_vs_app_inc_release(struct netns_ipvs *ipvs, struct ip_vs_app *inc)
net/netfilter/ipvs/ip_vs_app.c
128
if (!(pp = ip_vs_proto_get(inc->protocol)))
net/netfilter/ipvs/ip_vs_app.c
132
pp->unregister_app(ipvs, inc);
net/netfilter/ipvs/ip_vs_app.c
135
pp->name, inc->name, ntohs(inc->port));
net/netfilter/ipvs/ip_vs_app.c
137
list_del(&inc->a_list);
net/netfilter/ipvs/ip_vs_app.c
139
call_rcu(&inc->rcu_head, ip_vs_app_inc_rcu_free);
net/netfilter/ipvs/ip_vs_app.c
147
int ip_vs_app_inc_get(struct ip_vs_app *inc)
net/netfilter/ipvs/ip_vs_app.c
151
result = ip_vs_app_get(inc->app);
net/netfilter/ipvs/ip_vs_app.c
153
atomic_inc(&inc->usecnt);
net/netfilter/ipvs/ip_vs_app.c
161
void ip_vs_app_inc_put(struct ip_vs_app *inc)
net/netfilter/ipvs/ip_vs_app.c
163
atomic_dec(&inc->usecnt);
net/netfilter/ipvs/ip_vs_app.c
164
ip_vs_app_put(inc->app);
net/netfilter/ipvs/ip_vs_app.c
233
struct ip_vs_app *a, *anxt, *inc, *nxt;
net/netfilter/ipvs/ip_vs_app.c
240
list_for_each_entry_safe(inc, nxt, &a->incs_list, a_list) {
net/netfilter/ipvs/ip_vs_app.c
241
ip_vs_app_inc_release(ipvs, inc);
net/netfilter/ipvs/ip_vs_app.c
270
struct ip_vs_app *inc = cp->app;
net/netfilter/ipvs/ip_vs_app.c
272
if (!inc)
net/netfilter/ipvs/ip_vs_app.c
275
if (inc->unbind_conn)
net/netfilter/ipvs/ip_vs_app.c
276
inc->unbind_conn(inc, cp);
net/netfilter/ipvs/ip_vs_app.c
277
if (inc->done_conn)
net/netfilter/ipvs/ip_vs_app.c
278
inc->done_conn(inc, cp);
net/netfilter/ipvs/ip_vs_app.c
279
ip_vs_app_inc_put(inc);
net/netfilter/ipvs/ip_vs_app.c
520
struct ip_vs_app *app, *inc;
net/netfilter/ipvs/ip_vs_app.c
523
list_for_each_entry(inc, &app->incs_list, a_list) {
net/netfilter/ipvs/ip_vs_app.c
525
return inc;
net/netfilter/ipvs/ip_vs_app.c
544
struct ip_vs_app *inc, *app;
net/netfilter/ipvs/ip_vs_app.c
55
static void ip_vs_app_inc_destroy(struct ip_vs_app *inc)
net/netfilter/ipvs/ip_vs_app.c
553
inc = v;
net/netfilter/ipvs/ip_vs_app.c
554
app = inc->app;
net/netfilter/ipvs/ip_vs_app.c
556
if ((e = inc->a_list.next) != &app->incs_list)
net/netfilter/ipvs/ip_vs_app.c
562
list_for_each_entry(inc, &app->incs_list, a_list) {
net/netfilter/ipvs/ip_vs_app.c
563
return inc;
net/netfilter/ipvs/ip_vs_app.c
57
kfree(inc->timeout_table);
net/netfilter/ipvs/ip_vs_app.c
579
const struct ip_vs_app *inc = v;
net/netfilter/ipvs/ip_vs_app.c
58
kfree(inc);
net/netfilter/ipvs/ip_vs_app.c
582
ip_vs_proto_name(inc->protocol),
net/netfilter/ipvs/ip_vs_app.c
583
ntohs(inc->port),
net/netfilter/ipvs/ip_vs_app.c
584
atomic_read(&inc->usecnt),
net/netfilter/ipvs/ip_vs_app.c
585
inc->name);
net/netfilter/ipvs/ip_vs_app.c
63
struct ip_vs_app *inc = container_of(head, struct ip_vs_app, rcu_head);
net/netfilter/ipvs/ip_vs_app.c
65
ip_vs_app_inc_destroy(inc);
net/netfilter/ipvs/ip_vs_app.c
76
struct ip_vs_app *inc;
net/netfilter/ipvs/ip_vs_app.c
85
inc = kmemdup(app, sizeof(*inc), GFP_KERNEL);
net/netfilter/ipvs/ip_vs_app.c
86
if (!inc)
net/netfilter/ipvs/ip_vs_app.c
88
INIT_LIST_HEAD(&inc->p_list);
net/netfilter/ipvs/ip_vs_app.c
89
INIT_LIST_HEAD(&inc->incs_list);
net/netfilter/ipvs/ip_vs_app.c
90
inc->app = app;
net/netfilter/ipvs/ip_vs_app.c
91
inc->port = htons(port);
net/netfilter/ipvs/ip_vs_app.c
92
atomic_set(&inc->usecnt, 0);
net/netfilter/ipvs/ip_vs_app.c
95
inc->timeout_table =
net/netfilter/ipvs/ip_vs_app.c
98
if (!inc->timeout_table) {
net/netfilter/ipvs/ip_vs_proto_sctp.c
484
static int sctp_register_app(struct netns_ipvs *ipvs, struct ip_vs_app *inc)
net/netfilter/ipvs/ip_vs_proto_sctp.c
488
__be16 port = inc->port;
net/netfilter/ipvs/ip_vs_proto_sctp.c
500
list_add_rcu(&inc->p_list, &ipvs->sctp_apps[hash]);
net/netfilter/ipvs/ip_vs_proto_sctp.c
507
static void sctp_unregister_app(struct netns_ipvs *ipvs, struct ip_vs_app *inc)
net/netfilter/ipvs/ip_vs_proto_sctp.c
512
list_del_rcu(&inc->p_list);
net/netfilter/ipvs/ip_vs_proto_sctp.c
519
struct ip_vs_app *inc;
net/netfilter/ipvs/ip_vs_proto_sctp.c
528
list_for_each_entry_rcu(inc, &ipvs->sctp_apps[hash], p_list) {
net/netfilter/ipvs/ip_vs_proto_sctp.c
529
if (inc->port == cp->vport) {
net/netfilter/ipvs/ip_vs_proto_sctp.c
530
if (unlikely(!ip_vs_app_inc_get(inc)))
net/netfilter/ipvs/ip_vs_proto_sctp.c
540
inc->name, ntohs(inc->port));
net/netfilter/ipvs/ip_vs_proto_sctp.c
541
cp->app = inc;
net/netfilter/ipvs/ip_vs_proto_sctp.c
542
if (inc->init_conn)
net/netfilter/ipvs/ip_vs_proto_sctp.c
543
result = inc->init_conn(inc, cp);
net/netfilter/ipvs/ip_vs_proto_tcp.c
608
static int tcp_register_app(struct netns_ipvs *ipvs, struct ip_vs_app *inc)
net/netfilter/ipvs/ip_vs_proto_tcp.c
612
__be16 port = inc->port;
net/netfilter/ipvs/ip_vs_proto_tcp.c
624
list_add_rcu(&inc->p_list, &ipvs->tcp_apps[hash]);
net/netfilter/ipvs/ip_vs_proto_tcp.c
633
tcp_unregister_app(struct netns_ipvs *ipvs, struct ip_vs_app *inc)
net/netfilter/ipvs/ip_vs_proto_tcp.c
638
list_del_rcu(&inc->p_list);
net/netfilter/ipvs/ip_vs_proto_tcp.c
647
struct ip_vs_app *inc;
net/netfilter/ipvs/ip_vs_proto_tcp.c
657
list_for_each_entry_rcu(inc, &ipvs->tcp_apps[hash], p_list) {
net/netfilter/ipvs/ip_vs_proto_tcp.c
658
if (inc->port == cp->vport) {
net/netfilter/ipvs/ip_vs_proto_tcp.c
659
if (unlikely(!ip_vs_app_inc_get(inc)))
net/netfilter/ipvs/ip_vs_proto_tcp.c
669
inc->name, ntohs(inc->port));
net/netfilter/ipvs/ip_vs_proto_tcp.c
671
cp->app = inc;
net/netfilter/ipvs/ip_vs_proto_tcp.c
672
if (inc->init_conn)
net/netfilter/ipvs/ip_vs_proto_tcp.c
673
result = inc->init_conn(inc, cp);
net/netfilter/ipvs/ip_vs_proto_udp.c
354
static int udp_register_app(struct netns_ipvs *ipvs, struct ip_vs_app *inc)
net/netfilter/ipvs/ip_vs_proto_udp.c
358
__be16 port = inc->port;
net/netfilter/ipvs/ip_vs_proto_udp.c
370
list_add_rcu(&inc->p_list, &ipvs->udp_apps[hash]);
net/netfilter/ipvs/ip_vs_proto_udp.c
379
udp_unregister_app(struct netns_ipvs *ipvs, struct ip_vs_app *inc)
net/netfilter/ipvs/ip_vs_proto_udp.c
384
list_del_rcu(&inc->p_list);
net/netfilter/ipvs/ip_vs_proto_udp.c
392
struct ip_vs_app *inc;
net/netfilter/ipvs/ip_vs_proto_udp.c
402
list_for_each_entry_rcu(inc, &ipvs->udp_apps[hash], p_list) {
net/netfilter/ipvs/ip_vs_proto_udp.c
403
if (inc->port == cp->vport) {
net/netfilter/ipvs/ip_vs_proto_udp.c
404
if (unlikely(!ip_vs_app_inc_get(inc)))
net/netfilter/ipvs/ip_vs_proto_udp.c
414
inc->name, ntohs(inc->port));
net/netfilter/ipvs/ip_vs_proto_udp.c
416
cp->app = inc;
net/netfilter/ipvs/ip_vs_proto_udp.c
417
if (inc->init_conn)
net/netfilter/ipvs/ip_vs_proto_udp.c
418
result = inc->init_conn(inc, cp);
net/rds/af_rds.c
733
struct rds_incoming *inc;
net/rds/af_rds.c
748
list_for_each_entry(inc, &rs->rs_recv_queue, i_item) {
net/rds/af_rds.c
751
rds_inc_info_copy(inc, iter,
net/rds/af_rds.c
752
inc->i_saddr.s6_addr32[3],
net/rds/af_rds.c
771
struct rds_incoming *inc;
net/rds/af_rds.c
782
list_for_each_entry(inc, &rs->rs_recv_queue, i_item) {
net/rds/af_rds.c
785
rds6_inc_info_copy(inc, iter, &inc->i_saddr,
net/rds/connection.c
553
static void __rds_inc_msg_cp(struct rds_incoming *inc,
net/rds/connection.c
559
rds6_inc_info_copy(inc, iter, saddr, daddr, flip);
net/rds/connection.c
562
rds_inc_info_copy(inc, iter, *(__be32 *)saddr,
net/rds/ib.h
398
void rds_ib_inc_free(struct rds_incoming *inc);
net/rds/ib.h
399
int rds_ib_inc_copy_to_user(struct rds_incoming *inc, struct iov_iter *to);
net/rds/ib_recv.c
158
struct rds_ib_incoming *inc;
net/rds/ib_recv.c
168
list_for_each_entry_safe(inc, inc_tmp, &list, ii_cache_entry) {
net/rds/ib_recv.c
169
list_del(&inc->ii_cache_entry);
net/rds/ib_recv.c
170
WARN_ON(!list_empty(&inc->ii_frags));
net/rds/ib_recv.c
171
kmem_cache_free(rds_ib_incoming_slab, inc);
net/rds/ib_recv.c
204
void rds_ib_inc_free(struct rds_incoming *inc)
net/rds/ib_recv.c
209
struct rds_ib_connection *ic = inc->i_conn->c_transport_data;
net/rds/ib_recv.c
211
ibinc = container_of(inc, struct rds_ib_incoming, ii_inc);
net/rds/ib_recv.c
220
rdsdebug("freeing ibinc %p inc %p\n", ibinc, inc);
net/rds/ib_recv.c
534
int rds_ib_inc_copy_to_user(struct rds_incoming *inc, struct iov_iter *to)
net/rds/ib_recv.c
544
ibinc = container_of(inc, struct rds_ib_incoming, ii_inc);
net/rds/ib_recv.c
546
len = be32_to_cpu(inc->i_hdr.h_len);
net/rds/loop.c
111
static void rds_loop_inc_free(struct rds_incoming *inc)
net/rds/loop.c
113
struct rds_message *rm = container_of(inc, struct rds_message, m_inc);
net/rds/message.c
504
int rds_message_inc_copy_to_user(struct rds_incoming *inc, struct iov_iter *to)
net/rds/message.c
514
rm = container_of(inc, struct rds_message, m_inc);
net/rds/rds.h
589
int (*inc_copy_to_user)(struct rds_incoming *inc, struct iov_iter *to);
net/rds/rds.h
590
void (*inc_free)(struct rds_incoming *inc);
net/rds/rds.h
899
int rds_message_inc_copy_to_user(struct rds_incoming *inc, struct iov_iter *to);
net/rds/rds.h
924
void rds_inc_init(struct rds_incoming *inc, struct rds_connection *conn,
net/rds/rds.h
926
void rds_inc_path_init(struct rds_incoming *inc, struct rds_conn_path *conn,
net/rds/rds.h
928
void rds_inc_put(struct rds_incoming *inc);
net/rds/rds.h
931
struct rds_incoming *inc, gfp_t gfp);
net/rds/rds.h
936
void rds_inc_info_copy(struct rds_incoming *inc,
net/rds/rds.h
939
void rds6_inc_info_copy(struct rds_incoming *inc,
net/rds/recv.c
166
static void rds_recv_incoming_exthdrs(struct rds_incoming *inc, struct rds_sock *rs)
net/rds/recv.c
168
struct rds_header *hdr = &inc->i_hdr;
net/rds/recv.c
190
inc->i_usercopy.rdma_cookie = rds_rdma_make_cookie(
net/rds/recv.c
318
struct rds_incoming *inc, gfp_t gfp)
net/rds/recv.c
325
inc->i_conn = conn;
net/rds/recv.c
326
inc->i_rx_jiffies = jiffies;
net/rds/recv.c
328
cp = inc->i_conn_path;
net/rds/recv.c
335
inc,
net/rds/recv.c
336
(unsigned long long)be64_to_cpu(inc->i_hdr.h_sequence),
net/rds/recv.c
337
be32_to_cpu(inc->i_hdr.h_len),
net/rds/recv.c
338
be16_to_cpu(inc->i_hdr.h_sport),
net/rds/recv.c
339
be16_to_cpu(inc->i_hdr.h_dport),
net/rds/recv.c
340
inc->i_hdr.h_flags,
net/rds/recv.c
341
inc->i_rx_jiffies);
net/rds/recv.c
363
if (be64_to_cpu(inc->i_hdr.h_sequence) < cp->cp_next_rx_seq &&
net/rds/recv.c
364
(inc->i_hdr.h_flags & RDS_FLAG_RETRANSMITTED)) {
net/rds/recv.c
368
cp->cp_next_rx_seq = be64_to_cpu(inc->i_hdr.h_sequence) + 1;
net/rds/recv.c
370
if (rds_sysctl_ping_enable && inc->i_hdr.h_dport == 0) {
net/rds/recv.c
371
if (inc->i_hdr.h_sport == 0) {
net/rds/recv.c
377
rds_send_pong(cp, inc->i_hdr.h_sport);
net/rds/recv.c
379
if (RDS_HS_PROBE(be16_to_cpu(inc->i_hdr.h_sport),
net/rds/recv.c
380
be16_to_cpu(inc->i_hdr.h_dport))) {
net/rds/recv.c
381
rds_recv_hs_exthdrs(&inc->i_hdr, cp->cp_conn);
net/rds/recv.c
387
if (be16_to_cpu(inc->i_hdr.h_dport) == RDS_FLAG_PROBE_PORT &&
net/rds/recv.c
388
inc->i_hdr.h_sport == 0) {
net/rds/recv.c
389
rds_recv_hs_exthdrs(&inc->i_hdr, cp->cp_conn);
net/rds/recv.c
396
rs = rds_find_bound(daddr, inc->i_hdr.h_dport, conn->c_bound_if);
net/rds/recv.c
403
rds_recv_incoming_exthdrs(inc, rs);
net/rds/recv.c
411
rdsdebug("adding inc %p to rs %p's recv queue\n", inc, rs);
net/rds/recv.c
413
rds_recv_rcvbuf_delta(rs, sk, inc->i_conn->c_lcong,
net/rds/recv.c
414
be32_to_cpu(inc->i_hdr.h_len),
net/rds/recv.c
415
inc->i_hdr.h_dport);
net/rds/recv.c
417
inc->i_usercopy.rx_tstamp = ktime_get_real();
net/rds/recv.c
418
rds_inc_addref(inc);
net/rds/recv.c
419
inc->i_rx_lat_trace[RDS_MSG_RX_END] = local_clock();
net/rds/recv.c
420
list_add_tail(&inc->i_item, &rs->rs_recv_queue);
net/rds/recv.c
437
static int rds_next_incoming(struct rds_sock *rs, struct rds_incoming **inc)
net/rds/recv.c
44
void rds_inc_init(struct rds_incoming *inc, struct rds_connection *conn,
net/rds/recv.c
441
if (!*inc) {
net/rds/recv.c
444
*inc = list_entry(rs->rs_recv_queue.next,
net/rds/recv.c
447
rds_inc_addref(*inc);
net/rds/recv.c
452
return *inc != NULL;
net/rds/recv.c
455
static int rds_still_queued(struct rds_sock *rs, struct rds_incoming *inc,
net/rds/recv.c
464
if (!list_empty(&inc->i_item)) {
net/rds/recv.c
468
rds_recv_rcvbuf_delta(rs, sk, inc->i_conn->c_lcong,
net/rds/recv.c
469
-be32_to_cpu(inc->i_hdr.h_len),
net/rds/recv.c
47
refcount_set(&inc->i_refcount, 1);
net/rds/recv.c
470
inc->i_hdr.h_dport);
net/rds/recv.c
471
list_del_init(&inc->i_item);
net/rds/recv.c
472
to_drop = inc;
net/rds/recv.c
48
INIT_LIST_HEAD(&inc->i_item);
net/rds/recv.c
480
rdsdebug("inc %p rs %p still %d dropped %d\n", inc, rs, ret, drop);
net/rds/recv.c
49
inc->i_conn = conn;
net/rds/recv.c
50
inc->i_saddr = *saddr;
net/rds/recv.c
51
inc->i_usercopy.rdma_cookie = 0;
net/rds/recv.c
52
inc->i_usercopy.rx_tstamp = ktime_set(0, 0);
net/rds/recv.c
54
memset(inc->i_rx_lat_trace, 0, sizeof(inc->i_rx_lat_trace));
net/rds/recv.c
577
static int rds_cmsg_recv(struct rds_incoming *inc, struct msghdr *msg,
net/rds/recv.c
58
void rds_inc_path_init(struct rds_incoming *inc, struct rds_conn_path *cp,
net/rds/recv.c
582
if (inc->i_usercopy.rdma_cookie) {
net/rds/recv.c
584
sizeof(inc->i_usercopy.rdma_cookie),
net/rds/recv.c
585
&inc->i_usercopy.rdma_cookie);
net/rds/recv.c
590
if ((inc->i_usercopy.rx_tstamp != 0) &&
net/rds/recv.c
593
ns_to_kernel_old_timeval(inc->i_usercopy.rx_tstamp);
net/rds/recv.c
61
refcount_set(&inc->i_refcount, 1);
net/rds/recv.c
617
inc->i_rx_lat_trace[RDS_MSG_RX_CMSG] = local_clock();
net/rds/recv.c
62
INIT_LIST_HEAD(&inc->i_item);
net/rds/recv.c
622
t.rx_trace[i] = inc->i_rx_lat_trace[j + 1] -
net/rds/recv.c
623
inc->i_rx_lat_trace[j];
net/rds/recv.c
63
inc->i_conn = cp->cp_conn;
net/rds/recv.c
64
inc->i_conn_path = cp;
net/rds/recv.c
65
inc->i_saddr = *saddr;
net/rds/recv.c
66
inc->i_usercopy.rdma_cookie = 0;
net/rds/recv.c
67
inc->i_usercopy.rx_tstamp = ktime_set(0, 0);
net/rds/recv.c
680
struct rds_incoming *inc = NULL;
net/rds/recv.c
704
if (!rds_next_incoming(rs, &inc)) {
net/rds/recv.c
71
static void rds_inc_addref(struct rds_incoming *inc)
net/rds/recv.c
715
rds_next_incoming(rs, &inc)), timeo);
net/rds/recv.c
716
rdsdebug("recvmsg woke inc %p timeo %ld\n", inc,
net/rds/recv.c
727
rdsdebug("copying inc %p from %pI6c:%u to user\n", inc,
net/rds/recv.c
728
&inc->i_conn->c_faddr,
net/rds/recv.c
729
ntohs(inc->i_hdr.h_sport));
net/rds/recv.c
73
rdsdebug("addref inc %p ref %d\n", inc, refcount_read(&inc->i_refcount));
net/rds/recv.c
730
ret = inc->i_conn->c_trans->inc_copy_to_user(inc, &msg->msg_iter);
net/rds/recv.c
739
if (!rds_still_queued(rs, inc, !(msg_flags & MSG_PEEK))) {
net/rds/recv.c
74
refcount_inc(&inc->i_refcount);
net/rds/recv.c
740
rds_inc_put(inc);
net/rds/recv.c
741
inc = NULL;
net/rds/recv.c
747
if (ret < be32_to_cpu(inc->i_hdr.h_len)) {
net/rds/recv.c
749
ret = be32_to_cpu(inc->i_hdr.h_len);
net/rds/recv.c
753
if (rds_cmsg_recv(inc, msg, rs)) {
net/rds/recv.c
762
if (ipv6_addr_v4mapped(&inc->i_saddr)) {
net/rds/recv.c
764
sin->sin_port = inc->i_hdr.h_sport;
net/rds/recv.c
766
inc->i_saddr.s6_addr32[3];
net/rds/recv.c
77
void rds_inc_put(struct rds_incoming *inc)
net/rds/recv.c
771
sin6->sin6_port = inc->i_hdr.h_sport;
net/rds/recv.c
772
sin6->sin6_addr = inc->i_saddr;
net/rds/recv.c
781
if (inc)
net/rds/recv.c
782
rds_inc_put(inc);
net/rds/recv.c
79
rdsdebug("put inc %p ref %d\n", inc, refcount_read(&inc->i_refcount));
net/rds/recv.c
796
struct rds_incoming *inc, *tmp;
net/rds/recv.c
80
if (refcount_dec_and_test(&inc->i_refcount)) {
net/rds/recv.c
801
list_for_each_entry_safe(inc, tmp, &rs->rs_recv_queue, i_item) {
net/rds/recv.c
802
rds_recv_rcvbuf_delta(rs, sk, inc->i_conn->c_lcong,
net/rds/recv.c
803
-be32_to_cpu(inc->i_hdr.h_len),
net/rds/recv.c
804
inc->i_hdr.h_dport);
net/rds/recv.c
805
list_move(&inc->i_item, &to_drop);
net/rds/recv.c
809
list_for_each_entry_safe(inc, tmp, &to_drop, i_item) {
net/rds/recv.c
81
BUG_ON(!list_empty(&inc->i_item));
net/rds/recv.c
810
list_del_init(&inc->i_item);
net/rds/recv.c
811
rds_inc_put(inc);
net/rds/recv.c
819
void rds_inc_info_copy(struct rds_incoming *inc,
net/rds/recv.c
825
minfo.seq = be64_to_cpu(inc->i_hdr.h_sequence);
net/rds/recv.c
826
minfo.len = be32_to_cpu(inc->i_hdr.h_len);
net/rds/recv.c
827
minfo.tos = inc->i_conn->c_tos;
net/rds/recv.c
83
inc->i_conn->c_trans->inc_free(inc);
net/rds/recv.c
832
minfo.lport = inc->i_hdr.h_dport;
net/rds/recv.c
833
minfo.fport = inc->i_hdr.h_sport;
net/rds/recv.c
837
minfo.lport = inc->i_hdr.h_sport;
net/rds/recv.c
838
minfo.fport = inc->i_hdr.h_dport;
net/rds/recv.c
847
void rds6_inc_info_copy(struct rds_incoming *inc,
net/rds/recv.c
854
minfo6.seq = be64_to_cpu(inc->i_hdr.h_sequence);
net/rds/recv.c
855
minfo6.len = be32_to_cpu(inc->i_hdr.h_len);
net/rds/recv.c
856
minfo6.tos = inc->i_conn->c_tos;
net/rds/recv.c
861
minfo6.lport = inc->i_hdr.h_dport;
net/rds/recv.c
862
minfo6.fport = inc->i_hdr.h_sport;
net/rds/recv.c
866
minfo6.lport = inc->i_hdr.h_sport;
net/rds/recv.c
867
minfo6.fport = inc->i_hdr.h_dport;
net/rds/tcp.h
103
void rds_tcp_inc_free(struct rds_incoming *inc);
net/rds/tcp.h
104
int rds_tcp_inc_copy_to_user(struct rds_incoming *inc, struct iov_iter *to);
net/rds/tcp_recv.c
43
static void rds_tcp_inc_purge(struct rds_incoming *inc)
net/rds/tcp_recv.c
46
tinc = container_of(inc, struct rds_tcp_incoming, ti_inc);
net/rds/tcp_recv.c
47
rdsdebug("purging tinc %p inc %p\n", tinc, inc);
net/rds/tcp_recv.c
51
void rds_tcp_inc_free(struct rds_incoming *inc)
net/rds/tcp_recv.c
54
tinc = container_of(inc, struct rds_tcp_incoming, ti_inc);
net/rds/tcp_recv.c
55
rds_tcp_inc_purge(inc);
net/rds/tcp_recv.c
56
rdsdebug("freeing tinc %p inc %p\n", tinc, inc);
net/rds/tcp_recv.c
63
int rds_tcp_inc_copy_to_user(struct rds_incoming *inc, struct iov_iter *to)
net/rds/tcp_recv.c
72
tinc = container_of(inc, struct rds_tcp_incoming, ti_inc);
net/sctp/tsnmap.c
344
unsigned long inc;
net/sctp/tsnmap.c
350
inc = ALIGN((size - map->len), BITS_PER_LONG) + SCTP_TSN_MAP_INCREMENT;
net/sctp/tsnmap.c
351
len = min_t(u16, map->len + inc, SCTP_TSN_MAP_SIZE);
scripts/mod/file2alias.c
252
int inc,
scripts/mod/file2alias.c
261
*bcd += inc;
scripts/mod/file2alias.c
275
dec += inc;
sound/core/seq/seq_timer.h
95
static inline void snd_seq_inc_real_time(snd_seq_real_time_t *tm, snd_seq_real_time_t *inc)
sound/core/seq/seq_timer.h
97
tm->tv_sec += inc->tv_sec;
sound/core/seq/seq_timer.h
98
tm->tv_nsec += inc->tv_nsec;
sound/mips/hal2.c
326
codec->inc = 4;
sound/mips/hal2.c
336
int inc = hal2->dac.inc;
sound/mips/hal2.c
341
((0xffff & (inc - mod - 1)) << 16) | inc);
sound/mips/hal2.c
347
int inc = hal2->adc.inc;
sound/mips/hal2.c
352
((0xffff & (inc - mod - 1)) << 16) | inc);
sound/mips/hal2.c
69
unsigned short inc; /* INC value */
sound/pcmcia/pdaudiocf/pdaudiocf_pcm.c
38
int inc, ret = 0, rate;
sound/pcmcia/pdaudiocf/pdaudiocf_pcm.c
53
inc = 1;
sound/pcmcia/pdaudiocf/pdaudiocf_pcm.c
61
inc = -1;
sound/pcmcia/pdaudiocf/pdaudiocf_pcm.c
68
chip->pcm_running += inc;
sound/pcmcia/pdaudiocf/pdaudiocf_pcm.c
72
chip->pcm_running -= inc;
sound/soc/renesas/rcar/src.c
209
int inc, cnt, rate;
sound/soc/renesas/rcar/src.c
235
inc = fout / 100;
sound/soc/renesas/rcar/src.c
236
cnt = abs(new_rate - fout) / inc;
sound/soc/renesas/rcar/src.c
238
inc *= -1;
sound/soc/renesas/rcar/src.c
248
rate += inc;
tools/arch/x86/include/asm/asm.h
36
#define _ASM_INC __ASM_SIZE(inc)
tools/build/tests/ex/ex.c
19
inc();
tools/build/tests/ex/ex.c
9
int inc(void);
tools/hv/vmbus_bufring.c
45
static inline uint32_t vmbus_br_idxinc(uint32_t idx, uint32_t inc, uint32_t sz)
tools/hv/vmbus_bufring.c
47
idx += inc;
tools/include/nolibc/sys.h
106
void *sbrk(intptr_t inc)
tools/include/nolibc/sys.h
111
if (ret && sys_brk(ret + inc) == ret + inc)
tools/include/nolibc/sys.h
112
return ret + inc;
tools/perf/bench/epoll-ctl.c
207
int inc;
tools/perf/bench/epoll-ctl.c
213
inc = 100/pct;
tools/perf/bench/epoll-ctl.c
214
for (i = 0; i < nfds; i+=inc) {
tools/perf/ui/browsers/hists.c
1670
static int advance_hpp_check(struct perf_hpp *hpp, int inc)
tools/perf/ui/browsers/hists.c
1672
advance_hpp(hpp, inc);
tools/perf/util/blake2s.c
43
static inline void blake2s_increment_counter(struct blake2s_ctx *ctx, u32 inc)
tools/perf/util/blake2s.c
45
ctx->t[0] += inc;
tools/perf/util/blake2s.c
46
ctx->t[1] += (ctx->t[0] < inc);
tools/perf/util/blake2s.c
50
const u8 *data, size_t nblocks, u32 inc)
tools/perf/util/blake2s.c
57
blake2s_increment_counter(ctx, inc);
tools/perf/util/hist.h
671
static inline void advance_hpp(struct perf_hpp *hpp, int inc)
tools/perf/util/hist.h
673
hpp->buf += inc;
tools/perf/util/hist.h
674
hpp->size -= inc;
tools/perf/util/map.h
280
static inline void map__add_pgoff(struct map *map, u64 inc)
tools/perf/util/map.h
282
RC_CHK_ACCESS(map)->pgoff += inc;
tools/testing/selftests/rseq/rseq-or1k-bits.h
233
int RSEQ_TEMPLATE_IDENTIFIER(rseq_offset_deref_addv)(intptr_t *ptr, off_t off, intptr_t inc,
tools/testing/selftests/rseq/rseq-or1k-bits.h
248
RSEQ_ASM_OP_R_DEREF_ADDV(ptr, off, inc, 3)
tools/testing/selftests/rseq/rseq-or1k-bits.h
257
[inc] "r" (inc)
tools/testing/selftests/rseq/rseq-or1k.h
144
#define RSEQ_ASM_OP_R_DEREF_ADDV(ptr, off, inc, post_commit_label) \
tools/testing/selftests/rseq/rseq-or1k.h
148
RSEQ_ASM_OP_R_ADD(inc) \
tools/testing/selftests/rseq/rseq-riscv-bits.h
232
int RSEQ_TEMPLATE_IDENTIFIER(rseq_offset_deref_addv)(intptr_t *ptr, off_t off, intptr_t inc, int cpu)
tools/testing/selftests/rseq/rseq-riscv-bits.h
246
RSEQ_ASM_OP_R_DEREF_ADDV(ptr, off, inc, 3)
tools/testing/selftests/rseq/rseq-riscv-bits.h
255
[inc] "r" (inc)
tools/testing/selftests/rseq/rseq-riscv.h
160
#define RSEQ_ASM_OP_R_DEREF_ADDV(ptr, off, inc, post_commit_label) \
tools/testing/selftests/rseq/rseq-riscv.h
164
RSEQ_ASM_OP_R_ADD(inc) \
tools/testing/selftests/rseq/rseq-x86-bits.h
204
int RSEQ_TEMPLATE_IDENTIFIER(rseq_offset_deref_addv)(intptr_t *ptr, long off, intptr_t inc, int cpu)
tools/testing/selftests/rseq/rseq-x86-bits.h
236
[inc] "er" (inc)
tools/testing/selftests/rseq/rseq.h
305
intptr_t *ptr, long off, intptr_t inc, int cpu)
tools/testing/selftests/rseq/rseq.h
311
return rseq_offset_deref_addv_relaxed_cpu_id(ptr, off, inc, cpu);
tools/testing/selftests/rseq/rseq.h
313
return rseq_offset_deref_addv_relaxed_mm_cid(ptr, off, inc, cpu);