Symbol: rh
arch/arm/vfp/vfp.h
105
u64 rh, rl;
arch/arm/vfp/vfp.h
106
mul64to128(&rh, &rl, n, m);
arch/arm/vfp/vfp.h
107
return rh | (rl != 0);
arch/arm/vfp/vfp.h
73
u64 rh, rma, rmb, rl;
arch/arm/vfp/vfp.h
86
rh = (u64)nh * mh;
arch/arm/vfp/vfp.h
87
rh += ((u64)(rma < rmb) << 32) + (rma >> 32);
arch/arm/vfp/vfp.h
91
rh += (rl < rma);
arch/arm/vfp/vfp.h
94
*resh = rh;
arch/arm64/include/asm/atomic_ll_sc.h
323
[rl] "=&r" (r.low), [rh] "=&r" (r.high), \
arch/powerpc/kernel/rethook.c
31
void arch_rethook_prepare(struct rethook_node *rh, struct pt_regs *regs, bool mcount)
arch/powerpc/kernel/rethook.c
33
rh->ret_addr = regs->link;
arch/powerpc/kernel/rethook.c
34
rh->frame = regs->gpr[1];
arch/s390/kernel/rethook.c
6
void arch_rethook_prepare(struct rethook_node *rh, struct pt_regs *regs, bool mcount)
arch/s390/kernel/rethook.c
8
rh->ret_addr = regs->gprs[14];
arch/s390/kernel/rethook.c
9
rh->frame = regs->gprs[15];
arch/x86/kernel/rethook.c
117
void arch_rethook_prepare(struct rethook_node *rh, struct pt_regs *regs, bool mcount)
arch/x86/kernel/rethook.c
121
rh->ret_addr = stack[0];
arch/x86/kernel/rethook.c
122
rh->frame = regs->sp;
drivers/crypto/cavium/nitrox/nitrox_csr.h
489
u64 rh : 1;
drivers/crypto/cavium/nitrox/nitrox_csr.h
495
u64 rh : 1;
drivers/crypto/cavium/nitrox/nitrox_hal.c
206
pkt_slc_ctl.s.rh = 1;
drivers/dma/bestcomm/sram.c
103
rh_attach_region(bcom_sram->rh, zbase - bcom_sram->base_phys, regaddr_p[1]);
drivers/dma/bestcomm/sram.c
128
rh_destroy(bcom_sram->rh);
drivers/dma/bestcomm/sram.c
142
offset = rh_alloc_align(bcom_sram->rh, size, align, NULL);
drivers/dma/bestcomm/sram.c
163
rh_free(bcom_sram->rh, offset);
drivers/dma/bestcomm/sram.c
90
bcom_sram->rh = rh_create(4);
drivers/dma/bestcomm/sram.c
98
rh_attach_region(bcom_sram->rh, 0, bcom_sram->size);
drivers/gpio/gpiolib.c
140
static void desc_free_label(struct rcu_head *rh)
drivers/gpio/gpiolib.c
142
kfree(container_of(rh, struct gpio_desc_label, rh));
drivers/gpio/gpiolib.c
159
call_srcu(&desc->gdev->desc_srcu, &old->rh, desc_free_label);
drivers/gpio/gpiolib.h
165
struct rcu_head rh;
drivers/i3c/master/mipi-i3c-hci/dma.c
151
struct hci_rh_data *rh;
drivers/i3c/master/mipi-i3c-hci/dma.c
158
rh = &rings->headers[i];
drivers/i3c/master/mipi-i3c-hci/dma.c
167
rh = &rings->headers[i];
drivers/i3c/master/mipi-i3c-hci/dma.c
180
struct hci_rh_data *rh;
drivers/i3c/master/mipi-i3c-hci/dma.c
186
rh = &rings->headers[i];
drivers/i3c/master/mipi-i3c-hci/dma.c
188
if (rh->xfer)
drivers/i3c/master/mipi-i3c-hci/dma.c
190
rh->xfer_struct_sz * rh->xfer_entries,
drivers/i3c/master/mipi-i3c-hci/dma.c
191
rh->xfer, rh->xfer_dma);
drivers/i3c/master/mipi-i3c-hci/dma.c
192
if (rh->resp)
drivers/i3c/master/mipi-i3c-hci/dma.c
194
rh->resp_struct_sz * rh->xfer_entries,
drivers/i3c/master/mipi-i3c-hci/dma.c
195
rh->resp, rh->resp_dma);
drivers/i3c/master/mipi-i3c-hci/dma.c
196
kfree(rh->src_xfers);
drivers/i3c/master/mipi-i3c-hci/dma.c
197
if (rh->ibi_status)
drivers/i3c/master/mipi-i3c-hci/dma.c
199
rh->ibi_status_sz * rh->ibi_status_entries,
drivers/i3c/master/mipi-i3c-hci/dma.c
200
rh->ibi_status, rh->ibi_status_dma);
drivers/i3c/master/mipi-i3c-hci/dma.c
201
if (rh->ibi_data_dma)
drivers/i3c/master/mipi-i3c-hci/dma.c
202
dma_unmap_single(rings->sysdev, rh->ibi_data_dma,
drivers/i3c/master/mipi-i3c-hci/dma.c
203
rh->ibi_chunk_sz * rh->ibi_chunks_total,
drivers/i3c/master/mipi-i3c-hci/dma.c
205
kfree(rh->ibi_data);
drivers/i3c/master/mipi-i3c-hci/dma.c
212
static void hci_dma_init_rh(struct i3c_hci *hci, struct hci_rh_data *rh, int i)
drivers/i3c/master/mipi-i3c-hci/dma.c
216
rh_reg_write(CMD_RING_BASE_LO, lower_32_bits(rh->xfer_dma));
drivers/i3c/master/mipi-i3c-hci/dma.c
217
rh_reg_write(CMD_RING_BASE_HI, upper_32_bits(rh->xfer_dma));
drivers/i3c/master/mipi-i3c-hci/dma.c
218
rh_reg_write(RESP_RING_BASE_LO, lower_32_bits(rh->resp_dma));
drivers/i3c/master/mipi-i3c-hci/dma.c
219
rh_reg_write(RESP_RING_BASE_HI, upper_32_bits(rh->resp_dma));
drivers/i3c/master/mipi-i3c-hci/dma.c
221
regval = FIELD_PREP(CR_RING_SIZE, rh->xfer_entries);
drivers/i3c/master/mipi-i3c-hci/dma.c
235
rh_reg_write(IBI_STATUS_RING_BASE_LO, lower_32_bits(rh->ibi_status_dma));
drivers/i3c/master/mipi-i3c-hci/dma.c
236
rh_reg_write(IBI_STATUS_RING_BASE_HI, upper_32_bits(rh->ibi_status_dma));
drivers/i3c/master/mipi-i3c-hci/dma.c
237
rh_reg_write(IBI_DATA_RING_BASE_LO, lower_32_bits(rh->ibi_data_dma));
drivers/i3c/master/mipi-i3c-hci/dma.c
238
rh_reg_write(IBI_DATA_RING_BASE_HI, upper_32_bits(rh->ibi_data_dma));
drivers/i3c/master/mipi-i3c-hci/dma.c
240
regval = FIELD_PREP(IBI_STATUS_RING_SIZE, rh->ibi_status_entries) |
drivers/i3c/master/mipi-i3c-hci/dma.c
241
FIELD_PREP(IBI_DATA_CHUNK_SIZE, ilog2(rh->ibi_chunk_sz) - 2) |
drivers/i3c/master/mipi-i3c-hci/dma.c
242
FIELD_PREP(IBI_DATA_CHUNK_COUNT, rh->ibi_chunks_total);
drivers/i3c/master/mipi-i3c-hci/dma.c
261
rh->done_ptr = 0;
drivers/i3c/master/mipi-i3c-hci/dma.c
262
rh->ibi_chunk_ptr = 0;
drivers/i3c/master/mipi-i3c-hci/dma.c
263
rh->xfer_space = rh->xfer_entries;
drivers/i3c/master/mipi-i3c-hci/dma.c
284
struct hci_rh_data *rh = &rings->headers[i];
drivers/i3c/master/mipi-i3c-hci/dma.c
304
struct hci_rh_data *rh;
drivers/i3c/master/mipi-i3c-hci/dma.c
345
rh = &rings->headers[i];
drivers/i3c/master/mipi-i3c-hci/dma.c
346
rh->regs = hci->base_regs + offset;
drivers/i3c/master/mipi-i3c-hci/dma.c
347
init_completion(&rh->op_done);
drivers/i3c/master/mipi-i3c-hci/dma.c
349
rh->xfer_entries = XFER_RING_ENTRIES;
drivers/i3c/master/mipi-i3c-hci/dma.c
352
rh->xfer_struct_sz = FIELD_GET(CR_XFER_STRUCT_SIZE, regval);
drivers/i3c/master/mipi-i3c-hci/dma.c
353
rh->resp_struct_sz = FIELD_GET(CR_RESP_STRUCT_SIZE, regval);
drivers/i3c/master/mipi-i3c-hci/dma.c
356
rh->xfer_struct_sz, rh->resp_struct_sz);
drivers/i3c/master/mipi-i3c-hci/dma.c
357
xfers_sz = rh->xfer_struct_sz * rh->xfer_entries;
drivers/i3c/master/mipi-i3c-hci/dma.c
358
resps_sz = rh->resp_struct_sz * rh->xfer_entries;
drivers/i3c/master/mipi-i3c-hci/dma.c
360
rh->xfer = dma_alloc_coherent(rings->sysdev, xfers_sz,
drivers/i3c/master/mipi-i3c-hci/dma.c
361
&rh->xfer_dma, GFP_KERNEL);
drivers/i3c/master/mipi-i3c-hci/dma.c
362
rh->resp = dma_alloc_coherent(rings->sysdev, resps_sz,
drivers/i3c/master/mipi-i3c-hci/dma.c
363
&rh->resp_dma, GFP_KERNEL);
drivers/i3c/master/mipi-i3c-hci/dma.c
364
rh->src_xfers =
drivers/i3c/master/mipi-i3c-hci/dma.c
365
kmalloc_objs(*rh->src_xfers, rh->xfer_entries);
drivers/i3c/master/mipi-i3c-hci/dma.c
367
if (!rh->xfer || !rh->resp || !rh->src_xfers)
drivers/i3c/master/mipi-i3c-hci/dma.c
376
rh->ibi_status_sz = FIELD_GET(IBI_STATUS_STRUCT_SIZE, regval);
drivers/i3c/master/mipi-i3c-hci/dma.c
377
rh->ibi_status_entries = IBI_STATUS_RING_ENTRIES;
drivers/i3c/master/mipi-i3c-hci/dma.c
378
rh->ibi_chunks_total = IBI_CHUNK_POOL_SIZE;
drivers/i3c/master/mipi-i3c-hci/dma.c
380
rh->ibi_chunk_sz = dma_get_cache_alignment();
drivers/i3c/master/mipi-i3c-hci/dma.c
381
rh->ibi_chunk_sz *= IBI_CHUNK_CACHELINES;
drivers/i3c/master/mipi-i3c-hci/dma.c
387
rh->ibi_chunk_sz = umax(4, rh->ibi_chunk_sz);
drivers/i3c/master/mipi-i3c-hci/dma.c
388
rh->ibi_chunk_sz = roundup_pow_of_two(rh->ibi_chunk_sz);
drivers/i3c/master/mipi-i3c-hci/dma.c
389
if (rh->ibi_chunk_sz > 256) {
drivers/i3c/master/mipi-i3c-hci/dma.c
394
ibi_status_ring_sz = rh->ibi_status_sz * rh->ibi_status_entries;
drivers/i3c/master/mipi-i3c-hci/dma.c
395
ibi_data_ring_sz = rh->ibi_chunk_sz * rh->ibi_chunks_total;
drivers/i3c/master/mipi-i3c-hci/dma.c
397
rh->ibi_status =
drivers/i3c/master/mipi-i3c-hci/dma.c
399
&rh->ibi_status_dma, GFP_KERNEL);
drivers/i3c/master/mipi-i3c-hci/dma.c
400
rh->ibi_data = kmalloc(ibi_data_ring_sz, GFP_KERNEL);
drivers/i3c/master/mipi-i3c-hci/dma.c
402
if (!rh->ibi_status || !rh->ibi_data)
drivers/i3c/master/mipi-i3c-hci/dma.c
404
rh->ibi_data_dma =
drivers/i3c/master/mipi-i3c-hci/dma.c
405
dma_map_single(rings->sysdev, rh->ibi_data,
drivers/i3c/master/mipi-i3c-hci/dma.c
407
if (dma_mapping_error(rings->sysdev, rh->ibi_data_dma)) {
drivers/i3c/master/mipi-i3c-hci/dma.c
408
rh->ibi_data_dma = 0;
drivers/i3c/master/mipi-i3c-hci/dma.c
472
struct hci_rh_data *rh;
drivers/i3c/master/mipi-i3c-hci/dma.c
483
rh = &rings->headers[ring];
drivers/i3c/master/mipi-i3c-hci/dma.c
487
if (n > rh->xfer_space) {
drivers/i3c/master/mipi-i3c-hci/dma.c
497
u32 *ring_data = rh->xfer + rh->xfer_struct_sz * enqueue_ptr;
drivers/i3c/master/mipi-i3c-hci/dma.c
524
rh->src_xfers[enqueue_ptr] = xfer;
drivers/i3c/master/mipi-i3c-hci/dma.c
529
enqueue_ptr = (enqueue_ptr + 1) % rh->xfer_entries;
drivers/i3c/master/mipi-i3c-hci/dma.c
532
rh->xfer_space -= n;
drivers/i3c/master/mipi-i3c-hci/dma.c
546
struct hci_rh_data *rh = &rings->headers[xfer_list[0].ring_number];
drivers/i3c/master/mipi-i3c-hci/dma.c
55
#define rh_reg_read(r) readl(rh->regs + (RH_##r))
drivers/i3c/master/mipi-i3c-hci/dma.c
556
reinit_completion(&rh->op_done);
drivers/i3c/master/mipi-i3c-hci/dma.c
558
wait_for_completion_timeout(&rh->op_done, HZ);
drivers/i3c/master/mipi-i3c-hci/dma.c
56
#define rh_reg_write(r, v) writel(v, rh->regs + (RH_##r))
drivers/i3c/master/mipi-i3c-hci/dma.c
582
u32 *ring_data = rh->xfer + rh->xfer_struct_sz * idx;
drivers/i3c/master/mipi-i3c-hci/dma.c
593
rh->src_xfers[idx] = NULL;
drivers/i3c/master/mipi-i3c-hci/dma.c
617
static void hci_dma_xfer_done(struct i3c_hci *hci, struct hci_rh_data *rh)
drivers/i3c/master/mipi-i3c-hci/dma.c
620
unsigned int tid, done_ptr = rh->done_ptr;
drivers/i3c/master/mipi-i3c-hci/dma.c
629
ring_resp = rh->resp + rh->resp_struct_sz * done_ptr;
drivers/i3c/master/mipi-i3c-hci/dma.c
634
xfer = rh->src_xfers[done_ptr];
drivers/i3c/master/mipi-i3c-hci/dma.c
639
rh->src_xfers[done_ptr] = NULL;
drivers/i3c/master/mipi-i3c-hci/dma.c
652
done_ptr = (done_ptr + 1) % rh->xfer_entries;
drivers/i3c/master/mipi-i3c-hci/dma.c
653
rh->done_ptr = done_ptr;
drivers/i3c/master/mipi-i3c-hci/dma.c
657
rh->xfer_space += done_cnt;
drivers/i3c/master/mipi-i3c-hci/dma.c
705
static void hci_dma_process_ibi(struct i3c_hci *hci, struct hci_rh_data *rh)
drivers/i3c/master/mipi-i3c-hci/dma.c
733
ptr = (ptr + 1) % rh->ibi_status_entries) {
drivers/i3c/master/mipi-i3c-hci/dma.c
737
ring_ibi_status = rh->ibi_status + rh->ibi_status_sz * ptr;
drivers/i3c/master/mipi-i3c-hci/dma.c
755
ibi_size += chunks * rh->ibi_chunk_sz;
drivers/i3c/master/mipi-i3c-hci/dma.c
773
deq_ptr %= rh->ibi_status_entries;
drivers/i3c/master/mipi-i3c-hci/dma.c
813
ibi_data_offset = rh->ibi_chunk_sz * rh->ibi_chunk_ptr;
drivers/i3c/master/mipi-i3c-hci/dma.c
814
ring_ibi_data = rh->ibi_data + ibi_data_offset;
drivers/i3c/master/mipi-i3c-hci/dma.c
815
ring_ibi_data_dma = rh->ibi_data_dma + ibi_data_offset;
drivers/i3c/master/mipi-i3c-hci/dma.c
816
first_part = (rh->ibi_chunks_total - rh->ibi_chunk_ptr)
drivers/i3c/master/mipi-i3c-hci/dma.c
817
* rh->ibi_chunk_sz;
drivers/i3c/master/mipi-i3c-hci/dma.c
827
ring_ibi_data = rh->ibi_data;
drivers/i3c/master/mipi-i3c-hci/dma.c
828
ring_ibi_data_dma = rh->ibi_data_dma;
drivers/i3c/master/mipi-i3c-hci/dma.c
847
rh->ibi_chunk_ptr += ibi_chunks;
drivers/i3c/master/mipi-i3c-hci/dma.c
848
rh->ibi_chunk_ptr %= rh->ibi_chunks_total;
drivers/i3c/master/mipi-i3c-hci/dma.c
861
struct hci_rh_data *rh;
drivers/i3c/master/mipi-i3c-hci/dma.c
864
rh = &rings->headers[i];
drivers/i3c/master/mipi-i3c-hci/dma.c
873
hci_dma_process_ibi(hci, rh);
drivers/i3c/master/mipi-i3c-hci/dma.c
875
hci_dma_xfer_done(hci, rh);
drivers/i3c/master/mipi-i3c-hci/dma.c
877
complete(&rh->op_done);
drivers/md/dm-raid1.c
1123
r = dm_set_target_max_io_len(ti, dm_rh_get_region_size(ms->rh));
drivers/md/dm-raid1.c
1201
struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
drivers/md/dm-raid1.c
1209
bio_record->write_region = dm_rh_bio_to_region(ms->rh, bio);
drivers/md/dm-raid1.c
1214
r = log->type->in_sync(log, dm_rh_bio_to_region(ms->rh, bio), 0);
drivers/md/dm-raid1.c
1261
dm_rh_dec(ms->rh, bio_record->write_region);
drivers/md/dm-raid1.c
1315
struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
drivers/md/dm-raid1.c
1340
dm_rh_stop_recovery(ms->rh);
drivers/md/dm-raid1.c
1343
!dm_rh_recovery_in_flight(ms->rh));
drivers/md/dm-raid1.c
1361
struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
drivers/md/dm-raid1.c
1371
struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
drivers/md/dm-raid1.c
1377
dm_rh_start_recovery(ms->rh);
drivers/md/dm-raid1.c
1411
struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
drivers/md/dm-raid1.c
341
sector_t region_size = dm_rh_get_region_size(ms->rh);
drivers/md/dm-raid1.c
346
from.sector = m->offset + dm_rh_region_to_sector(ms->rh, key);
drivers/md/dm-raid1.c
365
dest->sector = m->offset + dm_rh_region_to_sector(ms->rh, key);
drivers/md/dm-raid1.c
392
struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
drivers/md/dm-raid1.c
397
dm_rh_recovery_prepare(ms->rh);
drivers/md/dm-raid1.c
402
while ((reg = dm_rh_recovery_start(ms->rh)))
drivers/md/dm-raid1.c
446
struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
drivers/md/dm-raid1.c
447
region_t region = dm_rh_bio_to_region(ms->rh, bio);
drivers/md/dm-raid1.c
562
int state = dm_rh_get_state(ms->rh, region, may_block);
drivers/md/dm-raid1.c
573
region = dm_rh_bio_to_region(ms->rh, bio);
drivers/md/dm-raid1.c
692
struct dm_dirty_log *log = dm_rh_dirty_log(ms->rh);
drivers/md/dm-raid1.c
70
struct dm_region_hash *rh;
drivers/md/dm-raid1.c
713
region = dm_rh_bio_to_region(ms->rh, bio);
drivers/md/dm-raid1.c
721
state = dm_rh_get_state(ms->rh, region, 1);
drivers/md/dm-raid1.c
756
dm_rh_inc_pending(ms->rh, &sync);
drivers/md/dm-raid1.c
757
dm_rh_inc_pending(ms->rh, &nosync);
drivers/md/dm-raid1.c
764
ms->log_failure = dm_rh_flush(ms->rh) ? 1 : ms->log_failure;
drivers/md/dm-raid1.c
779
dm_rh_delay(ms->rh, bio);
drivers/md/dm-raid1.c
821
dm_rh_mark_nosync(ms->rh, bio);
drivers/md/dm-raid1.c
875
dm_rh_update_states(ms->rh, errors_handled(ms));
drivers/md/dm-raid1.c
922
ms->rh = dm_region_hash_create(ms, dispatch_bios, wakeup_mirrord,
drivers/md/dm-raid1.c
926
if (IS_ERR(ms->rh)) {
drivers/md/dm-raid1.c
943
dm_region_hash_destroy(ms->rh);
drivers/md/dm-region-hash.c
106
struct dm_region_hash *rh; /* FIXME: can we get rid of this ? */
drivers/md/dm-region-hash.c
120
static region_t dm_rh_sector_to_region(struct dm_region_hash *rh, sector_t sector)
drivers/md/dm-region-hash.c
122
return sector >> rh->region_shift;
drivers/md/dm-region-hash.c
125
sector_t dm_rh_region_to_sector(struct dm_region_hash *rh, region_t region)
drivers/md/dm-region-hash.c
127
return region << rh->region_shift;
drivers/md/dm-region-hash.c
131
region_t dm_rh_bio_to_region(struct dm_region_hash *rh, struct bio *bio)
drivers/md/dm-region-hash.c
133
return dm_rh_sector_to_region(rh, bio->bi_iter.bi_sector -
drivers/md/dm-region-hash.c
134
rh->target_begin);
drivers/md/dm-region-hash.c
140
return reg->rh->context;
drivers/md/dm-region-hash.c
150
sector_t dm_rh_get_region_size(struct dm_region_hash *rh)
drivers/md/dm-region-hash.c
152
return rh->region_size;
drivers/md/dm-region-hash.c
173
struct dm_region_hash *rh;
drivers/md/dm-region-hash.c
187
rh = kzalloc_obj(*rh);
drivers/md/dm-region-hash.c
188
if (!rh) {
drivers/md/dm-region-hash.c
193
rh->context = context;
drivers/md/dm-region-hash.c
194
rh->dispatch_bios = dispatch_bios;
drivers/md/dm-region-hash.c
195
rh->wakeup_workers = wakeup_workers;
drivers/md/dm-region-hash.c
196
rh->wakeup_all_recovery_waiters = wakeup_all_recovery_waiters;
drivers/md/dm-region-hash.c
197
rh->target_begin = target_begin;
drivers/md/dm-region-hash.c
198
rh->max_recovery = max_recovery;
drivers/md/dm-region-hash.c
199
rh->log = log;
drivers/md/dm-region-hash.c
200
rh->region_size = region_size;
drivers/md/dm-region-hash.c
201
rh->region_shift = __ffs(region_size);
drivers/md/dm-region-hash.c
202
rwlock_init(&rh->hash_lock);
drivers/md/dm-region-hash.c
203
rh->mask = nr_buckets - 1;
drivers/md/dm-region-hash.c
204
rh->nr_buckets = nr_buckets;
drivers/md/dm-region-hash.c
206
rh->shift = RH_HASH_SHIFT;
drivers/md/dm-region-hash.c
207
rh->prime = RH_HASH_MULT;
drivers/md/dm-region-hash.c
209
rh->buckets = vmalloc_array(nr_buckets, sizeof(*rh->buckets));
drivers/md/dm-region-hash.c
210
if (!rh->buckets) {
drivers/md/dm-region-hash.c
212
kfree(rh);
drivers/md/dm-region-hash.c
217
INIT_LIST_HEAD(rh->buckets + i);
drivers/md/dm-region-hash.c
219
spin_lock_init(&rh->region_lock);
drivers/md/dm-region-hash.c
220
sema_init(&rh->recovery_count, 0);
drivers/md/dm-region-hash.c
221
atomic_set(&rh->recovery_in_flight, 0);
drivers/md/dm-region-hash.c
222
INIT_LIST_HEAD(&rh->clean_regions);
drivers/md/dm-region-hash.c
223
INIT_LIST_HEAD(&rh->quiesced_regions);
drivers/md/dm-region-hash.c
224
INIT_LIST_HEAD(&rh->recovered_regions);
drivers/md/dm-region-hash.c
225
INIT_LIST_HEAD(&rh->failed_recovered_regions);
drivers/md/dm-region-hash.c
226
rh->flush_failure = 0;
drivers/md/dm-region-hash.c
228
ret = mempool_init_kmalloc_pool(&rh->region_pool, MIN_REGIONS,
drivers/md/dm-region-hash.c
231
vfree(rh->buckets);
drivers/md/dm-region-hash.c
232
kfree(rh);
drivers/md/dm-region-hash.c
233
rh = ERR_PTR(-ENOMEM);
drivers/md/dm-region-hash.c
236
return rh;
drivers/md/dm-region-hash.c
240
void dm_region_hash_destroy(struct dm_region_hash *rh)
drivers/md/dm-region-hash.c
245
BUG_ON(!list_empty(&rh->quiesced_regions));
drivers/md/dm-region-hash.c
246
for (h = 0; h < rh->nr_buckets; h++) {
drivers/md/dm-region-hash.c
247
list_for_each_entry_safe(reg, nreg, rh->buckets + h,
drivers/md/dm-region-hash.c
250
mempool_free(reg, &rh->region_pool);
drivers/md/dm-region-hash.c
254
if (rh->log)
drivers/md/dm-region-hash.c
255
dm_dirty_log_destroy(rh->log);
drivers/md/dm-region-hash.c
257
mempool_exit(&rh->region_pool);
drivers/md/dm-region-hash.c
258
vfree(rh->buckets);
drivers/md/dm-region-hash.c
259
kfree(rh);
drivers/md/dm-region-hash.c
263
struct dm_dirty_log *dm_rh_dirty_log(struct dm_region_hash *rh)
drivers/md/dm-region-hash.c
265
return rh->log;
drivers/md/dm-region-hash.c
269
static unsigned int rh_hash(struct dm_region_hash *rh, region_t region)
drivers/md/dm-region-hash.c
271
return (unsigned int) ((region * rh->prime) >> rh->shift) & rh->mask;
drivers/md/dm-region-hash.c
274
static struct dm_region *__rh_lookup(struct dm_region_hash *rh, region_t region)
drivers/md/dm-region-hash.c
277
struct list_head *bucket = rh->buckets + rh_hash(rh, region);
drivers/md/dm-region-hash.c
286
static void __rh_insert(struct dm_region_hash *rh, struct dm_region *reg)
drivers/md/dm-region-hash.c
288
list_add(&reg->hash_list, rh->buckets + rh_hash(rh, reg->key));
drivers/md/dm-region-hash.c
291
static struct dm_region *__rh_alloc(struct dm_region_hash *rh, region_t region)
drivers/md/dm-region-hash.c
295
nreg = mempool_alloc(&rh->region_pool, GFP_ATOMIC);
drivers/md/dm-region-hash.c
299
nreg->state = rh->log->type->in_sync(rh->log, region, 1) ?
drivers/md/dm-region-hash.c
301
nreg->rh = rh;
drivers/md/dm-region-hash.c
307
write_lock_irq(&rh->hash_lock);
drivers/md/dm-region-hash.c
308
reg = __rh_lookup(rh, region);
drivers/md/dm-region-hash.c
311
mempool_free(nreg, &rh->region_pool);
drivers/md/dm-region-hash.c
313
__rh_insert(rh, nreg);
drivers/md/dm-region-hash.c
315
spin_lock(&rh->region_lock);
drivers/md/dm-region-hash.c
316
list_add(&nreg->list, &rh->clean_regions);
drivers/md/dm-region-hash.c
317
spin_unlock(&rh->region_lock);
drivers/md/dm-region-hash.c
322
write_unlock_irq(&rh->hash_lock);
drivers/md/dm-region-hash.c
327
static struct dm_region *__rh_find(struct dm_region_hash *rh, region_t region)
drivers/md/dm-region-hash.c
331
reg = __rh_lookup(rh, region);
drivers/md/dm-region-hash.c
333
read_unlock(&rh->hash_lock);
drivers/md/dm-region-hash.c
334
reg = __rh_alloc(rh, region);
drivers/md/dm-region-hash.c
335
read_lock(&rh->hash_lock);
drivers/md/dm-region-hash.c
341
int dm_rh_get_state(struct dm_region_hash *rh, region_t region, int may_block)
drivers/md/dm-region-hash.c
346
read_lock(&rh->hash_lock);
drivers/md/dm-region-hash.c
347
reg = __rh_lookup(rh, region);
drivers/md/dm-region-hash.c
348
read_unlock(&rh->hash_lock);
drivers/md/dm-region-hash.c
357
r = rh->log->type->in_sync(rh->log, region, may_block);
drivers/md/dm-region-hash.c
369
struct dm_region_hash *rh = reg->rh;
drivers/md/dm-region-hash.c
371
rh->log->type->set_region_sync(rh->log, reg->key, success);
drivers/md/dm-region-hash.c
382
rh->dispatch_bios(rh->context, &reg->delayed_bios);
drivers/md/dm-region-hash.c
383
if (atomic_dec_and_test(&rh->recovery_in_flight))
drivers/md/dm-region-hash.c
384
rh->wakeup_all_recovery_waiters(rh->context);
drivers/md/dm-region-hash.c
385
up(&rh->recovery_count);
drivers/md/dm-region-hash.c
398
void dm_rh_mark_nosync(struct dm_region_hash *rh, struct bio *bio)
drivers/md/dm-region-hash.c
401
struct dm_dirty_log *log = rh->log;
drivers/md/dm-region-hash.c
403
region_t region = dm_rh_bio_to_region(rh, bio);
drivers/md/dm-region-hash.c
407
rh->flush_failure = 1;
drivers/md/dm-region-hash.c
417
read_lock(&rh->hash_lock);
drivers/md/dm-region-hash.c
418
reg = __rh_find(rh, region);
drivers/md/dm-region-hash.c
419
read_unlock(&rh->hash_lock);
drivers/md/dm-region-hash.c
425
spin_lock_irqsave(&rh->region_lock, flags);
drivers/md/dm-region-hash.c
436
spin_unlock_irqrestore(&rh->region_lock, flags);
drivers/md/dm-region-hash.c
443
void dm_rh_update_states(struct dm_region_hash *rh, int errors_handled)
drivers/md/dm-region-hash.c
454
write_lock_irq(&rh->hash_lock);
drivers/md/dm-region-hash.c
455
spin_lock(&rh->region_lock);
drivers/md/dm-region-hash.c
456
if (!list_empty(&rh->clean_regions)) {
drivers/md/dm-region-hash.c
457
list_splice_init(&rh->clean_regions, &clean);
drivers/md/dm-region-hash.c
463
if (!list_empty(&rh->recovered_regions)) {
drivers/md/dm-region-hash.c
464
list_splice_init(&rh->recovered_regions, &recovered);
drivers/md/dm-region-hash.c
470
if (!list_empty(&rh->failed_recovered_regions)) {
drivers/md/dm-region-hash.c
471
list_splice_init(&rh->failed_recovered_regions,
drivers/md/dm-region-hash.c
478
spin_unlock(&rh->region_lock);
drivers/md/dm-region-hash.c
479
write_unlock_irq(&rh->hash_lock);
drivers/md/dm-region-hash.c
487
rh->log->type->clear_region(rh->log, reg->key);
drivers/md/dm-region-hash.c
489
mempool_free(reg, &rh->region_pool);
drivers/md/dm-region-hash.c
494
mempool_free(reg, &rh->region_pool);
drivers/md/dm-region-hash.c
498
rh->log->type->clear_region(rh->log, reg->key);
drivers/md/dm-region-hash.c
499
mempool_free(reg, &rh->region_pool);
drivers/md/dm-region-hash.c
502
rh->log->type->flush(rh->log);
drivers/md/dm-region-hash.c
506
static void rh_inc(struct dm_region_hash *rh, region_t region)
drivers/md/dm-region-hash.c
510
read_lock(&rh->hash_lock);
drivers/md/dm-region-hash.c
511
reg = __rh_find(rh, region);
drivers/md/dm-region-hash.c
513
spin_lock_irq(&rh->region_lock);
drivers/md/dm-region-hash.c
519
spin_unlock_irq(&rh->region_lock);
drivers/md/dm-region-hash.c
521
rh->log->type->mark_region(rh->log, reg->key);
drivers/md/dm-region-hash.c
523
spin_unlock_irq(&rh->region_lock);
drivers/md/dm-region-hash.c
526
read_unlock(&rh->hash_lock);
drivers/md/dm-region-hash.c
529
void dm_rh_inc_pending(struct dm_region_hash *rh, struct bio_list *bios)
drivers/md/dm-region-hash.c
536
rh_inc(rh, dm_rh_bio_to_region(rh, bio));
drivers/md/dm-region-hash.c
541
void dm_rh_dec(struct dm_region_hash *rh, region_t region)
drivers/md/dm-region-hash.c
547
read_lock(&rh->hash_lock);
drivers/md/dm-region-hash.c
548
reg = __rh_lookup(rh, region);
drivers/md/dm-region-hash.c
549
read_unlock(&rh->hash_lock);
drivers/md/dm-region-hash.c
551
spin_lock_irqsave(&rh->region_lock, flags);
drivers/md/dm-region-hash.c
565
if (unlikely(rh->flush_failure)) {
drivers/md/dm-region-hash.c
573
list_add_tail(&reg->list, &rh->quiesced_regions);
drivers/md/dm-region-hash.c
576
list_add(&reg->list, &rh->clean_regions);
drivers/md/dm-region-hash.c
580
spin_unlock_irqrestore(&rh->region_lock, flags);
drivers/md/dm-region-hash.c
583
rh->wakeup_workers(rh->context);
drivers/md/dm-region-hash.c
590
static int __rh_recovery_prepare(struct dm_region_hash *rh)
drivers/md/dm-region-hash.c
599
r = rh->log->type->get_resync_work(rh->log, &region);
drivers/md/dm-region-hash.c
607
read_lock(&rh->hash_lock);
drivers/md/dm-region-hash.c
608
reg = __rh_find(rh, region);
drivers/md/dm-region-hash.c
609
read_unlock(&rh->hash_lock);
drivers/md/dm-region-hash.c
611
spin_lock_irq(&rh->region_lock);
drivers/md/dm-region-hash.c
618
list_move(&reg->list, &rh->quiesced_regions);
drivers/md/dm-region-hash.c
620
spin_unlock_irq(&rh->region_lock);
drivers/md/dm-region-hash.c
625
void dm_rh_recovery_prepare(struct dm_region_hash *rh)
drivers/md/dm-region-hash.c
628
atomic_inc(&rh->recovery_in_flight);
drivers/md/dm-region-hash.c
630
while (!down_trylock(&rh->recovery_count)) {
drivers/md/dm-region-hash.c
631
atomic_inc(&rh->recovery_in_flight);
drivers/md/dm-region-hash.c
632
if (__rh_recovery_prepare(rh) <= 0) {
drivers/md/dm-region-hash.c
633
atomic_dec(&rh->recovery_in_flight);
drivers/md/dm-region-hash.c
634
up(&rh->recovery_count);
drivers/md/dm-region-hash.c
640
if (atomic_dec_and_test(&rh->recovery_in_flight))
drivers/md/dm-region-hash.c
641
rh->wakeup_all_recovery_waiters(rh->context);
drivers/md/dm-region-hash.c
648
struct dm_region *dm_rh_recovery_start(struct dm_region_hash *rh)
drivers/md/dm-region-hash.c
652
spin_lock_irq(&rh->region_lock);
drivers/md/dm-region-hash.c
653
if (!list_empty(&rh->quiesced_regions)) {
drivers/md/dm-region-hash.c
654
reg = list_entry(rh->quiesced_regions.next,
drivers/md/dm-region-hash.c
658
spin_unlock_irq(&rh->region_lock);
drivers/md/dm-region-hash.c
666
struct dm_region_hash *rh = reg->rh;
drivers/md/dm-region-hash.c
668
spin_lock_irq(&rh->region_lock);
drivers/md/dm-region-hash.c
670
list_add(&reg->list, &reg->rh->recovered_regions);
drivers/md/dm-region-hash.c
672
list_add(&reg->list, &reg->rh->failed_recovered_regions);
drivers/md/dm-region-hash.c
674
spin_unlock_irq(&rh->region_lock);
drivers/md/dm-region-hash.c
676
rh->wakeup_workers(rh->context);
drivers/md/dm-region-hash.c
681
int dm_rh_recovery_in_flight(struct dm_region_hash *rh)
drivers/md/dm-region-hash.c
683
return atomic_read(&rh->recovery_in_flight);
drivers/md/dm-region-hash.c
687
int dm_rh_flush(struct dm_region_hash *rh)
drivers/md/dm-region-hash.c
689
return rh->log->type->flush(rh->log);
drivers/md/dm-region-hash.c
693
void dm_rh_delay(struct dm_region_hash *rh, struct bio *bio)
drivers/md/dm-region-hash.c
697
read_lock(&rh->hash_lock);
drivers/md/dm-region-hash.c
698
reg = __rh_find(rh, dm_rh_bio_to_region(rh, bio));
drivers/md/dm-region-hash.c
700
read_unlock(&rh->hash_lock);
drivers/md/dm-region-hash.c
704
void dm_rh_stop_recovery(struct dm_region_hash *rh)
drivers/md/dm-region-hash.c
709
for (i = 0; i < rh->max_recovery; i++)
drivers/md/dm-region-hash.c
710
down(&rh->recovery_count);
drivers/md/dm-region-hash.c
714
void dm_rh_start_recovery(struct dm_region_hash *rh)
drivers/md/dm-region-hash.c
718
for (i = 0; i < rh->max_recovery; i++)
drivers/md/dm-region-hash.c
719
up(&rh->recovery_count);
drivers/md/dm-region-hash.c
721
rh->wakeup_workers(rh->context);
drivers/media/usb/uvc/uvc_v4l2.c
245
u16 rw, rh;
drivers/media/usb/uvc/uvc_v4l2.c
282
rh = fmt->fmt.pix.height;
drivers/media/usb/uvc/uvc_v4l2.c
289
d = min(w, rw) * min(h, rh);
drivers/media/usb/uvc/uvc_v4l2.c
290
d = w*h + rw*rh - 2*d;
drivers/net/ethernet/broadcom/b44.c
637
struct rx_header *rh;
drivers/net/ethernet/broadcom/b44.c
681
rh = (struct rx_header *) skb->data;
drivers/net/ethernet/broadcom/b44.c
683
rh->len = 0;
drivers/net/ethernet/broadcom/b44.c
684
rh->flags = 0;
drivers/net/ethernet/broadcom/b44.c
712
struct rx_header *rh;
drivers/net/ethernet/broadcom/b44.c
723
rh = (struct rx_header *) src_map->skb->data;
drivers/net/ethernet/broadcom/b44.c
724
rh->len = 0;
drivers/net/ethernet/broadcom/b44.c
725
rh->flags = 0;
drivers/net/ethernet/broadcom/b44.c
768
struct rx_header *rh;
drivers/net/ethernet/broadcom/b44.c
774
rh = (struct rx_header *) skb->data;
drivers/net/ethernet/broadcom/b44.c
775
len = le16_to_cpu(rh->len);
drivers/net/ethernet/broadcom/b44.c
777
(rh->flags & cpu_to_le16(RX_FLAG_ERRORS))) {
drivers/net/ethernet/broadcom/b44.c
791
len = le16_to_cpu(rh->len);
drivers/net/ethernet/cavium/liquidio/lio_core.c
580
union octeon_rh *rh,
drivers/net/ethernet/cavium/liquidio/lio_core.c
635
r_dh_off = (rh->r_dh.len - 1) * BYTES_PER_DHLEN_UNIT;
drivers/net/ethernet/cavium/liquidio/lio_core.c
638
if (rh->r_dh.has_hwtstamp) {
drivers/net/ethernet/cavium/liquidio/lio_core.c
659
if (rh->r_dh.has_hash) {
drivers/net/ethernet/cavium/liquidio/lio_core.c
667
skb_pull(skb, rh->r_dh.len * BYTES_PER_DHLEN_UNIT);
drivers/net/ethernet/cavium/liquidio/lio_core.c
671
(((rh->r_dh.encap_on) &&
drivers/net/ethernet/cavium/liquidio/lio_core.c
672
(rh->r_dh.csum_verified & CNNIC_TUN_CSUM_VERIFIED)) ||
drivers/net/ethernet/cavium/liquidio/lio_core.c
673
(!(rh->r_dh.encap_on) &&
drivers/net/ethernet/cavium/liquidio/lio_core.c
674
((rh->r_dh.csum_verified & CNNIC_CSUM_VERIFIED) ==
drivers/net/ethernet/cavium/liquidio/lio_core.c
684
if (rh->r_dh.encap_on) {
drivers/net/ethernet/cavium/liquidio/lio_core.c
692
rh->r_dh.vlan) {
drivers/net/ethernet/cavium/liquidio/lio_core.c
693
u16 priority = rh->r_dh.priority;
drivers/net/ethernet/cavium/liquidio/lio_core.c
694
u16 vid = rh->r_dh.vlan;
drivers/net/ethernet/cavium/liquidio/lio_core.c
703
rh->r_dh.len * BYTES_PER_DHLEN_UNIT;
drivers/net/ethernet/cavium/liquidio/lio_ethtool.c
38
u64 rh;
drivers/net/ethernet/cavium/liquidio/lio_ethtool.c
44
u64 rh;
drivers/net/ethernet/cavium/liquidio/lio_main.c
3293
recv_pkt->rh.r_nic_info.gmxport);
drivers/net/ethernet/cavium/liquidio/lio_main.c
3297
gmxport = recv_pkt->rh.r_nic_info.gmxport;
drivers/net/ethernet/cavium/liquidio/lio_main.c
3990
notice = recv_pkt->rh.r.ossp;
drivers/net/ethernet/cavium/liquidio/lio_main.c
96
u64 rh;
drivers/net/ethernet/cavium/liquidio/lio_vf_main.c
1890
recv_pkt->rh.r_nic_info.gmxport);
drivers/net/ethernet/cavium/liquidio/lio_vf_main.c
1894
gmxport = recv_pkt->rh.r_nic_info.gmxport;
drivers/net/ethernet/cavium/liquidio/lio_vf_main.c
43
u64 rh;
drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c
299
union octeon_rh *rh;
drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c
308
rh = &recv_pkt->rh;
drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c
309
ifidx = rh->r.ossp;
drivers/net/ethernet/cavium/liquidio/lio_vf_rep.c
327
skb_pull(skb, rh->r_dh.len * BYTES_PER_DHLEN_UNIT);
drivers/net/ethernet/cavium/liquidio/liquidio_common.h
1035
u64 rh;
drivers/net/ethernet/cavium/liquidio/liquidio_common.h
904
static inline int opcode_slow_path(union octeon_rh *rh)
drivers/net/ethernet/cavium/liquidio/liquidio_common.h
908
subcode1 = OPCODE_SUBCODE((rh)->r.opcode, (rh)->r.subcode);
drivers/net/ethernet/cavium/liquidio/octeon_device.c
1222
(u32)recv_pkt->rh.r_core_drv_init.app_mode),
drivers/net/ethernet/cavium/liquidio/octeon_device.c
1224
oct->app_mode = (u32)recv_pkt->rh.r_core_drv_init.app_mode;
drivers/net/ethernet/cavium/liquidio/octeon_device.c
1225
if (recv_pkt->rh.r_core_drv_init.app_mode == CVM_DRV_NIC_APP) {
drivers/net/ethernet/cavium/liquidio/octeon_device.c
1227
(u32)recv_pkt->rh.r_core_drv_init.max_nic_ports;
drivers/net/ethernet/cavium/liquidio/octeon_device.c
1229
(u32)recv_pkt->rh.r_core_drv_init.num_gmx_ports;
drivers/net/ethernet/cavium/liquidio/octeon_device.c
1238
oct->fw_info.app_cap_flags = recv_pkt->rh.r_core_drv_init.app_cap_flags;
drivers/net/ethernet/cavium/liquidio/octeon_device.c
1239
oct->fw_info.app_mode = (u32)recv_pkt->rh.r_core_drv_init.app_mode;
drivers/net/ethernet/cavium/liquidio/octeon_device.c
1240
oct->pfvf_hsword.app_mode = (u32)recv_pkt->rh.r_core_drv_init.app_mode;
drivers/net/ethernet/cavium/liquidio/octeon_device.c
1242
oct->pfvf_hsword.pkind = recv_pkt->rh.r_core_drv_init.pkind;
drivers/net/ethernet/cavium/liquidio/octeon_droq.c
348
recv_pkt->rh = info->rh;
drivers/net/ethernet/cavium/liquidio/octeon_droq.c
527
union octeon_rh *rh,
drivers/net/ethernet/cavium/liquidio/octeon_droq.c
536
disp_fn = octeon_get_dispatch(oct, (u16)rh->r.opcode,
drivers/net/ethernet/cavium/liquidio/octeon_droq.c
537
(u16)rh->r.subcode);
drivers/net/ethernet/cavium/liquidio/octeon_droq.c
545
rinfo->recv_pkt->rh = *rh;
drivers/net/ethernet/cavium/liquidio/octeon_droq.c
553
(unsigned int)rh->r.opcode,
drivers/net/ethernet/cavium/liquidio/octeon_droq.c
554
(unsigned int)rh->r.subcode);
drivers/net/ethernet/cavium/liquidio/octeon_droq.c
596
union octeon_rh *rh;
drivers/net/ethernet/cavium/liquidio/octeon_droq.c
621
rh = &info->rh;
drivers/net/ethernet/cavium/liquidio/octeon_droq.c
624
rh->r_dh.len += (ROUNDUP8(OCT_DROQ_INFO_SIZE) / sizeof(u64));
drivers/net/ethernet/cavium/liquidio/octeon_droq.c
626
if (opcode_slow_path(rh)) {
drivers/net/ethernet/cavium/liquidio/octeon_droq.c
629
buf_cnt = octeon_droq_dispatch_pkt(oct, droq, rh, info);
drivers/net/ethernet/cavium/liquidio/octeon_droq.c
689
rh, &droq->napi,
drivers/net/ethernet/cavium/liquidio/octeon_droq.c
763
(u16)rdisp->rinfo->recv_pkt->rh.r.opcode,
drivers/net/ethernet/cavium/liquidio/octeon_droq.c
764
(u16)rdisp->rinfo->recv_pkt->rh.r.subcode));
drivers/net/ethernet/cavium/liquidio/octeon_droq.c
814
(u16)rdisp->rinfo->recv_pkt->rh.r.opcode,
drivers/net/ethernet/cavium/liquidio/octeon_droq.c
815
(u16)rdisp->rinfo->recv_pkt->rh.r.subcode));
drivers/net/ethernet/cavium/liquidio/octeon_droq.h
148
union octeon_rh rh;
drivers/net/ethernet/cavium/liquidio/octeon_droq.h
58
union octeon_rh rh;
drivers/net/ethernet/cavium/liquidio/octeon_network.h
39
u64 rh;
drivers/net/ethernet/cavium/liquidio/octeon_network.h
69
u64 rh;
drivers/net/ethernet/cavium/liquidio/octeon_network.h
75
u64 rh;
drivers/net/ethernet/cavium/liquidio/octeon_network.h
86
u64 rh;
drivers/net/ethernet/cisco/enic/vnic_dev.c
34
struct vnic_resource_header __iomem *rh;
drivers/net/ethernet/cisco/enic/vnic_dev.c
47
rh = bar->vaddr;
drivers/net/ethernet/cisco/enic/vnic_dev.c
49
if (!rh) {
drivers/net/ethernet/cisco/enic/vnic_dev.c
55
if ((ioread32(&rh->magic) != VNIC_RES_MAGIC) ||
drivers/net/ethernet/cisco/enic/vnic_dev.c
56
(ioread32(&rh->version) != VNIC_RES_VERSION)) {
drivers/net/ethernet/cisco/enic/vnic_dev.c
62
ioread32(&rh->magic), ioread32(&rh->version));
drivers/net/ethernet/cisco/enic/vnic_dev.c
70
r = (struct vnic_resource __iomem *)(rh + 1);
drivers/net/ethernet/microchip/enc28j60.c
292
int rl, rh;
drivers/net/ethernet/microchip/enc28j60.c
296
rh = spi_read_op(priv, ENC28J60_READ_CTRL_REG, address + 1);
drivers/net/ethernet/microchip/enc28j60.c
298
return (rh << 8) | rl;
drivers/net/ethernet/sun/niu.c
3427
struct rx_pkt_hdr1 *rh;
drivers/net/ethernet/sun/niu.c
3489
len += sizeof(*rh);
drivers/net/ethernet/sun/niu.c
3490
len = min_t(int, len, sizeof(*rh) + VLAN_ETH_HLEN);
drivers/net/ethernet/sun/niu.c
3493
rh = (struct rx_pkt_hdr1 *) skb->data;
drivers/net/ethernet/sun/niu.c
3496
((u32)rh->hashval2_0 << 24 |
drivers/net/ethernet/sun/niu.c
3497
(u32)rh->hashval2_1 << 16 |
drivers/net/ethernet/sun/niu.c
3498
(u32)rh->hashval1_1 << 8 |
drivers/net/ethernet/sun/niu.c
3499
(u32)rh->hashval1_2 << 0),
drivers/net/ethernet/sun/niu.c
3501
skb_pull(skb, sizeof(*rh));
drivers/net/wireless/realtek/rtw89/debug.c
4636
const struct __diag_mac_rule_header *rh)
drivers/net/wireless/realtek/rtw89/debug.c
4638
switch (u8_get_bits(rh->io_band, __DIAG_MAC_IO)) {
drivers/net/wireless/realtek/rtw89/debug.c
4662
const struct __diag_mac_rule_header *rh)
drivers/net/wireless/realtek/rtw89/debug.c
4668
has_band = u8_get_bits(rh->io_band, __DIAG_MAC_HAS_BAND);
drivers/net/wireless/realtek/rtw89/debug.c
4672
band = u8_get_bits(rh->io_band, __DIAG_MAC_N_BAND);
drivers/net/wireless/realtek/rtw89/debug.c
4714
const struct __diag_mac_rule_header *rh = rule;
drivers/net/wireless/realtek/rtw89/debug.c
4715
u8 sheet = rh->sheet;
drivers/net/wireless/realtek/rtw89/debug.c
4716
u8 seq = rh->seq_major;
drivers/net/wireless/realtek/rtw89/debug.c
4718
if (!rtw89_mac_diag_match_hci(rtwdev, rh) ||
drivers/net/wireless/realtek/rtw89/debug.c
4719
!rtw89_mac_diag_match_band(rtwdev, rh)) {
drivers/net/wireless/realtek/rtw89/debug.c
4739
switch (rh->cmd) {
drivers/net/wireless/realtek/rtw89/debug.c
4750
p += scnprintf(p, end - p, "unknown rule cmd %u\n", rh->cmd);
drivers/net/wireless/realtek/rtw89/debug.c
4756
rule += rh->len * 4;
drivers/scsi/fnic/vnic_dev.c
102
r = (struct vnic_resource __iomem *)(rh + 1);
drivers/scsi/fnic/vnic_dev.c
78
struct vnic_resource_header __iomem *rh;
drivers/scsi/fnic/vnic_dev.c
87
rh = bar->vaddr;
drivers/scsi/fnic/vnic_dev.c
88
if (!rh) {
drivers/scsi/fnic/vnic_dev.c
93
if (ioread32(&rh->magic) != VNIC_RES_MAGIC ||
drivers/scsi/fnic/vnic_dev.c
94
ioread32(&rh->version) != VNIC_RES_VERSION) {
drivers/scsi/fnic/vnic_dev.c
98
ioread32(&rh->magic), ioread32(&rh->version));
drivers/scsi/lpfc/lpfc_ct.c
3208
struct lpfc_fdmi_reg_hba *rh;
drivers/scsi/lpfc/lpfc_ct.c
3270
rh = (struct lpfc_fdmi_reg_hba *)&CtReq->un;
drivers/scsi/lpfc/lpfc_ct.c
3272
memcpy(&rh->hi.PortName, &phba->pport->fc_sparam.portName,
drivers/scsi/lpfc/lpfc_ct.c
3279
rh->rpl.EntryCnt = cpu_to_be32(1);
drivers/scsi/lpfc/lpfc_ct.c
3280
memcpy(&rh->rpl.pe.PortName,
drivers/scsi/lpfc/lpfc_ct.c
3286
ab = (struct lpfc_fdmi_attr_block *)((uint8_t *)rh + size);
drivers/scsi/lpfc/lpfc_ct.c
3300
addsz = func(vport, ((uint8_t *)rh + size));
drivers/scsi/lpfc/lpfc_ct.c
3331
rh = (struct lpfc_fdmi_reg_hba *)base;
drivers/scsi/lpfc/lpfc_ct.c
3333
memcpy(&rh->hi.PortName,
drivers/scsi/snic/vnic_dev.c
69
struct vnic_resource_header __iomem *rh;
drivers/scsi/snic/vnic_dev.c
82
rh = bar->vaddr;
drivers/scsi/snic/vnic_dev.c
83
if (!rh) {
drivers/scsi/snic/vnic_dev.c
89
if (ioread32(&rh->magic) != VNIC_RES_MAGIC ||
drivers/scsi/snic/vnic_dev.c
90
ioread32(&rh->version) != VNIC_RES_VERSION) {
drivers/scsi/snic/vnic_dev.c
93
ioread32(&rh->magic), ioread32(&rh->version));
drivers/scsi/snic/vnic_dev.c
98
r = (struct vnic_resource __iomem *)(rh + 1);
drivers/soc/qcom/cmd-db.c
145
const struct rsc_hdr **rh)
drivers/soc/qcom/cmd-db.c
168
if (rh)
drivers/soc/qcom/cmd-db.c
169
*rh = rsc_hdr;
drivers/usb/gadget/function/uvc_v4l2.c
147
u16 rw, u16 rh)
drivers/usb/gadget/function/uvc_v4l2.c
171
d = min(w, rw) * min(h, rh);
drivers/usb/gadget/function/uvc_v4l2.c
172
d = w*h + rw*rh - 2*d;
drivers/usb/gadget/function/uvc_v4l2.c
184
uvcg_dbg(&video->uvc->func, "Unsupported size %ux%u\n", rw, rh);
drivers/usb/host/ohci-hub.c
542
u32 rh = roothub_a (ohci);
drivers/usb/host/ohci-hub.c
546
desc->bPwrOn2PwrGood = (rh & RH_A_POTPGT) >> 24;
drivers/usb/host/ohci-hub.c
554
if (rh & RH_A_NPS) /* no power switching? */
drivers/usb/host/ohci-hub.c
556
if (rh & RH_A_PSM) /* per-port power switching? */
drivers/usb/host/ohci-hub.c
558
if (rh & RH_A_NOCP) /* no overcurrent reporting? */
drivers/usb/host/ohci-hub.c
560
else if (rh & RH_A_OCPM) /* per-port overcurrent reporting? */
drivers/usb/host/ohci-hub.c
565
rh = roothub_b (ohci);
drivers/usb/host/ohci-hub.c
568
desc->u.hs.DeviceRemovable[0] = rh & RH_B_DR;
drivers/usb/host/ohci-hub.c
570
desc->u.hs.DeviceRemovable[1] = (rh & RH_B_DR) >> 8;
drivers/usb/host/ohci-omap.c
144
u32 rh = roothub_a (ohci);
drivers/usb/host/ohci-omap.c
147
rh &= ~RH_A_NPS;
drivers/usb/host/ohci-omap.c
153
rh &= ~RH_A_NOCP;
drivers/usb/host/ohci-omap.c
161
ohci_writel(ohci, rh, &ohci->regs->roothub.a);
drivers/usb/host/r8a66597-hcd.c
1001
rh->port |= USB_PORT_STAT_CONNECTION;
drivers/usb/host/r8a66597-hcd.c
1003
rh->port &= ~USB_PORT_STAT_CONNECTION;
drivers/usb/host/r8a66597-hcd.c
1004
rh->port |= USB_PORT_STAT_C_CONNECTION << 16;
drivers/usb/host/r8a66597-hcd.c
1040
struct r8a66597_root_hub *rh = &r8a66597->root_hub[port];
drivers/usb/host/r8a66597-hcd.c
1042
rh->port &= ~(USB_PORT_STAT_HIGH_SPEED | USB_PORT_STAT_LOW_SPEED);
drivers/usb/host/r8a66597-hcd.c
1044
rh->port |= USB_PORT_STAT_HIGH_SPEED;
drivers/usb/host/r8a66597-hcd.c
1046
rh->port |= USB_PORT_STAT_LOW_SPEED;
drivers/usb/host/r8a66597-hcd.c
1048
rh->port &= ~USB_PORT_STAT_RESET;
drivers/usb/host/r8a66597-hcd.c
1049
rh->port |= USB_PORT_STAT_ENABLE;
drivers/usb/host/r8a66597-hcd.c
1686
struct r8a66597_root_hub *rh = &r8a66597->root_hub[port];
drivers/usb/host/r8a66597-hcd.c
1688
if (rh->port & USB_PORT_STAT_RESET) {
drivers/usb/host/r8a66597-hcd.c
1700
if (!(rh->port & USB_PORT_STAT_CONNECTION)) {
drivers/usb/host/r8a66597-hcd.c
1705
if (rh->scount > 0) {
drivers/usb/host/r8a66597-hcd.c
1707
if (tmp == rh->old_syssts) {
drivers/usb/host/r8a66597-hcd.c
1708
rh->scount--;
drivers/usb/host/r8a66597-hcd.c
1709
if (rh->scount == 0)
drivers/usb/host/r8a66597-hcd.c
1714
rh->scount = R8A66597_MAX_SAMPLING;
drivers/usb/host/r8a66597-hcd.c
1715
rh->old_syssts = tmp;
drivers/usb/host/r8a66597-hcd.c
2145
struct r8a66597_root_hub *rh = &r8a66597->root_hub[port];
drivers/usb/host/r8a66597-hcd.c
2170
rh->port &= ~USB_PORT_STAT_POWER;
drivers/usb/host/r8a66597-hcd.c
2186
rh->port &= ~(1 << wValue);
drivers/usb/host/r8a66597-hcd.c
2198
*(__le32 *)buf = cpu_to_le32(rh->port);
drivers/usb/host/r8a66597-hcd.c
2211
rh->port |= USB_PORT_STAT_POWER;
drivers/usb/host/r8a66597-hcd.c
2214
struct r8a66597_device *dev = rh->dev;
drivers/usb/host/r8a66597-hcd.c
2216
rh->port |= USB_PORT_STAT_RESET;
drivers/usb/host/r8a66597-hcd.c
2230
rh->port |= 1 << wValue;
drivers/usb/host/r8a66597-hcd.c
2251
struct r8a66597_root_hub *rh = &r8a66597->root_hub[port];
drivers/usb/host/r8a66597-hcd.c
2254
if (!(rh->port & USB_PORT_STAT_ENABLE))
drivers/usb/host/r8a66597-hcd.c
2257
dev_dbg(&rh->dev->udev->dev, "suspend port = %d\n", port);
drivers/usb/host/r8a66597-hcd.c
2259
rh->port |= USB_PORT_STAT_SUSPEND;
drivers/usb/host/r8a66597-hcd.c
2261
if (rh->dev->udev->do_remote_wakeup) {
drivers/usb/host/r8a66597-hcd.c
2282
struct r8a66597_root_hub *rh = &r8a66597->root_hub[port];
drivers/usb/host/r8a66597-hcd.c
2285
if (!(rh->port & USB_PORT_STAT_SUSPEND))
drivers/usb/host/r8a66597-hcd.c
2288
dev_dbg(&rh->dev->udev->dev, "resume port = %d\n", port);
drivers/usb/host/r8a66597-hcd.c
2289
rh->port &= ~USB_PORT_STAT_SUSPEND;
drivers/usb/host/r8a66597-hcd.c
2290
rh->port |= USB_PORT_STAT_C_SUSPEND << 16;
drivers/usb/host/r8a66597-hcd.c
2349
struct r8a66597_root_hub *rh = &r8a66597->root_hub[port];
drivers/usb/host/r8a66597-hcd.c
2351
rh->port = 0x00000000;
drivers/usb/host/r8a66597-hcd.c
996
struct r8a66597_root_hub *rh = &r8a66597->root_hub[port];
drivers/usb/host/r8a66597-hcd.c
998
rh->old_syssts = r8a66597_read(r8a66597, get_syssts_reg(port)) & LNST;
drivers/usb/host/r8a66597-hcd.c
999
rh->scount = R8A66597_MAX_SAMPLING;
fs/ntfs3/fslog.c
1466
static int next_log_lsn(struct ntfs_log *log, const struct LFS_RECORD_HDR *rh,
fs/ntfs3/fslog.c
1470
u64 this_lsn = le64_to_cpu(rh->this_lsn);
fs/ntfs3/fslog.c
1473
final_log_off(log, this_lsn, le32_to_cpu(rh->client_data_len));
fs/ntfs3/fslog.c
2253
const struct LFS_RECORD_HDR *rh, void *buffer)
fs/ntfs3/fslog.c
2257
u64 lsn = le64_to_cpu(rh->this_lsn);
fs/ntfs3/fslog.c
2260
u32 data_len = le32_to_cpu(rh->client_data_len);
fs/ntfs3/fslog.c
2325
struct LFS_RECORD_HDR *rh = NULL;
fs/ntfs3/fslog.c
2340
(struct RECORD_PAGE_HDR **)&rh, NULL);
fs/ntfs3/fslog.c
2345
lsnr = le64_to_cpu(rh->this_lsn);
fs/ntfs3/fslog.c
2354
len = le32_to_cpu(rh->client_data_len);
fs/ntfs3/fslog.c
2373
err = read_log_rec_buf(log, rh, rst);
fs/ntfs3/fslog.c
2381
kfree(rh);
fs/ntfs3/fslog.c
2390
struct LFS_RECORD_HDR *rh = lcb->lrh;
fs/ntfs3/fslog.c
2394
if (!rh) {
fs/ntfs3/fslog.c
2396
(struct RECORD_PAGE_HDR **)&rh, NULL);
fs/ntfs3/fslog.c
2398
lcb->lrh = rh;
fs/ntfs3/fslog.c
2407
if (lsn != le64_to_cpu(rh->this_lsn))
fs/ntfs3/fslog.c
2410
len = le32_to_cpu(rh->client_data_len);
fs/ntfs3/fslog.c
2424
if (rh->flags & LOG_RECORD_MULTI_PAGE) {
fs/ntfs3/fslog.c
2434
err = read_log_rec_buf(log, rh, lr);
fs/ntfs3/fslog.c
2444
lcb->log_rec = Add2Ptr(rh, sizeof(struct LFS_RECORD_HDR));
fs/ntfs3/fslog.c
3772
struct RESTART_HDR *rh;
fs/ntfs3/fslog.c
5145
rh = kzalloc(log->page_size, GFP_NOFS);
fs/ntfs3/fslog.c
5146
if (!rh) {
fs/ntfs3/fslog.c
5151
rh->rhdr.sign = NTFS_RSTR_SIGNATURE;
fs/ntfs3/fslog.c
5152
rh->rhdr.fix_off = cpu_to_le16(offsetof(struct RESTART_HDR, fixups));
fs/ntfs3/fslog.c
5154
rh->rhdr.fix_num = cpu_to_le16(t16);
fs/ntfs3/fslog.c
5155
rh->sys_page_size = cpu_to_le32(log->page_size);
fs/ntfs3/fslog.c
5156
rh->page_size = cpu_to_le32(log->page_size);
fs/ntfs3/fslog.c
5160
rh->ra_off = cpu_to_le16(t16);
fs/ntfs3/fslog.c
5161
rh->minor_ver = cpu_to_le16(1); // 0x1A:
fs/ntfs3/fslog.c
5162
rh->major_ver = cpu_to_le16(1); // 0x1C:
fs/ntfs3/fslog.c
5164
ra2 = Add2Ptr(rh, t16);
fs/ntfs3/fslog.c
5173
ntfs_fix_pre_write(&rh->rhdr, log->page_size);
fs/ntfs3/fslog.c
5175
err = ntfs_sb_write_run(sbi, &ni->file.run, 0, rh, log->page_size, 0);
fs/ntfs3/fslog.c
5178
rh, log->page_size, 0);
fs/ntfs3/fslog.c
5180
kfree(rh);
fs/xfs/xfs_log_recover.c
369
xlog_logrec_hblks(struct xlog *log, struct xlog_rec_header *rh)
fs/xfs/xfs_log_recover.c
372
int h_size = be32_to_cpu(rh->h_size);
fs/xfs/xfs_log_recover.c
374
if ((be32_to_cpu(rh->h_version) & XLOG_VERSION_2) &&
include/linux/dm-region-hash.h
100
int dm_rh_recovery_in_flight(struct dm_region_hash *rh);
include/linux/dm-region-hash.h
103
void dm_rh_start_recovery(struct dm_region_hash *rh);
include/linux/dm-region-hash.h
104
void dm_rh_stop_recovery(struct dm_region_hash *rh);
include/linux/dm-region-hash.h
46
void dm_region_hash_destroy(struct dm_region_hash *rh);
include/linux/dm-region-hash.h
48
struct dm_dirty_log *dm_rh_dirty_log(struct dm_region_hash *rh);
include/linux/dm-region-hash.h
53
region_t dm_rh_bio_to_region(struct dm_region_hash *rh, struct bio *bio);
include/linux/dm-region-hash.h
54
sector_t dm_rh_region_to_sector(struct dm_region_hash *rh, region_t region);
include/linux/dm-region-hash.h
60
sector_t dm_rh_get_region_size(struct dm_region_hash *rh);
include/linux/dm-region-hash.h
67
int dm_rh_get_state(struct dm_region_hash *rh, region_t region, int may_block);
include/linux/dm-region-hash.h
68
void dm_rh_set_state(struct dm_region_hash *rh, region_t region,
include/linux/dm-region-hash.h
72
void dm_rh_update_states(struct dm_region_hash *rh, int errors_handled);
include/linux/dm-region-hash.h
75
int dm_rh_flush(struct dm_region_hash *rh);
include/linux/dm-region-hash.h
78
void dm_rh_inc_pending(struct dm_region_hash *rh, struct bio_list *bios);
include/linux/dm-region-hash.h
79
void dm_rh_dec(struct dm_region_hash *rh, region_t region);
include/linux/dm-region-hash.h
82
void dm_rh_delay(struct dm_region_hash *rh, struct bio *bio);
include/linux/dm-region-hash.h
84
void dm_rh_mark_nosync(struct dm_region_hash *rh, struct bio *bio);
include/linux/dm-region-hash.h
91
void dm_rh_recovery_prepare(struct dm_region_hash *rh);
include/linux/dm-region-hash.h
94
struct dm_region *dm_rh_recovery_start(struct dm_region_hash *rh);
include/linux/fsl/bestcomm/sram.h
28
rh_info_t *rh;
include/linux/kprobes.h
154
struct rethook *rh;
include/linux/math64.h
215
} rl, rm, rn, rh, a0, b0;
include/linux/math64.h
224
rh.ll = mul_u32_u32(a0.l.high, b0.l.high);
include/linux/math64.h
232
rh.l.low = c = (c >> 32) + rm.l.high + rn.l.high + rh.l.low;
include/linux/math64.h
233
rh.l.high = (c >> 32) + rh.l.high;
include/linux/math64.h
242
return (rl.ll >> shift) | (rh.ll << (64 - shift));
include/linux/math64.h
243
return rh.ll >> (shift & 63);
include/linux/math64.h
279
} u, rl, rh;
include/linux/math64.h
283
rh.ll = mul_u32_u32(u.l.high, mul) + rl.l.high;
include/linux/math64.h
286
rl.l.high = do_div(rh.ll, divisor);
include/linux/math64.h
291
rl.l.high = rh.l.low;
include/linux/rethook.h
61
void rethook_stop(struct rethook *rh);
include/linux/rethook.h
62
void rethook_free(struct rethook *rh);
include/linux/rethook.h
63
struct rethook_node *rethook_try_get(struct rethook *rh);
include/math-emu/op-2.h
147
#define __FP_FRAC_ADD_2(rh, rl, xh, xl, yh, yl) \
include/math-emu/op-2.h
148
(rh = xh + yh + ((rl = xl + yl) < xl))
include/math-emu/op-2.h
151
#define __FP_FRAC_SUB_2(rh, rl, xh, xl, yh, yl) \
include/math-emu/op-2.h
152
(rh = xh - yh - ((rl = xl - yl) > xl))
kernel/fork.c
272
static void thread_stack_free_rcu(struct rcu_head *rh)
kernel/fork.c
274
struct vm_stack *vm_stack = container_of(rh, struct vm_stack, rcu);
kernel/fork.c
394
static void thread_stack_free_rcu(struct rcu_head *rh)
kernel/fork.c
396
__free_pages(virt_to_page(rh), THREAD_SIZE_ORDER);
kernel/fork.c
401
struct rcu_head *rh = tsk->stack;
kernel/fork.c
403
call_rcu(rh, thread_stack_free_rcu);
kernel/fork.c
428
static void thread_stack_free_rcu(struct rcu_head *rh)
kernel/fork.c
430
kmem_cache_free(thread_stack_cache, rh);
kernel/fork.c
435
struct rcu_head *rh = tsk->stack;
kernel/fork.c
437
call_rcu(rh, thread_stack_free_rcu);
kernel/kprobes.c
2183
rhn = rethook_try_get(rp->rh);
kernel/kprobes.c
2200
static void kretprobe_rethook_handler(struct rethook_node *rh, void *data,
kernel/kprobes.c
2216
ri = container_of(rh, struct kretprobe_instance, node);
kernel/kprobes.c
2288
rp->rh = rethook_alloc((void *)rp, kretprobe_rethook_handler,
kernel/kprobes.c
2291
if (IS_ERR(rp->rh))
kernel/kprobes.c
2292
return PTR_ERR(rp->rh);
kernel/kprobes.c
2298
rethook_free(rp->rh);
kernel/kprobes.c
2299
rp->rh = NULL;
kernel/kprobes.c
2360
rethook_free(rps[i]->rh);
kernel/rcu/rcuscale.c
714
struct rcu_head rh;
kernel/rcu/rcuscale.c
718
static void kfree_call_rcu(struct rcu_head *rh)
kernel/rcu/rcuscale.c
720
struct kfree_obj *obj = container_of(rh, struct kfree_obj, rh);
kernel/rcu/rcuscale.c
763
call_rcu(&(alloc_ptr->rh), kfree_call_rcu);
kernel/rcu/rcuscale.c
775
kfree_rcu(alloc_ptr, rh);
kernel/rcu/rcuscale.c
843
static void call_rcu_lazy_test1(struct rcu_head *rh)
kernel/rcu/rcutorture.c
3126
struct rcu_head rh;
kernel/rcu/rcutorture.c
3137
struct fwd_cb_state *fcsp = container_of(rhp, struct fwd_cb_state, rh);
kernel/rcu/rcutorture.c
3143
cur_ops->call(&fcsp->rh, rcu_torture_fwd_prog_cb);
kernel/rcu/rcutorture.c
3148
struct rcu_head rh;
kernel/rcu/rcutorture.c
3211
struct rcu_fwd_cb *rfcp = container_of(rhp, struct rcu_fwd_cb, rh);
kernel/rcu/rcutorture.c
3295
init_rcu_head_on_stack(&fcs.rh);
kernel/rcu/rcutorture.c
3304
cur_ops->call(&fcs.rh, rcu_torture_fwd_prog_cb);
kernel/rcu/rcutorture.c
3342
destroy_rcu_head_on_stack(&fcs.rh);
kernel/rcu/rcutorture.c
3415
cur_ops->call(&rfcp->rh, rcu_torture_fwd_cb_cr);
kernel/rcu/tasks.h
1453
struct rcu_head rh;
kernel/rcu/tasks.h
1476
container_of(rhp, struct rcu_tasks_test_desc, rh);
kernel/rcu/tasks.h
1490
call_rcu_tasks(&tests[0].rh, test_rcu_tasks_callback);
kernel/rcu/tasks.h
1502
call_rcu_tasks_trace(&tests[1].rh, test_rcu_tasks_callback);
kernel/rcu/update.c
602
struct rcu_head rh;
kernel/rcu/update.c
619
kfree_rcu(rhp, rh);
kernel/trace/rethook.c
106
struct rethook *rh;
kernel/trace/rethook.c
111
rh = kzalloc_obj(struct rethook);
kernel/trace/rethook.c
112
if (!rh)
kernel/trace/rethook.c
115
rh->data = data;
kernel/trace/rethook.c
116
rcu_assign_pointer(rh->handler, handler);
kernel/trace/rethook.c
119
if (objpool_init(&rh->pool, num, size, GFP_KERNEL, rh,
kernel/trace/rethook.c
121
kfree(rh);
kernel/trace/rethook.c
124
return rh;
kernel/trace/rethook.c
130
struct rethook *rh = node->rethook;
kernel/trace/rethook.c
132
objpool_drop(node, &rh->pool);
kernel/trace/rethook.c
161
struct rethook_node *rethook_try_get(struct rethook *rh)
kernel/trace/rethook.c
163
rethook_handler_t handler = rethook_get_handler(rh);
kernel/trace/rethook.c
180
return (struct rethook_node *)objpool_pop(&rh->pool);
kernel/trace/rethook.c
208
struct rethook_node *rh = NULL;
kernel/trace/rethook.c
217
rh = container_of(node, struct rethook_node, llist);
kernel/trace/rethook.c
218
if (rh->ret_addr != (unsigned long)arch_rethook_trampoline) {
kernel/trace/rethook.c
220
return rh->ret_addr;
kernel/trace/rethook.c
37
struct rethook *rh = container_of(head, struct rethook, rcu);
kernel/trace/rethook.c
38
objpool_fini(&rh->pool);
kernel/trace/rethook.c
49
void rethook_stop(struct rethook *rh)
kernel/trace/rethook.c
51
rcu_assign_pointer(rh->handler, NULL);
kernel/trace/rethook.c
64
void rethook_free(struct rethook *rh)
kernel/trace/rethook.c
66
rethook_stop(rh);
kernel/trace/rethook.c
68
call_rcu(&rh->rcu, rethook_free_rcu);
kernel/trace/rethook.c
85
static inline rethook_handler_t rethook_get_handler(struct rethook *rh)
kernel/trace/rethook.c
87
return (rethook_handler_t)rcu_dereference_check(rh->handler,
mm/cma.c
340
bool (*cmp)(struct cma_init_memrange *lh, struct cma_init_memrange *rh))
net/ipv6/exthdrs_core.c
224
struct ipv6_rt_hdr _rh, *rh;
net/ipv6/exthdrs_core.c
226
rh = skb_header_pointer(skb, start, sizeof(_rh),
net/ipv6/exthdrs_core.c
228
if (!rh)
net/ipv6/exthdrs_core.c
232
rh->segments_left == 0)
net/ipv6/netfilter/ip6t_rt.c
34
const struct ipv6_rt_hdr *rh;
net/ipv6/netfilter/ip6t_rt.c
51
rh = skb_header_pointer(skb, ptr, sizeof(_route), &_route);
net/ipv6/netfilter/ip6t_rt.c
52
if (rh == NULL) {
net/ipv6/netfilter/ip6t_rt.c
57
hdrlen = ipv6_optlen(rh);
net/ipv6/netfilter/ip6t_rt.c
64
rh->segments_left,
net/ipv6/netfilter/ip6t_rt.c
70
((rtinfo->rt_type == rh->type) ^
security/ipe/audit.c
72
static void audit_dmv_roothash(struct audit_buffer *ab, const void *rh)
security/ipe/audit.c
75
ipe_digest_audit(ab, rh);
sound/soc/codecs/rt1305.c
1057
regmap_read(rt1305->regmap, RT1305_PR_BASE + 0x55, &rh);
sound/soc/codecs/rt1305.c
1059
rhl = (rh << 16) | rl;
sound/soc/codecs/rt1305.c
1062
pr_debug("Left_rhl = 0x%x rh=0x%x rl=0x%x\n", rhl, rh, rl);
sound/soc/codecs/rt1305.c
1074
regmap_read(rt1305->regmap, RT1305_PR_BASE + 0x55, &rh);
sound/soc/codecs/rt1305.c
1076
rhl = (rh << 16) | rl;
sound/soc/codecs/rt1305.c
1079
pr_debug("Right_rhl = 0x%x rh=0x%x rl=0x%x\n", rhl, rh, rl);
sound/soc/codecs/rt1305.c
993
unsigned int rh, rl, rhl, r0ohm;
sound/usb/usx2y/usbus428ctldefs.h
67
rh,
tools/testing/selftests/kvm/x86/hyperv_clock.c
27
} rm, rn, rh, a0, b0;
tools/testing/selftests/kvm/x86/hyperv_clock.c
35
rh.ll = (u64)a0.l.high * b0.l.high;
tools/testing/selftests/kvm/x86/hyperv_clock.c
37
rh.l.low = c = rm.l.high + rn.l.high + rh.l.low;
tools/testing/selftests/kvm/x86/hyperv_clock.c
38
rh.l.high = (c >> 32) + rh.l.high;
tools/testing/selftests/kvm/x86/hyperv_clock.c
40
return rh.ll;