Symbol: seg
arch/arm/mach-omap1/board-osk.c
293
static void __init osk_init_cf(int seg)
arch/arm/mach-omap1/board-osk.c
299
switch (seg) {
arch/arm/mach-omap1/board-osk.c
313
osk5912_cf_device.dev.platform_data = (void *)(uintptr_t)seg;
arch/arm/mach-omap1/board-osk.c
320
seg, omap_readl(EMIFS_CCS(seg)), omap_readl(EMIFS_ACS(seg)));
arch/arm/mach-omap1/board-osk.c
321
omap_writel(0x0004a1b3, EMIFS_CCS(seg)); /* synch mode 4 etc */
arch/arm/mach-omap1/board-osk.c
322
omap_writel(0x00000000, EMIFS_ACS(seg)); /* OE hold/setup */
arch/loongarch/pci/acpi.c
159
u16 seg = root->segment;
arch/loongarch/pci/acpi.c
168
dev_err(dev, "%04x:%pR ECAM region not found, use default value\n", seg, bus_res);
arch/loongarch/pci/acpi.c
186
dev_err(dev, "%04x:%pR error %ld mapping ECAM\n", seg, bus_res, PTR_ERR(cfg));
arch/m68k/include/asm/tlbflush.h
142
unsigned char seg;
arch/m68k/include/asm/tlbflush.h
149
seg = sun3_get_segmap(i);
arch/m68k/include/asm/tlbflush.h
150
if (seg == SUN3_INVALID_PMEG)
arch/m68k/include/asm/tlbflush.h
154
pmeg_alloc[seg] = 0;
arch/m68k/include/asm/tlbflush.h
155
pmeg_ctx[seg] = 0;
arch/m68k/include/asm/tlbflush.h
156
pmeg_vaddr[seg] = 0;
arch/m68k/include/asm/tlbflush.h
189
unsigned char seg, oldctx;
arch/m68k/include/asm/tlbflush.h
198
if((seg = sun3_get_segmap(start)) == SUN3_INVALID_PMEG)
arch/m68k/include/asm/tlbflush.h
200
if(pmeg_ctx[seg] == mm->context) {
arch/m68k/include/asm/tlbflush.h
201
pmeg_alloc[seg] = 0;
arch/m68k/include/asm/tlbflush.h
202
pmeg_ctx[seg] = 0;
arch/m68k/include/asm/tlbflush.h
203
pmeg_vaddr[seg] = 0;
arch/m68k/sun3/mmu_emu.c
130
unsigned long seg, num;
arch/m68k/sun3/mmu_emu.c
153
for(seg = bootmem_end; seg < 0x0f800000; seg += SUN3_PMEG_SIZE) {
arch/m68k/sun3/mmu_emu.c
154
i = sun3_get_segmap(seg);
arch/m68k/sun3/mmu_emu.c
159
print_pte_vaddr (seg);
arch/m68k/sun3/mmu_emu.c
161
sun3_put_segmap(seg, SUN3_INVALID_PMEG);
arch/m68k/sun3/mmu_emu.c
166
for (num=0, seg=0x0F800000; seg<0x10000000; seg+=16*PAGE_SIZE) {
arch/m68k/sun3/mmu_emu.c
167
if (sun3_get_segmap (seg) != SUN3_INVALID_PMEG) {
arch/m68k/sun3/mmu_emu.c
171
print_pte_vaddr (seg + (i*PAGE_SIZE));
arch/m68k/sun3/mmu_emu.c
178
m68k_vmalloc_end = seg;
arch/m68k/sun3/mmu_emu.c
183
pmeg_alloc[sun3_get_segmap(seg)] = 2;
arch/m68k/sun3/mmu_emu.c
192
for(seg = 0; seg < PAGE_OFFSET; seg += SUN3_PMEG_SIZE)
arch/m68k/sun3/mmu_emu.c
193
sun3_put_segmap(seg, SUN3_INVALID_PMEG);
arch/m68k/sun3/mmu_emu.c
196
for(seg = 0; seg < 0x10000000; seg += SUN3_PMEG_SIZE) {
arch/m68k/sun3/mmu_emu.c
197
i = sun3_get_segmap(seg);
arch/m68k/sun3/mmu_emu.c
199
(*(romvec->pv_setctxt))(j, (void *)seg, i);
arch/s390/mm/extmem.c
167
query_segment_type (struct dcss_segment *seg)
arch/s390/mm/extmem.c
185
memcpy (qin->qname, seg->dcss_name, 8);
arch/s390/mm/extmem.c
205
seg->vm_segtype = qout->range[0].start & 0xff;
arch/s390/mm/extmem.c
224
seg->vm_segtype = SEG_TYPE_EWEN;
arch/s390/mm/extmem.c
228
seg->start_addr = qout->segstart;
arch/s390/mm/extmem.c
229
seg->end = qout->segend;
arch/s390/mm/extmem.c
231
memcpy (seg->range, qout->range, 6*sizeof(struct qrange));
arch/s390/mm/extmem.c
232
seg->segcnt = qout->segcnt;
arch/s390/mm/extmem.c
256
struct dcss_segment seg;
arch/s390/mm/extmem.c
261
dcss_mkname(name, seg.dcss_name);
arch/s390/mm/extmem.c
262
rc = query_segment_type (&seg);
arch/s390/mm/extmem.c
265
return seg.vm_segtype;
arch/s390/mm/extmem.c
273
segment_overlaps_others (struct dcss_segment *seg)
arch/s390/mm/extmem.c
281
if ((tmp->start_addr >> 20) > (seg->end >> 20))
arch/s390/mm/extmem.c
283
if ((tmp->end >> 20) < (seg->start_addr >> 20))
arch/s390/mm/extmem.c
285
if (seg == tmp)
arch/s390/mm/extmem.c
300
struct dcss_segment *seg;
arch/s390/mm/extmem.c
305
seg = kmalloc_obj(*seg, GFP_KERNEL | GFP_DMA);
arch/s390/mm/extmem.c
306
if (seg == NULL) {
arch/s390/mm/extmem.c
310
dcss_mkname (name, seg->dcss_name);
arch/s390/mm/extmem.c
311
rc = query_segment_type (seg);
arch/s390/mm/extmem.c
315
if (segment_overlaps_others(seg)) {
arch/s390/mm/extmem.c
320
seg->res = kzalloc_obj(struct resource);
arch/s390/mm/extmem.c
321
if (seg->res == NULL) {
arch/s390/mm/extmem.c
325
seg->res->flags = IORESOURCE_BUSY | IORESOURCE_MEM;
arch/s390/mm/extmem.c
326
seg->res->start = seg->start_addr;
arch/s390/mm/extmem.c
327
seg->res->end = seg->end;
arch/s390/mm/extmem.c
328
memcpy(&seg->res_name, seg->dcss_name, 8);
arch/s390/mm/extmem.c
329
EBCASC(seg->res_name, 8);
arch/s390/mm/extmem.c
330
seg->res_name[8] = '\0';
arch/s390/mm/extmem.c
331
strlcat(seg->res_name, " (DCSS)", sizeof(seg->res_name));
arch/s390/mm/extmem.c
332
seg->res->name = seg->res_name;
arch/s390/mm/extmem.c
333
segtype = seg->vm_segtype;
arch/s390/mm/extmem.c
336
seg->res->flags |= IORESOURCE_READONLY;
arch/s390/mm/extmem.c
339
if (request_resource(&iomem_resource, seg->res)) {
arch/s390/mm/extmem.c
344
rc = vmem_add_mapping(seg->start_addr, seg->end - seg->start_addr + 1);
arch/s390/mm/extmem.c
349
diag_cc = dcss_diag(&loadnsr_scode, seg->dcss_name,
arch/s390/mm/extmem.c
352
diag_cc = dcss_diag(&loadshr_scode, seg->dcss_name,
arch/s390/mm/extmem.c
355
dcss_diag(&purgeseg_scode, seg->dcss_name,
arch/s390/mm/extmem.c
363
dcss_diag(&purgeseg_scode, seg->dcss_name,
arch/s390/mm/extmem.c
367
seg->start_addr = start_addr;
arch/s390/mm/extmem.c
368
seg->end = end_addr;
arch/s390/mm/extmem.c
369
seg->do_nonshared = do_nonshared;
arch/s390/mm/extmem.c
370
refcount_set(&seg->ref_count, 1);
arch/s390/mm/extmem.c
371
list_add(&seg->list, &dcss_list);
arch/s390/mm/extmem.c
372
*addr = seg->start_addr;
arch/s390/mm/extmem.c
373
*end = seg->end;
arch/s390/mm/extmem.c
376
"exclusive-writable\n", name, (void*) seg->start_addr,
arch/s390/mm/extmem.c
377
(void*) seg->end, segtype_string[seg->vm_segtype]);
arch/s390/mm/extmem.c
380
"shared access mode\n", name, (void*) seg->start_addr,
arch/s390/mm/extmem.c
381
(void*) seg->end, segtype_string[seg->vm_segtype]);
arch/s390/mm/extmem.c
385
vmem_remove_mapping(seg->start_addr, seg->end - seg->start_addr + 1);
arch/s390/mm/extmem.c
387
release_resource(seg->res);
arch/s390/mm/extmem.c
389
kfree(seg->res);
arch/s390/mm/extmem.c
391
kfree(seg);
arch/s390/mm/extmem.c
418
struct dcss_segment *seg;
arch/s390/mm/extmem.c
425
seg = segment_by_name (name);
arch/s390/mm/extmem.c
426
if (seg == NULL)
arch/s390/mm/extmem.c
429
if (do_nonshared == seg->do_nonshared) {
arch/s390/mm/extmem.c
430
refcount_inc(&seg->ref_count);
arch/s390/mm/extmem.c
431
*addr = seg->start_addr;
arch/s390/mm/extmem.c
432
*end = seg->end;
arch/s390/mm/extmem.c
433
rc = seg->vm_segtype;
arch/s390/mm/extmem.c
459
struct dcss_segment *seg;
arch/s390/mm/extmem.c
465
seg = segment_by_name (name);
arch/s390/mm/extmem.c
466
if (seg == NULL) {
arch/s390/mm/extmem.c
470
if (do_nonshared == seg->do_nonshared) {
arch/s390/mm/extmem.c
476
if (refcount_read(&seg->ref_count) != 1) {
arch/s390/mm/extmem.c
481
release_resource(seg->res);
arch/s390/mm/extmem.c
483
seg->res->flags &= ~IORESOURCE_READONLY;
arch/s390/mm/extmem.c
485
if (seg->vm_segtype == SEG_TYPE_SR ||
arch/s390/mm/extmem.c
486
seg->vm_segtype == SEG_TYPE_ER)
arch/s390/mm/extmem.c
487
seg->res->flags |= IORESOURCE_READONLY;
arch/s390/mm/extmem.c
489
if (request_resource(&iomem_resource, seg->res)) {
arch/s390/mm/extmem.c
493
kfree(seg->res);
arch/s390/mm/extmem.c
497
dcss_diag(&purgeseg_scode, seg->dcss_name, &dummy, &dummy);
arch/s390/mm/extmem.c
499
diag_cc = dcss_diag(&loadnsr_scode, seg->dcss_name,
arch/s390/mm/extmem.c
502
diag_cc = dcss_diag(&loadshr_scode, seg->dcss_name,
arch/s390/mm/extmem.c
514
seg->start_addr = start_addr;
arch/s390/mm/extmem.c
515
seg->end = end_addr;
arch/s390/mm/extmem.c
516
seg->do_nonshared = do_nonshared;
arch/s390/mm/extmem.c
520
release_resource(seg->res);
arch/s390/mm/extmem.c
521
kfree(seg->res);
arch/s390/mm/extmem.c
523
vmem_remove_mapping(seg->start_addr, seg->end - seg->start_addr + 1);
arch/s390/mm/extmem.c
524
list_del(&seg->list);
arch/s390/mm/extmem.c
525
dcss_diag(&purgeseg_scode, seg->dcss_name, &dummy, &dummy);
arch/s390/mm/extmem.c
526
kfree(seg);
arch/s390/mm/extmem.c
534
struct dcss_segment *seg = (struct dcss_segment *)data;
arch/s390/mm/extmem.c
537
dcss_diag(&purgeseg_scode, seg->dcss_name, &dummy, &dummy);
arch/s390/mm/extmem.c
548
struct dcss_segment *seg;
arch/s390/mm/extmem.c
554
seg = segment_by_name (name);
arch/s390/mm/extmem.c
555
if (seg == NULL) {
arch/s390/mm/extmem.c
559
if (!refcount_dec_and_test(&seg->ref_count))
arch/s390/mm/extmem.c
561
release_resource(seg->res);
arch/s390/mm/extmem.c
562
kfree(seg->res);
arch/s390/mm/extmem.c
563
vmem_remove_mapping(seg->start_addr, seg->end - seg->start_addr + 1);
arch/s390/mm/extmem.c
564
list_del(&seg->list);
arch/s390/mm/extmem.c
572
smp_call_function_single(0, __dcss_diag_purge_on_cpu_0, seg, 1);
arch/s390/mm/extmem.c
573
kfree(seg);
arch/s390/mm/extmem.c
584
struct dcss_segment *seg;
arch/s390/mm/extmem.c
593
seg = segment_by_name (name);
arch/s390/mm/extmem.c
595
if (seg == NULL) {
arch/s390/mm/extmem.c
601
for (i=0; i<seg->segcnt; i++) {
arch/s390/mm/extmem.c
605
seg->range[i].start >> PAGE_SHIFT,
arch/s390/mm/extmem.c
606
seg->range[i].end >> PAGE_SHIFT,
arch/s390/mm/extmem.c
607
segtype_string[seg->range[i].start & 0xff]);
arch/x86/boot/boot.h
49
u16 seg;
arch/x86/boot/boot.h
50
asm("movw %%ds,%0" : "=rm" (seg));
arch/x86/boot/boot.h
51
return seg;
arch/x86/boot/boot.h
54
static inline void set_fs(u16 seg)
arch/x86/boot/boot.h
56
asm volatile("movw %0,%%fs" : : "rm" (seg));
arch/x86/boot/boot.h
60
u16 seg;
arch/x86/boot/boot.h
61
asm volatile("movw %%fs,%0" : "=rm" (seg));
arch/x86/boot/boot.h
62
return seg;
arch/x86/boot/boot.h
65
static inline void set_gs(u16 seg)
arch/x86/boot/boot.h
67
asm volatile("movw %0,%%gs" : : "rm" (seg));
arch/x86/boot/boot.h
71
u16 seg;
arch/x86/boot/boot.h
72
asm volatile("movw %%gs,%0" : "=rm" (seg));
arch/x86/boot/boot.h
73
return seg;
arch/x86/boot/compressed/cmdline.c
7
static inline void set_fs(unsigned long seg)
arch/x86/boot/compressed/cmdline.c
9
fs = seg << 4; /* shift it back */
arch/x86/boot/vesa.h
12
u16 off, seg;
arch/x86/boot/video-vesa.c
51
set_fs(vginfo.video_mode_ptr.seg);
arch/x86/hyperv/ivm.c
256
#define hv_populate_vmcb_seg(seg, gdtr_base) \
arch/x86/hyperv/ivm.c
258
if (seg.selector) { \
arch/x86/hyperv/ivm.c
259
seg.base = 0; \
arch/x86/hyperv/ivm.c
260
seg.limit = HV_AP_SEGMENT_LIMIT; \
arch/x86/hyperv/ivm.c
261
seg.attrib = *(u16 *)(gdtr_base + seg.selector + 5); \
arch/x86/hyperv/ivm.c
262
seg.attrib = (seg.attrib & 0xFF) | ((seg.attrib >> 4) & 0xF00); \
arch/x86/include/asm/desc.h
81
unsigned dpl, unsigned ist, unsigned seg)
arch/x86/include/asm/desc.h
95
gate->segment = seg;
arch/x86/include/asm/efi.h
278
#define __efi64_argmap_get_location(protocol, seg, bus, dev, func) \
arch/x86/include/asm/efi.h
279
((protocol), efi64_zero_upper(seg), efi64_zero_upper(bus), \
arch/x86/include/asm/kvm_host.h
1768
u64 (*get_segment_base)(struct kvm_vcpu *vcpu, int seg);
arch/x86/include/asm/kvm_host.h
1770
struct kvm_segment *var, int seg);
arch/x86/include/asm/kvm_host.h
1774
struct kvm_segment *var, int seg);
arch/x86/include/asm/kvm_host.h
2224
void kvm_get_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg);
arch/x86/include/asm/kvm_host.h
2225
void kvm_set_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg);
arch/x86/include/asm/kvm_host.h
2226
int kvm_load_segment_descriptor(struct kvm_vcpu *vcpu, u16 selector, int seg);
arch/x86/include/asm/pci_x86.h
184
extern int pci_mmconfig_insert(struct device *dev, u16 seg, u8 start, u8 end,
arch/x86/include/asm/pci_x86.h
186
extern int pci_mmconfig_delete(u16 seg, u8 start, u8 end);
arch/x86/include/asm/segment.h
260
[p] "=r" (p), [seg] "r" (__CPUNODE_SEG));
arch/x86/include/asm/segment.h
305
#define __loadsegment_simple(seg, value) \
arch/x86/include/asm/segment.h
310
"1: movl %k0,%%" #seg " \n" \
arch/x86/include/asm/segment.h
345
#define loadsegment(seg, value) __loadsegment_ ## seg (value)
arch/x86/include/asm/segment.h
350
#define savesegment(seg, value) \
arch/x86/include/asm/segment.h
351
asm("movl %%" #seg ",%k0" : "=r" (value) : : "memory")
arch/x86/kernel/ptrace.c
254
unsigned int seg;
arch/x86/kernel/ptrace.c
260
asm("movl %%fs,%0" : "=r" (seg));
arch/x86/kernel/ptrace.c
261
return seg;
arch/x86/kernel/ptrace.c
266
asm("movl %%gs,%0" : "=r" (seg));
arch/x86/kernel/ptrace.c
267
return seg;
arch/x86/kernel/ptrace.c
272
asm("movl %%ds,%0" : "=r" (seg));
arch/x86/kernel/ptrace.c
273
return seg;
arch/x86/kernel/ptrace.c
278
asm("movl %%es,%0" : "=r" (seg));
arch/x86/kernel/ptrace.c
279
return seg;
arch/x86/kernel/signal_32.c
208
#define get_user_seg(seg) ({ unsigned int v; savesegment(seg, v); v; })
arch/x86/kernel/step.c
16
unsigned long addr, seg;
arch/x86/kernel/step.c
19
seg = regs->cs;
arch/x86/kernel/step.c
21
addr = (addr & 0xffff) + (seg << 4);
arch/x86/kernel/step.c
32
if ((seg & SEGMENT_TI_MASK) == SEGMENT_LDT) {
arch/x86/kernel/step.c
36
seg >>= 3;
arch/x86/kernel/step.c
40
seg >= child->mm->context.ldt->nr_entries))
arch/x86/kernel/step.c
43
desc = &child->mm->context.ldt->entries[seg];
arch/x86/kvm/emulate.c
1522
u16 selector, int seg, u8 cpl,
arch/x86/kvm/emulate.c
1541
ctxt->ops->get_segment(ctxt, &dummy, &seg_desc, NULL, seg);
arch/x86/kvm/emulate.c
1544
} else if (seg <= VCPU_SREG_GS && ctxt->mode == X86EMUL_MODE_VM86) {
arch/x86/kvm/emulate.c
1558
if (seg == VCPU_SREG_TR && (selector & (1 << 2)))
arch/x86/kvm/emulate.c
1563
if (seg == VCPU_SREG_CS || seg == VCPU_SREG_TR)
arch/x86/kvm/emulate.c
1566
if (seg == VCPU_SREG_SS) {
arch/x86/kvm/emulate.c
1595
if (seg <= VCPU_SREG_GS && !seg_desc.s) {
arch/x86/kvm/emulate.c
1603
switch (seg) {
arch/x86/kvm/emulate.c
1691
err_vec = (seg == VCPU_SREG_SS) ? SS_VECTOR : NP_VECTOR;
arch/x86/kvm/emulate.c
1714
if (seg == VCPU_SREG_TR) {
arch/x86/kvm/emulate.c
1723
ctxt->ops->set_segment(ctxt, selector, &seg_desc, base3, seg);
arch/x86/kvm/emulate.c
1732
u16 selector, int seg)
arch/x86/kvm/emulate.c
1746
if (seg == VCPU_SREG_SS && selector == 3 &&
arch/x86/kvm/emulate.c
1750
return __load_segment_descriptor(ctxt, selector, seg, cpl,
arch/x86/kvm/emulate.c
1811
addr.seg = VCPU_SREG_SS;
arch/x86/kvm/emulate.c
1830
addr.seg = VCPU_SREG_SS;
arch/x86/kvm/emulate.c
1926
int seg = ctxt->src2.val;
arch/x86/kvm/emulate.c
1928
ctxt->src.val = get_segment_selector(ctxt, seg);
arch/x86/kvm/emulate.c
1939
int seg = ctxt->src2.val;
arch/x86/kvm/emulate.c
1947
if (seg == VCPU_SREG_SS)
arch/x86/kvm/emulate.c
1952
rc = load_segment_descriptor(ctxt, (u16)selector, seg);
arch/x86/kvm/emulate.c
2288
int seg = ctxt->src2.val;
arch/x86/kvm/emulate.c
2294
rc = load_segment_descriptor(ctxt, sel, seg);
arch/x86/kvm/emulate.c
4655
op->addr.mem.seg = VCPU_SREG_ES;
arch/x86/kvm/emulate.c
4709
op->addr.mem.seg = ctxt->seg_override;
arch/x86/kvm/emulate.c
4720
op->addr.mem.seg = ctxt->seg_override;
arch/x86/kvm/emulate.c
5171
ctxt->memop.addr.mem.seg = ctxt->seg_override;
arch/x86/kvm/emulate.c
523
static unsigned long seg_base(struct x86_emulate_ctxt *ctxt, int seg)
arch/x86/kvm/emulate.c
525
if (ctxt->mode == X86EMUL_MODE_PROT64 && seg < VCPU_SREG_FS)
arch/x86/kvm/emulate.c
528
return ctxt->ops->get_cached_segment_base(ctxt, seg);
arch/x86/kvm/emulate.c
578
static u16 get_segment_selector(struct x86_emulate_ctxt *ctxt, unsigned seg)
arch/x86/kvm/emulate.c
583
ctxt->ops->get_segment(ctxt, &selector, &desc, NULL, seg);
arch/x86/kvm/emulate.c
588
unsigned seg)
arch/x86/kvm/emulate.c
594
ctxt->ops->get_segment(ctxt, &dummy, &desc, &base3, seg);
arch/x86/kvm/emulate.c
595
ctxt->ops->set_segment(ctxt, selector, &desc, base3, seg);
arch/x86/kvm/emulate.c
650
la = seg_base(ctxt, addr.seg) + addr.ea;
arch/x86/kvm/emulate.c
666
addr.seg);
arch/x86/kvm/emulate.c
698
if (addr.seg == VCPU_SREG_SS)
arch/x86/kvm/emulate.c
719
struct segmented_address addr = { .seg = VCPU_SREG_CS,
arch/x86/kvm/emulate.c
848
struct segmented_address addr = { .seg = VCPU_SREG_CS,
arch/x86/kvm/kvm_emulate.h
202
struct desc_struct *desc, u32 *base3, int seg);
arch/x86/kvm/kvm_emulate.h
204
struct desc_struct *desc, u32 base3, int seg);
arch/x86/kvm/kvm_emulate.h
206
int seg);
arch/x86/kvm/kvm_emulate.h
263
unsigned seg;
arch/x86/kvm/smm.c
142
static u32 enter_smm_get_segment_flags(struct kvm_segment *seg)
arch/x86/kvm/smm.c
145
flags |= seg->g << 23;
arch/x86/kvm/smm.c
146
flags |= seg->db << 22;
arch/x86/kvm/smm.c
147
flags |= seg->l << 21;
arch/x86/kvm/smm.c
148
flags |= seg->avl << 20;
arch/x86/kvm/smm.c
149
flags |= seg->present << 15;
arch/x86/kvm/smm.c
150
flags |= seg->dpl << 13;
arch/x86/kvm/smm.c
151
flags |= seg->s << 12;
arch/x86/kvm/smm.c
152
flags |= seg->type << 8;
arch/x86/kvm/smm.c
160
struct kvm_segment seg;
arch/x86/kvm/smm.c
162
kvm_get_segment(vcpu, &seg, n);
arch/x86/kvm/smm.c
163
*selector = seg.selector;
arch/x86/kvm/smm.c
164
state->base = seg.base;
arch/x86/kvm/smm.c
165
state->limit = seg.limit;
arch/x86/kvm/smm.c
166
state->flags = enter_smm_get_segment_flags(&seg);
arch/x86/kvm/smm.c
174
struct kvm_segment seg;
arch/x86/kvm/smm.c
176
kvm_get_segment(vcpu, &seg, n);
arch/x86/kvm/smm.c
177
state->selector = seg.selector;
arch/x86/kvm/smm.c
178
state->attributes = enter_smm_get_segment_flags(&seg) >> 8;
arch/x86/kvm/smm.c
179
state->limit = seg.limit;
arch/x86/kvm/smm.c
180
state->base = seg.base;
arch/x86/kvm/svm/svm.c
1551
static struct vmcb_seg *svm_seg(struct kvm_vcpu *vcpu, int seg)
arch/x86/kvm/svm/svm.c
1556
switch (seg) {
arch/x86/kvm/svm/svm.c
1570
static u64 svm_get_segment_base(struct kvm_vcpu *vcpu, int seg)
arch/x86/kvm/svm/svm.c
1572
struct vmcb_seg *s = svm_seg(vcpu, seg);
arch/x86/kvm/svm/svm.c
1578
struct kvm_segment *var, int seg)
arch/x86/kvm/svm/svm.c
1580
struct vmcb_seg *s = svm_seg(vcpu, seg);
arch/x86/kvm/svm/svm.c
1609
switch (seg) {
arch/x86/kvm/svm/svm.c
1802
struct kvm_segment *var, int seg)
arch/x86/kvm/svm/svm.c
1805
struct vmcb_seg *s = svm_seg(vcpu, seg);
arch/x86/kvm/svm/svm.c
1825
if (seg == VCPU_SREG_SS)
arch/x86/kvm/svm/svm.c
964
static void init_seg(struct vmcb_seg *seg)
arch/x86/kvm/svm/svm.c
966
seg->selector = 0;
arch/x86/kvm/svm/svm.c
967
seg->attrib = SVM_SELECTOR_P_MASK | SVM_SELECTOR_S_MASK |
arch/x86/kvm/svm/svm.c
969
seg->limit = 0xffff;
arch/x86/kvm/svm/svm.c
970
seg->base = 0;
arch/x86/kvm/svm/svm.c
973
static void init_sys_seg(struct vmcb_seg *seg, uint32_t type)
arch/x86/kvm/svm/svm.c
975
seg->selector = 0;
arch/x86/kvm/svm/svm.c
976
seg->attrib = SVM_SELECTOR_P_MASK | type;
arch/x86/kvm/svm/svm.c
977
seg->limit = 0xffff;
arch/x86/kvm/svm/svm.c
978
seg->base = 0;
arch/x86/kvm/vmx/main.c
329
static u64 vt_get_segment_base(struct kvm_vcpu *vcpu, int seg)
arch/x86/kvm/vmx/main.c
334
return vmx_get_segment_base(vcpu, seg);
arch/x86/kvm/vmx/main.c
338
int seg)
arch/x86/kvm/vmx/main.c
345
vmx_get_segment(vcpu, var, seg);
arch/x86/kvm/vmx/main.c
349
int seg)
arch/x86/kvm/vmx/main.c
354
vmx_set_segment(vcpu, var, seg);
arch/x86/kvm/vmx/nested.c
4779
struct kvm_segment seg;
arch/x86/kvm/vmx/nested.c
4854
seg = (struct kvm_segment) {
arch/x86/kvm/vmx/nested.c
4864
seg.l = 1;
arch/x86/kvm/vmx/nested.c
4866
seg.db = 1;
arch/x86/kvm/vmx/nested.c
4867
__vmx_set_segment(vcpu, &seg, VCPU_SREG_CS);
arch/x86/kvm/vmx/nested.c
4868
seg = (struct kvm_segment) {
arch/x86/kvm/vmx/nested.c
4877
seg.selector = vmcs12->host_ds_selector;
arch/x86/kvm/vmx/nested.c
4878
__vmx_set_segment(vcpu, &seg, VCPU_SREG_DS);
arch/x86/kvm/vmx/nested.c
4879
seg.selector = vmcs12->host_es_selector;
arch/x86/kvm/vmx/nested.c
4880
__vmx_set_segment(vcpu, &seg, VCPU_SREG_ES);
arch/x86/kvm/vmx/nested.c
4881
seg.selector = vmcs12->host_ss_selector;
arch/x86/kvm/vmx/nested.c
4882
__vmx_set_segment(vcpu, &seg, VCPU_SREG_SS);
arch/x86/kvm/vmx/nested.c
4883
seg.selector = vmcs12->host_fs_selector;
arch/x86/kvm/vmx/nested.c
4884
seg.base = vmcs12->host_fs_base;
arch/x86/kvm/vmx/nested.c
4885
__vmx_set_segment(vcpu, &seg, VCPU_SREG_FS);
arch/x86/kvm/vmx/nested.c
4886
seg.selector = vmcs12->host_gs_selector;
arch/x86/kvm/vmx/nested.c
4887
seg.base = vmcs12->host_gs_base;
arch/x86/kvm/vmx/nested.c
4888
__vmx_set_segment(vcpu, &seg, VCPU_SREG_GS);
arch/x86/kvm/vmx/nested.c
4889
seg = (struct kvm_segment) {
arch/x86/kvm/vmx/nested.c
4896
__vmx_set_segment(vcpu, &seg, VCPU_SREG_TR);
arch/x86/kvm/vmx/nested.c
4898
memset(&seg, 0, sizeof(seg));
arch/x86/kvm/vmx/nested.c
4899
seg.unusable = 1;
arch/x86/kvm/vmx/nested.c
4900
__vmx_set_segment(vcpu, &seg, VCPU_SREG_LDTR);
arch/x86/kvm/vmx/vmx.c
3172
static void fix_pmode_seg(struct kvm_vcpu *vcpu, int seg,
arch/x86/kvm/vmx/vmx.c
3183
if (seg == VCPU_SREG_CS || seg == VCPU_SREG_SS)
arch/x86/kvm/vmx/vmx.c
3188
__vmx_set_segment(vcpu, save, seg);
arch/x86/kvm/vmx/vmx.c
3229
static void fix_rmode_seg(int seg, struct kvm_segment *save)
arch/x86/kvm/vmx/vmx.c
3231
const struct kvm_vmx_segment_field *sf = &kvm_vmx_segment_fields[seg];
arch/x86/kvm/vmx/vmx.c
3235
if (seg == VCPU_SREG_CS)
arch/x86/kvm/vmx/vmx.c
3252
"when entering protected mode (seg=%d)", seg);
arch/x86/kvm/vmx/vmx.c
3707
void vmx_get_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg)
arch/x86/kvm/vmx/vmx.c
3712
if (vmx->rmode.vm86_active && seg != VCPU_SREG_LDTR) {
arch/x86/kvm/vmx/vmx.c
3713
*var = vmx->rmode.segs[seg];
arch/x86/kvm/vmx/vmx.c
3714
if (seg == VCPU_SREG_TR
arch/x86/kvm/vmx/vmx.c
3715
|| var->selector == vmx_read_guest_seg_selector(vmx, seg))
arch/x86/kvm/vmx/vmx.c
3717
var->base = vmx_read_guest_seg_base(vmx, seg);
arch/x86/kvm/vmx/vmx.c
3718
var->selector = vmx_read_guest_seg_selector(vmx, seg);
arch/x86/kvm/vmx/vmx.c
3721
var->base = vmx_read_guest_seg_base(vmx, seg);
arch/x86/kvm/vmx/vmx.c
3722
var->limit = vmx_read_guest_seg_limit(vmx, seg);
arch/x86/kvm/vmx/vmx.c
3723
var->selector = vmx_read_guest_seg_selector(vmx, seg);
arch/x86/kvm/vmx/vmx.c
3724
ar = vmx_read_guest_seg_ar(vmx, seg);
arch/x86/kvm/vmx/vmx.c
3743
u64 vmx_get_segment_base(struct kvm_vcpu *vcpu, int seg)
arch/x86/kvm/vmx/vmx.c
3748
vmx_get_segment(vcpu, &s, seg);
arch/x86/kvm/vmx/vmx.c
3751
return vmx_read_guest_seg_base(to_vmx(vcpu), seg);
arch/x86/kvm/vmx/vmx.c
3796
void __vmx_set_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg)
arch/x86/kvm/vmx/vmx.c
3799
const struct kvm_vmx_segment_field *sf = &kvm_vmx_segment_fields[seg];
arch/x86/kvm/vmx/vmx.c
3803
if (vmx->rmode.vm86_active && seg != VCPU_SREG_LDTR) {
arch/x86/kvm/vmx/vmx.c
3804
vmx->rmode.segs[seg] = *var;
arch/x86/kvm/vmx/vmx.c
3805
if (seg == VCPU_SREG_TR)
arch/x86/kvm/vmx/vmx.c
3808
fix_rmode_seg(seg, &vmx->rmode.segs[seg]);
arch/x86/kvm/vmx/vmx.c
3827
if (is_unrestricted_guest(vcpu) && (seg != VCPU_SREG_LDTR))
arch/x86/kvm/vmx/vmx.c
3833
void vmx_set_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg)
arch/x86/kvm/vmx/vmx.c
3835
__vmx_set_segment(vcpu, var, seg);
arch/x86/kvm/vmx/vmx.c
3872
static bool rmode_segment_valid(struct kvm_vcpu *vcpu, int seg)
arch/x86/kvm/vmx/vmx.c
3877
vmx_get_segment(vcpu, &var, seg);
arch/x86/kvm/vmx/vmx.c
3879
if (seg == VCPU_SREG_CS)
arch/x86/kvm/vmx/vmx.c
3943
static bool data_segment_valid(struct kvm_vcpu *vcpu, int seg)
arch/x86/kvm/vmx/vmx.c
3948
vmx_get_segment(vcpu, &var, seg);
arch/x86/kvm/vmx/vmx.c
4128
static void seg_setup(int seg)
arch/x86/kvm/vmx/vmx.c
4130
const struct kvm_vmx_segment_field *sf = &kvm_vmx_segment_fields[seg];
arch/x86/kvm/vmx/vmx.c
4137
if (seg == VCPU_SREG_CS)
arch/x86/kvm/vmx/vmx.c
596
#define VMX_SEGMENT_FIELD(seg) \
arch/x86/kvm/vmx/vmx.c
597
[VCPU_SREG_##seg] = { \
arch/x86/kvm/vmx/vmx.c
598
.selector = GUEST_##seg##_SELECTOR, \
arch/x86/kvm/vmx/vmx.c
599
.base = GUEST_##seg##_BASE, \
arch/x86/kvm/vmx/vmx.c
600
.limit = GUEST_##seg##_LIMIT, \
arch/x86/kvm/vmx/vmx.c
601
.ar_bytes = GUEST_##seg##_AR_BYTES, \
arch/x86/kvm/vmx/vmx.c
876
static bool vmx_segment_cache_test_set(struct vcpu_vmx *vmx, unsigned seg,
arch/x86/kvm/vmx/vmx.c
880
u32 mask = 1 << (seg * SEG_FIELD_NR + field);
arch/x86/kvm/vmx/vmx.c
891
static u16 vmx_read_guest_seg_selector(struct vcpu_vmx *vmx, unsigned seg)
arch/x86/kvm/vmx/vmx.c
893
u16 *p = &vmx->segment_cache.seg[seg].selector;
arch/x86/kvm/vmx/vmx.c
895
if (!vmx_segment_cache_test_set(vmx, seg, SEG_FIELD_SEL))
arch/x86/kvm/vmx/vmx.c
896
*p = vmcs_read16(kvm_vmx_segment_fields[seg].selector);
arch/x86/kvm/vmx/vmx.c
900
static ulong vmx_read_guest_seg_base(struct vcpu_vmx *vmx, unsigned seg)
arch/x86/kvm/vmx/vmx.c
902
ulong *p = &vmx->segment_cache.seg[seg].base;
arch/x86/kvm/vmx/vmx.c
904
if (!vmx_segment_cache_test_set(vmx, seg, SEG_FIELD_BASE))
arch/x86/kvm/vmx/vmx.c
905
*p = vmcs_readl(kvm_vmx_segment_fields[seg].base);
arch/x86/kvm/vmx/vmx.c
909
static u32 vmx_read_guest_seg_limit(struct vcpu_vmx *vmx, unsigned seg)
arch/x86/kvm/vmx/vmx.c
911
u32 *p = &vmx->segment_cache.seg[seg].limit;
arch/x86/kvm/vmx/vmx.c
913
if (!vmx_segment_cache_test_set(vmx, seg, SEG_FIELD_LIMIT))
arch/x86/kvm/vmx/vmx.c
914
*p = vmcs_read32(kvm_vmx_segment_fields[seg].limit);
arch/x86/kvm/vmx/vmx.c
918
static u32 vmx_read_guest_seg_ar(struct vcpu_vmx *vmx, unsigned seg)
arch/x86/kvm/vmx/vmx.c
920
u32 *p = &vmx->segment_cache.seg[seg].ar;
arch/x86/kvm/vmx/vmx.c
922
if (!vmx_segment_cache_test_set(vmx, seg, SEG_FIELD_AR))
arch/x86/kvm/vmx/vmx.c
923
*p = vmcs_read32(kvm_vmx_segment_fields[seg].ar_bytes);
arch/x86/kvm/vmx/vmx.h
254
} seg[8];
arch/x86/kvm/vmx/vmx.h
360
void vmx_get_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg);
arch/x86/kvm/vmx/vmx.h
361
void __vmx_set_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg);
arch/x86/kvm/vmx/x86_ops.h
61
u64 vmx_get_segment_base(struct kvm_vcpu *vcpu, int seg);
arch/x86/kvm/vmx/x86_ops.h
62
void vmx_get_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg);
arch/x86/kvm/vmx/x86_ops.h
63
void vmx_set_segment(struct kvm_vcpu *vcpu, struct kvm_segment *var, int seg);
arch/x86/kvm/x86.c
7814
struct kvm_segment *var, int seg)
arch/x86/kvm/x86.c
7816
kvm_x86_call(set_segment)(vcpu, var, seg);
arch/x86/kvm/x86.c
7820
struct kvm_segment *var, int seg)
arch/x86/kvm/x86.c
7822
kvm_x86_call(get_segment)(vcpu, var, seg);
arch/x86/kvm/x86.c
8498
static unsigned long get_segment_base(struct kvm_vcpu *vcpu, int seg)
arch/x86/kvm/x86.c
8500
return kvm_x86_call(get_segment_base)(vcpu, seg);
arch/x86/kvm/x86.c
8640
struct x86_emulate_ctxt *ctxt, int seg)
arch/x86/kvm/x86.c
8642
return get_segment_base(emul_to_vcpu(ctxt), seg);
arch/x86/kvm/x86.c
8647
int seg)
arch/x86/kvm/x86.c
8651
kvm_get_segment(emul_to_vcpu(ctxt), &var, seg);
arch/x86/kvm/x86.c
8683
int seg)
arch/x86/kvm/x86.c
8707
kvm_set_segment(vcpu, &var, seg);
arch/x86/math-emu/fpu_system.h
23
static inline struct desc_struct FPU_get_ldt_descriptor(unsigned seg)
arch/x86/math-emu/fpu_system.h
29
seg >>= 3;
arch/x86/math-emu/fpu_system.h
31
if (current->mm->context.ldt && seg < current->mm->context.ldt->nr_entries)
arch/x86/math-emu/fpu_system.h
32
ret = current->mm->context.ldt->entries[seg];
arch/x86/pci/acpi.c
373
static int check_segment(u16 seg, struct device *dev, char *estr)
arch/x86/pci/acpi.c
375
if (seg) {
arch/x86/pci/acpi.c
394
int result, seg;
arch/x86/pci/acpi.c
403
seg = info->sd.domain;
arch/x86/pci/acpi.c
405
dev_dbg(dev, "%s(%04x %pR ECAM %pa)\n", __func__, seg,
arch/x86/pci/acpi.c
413
return check_segment(seg, dev, "MMCONFIG is disabled,");
arch/x86/pci/acpi.c
415
result = pci_mmconfig_insert(dev, seg, info->start_bus, info->end_bus,
arch/x86/pci/acpi.c
423
return check_segment(seg, dev,
arch/x86/pci/ce4100.c
257
static int ce4100_conf_read(unsigned int seg, unsigned int bus,
arch/x86/pci/ce4100.c
260
WARN_ON(seg);
arch/x86/pci/ce4100.c
269
return pci_direct_conf1.read(seg, bus, devfn, reg, len, value);
arch/x86/pci/ce4100.c
291
static int ce4100_conf_write(unsigned int seg, unsigned int bus,
arch/x86/pci/ce4100.c
294
WARN_ON(seg);
arch/x86/pci/ce4100.c
304
return pci_direct_conf1.write(seg, bus, devfn, reg, len, value);
arch/x86/pci/direct.c
101
WARN_ON(seg);
arch/x86/pci/direct.c
137
static int pci_conf2_write(unsigned int seg, unsigned int bus,
arch/x86/pci/direct.c
143
WARN_ON(seg);
arch/x86/pci/direct.c
21
static int pci_conf1_read(unsigned int seg, unsigned int bus,
arch/x86/pci/direct.c
26
if (seg || (bus > 255) || (devfn > 255) || (reg > 4095)) {
arch/x86/pci/direct.c
52
static int pci_conf1_write(unsigned int seg, unsigned int bus,
arch/x86/pci/direct.c
57
if (seg || (bus > 255) || (devfn > 255) || (reg > 4095))
arch/x86/pci/direct.c
95
static int pci_conf2_read(unsigned int seg, unsigned int bus,
arch/x86/pci/mmconfig-shared.c
783
int pci_mmconfig_insert(struct device *dev, u16 seg, u8 start, u8 end,
arch/x86/pci/mmconfig-shared.c
790
dev_dbg(dev, "%s(%04x [bus %02x-%02x])\n", __func__, seg, start, end);
arch/x86/pci/mmconfig-shared.c
799
cfg = pci_mmconfig_lookup(seg, start);
arch/x86/pci/mmconfig-shared.c
819
cfg = pci_mmconfig_alloc(seg, start, end, addr);
arch/x86/pci/mmconfig-shared.c
858
int pci_mmconfig_delete(u16 seg, u8 start, u8 end)
arch/x86/pci/mmconfig-shared.c
864
if (cfg->segment == seg && cfg->start_bus == start &&
arch/x86/pci/mmconfig_32.c
100
base = get_base_addr(seg, bus, devfn);
arch/x86/pci/mmconfig_32.c
27
static u32 get_base_addr(unsigned int seg, int bus, unsigned devfn)
arch/x86/pci/mmconfig_32.c
29
struct pci_mmcfg_region *cfg = pci_mmconfig_lookup(seg, bus);
arch/x86/pci/mmconfig_32.c
51
static int pci_mmcfg_read(unsigned int seg, unsigned int bus,
arch/x86/pci/mmconfig_32.c
63
base = get_base_addr(seg, bus, devfn);
arch/x86/pci/mmconfig_32.c
90
static int pci_mmcfg_write(unsigned int seg, unsigned int bus,
arch/x86/pci/mmconfig_64.c
19
static char __iomem *pci_dev_base(unsigned int seg, unsigned int bus, unsigned int devfn)
arch/x86/pci/mmconfig_64.c
21
struct pci_mmcfg_region *cfg = pci_mmconfig_lookup(seg, bus);
arch/x86/pci/mmconfig_64.c
28
static int pci_mmcfg_read(unsigned int seg, unsigned int bus,
arch/x86/pci/mmconfig_64.c
40
addr = pci_dev_base(seg, bus, devfn);
arch/x86/pci/mmconfig_64.c
62
static int pci_mmcfg_write(unsigned int seg, unsigned int bus,
arch/x86/pci/mmconfig_64.c
72
addr = pci_dev_base(seg, bus, devfn);
arch/x86/pci/numachip.c
19
static inline char __iomem *pci_dev_base(unsigned int seg, unsigned int bus, unsigned int devfn)
arch/x86/pci/numachip.c
21
struct pci_mmcfg_region *cfg = pci_mmconfig_lookup(seg, bus);
arch/x86/pci/numachip.c
28
static int pci_mmcfg_read_numachip(unsigned int seg, unsigned int bus,
arch/x86/pci/numachip.c
46
addr = pci_dev_base(seg, bus, devfn);
arch/x86/pci/numachip.c
68
static int pci_mmcfg_write_numachip(unsigned int seg, unsigned int bus,
arch/x86/pci/numachip.c
82
addr = pci_dev_base(seg, bus, devfn);
arch/x86/pci/olpc.c
197
static int pci_olpc_read(unsigned int seg, unsigned int bus,
arch/x86/pci/olpc.c
202
WARN_ON(seg);
arch/x86/pci/olpc.c
206
return pci_direct_conf1.read(seg, bus, devfn, reg, len, value);
arch/x86/pci/olpc.c
259
static int pci_olpc_write(unsigned int seg, unsigned int bus,
arch/x86/pci/olpc.c
262
WARN_ON(seg);
arch/x86/pci/olpc.c
266
return pci_direct_conf1.write(seg, bus, devfn, reg, len, value);
arch/x86/pci/pcbios.c
196
static int pci_bios_read(unsigned int seg, unsigned int bus,
arch/x86/pci/pcbios.c
204
WARN_ON(seg);
arch/x86/pci/pcbios.c
246
static int pci_bios_write(unsigned int seg, unsigned int bus,
arch/x86/pci/pcbios.c
254
WARN_ON(seg);
arch/x86/pci/xen.c
364
restore_ext.seg = pci_domain_nr(dev->bus);
arch/x86/realmode/init.c
127
u16 *seg = (u16 *) (base + *rel++);
arch/x86/realmode/init.c
128
*seg = real_mode_seg;
crypto/krb5/rfc6803_camellia.c
34
size_t bsize, offset, seg;
crypto/krb5/rfc6803_camellia.c
96
seg = min_t(size_t, result->len - offset, K.len);
crypto/krb5/rfc6803_camellia.c
97
memcpy(result->data + offset, K.data, seg);
crypto/krb5/rfc6803_camellia.c
98
offset += seg;
drivers/acpi/apei/einj-cxl.c
54
u64 seg = 0, bus;
drivers/acpi/apei/einj-cxl.c
63
seg = bridge->domain_nr;
drivers/acpi/apei/einj-cxl.c
66
*sbdf = (seg << 24) | (bus << 16) | (dport_dev->devfn << 8);
drivers/acpi/pci_mcfg.c
100
#define THUNDER_ECAM_QUIRK(rev, seg) \
drivers/acpi/pci_mcfg.c
101
{ "CAVIUM", "THUNDERX", rev, seg, MCFG_BUS_ANY, \
drivers/acpi/pci_mcfg.c
128
#define XGENE_V1_ECAM_MCFG(rev, seg) \
drivers/acpi/pci_mcfg.c
129
{"APM ", "XGENE ", rev, seg, MCFG_BUS_ANY, \
drivers/acpi/pci_mcfg.c
132
#define XGENE_V2_ECAM_MCFG(rev, seg) \
drivers/acpi/pci_mcfg.c
133
{"APM ", "XGENE ", rev, seg, MCFG_BUS_ANY, \
drivers/acpi/pci_mcfg.c
155
#define ALTRA_ECAM_QUIRK(rev, seg) \
drivers/acpi/pci_mcfg.c
156
{ "Ampere", "Altra ", rev, seg, MCFG_BUS_ANY, &pci_32b_read_ops }
drivers/acpi/pci_mcfg.c
177
#define LOONGSON_ECAM_MCFG(table_id, seg) \
drivers/acpi/pci_mcfg.c
178
{ "LOONGS", table_id, 1, seg, MCFG_BUS_ANY, &loongson_pci_ecam_ops }
drivers/acpi/pci_mcfg.c
250
u16 seg = root->segment;
drivers/acpi/pci_mcfg.c
262
if (e->segment == seg && e->bus_start <= bus_res->start &&
drivers/acpi/pci_mcfg.c
46
#define AL_ECAM(table_id, rev, seg, ops) \
drivers/acpi/pci_mcfg.c
47
{ "AMAZON", table_id, rev, seg, MCFG_BUS_ANY, ops }
drivers/acpi/pci_mcfg.c
58
#define QCOM_ECAM32(seg) \
drivers/acpi/pci_mcfg.c
59
{ "QCOM ", "QDF2432 ", 1, seg, MCFG_BUS_ANY, &pci_32b_ops }
drivers/acpi/pci_mcfg.c
70
#define HISI_QUAD_DOM(table_id, seg, ops) \
drivers/acpi/pci_mcfg.c
71
{ "HISI ", table_id, 0, (seg) + 0, MCFG_BUS_ANY, ops }, \
drivers/acpi/pci_mcfg.c
72
{ "HISI ", table_id, 0, (seg) + 1, MCFG_BUS_ANY, ops }, \
drivers/acpi/pci_mcfg.c
73
{ "HISI ", table_id, 0, (seg) + 2, MCFG_BUS_ANY, ops }, \
drivers/acpi/pci_mcfg.c
74
{ "HISI ", table_id, 0, (seg) + 3, MCFG_BUS_ANY, ops }
drivers/block/ublk_drv.c
4363
lim.seg_boundary_mask = ub->params.seg.seg_boundary_mask;
drivers/block/ublk_drv.c
4364
lim.max_segment_size = ub->params.seg.max_segment_size;
drivers/block/ublk_drv.c
4365
lim.max_segments = ub->params.seg.max_segments;
drivers/block/ublk_drv.c
907
const struct ublk_param_segment *p = &ub->params.seg;
drivers/block/xen-blkback/blkback.c
1093
dst->u.rw.seg[i] = src->u.rw.seg[i];
drivers/block/xen-blkback/blkback.c
1146
dst->u.rw.seg[i] = src->u.rw.seg[i];
drivers/block/xen-blkback/blkback.c
1294
struct seg_buf *seg = pending_req->seg;
drivers/block/xen-blkback/blkback.c
1366
pages[i]->gref = req->u.rw.seg[i].gref;
drivers/block/xen-blkback/blkback.c
1367
seg[i].nsec = req->u.rw.seg[i].last_sect -
drivers/block/xen-blkback/blkback.c
1368
req->u.rw.seg[i].first_sect + 1;
drivers/block/xen-blkback/blkback.c
1369
seg[i].offset = (req->u.rw.seg[i].first_sect << 9);
drivers/block/xen-blkback/blkback.c
1370
if ((req->u.rw.seg[i].last_sect >= (XEN_PAGE_SIZE >> 9)) ||
drivers/block/xen-blkback/blkback.c
1371
(req->u.rw.seg[i].last_sect <
drivers/block/xen-blkback/blkback.c
1372
req->u.rw.seg[i].first_sect))
drivers/block/xen-blkback/blkback.c
1374
preq.nr_sects += seg[i].nsec;
drivers/block/xen-blkback/blkback.c
1379
if (xen_blkbk_parse_indirect(req, pending_req, seg, &preq))
drivers/block/xen-blkback/blkback.c
1397
if (((int)preq.sector_number|(int)seg[i].nsec) &
drivers/block/xen-blkback/blkback.c
1431
seg[i].nsec << 9,
drivers/block/xen-blkback/blkback.c
1432
seg[i].offset) == 0)) {
drivers/block/xen-blkback/blkback.c
1442
preq.sector_number += seg[i].nsec;
drivers/block/xen-blkback/blkback.c
914
struct seg_buf seg[],
drivers/block/xen-blkback/blkback.c
953
seg[n].nsec = last_sect - first_sect + 1;
drivers/block/xen-blkback/blkback.c
954
seg[n].offset = first_sect << 9;
drivers/block/xen-blkback/blkback.c
955
preq->nr_sects += seg[n].nsec;
drivers/block/xen-blkback/common.h
138
struct blkif_request_segment seg[BLKIF_MAX_SEGMENTS_PER_REQUEST];
drivers/block/xen-blkback/common.h
355
struct seg_buf seg[MAX_INDIRECT_SEGMENTS];
drivers/block/xen-blkback/common.h
85
struct blkif_request_segment seg[BLKIF_MAX_SEGMENTS_PER_REQUEST];
drivers/block/xen-blkfront.c
667
ring_req->u.rw.seg[grant_idx] =
drivers/bluetooth/hci_aml.c
202
u8 *seg,
drivers/bluetooth/hci_aml.c
214
op_addr, (u32 *)seg, seg_size);
drivers/dma/xilinx/xilinx_dma.c
1029
struct xilinx_axidma_tx_segment *seg;
drivers/dma/xilinx/xilinx_dma.c
1035
seg = list_first_entry(&desc->segments, struct xilinx_axidma_tx_segment,
drivers/dma/xilinx/xilinx_dma.c
1038
hw = &seg->hw;
drivers/dma/xilinx/xilinx_dma.c
1758
struct xilinx_axidma_tx_segment *seg;
drivers/dma/xilinx/xilinx_dma.c
1760
seg = list_last_entry(&desc->segments,
drivers/dma/xilinx/xilinx_dma.c
1762
if (!(seg->hw.status & XILINX_DMA_BD_COMP_MASK) && chan->has_sg)
drivers/dma/xilinx/xilinx_dma.c
648
struct xilinx_axidma_tx_segment *seg;
drivers/dma/xilinx/xilinx_dma.c
651
seg = list_first_entry(&desc->segments,
drivers/dma/xilinx/xilinx_dma.c
653
return seg->hw.app;
drivers/edac/i10nm_base.c
440
d->pcu_cr3 = pci_get_dev_wrapper(d->seg, d->bus[res_cfg->pcu_cr3_bdf.bus],
drivers/edac/i10nm_base.c
502
d->sad_all = pci_get_dev_wrapper(d->seg, d->bus[res_cfg->sad_all_bdf.bus],
drivers/edac/i10nm_base.c
673
mdev = pci_get_dev_wrapper(d->seg,
drivers/edac/i10nm_base.c
739
mdev = pci_get_dev_wrapper(d->seg,
drivers/edac/i10nm_base.c
800
d->util_all = pci_get_dev_wrapper(d->seg, d->bus[res_cfg->util_all_bdf.bus],
drivers/edac/i10nm_base.c
806
d->uracu = pci_get_dev_wrapper(d->seg, d->bus[res_cfg->uracu_bdf.bus],
drivers/edac/i10nm_base.c
905
mdev = pci_get_dev_wrapper(d->seg, d->bus[res_cfg->hbm_mdev_bdf.bus],
drivers/edac/sb_edac.c
2466
int seg = 0;
drivers/edac/sb_edac.c
2497
seg = pci_domain_nr(pdev->bus);
drivers/edac/sb_edac.c
2501
sbridge_dev = get_sbridge_dev(seg, bus, dev_descr->dom,
drivers/edac/sb_edac.c
2515
sbridge_dev = alloc_sbridge_dev(seg, bus, dev_descr->dom, table);
drivers/edac/sb_edac.c
364
int seg;
drivers/edac/sb_edac.c
742
static struct sbridge_dev *get_sbridge_dev(int seg, u8 bus, enum domain dom,
drivers/edac/sb_edac.c
761
if ((sbridge_dev->seg == seg) && (sbridge_dev->bus == bus) &&
drivers/edac/sb_edac.c
769
static struct sbridge_dev *alloc_sbridge_dev(int seg, u8 bus, enum domain dom,
drivers/edac/sb_edac.c
785
sbridge_dev->seg = seg;
drivers/edac/skx_base.c
50
if (d->seg == pci_domain_nr(bus) && d->bus[idx] == bus->number)
drivers/edac/skx_common.c
365
d->seg = pci_domain_nr(pdev->bus);
drivers/edac/skx_common.c
369
d->seg = GET_BITFIELD(reg, 16, 23);
drivers/edac/skx_common.h
126
int seg;
drivers/firmware/google/memconsole-coreboot.c
48
} seg[2] = { {0}, {0} };
drivers/firmware/google/memconsole-coreboot.c
55
seg[0] = (struct seg){.phys = cursor, .len = size - cursor};
drivers/firmware/google/memconsole-coreboot.c
56
seg[1] = (struct seg){.phys = 0, .len = cursor};
drivers/firmware/google/memconsole-coreboot.c
58
seg[0] = (struct seg){.phys = 0, .len = min(cursor, size)};
drivers/firmware/google/memconsole-coreboot.c
61
for (i = 0; i < ARRAY_SIZE(seg) && count > done; i++) {
drivers/firmware/google/memconsole-coreboot.c
63
cbmem_console->body + seg[i].phys, seg[i].len);
drivers/firmware/google/memconsole-coreboot.c
64
pos -= seg[i].len;
drivers/gpu/drm/amd/amdgpu/amdgpu.h
145
u32 seg;
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
6976
gfx_v9_0_ras_fields[i].seg != reg->seg ||
drivers/gpu/drm/amd/amdgpu/gfx_v9_4.c
837
gfx_v9_4_ras_fields[i].seg != reg->seg ||
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_2.c
1473
gfx_v9_4_2_ras_fields[i].seg != reg->seg ||
drivers/gpu/drm/amd/amdgpu/nv.c
398
else if (reg_offset != (adev->reg_offset[en->hwip][en->inst][en->seg]
drivers/gpu/drm/amd/amdgpu/soc15.c
447
else if (reg_offset != (adev->reg_offset[en->hwip][en->inst][en->seg]
drivers/gpu/drm/amd/amdgpu/soc15.h
56
uint32_t seg;
drivers/gpu/drm/amd/amdgpu/soc15.h
63
uint32_t seg;
drivers/gpu/drm/amd/amdgpu/soc15.h
73
uint32_t seg;
drivers/gpu/drm/amd/amdgpu/soc15.h
82
uint32_t seg;
drivers/gpu/drm/amd/amdgpu/soc15.h
94
#define SOC15_REG_ENTRY_OFFSET(entry) (adev->reg_offset[entry.hwip][entry.inst][entry.seg] + entry.reg_offset)
drivers/gpu/drm/amd/amdgpu/soc15.h
98
(adev->reg_offset[entry.hwip][inst][entry.seg] + entry.reg_offset)
drivers/gpu/drm/amd/amdgpu/soc21.c
350
else if (reg_offset != (adev->reg_offset[en->hwip][en->inst][en->seg]
drivers/gpu/drm/amd/amdgpu/soc24.c
179
else if (reg_offset != (adev->reg_offset[en->hwip][en->inst][en->seg]
drivers/gpu/drm/amd/amdgpu/soc_v1_0.c
194
else if (reg_offset != (adev->reg_offset[en->hwip][en->inst][en->seg]
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn20/dcn20_clk_mgr.c
49
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn20/dcn20_clk_mgr.c
51
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn20/dcn20_clk_mgr.c
57
#define CLK_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn20/dcn20_clk_mgr.c
58
CLK_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn201/dcn201_clk_mgr.c
45
#define BASE_INNER(seg) DMU_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn201/dcn201_clk_mgr.c
47
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn201/dcn201_clk_mgr.c
53
#define CLK_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn201/dcn201_clk_mgr.c
54
CLK_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn30/dcn30_clk_mgr.c
55
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn30/dcn30_clk_mgr.c
57
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn314/dcn314_smu.c
49
#define BASE_INNER(seg) MP1_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn314/dcn314_smu.c
51
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn32/dcn32_clk_mgr.c
91
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn32/dcn32_clk_mgr.c
93
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn35/dcn351_clk_mgr.c
101
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn35/dcn351_clk_mgr.c
99
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn35/dcn35_clk_mgr.c
130
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn35/dcn35_clk_mgr.c
132
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn35/dcn35_smu.c
49
#define BASE_INNER(seg) MP1_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn35/dcn35_smu.c
51
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn401/dcn401_clk_mgr.c
49
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn401/dcn401_clk_mgr.c
51
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/dcn10/dcn10_dwb.h
28
#define BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/dcn10/dcn10_dwb.h
29
DCE_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/dcn10/dcn10_dwb.h
31
#define BASE(seg) \
drivers/gpu/drm/amd/display/dc/dcn10/dcn10_dwb.h
32
BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/gpio/dce120/hw_factory_dce120.c
53
#define BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/gpio/dce120/hw_factory_dce120.c
54
DCE_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/gpio/dce120/hw_factory_dce120.c
57
#define BASE(seg) \
drivers/gpu/drm/amd/display/dc/gpio/dce120/hw_factory_dce120.c
58
BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/gpio/dce120/hw_translate_dce120.c
44
#define BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/gpio/dce120/hw_translate_dce120.c
45
DCE_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/gpio/dce120/hw_translate_dce120.c
48
#define BASE(seg) \
drivers/gpu/drm/amd/display/dc/gpio/dce120/hw_translate_dce120.c
49
BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/gpio/dcn10/hw_factory_dcn10.c
50
#define BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/gpio/dcn10/hw_factory_dcn10.c
51
DCE_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/gpio/dcn10/hw_factory_dcn10.c
54
#define BASE(seg) \
drivers/gpu/drm/amd/display/dc/gpio/dcn10/hw_factory_dcn10.c
55
BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/gpio/dcn10/hw_translate_dcn10.c
44
#define BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/gpio/dcn10/hw_translate_dcn10.c
45
DCE_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/gpio/dcn10/hw_translate_dcn10.c
48
#define BASE(seg) \
drivers/gpu/drm/amd/display/dc/gpio/dcn10/hw_translate_dcn10.c
49
BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/gpio/dcn20/hw_factory_dcn20.c
53
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/gpio/dcn20/hw_factory_dcn20.c
55
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/gpio/dcn20/hw_translate_dcn20.c
50
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/gpio/dcn20/hw_translate_dcn20.c
52
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/gpio/dcn21/hw_factory_dcn21.c
51
#define BASE_INNER(seg) DMU_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/gpio/dcn21/hw_factory_dcn21.c
53
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/gpio/dcn21/hw_translate_dcn21.c
50
#define BASE_INNER(seg) DMU_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/gpio/dcn21/hw_translate_dcn21.c
52
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/gpio/dcn30/hw_factory_dcn30.c
60
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/gpio/dcn30/hw_factory_dcn30.c
62
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/gpio/dcn30/hw_translate_dcn30.c
55
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/gpio/dcn30/hw_translate_dcn30.c
57
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/gpio/dcn315/hw_factory_dcn315.c
57
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/gpio/dcn315/hw_factory_dcn315.c
59
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/gpio/dcn315/hw_translate_dcn315.c
50
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/gpio/dcn315/hw_translate_dcn315.c
52
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/gpio/dcn32/hw_factory_dcn32.c
53
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/gpio/dcn32/hw_factory_dcn32.c
55
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/gpio/dcn32/hw_translate_dcn32.c
48
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/gpio/dcn32/hw_translate_dcn32.c
50
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/gpio/dcn401/hw_factory_dcn401.c
33
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/gpio/dcn401/hw_factory_dcn401.c
35
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/gpio/dcn401/hw_translate_dcn401.c
23
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/gpio/dcn401/hw_translate_dcn401.c
25
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/irq/dce120/irq_service_dce120.c
65
#define BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/irq/dce120/irq_service_dce120.c
66
DCE_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/irq/dce120/irq_service_dce120.c
68
#define BASE(seg) \
drivers/gpu/drm/amd/display/dc/irq/dce120/irq_service_dce120.c
69
BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/irq/dcn10/irq_service_dcn10.c
162
#define BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/irq/dcn10/irq_service_dcn10.c
163
DCE_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/irq/dcn10/irq_service_dcn10.c
165
#define BASE(seg) \
drivers/gpu/drm/amd/display/dc/irq/dcn10/irq_service_dcn10.c
166
BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/irq/dcn20/irq_service_dcn20.c
164
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/irq/dcn20/irq_service_dcn20.c
167
#define BASE(seg) \
drivers/gpu/drm/amd/display/dc/irq/dcn20/irq_service_dcn20.c
168
BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/irq/dcn201/irq_service_dcn201.c
113
#define BASE_INNER(seg) DMU_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/irq/dcn201/irq_service_dcn201.c
115
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/irq/dcn201/irq_service_dcn201.c
118
#define BASE(seg) \
drivers/gpu/drm/amd/display/dc/irq/dcn201/irq_service_dcn201.c
119
BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/irq/dcn21/irq_service_dcn21.c
171
#define BASE_INNER(seg) DMU_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/irq/dcn21/irq_service_dcn21.c
174
#define BASE(seg) \
drivers/gpu/drm/amd/display/dc/irq/dcn21/irq_service_dcn21.c
175
BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/irq/dcn30/irq_service_dcn30.c
178
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/irq/dcn30/irq_service_dcn30.c
181
#define BASE(seg) \
drivers/gpu/drm/amd/display/dc/irq/dcn30/irq_service_dcn30.c
182
BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/irq/dcn302/irq_service_dcn302.c
165
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/irq/dcn302/irq_service_dcn302.c
168
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/irq/dcn303/irq_service_dcn303.c
111
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/irq/dcn303/irq_service_dcn303.c
114
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/irq/dcn31/irq_service_dcn31.c
167
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/irq/dcn31/irq_service_dcn31.c
170
#define BASE(seg) \
drivers/gpu/drm/amd/display/dc/irq/dcn31/irq_service_dcn31.c
171
BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/irq/dcn314/irq_service_dcn314.c
169
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/irq/dcn314/irq_service_dcn314.c
172
#define BASE(seg) \
drivers/gpu/drm/amd/display/dc/irq/dcn314/irq_service_dcn314.c
173
BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/irq/dcn315/irq_service_dcn315.c
174
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/irq/dcn315/irq_service_dcn315.c
177
#define BASE(seg) \
drivers/gpu/drm/amd/display/dc/irq/dcn315/irq_service_dcn315.c
178
BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/irq/dcn32/irq_service_dcn32.c
178
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/irq/dcn32/irq_service_dcn32.c
181
#define BASE(seg) \
drivers/gpu/drm/amd/display/dc/irq/dcn32/irq_service_dcn32.c
182
BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/irq/dcn35/irq_service_dcn35.c
166
#define BASE_INNER(seg) ctx->dcn_reg_offsets[seg]
drivers/gpu/drm/amd/display/dc/irq/dcn35/irq_service_dcn35.c
169
#define BASE(seg) \
drivers/gpu/drm/amd/display/dc/irq/dcn35/irq_service_dcn35.c
170
BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/irq/dcn351/irq_service_dcn351.c
145
#define BASE_INNER(seg) ctx->dcn_reg_offsets[seg]
drivers/gpu/drm/amd/display/dc/irq/dcn351/irq_service_dcn351.c
148
#define BASE(seg) \
drivers/gpu/drm/amd/display/dc/irq/dcn351/irq_service_dcn351.c
149
BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/irq/dcn36/irq_service_dcn36.c
144
#define BASE_INNER(seg) ctx->dcn_reg_offsets[seg]
drivers/gpu/drm/amd/display/dc/irq/dcn36/irq_service_dcn36.c
147
#define BASE(seg) \
drivers/gpu/drm/amd/display/dc/irq/dcn36/irq_service_dcn36.c
148
BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/irq/dcn401/irq_service_dcn401.c
158
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/irq/dcn401/irq_service_dcn401.c
161
#define BASE(seg) \
drivers/gpu/drm/amd/display/dc/irq/dcn401/irq_service_dcn401.c
162
BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dce120/dce120_resource.c
124
#define BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dce120/dce120_resource.c
125
DCE_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dce120/dce120_resource.c
127
#define NBIO_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dce120/dce120_resource.c
128
NBIF_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dce120/dce120_resource.c
130
#define NBIO_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dce120/dce120_resource.c
131
NBIO_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dce120/dce120_resource.c
134
#define BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dce120/dce120_resource.c
135
BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dce120/dce120_resource.c
146
#define MMHUB_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dce120/dce120_resource.c
147
MMHUB_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dce120/dce120_resource.c
149
#define MMHUB_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dce120/dce120_resource.c
150
MMHUB_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn10/dcn10_resource.c
107
#define BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn10/dcn10_resource.c
108
DCE_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn10/dcn10_resource.c
110
#define BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn10/dcn10_resource.c
111
BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn10/dcn10_resource.c
135
#define NBIO_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn10/dcn10_resource.c
136
NBIF_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn10/dcn10_resource.c
138
#define NBIO_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn10/dcn10_resource.c
139
NBIO_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn10/dcn10_resource.c
146
#define MMHUB_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn10/dcn10_resource.c
147
MMHUB_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn10/dcn10_resource.c
149
#define MMHUB_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn10/dcn10_resource.c
150
MMHUB_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
125
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
127
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
163
#define NBIO_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
164
NBIO_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
166
#define NBIO_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
167
NBIO_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
174
#define MMHUB_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
175
MMHUB_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
177
#define MMHUB_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn20/dcn20_resource.c
178
MMHUB_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
248
#define BASE_INNER(seg) DMU_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
250
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
280
#define NBIO_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
281
NBIO_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
283
#define NBIO_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
284
NBIO_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
291
#define MMHUB_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
292
MMHUB_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
294
#define MMHUB_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn201/dcn201_resource.c
295
MMHUB_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
100
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
127
#define NBIO_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
128
NBIF0_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
130
#define NBIO_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
131
NBIO_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
138
#define MMHUB_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
139
MMHUB_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
141
#define MMHUB_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
142
MMHUB_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn21/dcn21_resource.c
98
#define BASE_INNER(seg) DMU_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
114
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
116
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
158
#define NBIO_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
159
NBIO_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
161
#define NBIO_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
162
NBIO_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
169
#define MMHUB_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
170
MMHUB_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
172
#define MMHUB_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
173
MMHUB_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
180
#define CLK_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
181
CLK_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
183
#define CLK_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
184
CLK_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
113
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
115
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
162
#define NBIO_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
163
NBIO_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
165
#define NBIO_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
166
NBIO_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
173
#define MMHUB_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
174
MMHUB_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
176
#define MMHUB_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
177
MMHUB_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
184
#define CLK_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
185
CLK_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
187
#define CLK_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn301/dcn301_resource.c
188
CLK_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
165
#define NBIO_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
166
NBIO_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
168
#define NBIO_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
169
NBIO_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
176
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn302/dcn302_resource.c
178
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
161
#define NBIO_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
162
NBIO_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
164
#define NBIO_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
165
NBIO_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
172
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn303/dcn303_resource.c
174
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
125
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
127
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
169
#define NBIO_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
170
NBIO_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
172
#define NBIO_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
173
NBIO_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
180
#define MMHUB_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
181
MMHUB_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
183
#define MMHUB_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
184
MMHUB_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
191
#define CLK_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
192
CLK_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
194
#define CLK_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn31/dcn31_resource.c
195
CLK_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
142
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
144
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
186
#define NBIO_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
187
NBIO_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
189
#define NBIO_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
190
NBIO_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
197
#define MMHUB_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
198
MMHUB_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
200
#define MMHUB_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
201
MMHUB_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
208
#define CLK_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
209
CLK_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
211
#define CLK_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn314/dcn314_resource.c
212
CLK_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
159
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
161
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
203
#define NBIO_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
204
NBIO_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
206
#define NBIO_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn315/dcn315_resource.c
207
NBIO_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
145
#define BASE_INNER(seg) DCN_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
147
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
189
#define NBIO_BASE_INNER(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
190
NBIO_BASE__INST0_SEG ## seg
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
192
#define NBIO_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn316/dcn316_resource.c
193
NBIO_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
114
#define BASE_INNER(seg) ctx->dcn_reg_offsets[seg]
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
116
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
185
#define NBIO_BASE_INNER(seg) ctx->nbio_reg_offsets[seg]
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
187
#define NBIO_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn32/dcn32_resource.c
188
NBIO_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
114
#define BASE_INNER(seg) ctx->dcn_reg_offsets[seg]
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
116
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
185
#define NBIO_BASE_INNER(seg) ctx->nbio_reg_offsets[seg]
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
187
#define NBIO_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn321/dcn321_resource.c
188
NBIO_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
128
#define BASE_INNER(seg) ctx->dcn_reg_offsets[seg]
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
130
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
201
#define NBIO_BASE_INNER(seg) ctx->nbio_reg_offsets[seg]
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
203
#define NBIO_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn35/dcn35_resource.c
204
NBIO_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
108
#define BASE_INNER(seg) ctx->dcn_reg_offsets[seg]
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
110
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
181
#define NBIO_BASE_INNER(seg) ctx->nbio_reg_offsets[seg]
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
183
#define NBIO_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn351/dcn351_resource.c
184
NBIO_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
113
#define BASE_INNER(seg) ctx->dcn_reg_offsets[seg]
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
115
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
186
#define NBIO_BASE_INNER(seg) ctx->nbio_reg_offsets[seg]
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
188
#define NBIO_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn36/dcn36_resource.c
189
NBIO_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
172
#define NBIO_BASE_INNER(seg) ctx->nbio_reg_offsets[seg]
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
174
#define NBIO_BASE(seg) \
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
175
NBIO_BASE_INNER(seg)
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
95
#define BASE_INNER(seg) ctx->dcn_reg_offsets[seg]
drivers/gpu/drm/amd/display/dc/resource/dcn401/dcn401_resource.c
97
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn20.c
35
#define BASE_INNER(seg) DCN_BASE__INST0_SEG##seg
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn21.c
34
#define BASE_INNER(seg) DMU_BASE__INST0_SEG##seg
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn30.c
35
#define BASE_INNER(seg) DCN_BASE__INST0_SEG##seg
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn301.c
34
#define BASE_INNER(seg) DCN_BASE__INST0_SEG##seg
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn302.c
34
#define BASE_INNER(seg) DCN_BASE__INST0_SEG##seg
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn303.c
35
#define BASE_INNER(seg) DCN_BASE__INST0_SEG##seg
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn31.c
34
#define BASE_INNER(seg) DCN_BASE__INST0_SEG##seg
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn314.c
40
#define BASE_INNER(seg) DCN_BASE__INST0_SEG##seg
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn315.c
40
#define BASE_INNER(seg) DCN_BASE__INST0_SEG##seg
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn316.c
40
#define BASE_INNER(seg) DCN_BASE__INST0_SEG##seg
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn32.c
35
#define BASE_INNER(seg) ctx->dcn_reg_offsets[seg]
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn35.c
35
#define BASE_INNER(seg) ctx->dcn_reg_offsets[seg]
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn351.c
11
#define BASE_INNER(seg) ctx->dcn_reg_offsets[seg]
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn36.c
11
#define BASE_INNER(seg) ctx->dcn_reg_offsets[seg]
drivers/gpu/drm/amd/display/dmub/src/dmub_dcn401.c
14
#define BASE_INNER(seg) DCN_BASE__INST0_SEG##seg
drivers/gpu/drm/amd/display/dmub/src/dmub_reg.h
35
#define BASE(seg) BASE_INNER(seg)
drivers/gpu/drm/amd/pm/powerplay/hwmgr/common_baco.c
112
reg = adev->reg_offset[entry[i].hwip][entry[i].inst][entry[i].seg]
drivers/gpu/drm/amd/pm/powerplay/hwmgr/common_baco.h
49
uint32_t seg;
drivers/gpu/drm/nouveau/include/nvif/push.h
107
__p->seg++; \
drivers/gpu/drm/nouveau/include/nvif/push.h
129
PUSH_ASSERT(_p->cur + _s <= _p->seg, "segment overrun"); \
drivers/gpu/drm/nouveau/include/nvif/push.h
43
u32 *seg;
drivers/gpu/drm/nouveau/include/nvif/push.h
56
push->seg = push->cur + size;
drivers/gpu/drm/nouveau/include/nvif/push.h
95
PUSH_ASSERT(_p->cur < _p->seg, "segment overrun"); \
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/anx9805.c
51
u8 seg = 0x00, off = 0x00, tmp;
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/anx9805.c
62
nvkm_wri2cr(adap, bus->addr, 0x41, seg);
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/anx9805.c
82
seg = msg->buf[0];
drivers/gpu/drm/sun4i/sun4i_hdmi.h
116
#define SUN4I_HDMI_DDC_ADDR_SEGMENT(seg) (((seg) & 0xff) << 24)
drivers/gpu/drm/sun4i/sun4i_hdmi.h
173
#define SUN6I_HDMI_DDC_ADDR_SEGMENT(seg) (((seg) & 0xff) << 24)
drivers/infiniband/core/mad.c
820
struct ib_rmpp_segment *seg = NULL;
drivers/infiniband/core/mad.c
830
seg = kmalloc(sizeof(*seg) + seg_size, gfp_mask);
drivers/infiniband/core/mad.c
831
if (!seg) {
drivers/infiniband/core/mad.c
835
seg->num = ++send_buf->seg_count;
drivers/infiniband/core/mad.c
836
list_add_tail(&seg->list, &send_wr->rmpp_list);
drivers/infiniband/core/mad.c
841
memset(seg->data + seg_size - pad, 0, pad);
drivers/infiniband/core/mad_rmpp.c
369
static inline int get_last_flag(struct ib_mad_recv_buf *seg)
drivers/infiniband/core/mad_rmpp.c
373
rmpp_mad = (struct ib_rmpp_mad *) seg->mad;
drivers/infiniband/core/mad_rmpp.c
377
static inline int get_seg_num(struct ib_mad_recv_buf *seg)
drivers/infiniband/core/mad_rmpp.c
381
rmpp_mad = (struct ib_rmpp_mad *) seg->mad;
drivers/infiniband/core/mad_rmpp.c
386
struct ib_mad_recv_buf *seg)
drivers/infiniband/core/mad_rmpp.c
388
if (seg->list.next == rmpp_list)
drivers/infiniband/core/mad_rmpp.c
391
return container_of(seg->list.next, struct ib_mad_recv_buf, list);
drivers/infiniband/core/user_mad.c
444
int left, seg;
drivers/infiniband/core/user_mad.c
453
for (seg = 1, left = msg->data_len, buf += msg->hdr_len; left > 0;
drivers/infiniband/core/user_mad.c
454
seg++, left -= msg->seg_size, buf += msg->seg_size) {
drivers/infiniband/core/user_mad.c
455
if (copy_from_user(ib_get_rmpp_segment(msg, seg), buf,
drivers/infiniband/hw/mlx5/umr.c
252
void *cur_edge, *seg;
drivers/infiniband/hw/mlx5/umr.c
262
err = mlx5r_begin_wqe(qp, &seg, &ctrl, &idx, &size, &cur_edge, 0,
drivers/infiniband/hw/mlx5/umr.c
269
mlx5r_memcpy_send_wqe(&qp->sq, &cur_edge, &seg, &size, wqe, wqe_size);
drivers/infiniband/hw/mlx5/umr.c
272
mlx5r_finish_wqe(qp, ctrl, seg, size, cur_edge, idx, id.wr_id, 0,
drivers/infiniband/hw/mlx5/umr.c
443
struct mlx5_mkey_seg *seg,
drivers/infiniband/hw/mlx5/umr.c
450
MLX5_SET(mkc, seg, a, !!(access_flags & IB_ACCESS_REMOTE_ATOMIC));
drivers/infiniband/hw/mlx5/umr.c
451
MLX5_SET(mkc, seg, rw, !!(access_flags & IB_ACCESS_REMOTE_WRITE));
drivers/infiniband/hw/mlx5/umr.c
452
MLX5_SET(mkc, seg, rr, !!(access_flags & IB_ACCESS_REMOTE_READ));
drivers/infiniband/hw/mlx5/umr.c
453
MLX5_SET(mkc, seg, lw, !!(access_flags & IB_ACCESS_LOCAL_WRITE));
drivers/infiniband/hw/mlx5/umr.c
454
MLX5_SET(mkc, seg, lr, 1);
drivers/infiniband/hw/mlx5/umr.c
455
MLX5_SET(mkc, seg, relaxed_ordering_write,
drivers/infiniband/hw/mlx5/umr.c
457
MLX5_SET(mkc, seg, relaxed_ordering_read, ro_read);
drivers/infiniband/hw/mlx5/wr.c
1001
handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
drivers/infiniband/hw/mlx5/wr.c
1005
void **seg, int *size, void **cur_edge)
drivers/infiniband/hw/mlx5/wr.c
1007
set_datagram_seg(*seg, wr);
drivers/infiniband/hw/mlx5/wr.c
1008
*seg += sizeof(struct mlx5_wqe_datagram_seg);
drivers/infiniband/hw/mlx5/wr.c
1010
handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
drivers/infiniband/hw/mlx5/wr.c
1016
pad = *seg;
drivers/infiniband/hw/mlx5/wr.c
1018
*seg += sizeof(struct mlx5_wqe_eth_pad);
drivers/infiniband/hw/mlx5/wr.c
1020
set_eth_seg(wr, qp, seg, size, cur_edge);
drivers/infiniband/hw/mlx5/wr.c
1021
handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
drivers/infiniband/hw/mlx5/wr.c
1065
void *seg;
drivers/infiniband/hw/mlx5/wr.c
1098
err = begin_wqe(qp, &seg, &ctrl, wr, &idx, &size, &cur_edge,
drivers/infiniband/hw/mlx5/wr.c
1124
xrc = seg;
drivers/infiniband/hw/mlx5/wr.c
1125
seg += sizeof(*xrc);
drivers/infiniband/hw/mlx5/wr.c
1129
err = handle_qpt_rc(dev, qp, wr, &ctrl, &seg, &size,
drivers/infiniband/hw/mlx5/wr.c
1141
handle_qpt_uc(wr, &seg, &size);
drivers/infiniband/hw/mlx5/wr.c
1152
handle_qpt_hw_gsi(qp, wr, &seg, &size, &cur_edge);
drivers/infiniband/hw/mlx5/wr.c
1155
handle_qpt_ud(qp, wr, &seg, &size, &cur_edge);
drivers/infiniband/hw/mlx5/wr.c
1163
err = set_data_inl_seg(qp, wr, &seg, &size, &cur_edge);
drivers/infiniband/hw/mlx5/wr.c
1171
handle_post_send_edge(&qp->sq, &seg, size,
drivers/infiniband/hw/mlx5/wr.c
1177
(struct mlx5_wqe_data_seg *)seg,
drivers/infiniband/hw/mlx5/wr.c
1180
seg += sizeof(struct mlx5_wqe_data_seg);
drivers/infiniband/hw/mlx5/wr.c
1185
mlx5r_finish_wqe(qp, ctrl, seg, size, cur_edge, idx, wr->wr_id,
drivers/infiniband/hw/mlx5/wr.c
189
static void set_reg_mkey_seg(struct mlx5_mkey_seg *seg,
drivers/infiniband/hw/mlx5/wr.c
195
memset(seg, 0, sizeof(*seg));
drivers/infiniband/hw/mlx5/wr.c
198
seg->log2_page_size = ilog2(mr->ibmr.page_size);
drivers/infiniband/hw/mlx5/wr.c
203
seg->flags = get_umr_flags(access) | mr->access_mode;
drivers/infiniband/hw/mlx5/wr.c
204
seg->qpn_mkey7_0 = cpu_to_be32((key & 0xff) | 0xffffff00);
drivers/infiniband/hw/mlx5/wr.c
205
seg->flags_pd = cpu_to_be32(MLX5_MKEY_REMOTE_INVAL);
drivers/infiniband/hw/mlx5/wr.c
206
seg->start_addr = cpu_to_be64(mr->ibmr.iova);
drivers/infiniband/hw/mlx5/wr.c
207
seg->len = cpu_to_be64(mr->ibmr.length);
drivers/infiniband/hw/mlx5/wr.c
208
seg->xlt_oct_size = cpu_to_be32(ndescs);
drivers/infiniband/hw/mlx5/wr.c
211
static void set_linv_mkey_seg(struct mlx5_mkey_seg *seg)
drivers/infiniband/hw/mlx5/wr.c
213
memset(seg, 0, sizeof(*seg));
drivers/infiniband/hw/mlx5/wr.c
214
seg->status = MLX5_MKEY_STATUS_FREE;
drivers/infiniband/hw/mlx5/wr.c
263
struct mlx5_wqe_inline_seg *seg;
drivers/infiniband/hw/mlx5/wr.c
268
seg = *wqe;
drivers/infiniband/hw/mlx5/wr.c
269
*wqe += sizeof(*seg);
drivers/infiniband/hw/mlx5/wr.c
270
offset = sizeof(*seg);
drivers/infiniband/hw/mlx5/wr.c
300
seg->byte_count = cpu_to_be32(inl | MLX5_INLINE_SEG);
drivers/infiniband/hw/mlx5/wr.c
302
*wqe_sz += ALIGN(inl + sizeof(seg->byte_count), 16) / 16;
drivers/infiniband/hw/mlx5/wr.c
419
struct mlx5_ib_qp *qp, void **seg, int *size,
drivers/infiniband/hw/mlx5/wr.c
457
struct mlx5_klm *data_klm = *seg;
drivers/infiniband/hw/mlx5/wr.c
483
sblock_ctrl = *seg;
drivers/infiniband/hw/mlx5/wr.c
512
*seg += wqe_size;
drivers/infiniband/hw/mlx5/wr.c
514
handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
drivers/infiniband/hw/mlx5/wr.c
516
bsf = *seg;
drivers/infiniband/hw/mlx5/wr.c
521
*seg += sizeof(*bsf);
drivers/infiniband/hw/mlx5/wr.c
523
handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
drivers/infiniband/hw/mlx5/wr.c
528
static void set_sig_mkey_segment(struct mlx5_mkey_seg *seg,
drivers/infiniband/hw/mlx5/wr.c
535
memset(seg, 0, sizeof(*seg));
drivers/infiniband/hw/mlx5/wr.c
537
seg->flags = get_umr_flags(access_flags) | MLX5_MKC_ACCESS_MODE_KLMS;
drivers/infiniband/hw/mlx5/wr.c
538
seg->qpn_mkey7_0 = cpu_to_be32((sig_key & 0xff) | 0xffffff00);
drivers/infiniband/hw/mlx5/wr.c
539
seg->flags_pd = cpu_to_be32(MLX5_MKEY_REMOTE_INVAL | sigerr << 26 |
drivers/infiniband/hw/mlx5/wr.c
541
seg->len = cpu_to_be64(length);
drivers/infiniband/hw/mlx5/wr.c
542
seg->xlt_oct_size = cpu_to_be32(mlx5r_umr_get_xlt_octo(size));
drivers/infiniband/hw/mlx5/wr.c
543
seg->bsfs_octo_size = cpu_to_be32(MLX5_MKEY_BSF_OCTO_SIZE);
drivers/infiniband/hw/mlx5/wr.c
55
void **seg, int *size, void **cur_edge)
drivers/infiniband/hw/mlx5/wr.c
558
struct mlx5_ib_qp *qp, void **seg, int *size,
drivers/infiniband/hw/mlx5/wr.c
57
struct mlx5_wqe_eth_seg *eseg = *seg;
drivers/infiniband/hw/mlx5/wr.c
588
set_sig_umr_segment(*seg, xlt_size);
drivers/infiniband/hw/mlx5/wr.c
589
*seg += sizeof(struct mlx5_wqe_umr_ctrl_seg);
drivers/infiniband/hw/mlx5/wr.c
591
handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
drivers/infiniband/hw/mlx5/wr.c
593
set_sig_mkey_segment(*seg, wr->mr, wr->access, xlt_size, region_len,
drivers/infiniband/hw/mlx5/wr.c
595
*seg += sizeof(struct mlx5_mkey_seg);
drivers/infiniband/hw/mlx5/wr.c
597
handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
drivers/infiniband/hw/mlx5/wr.c
599
ret = set_sig_data_segment(send_wr, wr->mr, sig_attrs, qp, seg, size,
drivers/infiniband/hw/mlx5/wr.c
609
u32 psv_idx, void **seg, int *size)
drivers/infiniband/hw/mlx5/wr.c
611
struct mlx5_seg_set_psv *psv_seg = *seg;
drivers/infiniband/hw/mlx5/wr.c
629
*seg += sizeof(*psv_seg);
drivers/infiniband/hw/mlx5/wr.c
637
void **seg, int *size, void **cur_edge,
drivers/infiniband/hw/mlx5/wr.c
670
set_reg_umr_seg(*seg, mr, flags, atomic);
drivers/infiniband/hw/mlx5/wr.c
671
*seg += sizeof(struct mlx5_wqe_umr_ctrl_seg);
drivers/infiniband/hw/mlx5/wr.c
673
handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
drivers/infiniband/hw/mlx5/wr.c
675
set_reg_mkey_seg(*seg, mr, wr->key, wr->access);
drivers/infiniband/hw/mlx5/wr.c
676
*seg += sizeof(struct mlx5_mkey_seg);
drivers/infiniband/hw/mlx5/wr.c
678
handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
drivers/infiniband/hw/mlx5/wr.c
681
mlx5r_memcpy_send_wqe(&qp->sq, cur_edge, seg, size, mr->descs,
drivers/infiniband/hw/mlx5/wr.c
685
set_reg_data_seg(*seg, mr, pd);
drivers/infiniband/hw/mlx5/wr.c
686
*seg += sizeof(struct mlx5_wqe_data_seg);
drivers/infiniband/hw/mlx5/wr.c
692
static void set_linv_wr(struct mlx5_ib_qp *qp, void **seg, int *size,
drivers/infiniband/hw/mlx5/wr.c
695
set_linv_umr_seg(*seg);
drivers/infiniband/hw/mlx5/wr.c
696
*seg += sizeof(struct mlx5_wqe_umr_ctrl_seg);
drivers/infiniband/hw/mlx5/wr.c
698
handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
drivers/infiniband/hw/mlx5/wr.c
699
set_linv_mkey_seg(*seg);
drivers/infiniband/hw/mlx5/wr.c
700
*seg += sizeof(struct mlx5_mkey_seg);
drivers/infiniband/hw/mlx5/wr.c
702
handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
drivers/infiniband/hw/mlx5/wr.c
724
int mlx5r_begin_wqe(struct mlx5_ib_qp *qp, void **seg,
drivers/infiniband/hw/mlx5/wr.c
733
*seg = mlx5_frag_buf_get_wqe(&qp->sq.fbc, *idx);
drivers/infiniband/hw/mlx5/wr.c
734
*ctrl = *seg;
drivers/infiniband/hw/mlx5/wr.c
735
*(uint32_t *)(*seg + 8) = 0;
drivers/infiniband/hw/mlx5/wr.c
741
*seg += sizeof(**ctrl);
drivers/infiniband/hw/mlx5/wr.c
748
static int begin_wqe(struct mlx5_ib_qp *qp, void **seg,
drivers/infiniband/hw/mlx5/wr.c
753
return mlx5r_begin_wqe(qp, seg, ctrl, idx, size, cur_edge, nreq,
drivers/infiniband/hw/mlx5/wr.c
759
void *seg, u8 size, void *cur_edge, unsigned int idx,
drivers/infiniband/hw/mlx5/wr.c
780
seg = PTR_ALIGN(seg, MLX5_SEND_WQE_BB);
drivers/infiniband/hw/mlx5/wr.c
781
qp->sq.cur_edge = (unlikely(seg == cur_edge)) ?
drivers/infiniband/hw/mlx5/wr.c
787
static void handle_rdma_op(const struct ib_send_wr *wr, void **seg, int *size)
drivers/infiniband/hw/mlx5/wr.c
789
set_raddr_seg(*seg, rdma_wr(wr)->remote_addr, rdma_wr(wr)->rkey);
drivers/infiniband/hw/mlx5/wr.c
790
*seg += sizeof(struct mlx5_wqe_raddr_seg);
drivers/infiniband/hw/mlx5/wr.c
795
struct mlx5_wqe_ctrl_seg **ctrl, void **seg,
drivers/infiniband/hw/mlx5/wr.c
800
set_linv_wr(qp, seg, size, cur_edge);
drivers/infiniband/hw/mlx5/wr.c
804
struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size,
drivers/infiniband/hw/mlx5/wr.c
809
return set_reg_wr(qp, reg_wr(wr), seg, size, cur_edge, true);
drivers/infiniband/hw/mlx5/wr.c
814
struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size,
drivers/infiniband/hw/mlx5/wr.c
824
err = mlx5r_begin_wqe(qp, seg, ctrl, idx, size, cur_edge, nreq,
drivers/infiniband/hw/mlx5/wr.c
831
err = set_psv_wr(domain, psv_index, seg, size);
drivers/infiniband/hw/mlx5/wr.c
836
mlx5r_finish_wqe(qp, *ctrl, *seg, *size, *cur_edge, *idx, wr->wr_id,
drivers/infiniband/hw/mlx5/wr.c
846
struct mlx5_wqe_ctrl_seg **ctrl, void **seg,
drivers/infiniband/hw/mlx5/wr.c
85
*seg += stride;
drivers/infiniband/hw/mlx5/wr.c
873
err = set_reg_wr(qp, &reg_pi_wr, seg, size, cur_edge, false);
drivers/infiniband/hw/mlx5/wr.c
877
mlx5r_finish_wqe(qp, *ctrl, *seg, *size, *cur_edge, *idx,
drivers/infiniband/hw/mlx5/wr.c
88
handle_post_send_edge(&qp->sq, seg, *size, cur_edge);
drivers/infiniband/hw/mlx5/wr.c
880
err = begin_wqe(qp, seg, ctrl, wr, idx, size, cur_edge, nreq);
drivers/infiniband/hw/mlx5/wr.c
904
err = set_pi_umr_wr(wr, qp, seg, size, cur_edge);
drivers/infiniband/hw/mlx5/wr.c
909
mlx5r_finish_wqe(qp, *ctrl, *seg, *size, *cur_edge, *idx, wr->wr_id,
drivers/infiniband/hw/mlx5/wr.c
91
mlx5r_memcpy_send_wqe(&qp->sq, cur_edge, seg, size,
drivers/infiniband/hw/mlx5/wr.c
913
err = handle_psv(dev, qp, wr, ctrl, seg, size, cur_edge, idx, nreq,
drivers/infiniband/hw/mlx5/wr.c
919
err = handle_psv(dev, qp, wr, ctrl, seg, size, cur_edge, idx, nreq,
drivers/infiniband/hw/mlx5/wr.c
933
struct mlx5_wqe_ctrl_seg **ctrl, void **seg, int *size,
drivers/infiniband/hw/mlx5/wr.c
943
handle_rdma_op(wr, seg, size);
drivers/infiniband/hw/mlx5/wr.c
954
handle_local_inv(qp, wr, ctrl, seg, size, cur_edge, *idx);
drivers/infiniband/hw/mlx5/wr.c
959
err = handle_reg_mr(qp, wr, ctrl, seg, size, cur_edge, *idx);
drivers/infiniband/hw/mlx5/wr.c
966
err = handle_reg_mr_integrity(dev, qp, wr, ctrl, seg, size,
drivers/infiniband/hw/mlx5/wr.c
98
*seg += sizeof(struct mlx5_wqe_eth_seg);
drivers/infiniband/hw/mlx5/wr.c
982
static void handle_qpt_uc(const struct ib_send_wr *wr, void **seg, int *size)
drivers/infiniband/hw/mlx5/wr.c
987
handle_rdma_op(wr, seg, size);
drivers/infiniband/hw/mlx5/wr.c
995
const struct ib_send_wr *wr, void **seg,
drivers/infiniband/hw/mlx5/wr.c
998
set_datagram_seg(*seg, wr);
drivers/infiniband/hw/mlx5/wr.c
999
*seg += sizeof(struct mlx5_wqe_datagram_seg);
drivers/infiniband/hw/mlx5/wr.h
100
void *seg, u8 size, void *cur_edge, unsigned int idx,
drivers/infiniband/hw/mlx5/wr.h
51
static inline void handle_post_send_edge(struct mlx5_ib_wq *sq, void **seg,
drivers/infiniband/hw/mlx5/wr.h
56
if (likely(*seg != *cur_edge))
drivers/infiniband/hw/mlx5/wr.h
62
*seg = mlx5_frag_buf_get_wqe(&sq->fbc, idx);
drivers/infiniband/hw/mlx5/wr.h
75
void **seg, u32 *wqe_sz,
drivers/infiniband/hw/mlx5/wr.h
79
size_t leftlen = *cur_edge - *seg;
drivers/infiniband/hw/mlx5/wr.h
83
memcpy(*seg, src, copysz);
drivers/infiniband/hw/mlx5/wr.h
88
*seg += stride;
drivers/infiniband/hw/mlx5/wr.h
90
handle_post_send_edge(sq, seg, *wqe_sz, cur_edge);
drivers/infiniband/hw/mlx5/wr.h
95
int mlx5r_begin_wqe(struct mlx5_ib_qp *qp, void **seg,
drivers/infiniband/hw/mthca/mthca_mr.c
104
__clear_bit(seg, buddy->bits[o]);
drivers/infiniband/hw/mthca/mthca_mr.c
109
seg <<= 1;
drivers/infiniband/hw/mthca/mthca_mr.c
110
__set_bit(seg ^ 1, buddy->bits[o]);
drivers/infiniband/hw/mthca/mthca_mr.c
116
seg <<= order;
drivers/infiniband/hw/mthca/mthca_mr.c
118
return seg;
drivers/infiniband/hw/mthca/mthca_mr.c
121
static void mthca_buddy_free(struct mthca_buddy *buddy, u32 seg, int order)
drivers/infiniband/hw/mthca/mthca_mr.c
123
seg >>= order;
drivers/infiniband/hw/mthca/mthca_mr.c
127
while (test_bit(seg ^ 1, buddy->bits[order])) {
drivers/infiniband/hw/mthca/mthca_mr.c
128
__clear_bit(seg ^ 1, buddy->bits[order]);
drivers/infiniband/hw/mthca/mthca_mr.c
130
seg >>= 1;
drivers/infiniband/hw/mthca/mthca_mr.c
134
__set_bit(seg, buddy->bits[order]);
drivers/infiniband/hw/mthca/mthca_mr.c
190
u32 seg = mthca_buddy_alloc(buddy, order);
drivers/infiniband/hw/mthca/mthca_mr.c
192
if (seg == -1)
drivers/infiniband/hw/mthca/mthca_mr.c
196
if (mthca_table_get_range(dev, dev->mr_table.mtt_table, seg,
drivers/infiniband/hw/mthca/mthca_mr.c
197
seg + (1 << order) - 1)) {
drivers/infiniband/hw/mthca/mthca_mr.c
198
mthca_buddy_free(buddy, seg, order);
drivers/infiniband/hw/mthca/mthca_mr.c
199
seg = -1;
drivers/infiniband/hw/mthca/mthca_mr.c
202
return seg;
drivers/infiniband/hw/mthca/mthca_mr.c
88
u32 seg;
drivers/infiniband/hw/mthca/mthca_mr.c
95
seg = find_first_bit(buddy->bits[o], m);
drivers/infiniband/hw/mthca/mthca_mr.c
96
if (seg < m)
drivers/infiniband/sw/siw/siw_qp_tx.c
445
int seg = 0, do_crc = c_tx->do_crc, is_kva = 0, rv;
drivers/infiniband/sw/siw/siw_qp_tx.c
463
seg = 1;
drivers/infiniband/sw/siw/siw_qp_tx.c
485
iov[seg].iov_base =
drivers/infiniband/sw/siw/siw_qp_tx.c
487
iov[seg].iov_len = sge_len;
drivers/infiniband/sw/siw/siw_qp_tx.c
491
iov[seg].iov_base, sge_len);
drivers/infiniband/sw/siw/siw_qp_tx.c
494
seg++;
drivers/infiniband/sw/siw/siw_qp_tx.c
507
siw_unmap_pages(iov, kmap_mask, seg);
drivers/infiniband/sw/siw/siw_qp_tx.c
512
page_array[seg] = p;
drivers/infiniband/sw/siw/siw_qp_tx.c
518
kmap_mask |= BIT(seg);
drivers/infiniband/sw/siw/siw_qp_tx.c
519
iov[seg].iov_base = kaddr + fp_off;
drivers/infiniband/sw/siw/siw_qp_tx.c
520
iov[seg].iov_len = plen;
drivers/infiniband/sw/siw/siw_qp_tx.c
525
iov[seg].iov_base,
drivers/infiniband/sw/siw/siw_qp_tx.c
540
page_array[seg] = ib_virt_dma_to_page(va);
drivers/infiniband/sw/siw/siw_qp_tx.c
552
if (++seg >= (int)MAX_ARRAY) {
drivers/infiniband/sw/siw/siw_qp_tx.c
554
siw_unmap_pages(iov, kmap_mask, seg-1);
drivers/infiniband/sw/siw/siw_qp_tx.c
571
iov[seg].iov_base = &c_tx->trailer.pad[4 - c_tx->pad];
drivers/infiniband/sw/siw/siw_qp_tx.c
572
iov[seg].iov_len = trl_len = MAX_TRAILER - (4 - c_tx->pad);
drivers/infiniband/sw/siw/siw_qp_tx.c
574
iov[seg].iov_base = &c_tx->trailer.pad[c_tx->ctrl_sent];
drivers/infiniband/sw/siw/siw_qp_tx.c
575
iov[seg].iov_len = trl_len = MAX_TRAILER - c_tx->ctrl_sent;
drivers/infiniband/sw/siw/siw_qp_tx.c
597
rv = siw_sendmsg(s, MSG_DONTWAIT | MSG_EOR, &iov[seg],
drivers/infiniband/sw/siw/siw_qp_tx.c
605
rv = siw_sendmsg(s, MSG_DONTWAIT | MSG_EOR, iov, seg + 1,
drivers/infiniband/sw/siw/siw_qp_tx.c
607
siw_unmap_pages(iov, kmap_mask, seg);
drivers/input/touch-overlay.c
125
static bool touch_overlay_segment_event(struct touch_overlay_segment *seg,
drivers/input/touch-overlay.c
128
if (pos->x >= seg->x_origin && pos->x < (seg->x_origin + seg->x_size) &&
drivers/input/touch-overlay.c
129
pos->y >= seg->y_origin && pos->y < (seg->y_origin + seg->y_size))
drivers/iommu/amd/amd_iommu.h
144
int seg = pci_domain_nr(pdev->bus);
drivers/iommu/amd/amd_iommu.h
147
return PCI_SEG_DEVID_TO_SBDF(seg, devid);
drivers/iommu/amd/amd_iommu_types.h
444
#define PCI_SEG_DEVID_TO_SBDF(seg, devid) ((((u32)(seg) & 0xffff) << 16) | \
drivers/iommu/amd/debugfs.c
142
int seg, bus, slot, func;
drivers/iommu/amd/debugfs.c
157
i = sscanf(srcid_ptr, "%x:%x:%x.%x", &seg, &bus, &slot, &func);
drivers/iommu/amd/debugfs.c
164
seg = 0;
drivers/iommu/amd/debugfs.c
171
if (pci_seg->id != seg)
drivers/iommu/amd/debugfs.c
185
if (pci_seg->id != seg) {
drivers/iommu/amd/debugfs.c
190
sbdf = PCI_SEG_DEVID_TO_SBDF(seg, devid);
drivers/iommu/amd/debugfs.c
239
u16 seg, devid;
drivers/iommu/amd/debugfs.c
245
seg = PCI_SBDF_TO_SEGID(sbdf);
drivers/iommu/amd/debugfs.c
249
if (pci_seg->id != seg)
drivers/iommu/amd/debugfs.c
338
u16 devid, seg;
drivers/iommu/amd/debugfs.c
350
seg = PCI_SBDF_TO_SEGID(sbdf);
drivers/iommu/amd/debugfs.c
354
if (pci_seg->id != seg)
drivers/iommu/amd/init.c
3699
u32 seg = 0, bus, dev, fn;
drivers/iommu/amd/init.c
3704
sscanf(str, "=%d@%x:%x:%x.%x", &id, &seg, &bus, &dev, &fn) == 5)
drivers/iommu/amd/init.c
3708
sscanf(str, "[%d]=%x:%x:%x.%x", &id, &seg, &bus, &dev, &fn) == 5) {
drivers/iommu/amd/init.c
3710
str, id, seg, bus, dev, fn);
drivers/iommu/amd/init.c
3724
devid = IVRS_GET_SBDF_ID(seg, bus, dev, fn);
drivers/iommu/amd/init.c
3737
u32 seg = 0, bus, dev, fn;
drivers/iommu/amd/init.c
3742
sscanf(str, "=%d@%x:%x:%x.%x", &id, &seg, &bus, &dev, &fn) == 5)
drivers/iommu/amd/init.c
3746
sscanf(str, "[%d]=%x:%x:%x.%x", &id, &seg, &bus, &dev, &fn) == 5) {
drivers/iommu/amd/init.c
3748
str, id, seg, bus, dev, fn);
drivers/iommu/amd/init.c
3762
devid = IVRS_GET_SBDF_ID(seg, bus, dev, fn);
drivers/iommu/amd/init.c
3777
u32 seg = 0, bus, dev, fn;
drivers/iommu/amd/init.c
3794
sscanf(str, "[%x:%x:%x.%x]=%s", &seg, &bus, &dev, &fn, acpiid) == 5) {
drivers/iommu/amd/init.c
3796
str, acpiid, seg, bus, dev, fn);
drivers/iommu/amd/init.c
3812
sscanf(addr, "%x:%x:%x.%x", &seg, &bus, &dev, &fn) == 4)
drivers/iommu/amd/init.c
3847
early_acpihid_map[i].devid = IVRS_GET_SBDF_ID(seg, bus, dev, fn);
drivers/iommu/amd/init.c
87
#define IVRS_GET_SBDF_ID(seg, bus, dev, fn) (((seg & 0xffff) << 16) | ((bus & 0xff) << 8) \
drivers/iommu/amd/iommu.c
326
u16 seg;
drivers/iommu/amd/iommu.c
331
seg = pci_domain_nr(pdev->bus);
drivers/iommu/amd/iommu.c
335
seg = PCI_SBDF_TO_SEGID(devid);
drivers/iommu/amd/iommu.c
338
return seg;
drivers/iommu/amd/iommu.c
349
static struct amd_iommu *__rlookup_amd_iommu(u16 seg, u16 devid)
drivers/iommu/amd/iommu.c
354
if (pci_seg->id == seg)
drivers/iommu/amd/iommu.c
362
u16 seg = get_device_segment(dev);
drivers/iommu/amd/iommu.c
367
return __rlookup_amd_iommu(seg, PCI_SBDF_TO_DEVID(devid));
drivers/iommu/amd/iommu.c
3744
int i, ret, devid, seg, sbdf;
drivers/iommu/amd/iommu.c
3756
seg = PCI_SBDF_TO_SEGID(sbdf);
drivers/iommu/amd/iommu.c
3758
iommu = __rlookup_amd_iommu(seg, devid);
drivers/iommu/intel/debugfs.c
352
u16 devfn, bus, seg;
drivers/iommu/intel/debugfs.c
356
seg = info->segment;
drivers/iommu/intel/debugfs.c
364
if (seg != iommu->segment)
drivers/iommu/intel/dmar.c
157
info->seg = pci_domain_nr(dev->bus);
drivers/iommu/intel/dmar.c
225
if (segment != info->seg)
drivers/iommu/intel/dmar.c
279
if (info->seg != segment)
drivers/isdn/mISDN/dsp_audio.c
48
int seg;
drivers/isdn/mISDN/dsp_audio.c
65
for (seg = 0; seg < 8; seg++) {
drivers/isdn/mISDN/dsp_audio.c
66
if (pcm_val <= seg_end[seg])
drivers/isdn/mISDN/dsp_audio.c
70
return ((seg << 4) |
drivers/isdn/mISDN/dsp_audio.c
71
((pcm_val >> ((seg) ? (seg + 3) : 4)) & 0x0F)) ^ mask;
drivers/isdn/mISDN/dsp_audio.c
78
int seg;
drivers/isdn/mISDN/dsp_audio.c
82
seg = (((int) alaw & 0x70) >> 4);
drivers/isdn/mISDN/dsp_audio.c
83
if (seg)
drivers/isdn/mISDN/dsp_audio.c
84
i = (i + 0x100) << (seg - 1);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1029
for (i = 0; i < ARRAY_SIZE(instance->seg); i++)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1030
if (instance->seg[i].va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1031
mtk_vcodec_mem_free(ctx, &instance->seg[i]);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1155
static void vdec_av1_slice_setup_seg(struct vdec_av1_slice_seg *seg,
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1160
seg->segmentation_enabled = SEGMENTATION_FLAG(ctrl_seg, ENABLED);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1161
seg->segmentation_update_map = SEGMENTATION_FLAG(ctrl_seg, UPDATE_MAP);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1162
seg->segmentation_temporal_update = SEGMENTATION_FLAG(ctrl_seg, TEMPORAL_UPDATE);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1163
seg->segmentation_update_data = SEGMENTATION_FLAG(ctrl_seg, UPDATE_DATA);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1164
seg->segid_preskip = SEGMENTATION_FLAG(ctrl_seg, SEG_ID_PRE_SKIP);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1165
seg->last_active_segid = ctrl_seg->last_active_seg_id;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1168
seg->feature_enabled_mask[i] = ctrl_seg->feature_enabled[i];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1170
seg->feature_data[i][j] = ctrl_seg->feature_data[i][j];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1192
struct vdec_av1_slice_seg *seg = &uh->seg;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1196
if (seg->segmentation_enabled &&
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1197
(seg->feature_enabled_mask[segmentation_id] & BIT(SEG_LVL_ALT_Q))) {
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1198
data = seg->feature_data[segmentation_id][SEG_LVL_ALT_Q];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1358
vdec_av1_slice_setup_seg(&uh->seg, &ctrl_fh->segmentation);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1621
work_buffer[i].segid_addr.buf = instance->seg[i].dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1622
work_buffer[i].segid_addr.size = instance->seg[i].size;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1647
if (uh->primary_ref_frame == AV1_PRIMARY_REF_NONE || !uh->seg.segmentation_enabled) {
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
1650
buf = &instance->seg[vsi->slot_id];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
482
struct vdec_av1_slice_seg seg;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
716
struct mtk_vcodec_mem seg[AV1_MAX_FRAME_BUF_COUNT];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
980
if (instance->seg[i].va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
981
mtk_vcodec_mem_free(ctx, &instance->seg[i]);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
982
instance->seg[i].size = max_sb_w * max_sb_h * 512;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_av1_req_lat_if.c
983
ret = mtk_vcodec_mem_alloc(ctx, &instance->seg[i]);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1173
vdec_vp9_slice_setup_seg_buffer(instance, vsi, &instance->seg[0]);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1747
vsi->seg[i].dma_addr = instance->seg[i].dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1748
vsi->seg[i].size = instance->seg[i].size;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1781
vdec_vp9_slice_setup_seg_buffer(instance, vsi, &instance->seg[1]);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
1810
vdec_vp9_slice_setup_seg_buffer(instance, vsi, &instance->seg[0]);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
30
#define SEG_FLAG(x) (!!((seg)->flags & V4L2_VP9_SEGMENTATION_FLAG_##x))
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
360
struct vdec_vp9_slice_mem seg[2];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
44
u8 seg[32]; /* ignore */
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
472
struct mtk_vcodec_mem seg[2];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
601
if (instance->seg[i].va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
602
mtk_vcodec_mem_free(ctx, &instance->seg[i]);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
603
instance->seg[i].size = size;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
604
if (mtk_vcodec_mem_alloc(ctx, &instance->seg[i]))
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
643
for (i = 0; i < ARRAY_SIZE(instance->seg); i++) {
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
644
if (instance->seg[i].va)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
645
mtk_vcodec_mem_free(ctx, &instance->seg[i]);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
857
struct v4l2_vp9_segmentation *seg)
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
865
uh->segmentation_tree_probs[i] = seg->tree_probs[i];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
868
uh->segmentation_pred_prob[i] = seg->pred_probs[i];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
872
uh->feature_enabled[i] = seg->feature_enabled[i];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
874
uh->feature_value[i][j] = seg->feature_data[i][j];
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
962
vdec_vp9_slice_setup_segmentation(uh, &hdr->seg);
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
994
vsi->seg[i].dma_addr = instance->seg[i].dma_addr;
drivers/media/platform/mediatek/vcodec/decoder/vdec/vdec_vp9_req_lat_if.c
995
vsi->seg[i].size = instance->seg[i].size;
drivers/media/platform/rockchip/rkvdec/rkvdec-vp9.c
153
struct v4l2_vp9_segmentation seg;
drivers/media/platform/rockchip/rkvdec/rkvdec-vp9.c
318
const struct v4l2_vp9_segmentation *seg;
drivers/media/platform/rockchip/rkvdec/rkvdec-vp9.c
324
seg = &dec_params->seg;
drivers/media/platform/rockchip/rkvdec/rkvdec-vp9.c
337
memcpy(rkprobs->pred, seg->pred_probs, sizeof(rkprobs->pred));
drivers/media/platform/rockchip/rkvdec/rkvdec-vp9.c
338
memcpy(rkprobs->tree, seg->tree_probs, sizeof(rkprobs->tree));
drivers/media/platform/rockchip/rkvdec/rkvdec-vp9.c
415
const struct v4l2_vp9_segmentation *seg;
drivers/media/platform/rockchip/rkvdec/rkvdec-vp9.c
419
seg = vp9_ctx->last.valid ? &vp9_ctx->last.seg : &vp9_ctx->cur.seg;
drivers/media/platform/rockchip/rkvdec/rkvdec-vp9.c
421
if (v4l2_vp9_seg_feat_enabled(seg->feature_enabled, feature_id, segid)) {
drivers/media/platform/rockchip/rkvdec/rkvdec-vp9.c
422
feature_val = seg->feature_data[segid][feature_id];
drivers/media/platform/rockchip/rkvdec/rkvdec-vp9.c
428
if (v4l2_vp9_seg_feat_enabled(seg->feature_enabled, feature_id, segid)) {
drivers/media/platform/rockchip/rkvdec/rkvdec-vp9.c
429
feature_val = seg->feature_data[segid][feature_id];
drivers/media/platform/rockchip/rkvdec/rkvdec-vp9.c
435
if (v4l2_vp9_seg_feat_enabled(seg->feature_enabled, feature_id, segid)) {
drivers/media/platform/rockchip/rkvdec/rkvdec-vp9.c
436
feature_val = seg->feature_data[segid][feature_id];
drivers/media/platform/rockchip/rkvdec/rkvdec-vp9.c
443
v4l2_vp9_seg_feat_enabled(seg->feature_enabled, feature_id, segid);
drivers/media/platform/rockchip/rkvdec/rkvdec-vp9.c
446
(seg->flags & V4L2_VP9_SEGMENTATION_FLAG_ABS_OR_DELTA_UPDATE);
drivers/media/platform/rockchip/rkvdec/rkvdec-vp9.c
466
vp9_ctx->cur.seg = dec_params->seg;
drivers/media/platform/rockchip/rkvdec/rkvdec-vp9.c
484
const struct v4l2_vp9_segmentation *seg;
drivers/media/platform/rockchip/rkvdec/rkvdec-vp9.c
503
seg = &dec_params->seg;
drivers/media/platform/rockchip/rkvdec/rkvdec-vp9.c
541
(!(seg->flags & V4L2_VP9_SEGMENTATION_FLAG_ENABLED) ||
drivers/media/platform/rockchip/rkvdec/rkvdec-vp9.c
542
(seg->flags & V4L2_VP9_SEGMENTATION_FLAG_UPDATE_MAP)))
drivers/media/platform/rockchip/rkvdec/rkvdec-vp9.c
596
vp9_ctx->last.seg.flags & V4L2_VP9_SEGMENTATION_FLAG_ENABLED;
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
137
const struct v4l2_vp8_segment *seg = &hdr->segment;
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
143
if (!(seg->flags & V4L2_VP8_SEGMENT_FLAG_ENABLED)) {
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
145
} else if (seg->flags & V4L2_VP8_SEGMENT_FLAG_DELTA_VALUE_MODE) {
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
147
u32 lf_level = clamp(lf->level + seg->lf_update[i],
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
155
seg->lf_update[i]);
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
180
const struct v4l2_vp8_segment *seg = &hdr->segment;
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
184
if (!(seg->flags & V4L2_VP8_SEGMENT_FLAG_ENABLED)) {
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
186
} else if (seg->flags & V4L2_VP8_SEGMENT_FLAG_DELTA_VALUE_MODE) {
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
188
u32 quant = clamp(q->y_ac_qi + seg->quant_update[i],
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
196
seg->quant_update[i]);
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
410
const struct v4l2_vp8_segment *seg = &hdr->segment;
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
421
if (seg->flags & V4L2_VP8_SEGMENT_FLAG_ENABLED) {
drivers/media/platform/verisilicon/hantro_g1_vp8_dec.c
423
if (seg->flags & V4L2_VP8_SEGMENT_FLAG_UPDATE_MAP)
drivers/media/platform/verisilicon/hantro_g2_vp9_dec.c
363
const struct v4l2_vp9_segmentation *seg,
drivers/media/platform/verisilicon/hantro_g2_vp9_dec.c
369
vp9_ctx->feature_data[segid][feature] = seg->feature_data[segid][feature];
drivers/media/platform/verisilicon/hantro_g2_vp9_dec.c
371
vp9_ctx->feature_enabled[segid] |= (seg->feature_enabled[segid] & mask);
drivers/media/platform/verisilicon/hantro_g2_vp9_dec.c
390
const struct v4l2_vp9_segmentation *seg;
drivers/media/platform/verisilicon/hantro_g2_vp9_dec.c
407
segment_enabled = !!(dec_params->seg.flags & V4L2_VP9_SEGMENTATION_FLAG_ENABLED);
drivers/media/platform/verisilicon/hantro_g2_vp9_dec.c
410
!!(dec_params->seg.flags & V4L2_VP9_SEGMENTATION_FLAG_UPDATE_MAP));
drivers/media/platform/verisilicon/hantro_g2_vp9_dec.c
412
!!(dec_params->seg.flags & V4L2_VP9_SEGMENTATION_FLAG_TEMPORAL_UPDATE));
drivers/media/platform/verisilicon/hantro_g2_vp9_dec.c
414
seg = &dec_params->seg;
drivers/media/platform/verisilicon/hantro_g2_vp9_dec.c
415
absolute = !!(seg->flags & V4L2_VP9_SEGMENTATION_FLAG_ABS_OR_DELTA_UPDATE);
drivers/media/platform/verisilicon/hantro_g2_vp9_dec.c
416
update_data = !!(seg->flags & V4L2_VP9_SEGMENTATION_FLAG_UPDATE_DATA);
drivers/media/platform/verisilicon/hantro_g2_vp9_dec.c
424
update_feat_and_flag(vp9_ctx, seg, feat_id, segid);
drivers/media/platform/verisilicon/hantro_g2_vp9_dec.c
437
update_feat_and_flag(vp9_ctx, seg, feat_id, segid);
drivers/media/platform/verisilicon/hantro_g2_vp9_dec.c
450
update_feat_and_flag(vp9_ctx, seg, feat_id, segid);
drivers/media/platform/verisilicon/hantro_g2_vp9_dec.c
462
update_feat_and_flag(vp9_ctx, seg, feat_id, segid);
drivers/media/platform/verisilicon/hantro_g2_vp9_dec.c
639
const struct v4l2_vp9_segmentation *seg = &dec_params->seg;
drivers/media/platform/verisilicon/hantro_g2_vp9_dec.c
653
memcpy(all_probs->mb_segment_tree_probs, seg->tree_probs,
drivers/media/platform/verisilicon/hantro_g2_vp9_dec.c
656
memcpy(all_probs->segment_pred_probs, seg->pred_probs,
drivers/media/platform/verisilicon/hantro_g2_vp9_dec.c
814
const struct v4l2_vp9_segmentation *seg;
drivers/media/platform/verisilicon/hantro_g2_vp9_dec.c
827
seg = &dec_params->seg;
drivers/media/platform/verisilicon/hantro_g2_vp9_dec.c
885
seg->flags & V4L2_VP9_SEGMENTATION_FLAG_UPDATE_MAP);
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
278
const struct v4l2_vp8_segment *seg = &hdr->segment;
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
284
if (!(seg->flags & V4L2_VP8_SEGMENT_FLAG_ENABLED)) {
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
286
} else if (seg->flags & V4L2_VP8_SEGMENT_FLAG_DELTA_VALUE_MODE) {
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
288
u32 lf_level = clamp(lf->level + seg->lf_update[i],
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
296
seg->lf_update[i]);
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
318
const struct v4l2_vp8_segment *seg = &hdr->segment;
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
322
if (!(seg->flags & V4L2_VP8_SEGMENT_FLAG_ENABLED)) {
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
324
} else if (seg->flags & V4L2_VP8_SEGMENT_FLAG_DELTA_VALUE_MODE) {
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
326
u32 quant = clamp(q->y_ac_qi + seg->quant_update[i],
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
334
seg->quant_update[i]);
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
486
const struct v4l2_vp8_segment *seg = &hdr->segment;
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
497
if (seg->flags & V4L2_VP8_SEGMENT_FLAG_ENABLED) {
drivers/media/platform/verisilicon/rockchip_vpu2_hw_vp8_dec.c
499
if (seg->flags & V4L2_VP8_SEGMENT_FLAG_UPDATE_MAP)
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
802
const struct v4l2_av1_segmentation *seg = &frame->segmentation;
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
807
if (!!(seg->flags & V4L2_AV1_SEGMENTATION_FLAG_ENABLED) &&
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
813
struct hantro_decoded_buffer *seg;
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
816
seg = vb2_to_hantro_decoded_buf(&av1_dec->frame_refs[idx].vb2_ref->vb2_buf);
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
817
luma_addr = hantro_get_dec_buf_addr(ctx, &seg->base.vb.vb2_buf);
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
826
!!(seg->flags & V4L2_AV1_SEGMENTATION_FLAG_TEMPORAL_UPDATE));
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
828
!!(seg->flags & V4L2_AV1_SEGMENTATION_FLAG_UPDATE_MAP));
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
830
!!(seg->flags & V4L2_AV1_SEGMENTATION_FLAG_ENABLED));
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
840
if (seg->flags & V4L2_AV1_SEGMENTATION_FLAG_ENABLED) {
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
844
if (seg->feature_enabled[s] &
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
847
clamp(abs(seg->feature_data[s][V4L2_AV1_SEG_LVL_ALT_Q]),
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
850
(seg->feature_data[s][V4L2_AV1_SEG_LVL_ALT_Q] < 0) << s;
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
853
if (seg->feature_enabled[s] &
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
856
clamp(abs(seg->feature_data[s][V4L2_AV1_SEG_LVL_ALT_LF_Y_V]),
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
859
if (seg->feature_enabled[s] &
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
862
clamp(abs(seg->feature_data[s][V4L2_AV1_SEG_LVL_ALT_LF_Y_H]),
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
865
if (seg->feature_enabled[s] &
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
868
clamp(abs(seg->feature_data[s][V4L2_AV1_SEG_LVL_ALT_LF_U]),
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
871
if (seg->feature_enabled[s] &
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
874
clamp(abs(seg->feature_data[s][V4L2_AV1_SEG_LVL_ALT_LF_V]),
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
877
if (frame->frame_type && seg->feature_enabled[s] &
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
881
if (seg->feature_enabled[s] &
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
885
if (seg->feature_enabled[s] &
drivers/media/platform/verisilicon/rockchip_vpu981_hw_av1_dec.c
893
if (seg->feature_enabled[i]
drivers/media/test-drivers/visl/visl-trace-vp9.h
65
__entry->f.seg.feature_data,
drivers/media/test-drivers/visl/visl-trace-vp9.h
66
sizeof(__entry->f.seg.feature_data),
drivers/media/test-drivers/visl/visl-trace-vp9.h
68
__print_array(__entry->f.seg.feature_enabled,
drivers/media/test-drivers/visl/visl-trace-vp9.h
69
ARRAY_SIZE(__entry->f.seg.feature_enabled),
drivers/media/test-drivers/visl/visl-trace-vp9.h
70
sizeof(__entry->f.seg.feature_enabled[0])),
drivers/media/test-drivers/visl/visl-trace-vp9.h
71
__print_array(__entry->f.seg.tree_probs,
drivers/media/test-drivers/visl/visl-trace-vp9.h
72
ARRAY_SIZE(__entry->f.seg.tree_probs),
drivers/media/test-drivers/visl/visl-trace-vp9.h
73
sizeof(__entry->f.seg.tree_probs[0])),
drivers/media/test-drivers/visl/visl-trace-vp9.h
74
__print_array(__entry->f.seg.pred_probs,
drivers/media/test-drivers/visl/visl-trace-vp9.h
75
ARRAY_SIZE(__entry->f.seg.pred_probs),
drivers/media/test-drivers/visl/visl-trace-vp9.h
76
sizeof(__entry->f.seg.pred_probs[0])),
drivers/media/test-drivers/visl/visl-trace-vp9.h
77
__print_flags(__entry->f.seg.flags, "|",
drivers/media/v4l2-core/v4l2-ctrls-core.c
532
validate_vp9_seg_params(struct v4l2_vp9_segmentation *seg)
drivers/media/v4l2-core/v4l2-ctrls-core.c
536
if (seg->flags & ~(V4L2_VP9_SEGMENTATION_FLAG_ENABLED |
drivers/media/v4l2-core/v4l2-ctrls-core.c
543
for (i = 0; i < ARRAY_SIZE(seg->feature_enabled); i++) {
drivers/media/v4l2-core/v4l2-ctrls-core.c
544
if (seg->feature_enabled[i] &
drivers/media/v4l2-core/v4l2-ctrls-core.c
549
for (i = 0; i < ARRAY_SIZE(seg->feature_data); i++) {
drivers/media/v4l2-core/v4l2-ctrls-core.c
552
for (j = 0; j < ARRAY_SIZE(seg->feature_data[j]); j++) {
drivers/media/v4l2-core/v4l2-ctrls-core.c
553
if (seg->feature_data[i][j] < -range[j] ||
drivers/media/v4l2-core/v4l2-ctrls-core.c
554
seg->feature_data[i][j] > range[j])
drivers/media/v4l2-core/v4l2-ctrls-core.c
559
zero_reserved(*seg);
drivers/media/v4l2-core/v4l2-ctrls-core.c
645
ret = validate_vp9_seg_params(&frame->seg);
drivers/net/ethernet/asix/ax88796c_main.c
203
info->seg.flags_seqnum_seglen = TX_HDR_SEG_FS | TX_HDR_SEG_LS
drivers/net/ethernet/asix/ax88796c_main.c
206
info->seg.eo_so_seglenbar = pkt_len_bar;
drivers/net/ethernet/asix/ax88796c_main.c
208
cpu_to_be16s(&info->seg.flags_seqnum_seglen);
drivers/net/ethernet/asix/ax88796c_main.c
209
cpu_to_be16s(&info->seg.eo_so_seglenbar);
drivers/net/ethernet/asix/ax88796c_main.h
554
struct tx_segment_header seg;
drivers/net/ethernet/broadcom/tg3.c
7926
struct sk_buff *segs, *seg, *next;
drivers/net/ethernet/broadcom/tg3.c
7951
skb_list_walk_safe(segs, seg, next) {
drivers/net/ethernet/broadcom/tg3.c
7952
skb_mark_not_on_list(seg);
drivers/net/ethernet/broadcom/tg3.c
7953
__tg3_start_xmit(seg, tp->dev);
drivers/net/ethernet/google/gve/gve.h
352
struct gve_tx_seg_desc seg; /* subsequent descs for a packet */
drivers/net/ethernet/google/gve/gve_tx.c
539
seg_desc->seg.type_flags = GVE_TXD_SEG;
drivers/net/ethernet/google/gve/gve_tx.c
542
seg_desc->seg.type_flags |= GVE_TXSF_IPV6;
drivers/net/ethernet/google/gve/gve_tx.c
543
seg_desc->seg.l3_offset = l3_offset >> 1;
drivers/net/ethernet/google/gve/gve_tx.c
544
seg_desc->seg.mss = cpu_to_be16(gso_size);
drivers/net/ethernet/google/gve/gve_tx.c
546
seg_desc->seg.seg_len = cpu_to_be16(len);
drivers/net/ethernet/google/gve/gve_tx.c
547
seg_desc->seg.seg_addr = cpu_to_be64(addr);
drivers/net/ethernet/huawei/hinic/hinic_hw_mbox.c
677
struct hinic_send_mbox *mbox, void *seg,
drivers/net/ethernet/huawei/hinic/hinic_hw_mbox.c
682
u32 *data = seg;
drivers/net/ethernet/huawei/hinic/hinic_hw_mbox.c
687
memcpy(mbox_max_buf, seg, seg_len);
drivers/net/ethernet/huawei/hinic/hinic_hw_mbox.c
797
u64 header, u16 dst_func, void *seg, u16 seg_len,
drivers/net/ethernet/huawei/hinic/hinic_hw_mbox.c
820
mbox_copy_send_data(hwdev, send_mbox, seg, seg_len);
drivers/net/ethernet/huawei/hinic3/hinic3_mbox.c
571
struct hinic3_send_mbox *mbox, void *seg,
drivers/net/ethernet/huawei/hinic3/hinic3_mbox.c
576
const __le32 *src = seg;
drivers/net/ethernet/huawei/hinic3/hinic3_mbox.c
643
u16 dst_func, void *seg, u32 seg_len, void *msg_info)
drivers/net/ethernet/huawei/hinic3/hinic3_mbox.c
663
mbox_copy_send_data(hwdev, send_mbox, seg, seg_len);
drivers/net/ethernet/intel/ice/ice_ddp.c
1362
struct ice_generic_seg_hdr *seg;
drivers/net/ethernet/intel/ice/ice_ddp.c
1364
seg = ice_get_pkg_seg_by_idx(pkg_hdr, idx);
drivers/net/ethernet/intel/ice/ice_ddp.c
1365
if (!seg)
drivers/net/ethernet/intel/ice/ice_ddp.c
1368
return le32_to_cpu(seg->seg_type) == SEGMENT_TYPE_SIGNING;
drivers/net/ethernet/intel/ice/ice_ddp.c
1384
struct ice_sign_seg *seg;
drivers/net/ethernet/intel/ice/ice_ddp.c
1389
seg = (struct ice_sign_seg *)ice_get_pkg_seg_by_idx(pkg_hdr, idx);
drivers/net/ethernet/intel/ice/ice_ddp.c
1391
if (seg && le32_to_cpu(seg->seg_id) == seg_id &&
drivers/net/ethernet/intel/ice/ice_ddp.c
1392
le32_to_cpu(seg->sign_type) == sign_type)
drivers/net/ethernet/intel/ice/ice_ddp.c
1404
ice_download_pkg_sig_seg(struct ice_ddp_send_ctx *ctx, struct ice_sign_seg *seg)
drivers/net/ethernet/intel/ice/ice_ddp.c
1406
return ice_dwnld_cfg_bufs_no_lock(ctx, seg->buf_tbl.buf_array, 0,
drivers/net/ethernet/intel/ice/ice_ddp.c
1407
le32_to_cpu(seg->buf_tbl.buf_count));
drivers/net/ethernet/intel/ice/ice_ddp.c
1426
struct ice_seg *seg;
drivers/net/ethernet/intel/ice/ice_ddp.c
1429
seg = (struct ice_seg *)ice_get_pkg_seg_by_idx(pkg_hdr, idx);
drivers/net/ethernet/intel/ice/ice_ddp.c
1430
if (!seg)
drivers/net/ethernet/intel/ice/ice_ddp.c
1433
bufs = ice_find_buf_table(seg);
drivers/net/ethernet/intel/ice/ice_ddp.c
1462
struct ice_sign_seg *seg;
drivers/net/ethernet/intel/ice/ice_ddp.c
1464
seg = (struct ice_sign_seg *)ice_get_pkg_seg_by_idx(pkg_hdr, idx);
drivers/net/ethernet/intel/ice/ice_ddp.c
1465
if (!seg) {
drivers/net/ethernet/intel/ice/ice_ddp.c
1471
count = le32_to_cpu(seg->signed_buf_count);
drivers/net/ethernet/intel/ice/ice_ddp.c
1472
state = ice_download_pkg_sig_seg(ctx, seg);
drivers/net/ethernet/intel/ice/ice_ddp.c
1476
conf_idx = le32_to_cpu(seg->signed_seg_idx);
drivers/net/ethernet/intel/ice/ice_ddp.c
1477
start = le32_to_cpu(seg->signed_buf_start);
drivers/net/ethernet/intel/ice/ice_ddp.c
1485
flags = le32_to_cpu(seg->flags);
drivers/net/ethernet/intel/ice/ice_ddp.c
1841
const struct ice_generic_seg_hdr *seg;
drivers/net/ethernet/intel/ice/ice_ddp.c
1843
seg = (void *)pkg_hdr + le32_to_cpu(pkg_hdr->seg_offset[i]);
drivers/net/ethernet/intel/ice/ice_ddp.c
1845
if (le32_to_cpu(seg->seg_type) == seg_type)
drivers/net/ethernet/intel/ice/ice_ddp.c
1846
return seg;
drivers/net/ethernet/intel/ice/ice_ddp.c
2043
struct ice_seg **seg)
drivers/net/ethernet/intel/ice/ice_ddp.c
2059
*seg = (struct ice_seg *)ice_find_seg_in_pkg(hw, hw->pkg_seg_id,
drivers/net/ethernet/intel/ice/ice_ddp.c
2061
if (!*seg) {
drivers/net/ethernet/intel/ice/ice_ddp.c
2074
if ((*seg)->hdr.seg_format_ver.major !=
drivers/net/ethernet/intel/ice/ice_ddp.c
2076
(*seg)->hdr.seg_format_ver.minor >
drivers/net/ethernet/intel/ice/ice_ddp.c
2155
struct ice_seg *seg = hw->seg;
drivers/net/ethernet/intel/ice/ice_ddp.c
2159
if (!seg)
drivers/net/ethernet/intel/ice/ice_ddp.c
2165
tcam = ice_pkg_enum_entry(seg, &state,
drivers/net/ethernet/intel/ice/ice_ddp.c
2173
seg = NULL;
drivers/net/ethernet/intel/ice/ice_ddp.c
2207
struct ice_seg *seg;
drivers/net/ethernet/intel/ice/ice_ddp.c
2233
state = ice_chk_pkg_compat(hw, pkg, &seg);
drivers/net/ethernet/intel/ice/ice_ddp.c
2238
ice_init_pkg_hints(hw, seg);
drivers/net/ethernet/intel/ice/ice_ddp.c
2239
state = ice_download_pkg(hw, pkg, seg);
drivers/net/ethernet/intel/ice/ice_ddp.c
2256
hw->seg = seg;
drivers/net/ethernet/intel/ice/ice_ddp.c
2391
const struct ice_run_time_cfg_seg *seg;
drivers/net/ethernet/intel/ice/ice_ddp.c
2460
seg = (const struct ice_run_time_cfg_seg *)
drivers/net/ethernet/intel/ice/ice_ddp.c
2462
if (!seg) {
drivers/net/ethernet/intel/ice/ice_ddp.c
2467
if (le32_to_cpu(seg->buf_table.buf_count) < ICE_MIN_S_COUNT) {
drivers/net/ethernet/intel/ice/ice_ddp.c
2469
seg->buf_table.buf_count);
drivers/net/ethernet/intel/ice/ice_ddp.c
2473
section = ice_pkg_val_buf(seg->buf_table.buf_array);
drivers/net/ethernet/intel/ice/ice_ddp.c
381
if (!hw->seg)
drivers/net/ethernet/intel/ice/ice_ddp.c
384
ice_seg = hw->seg;
drivers/net/ethernet/intel/ice/ice_ddp.c
60
const struct ice_generic_seg_hdr *seg;
drivers/net/ethernet/intel/ice/ice_ddp.c
63
if (len < off + sizeof(*seg))
drivers/net/ethernet/intel/ice/ice_ddp.c
66
seg = (void *)pkg + off;
drivers/net/ethernet/intel/ice/ice_ddp.c
69
if (len < off + le32_to_cpu(seg->seg_size))
drivers/net/ethernet/intel/ice/ice_ddp.c
785
ice_seg = hw->seg;
drivers/net/ethernet/intel/ice/ice_ddp.c
830
if (!lkups->n_val_words || !hw->seg)
drivers/net/ethernet/intel/ice/ice_ddp.c
833
ice_seg = hw->seg;
drivers/net/ethernet/intel/ice/ice_ddp.c
90
hw->seg = NULL;
drivers/net/ethernet/intel/ice/ice_ddp.c
900
if (!hw->seg)
drivers/net/ethernet/intel/ice/ice_ddp.c
903
ice_seg = hw->seg;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1010
ice_set_fdir_ip4_usr_seg(struct ice_flow_seg_info *seg,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1029
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_IPV4);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1033
ice_flow_set_fld(seg, ICE_FLOW_FIELD_IDX_IPV4_SA,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1043
ice_flow_set_fld(seg, ICE_FLOW_FIELD_IDX_IPV4_DA,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1066
ice_set_fdir_ip6_seg(struct ice_flow_seg_info *seg,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1098
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_IPV6 | l4_proto);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1102
ice_flow_set_fld(seg, ICE_FLOW_FIELD_IDX_IPV6_SA,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1113
ice_flow_set_fld(seg, ICE_FLOW_FIELD_IDX_IPV6_DA,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1124
ice_flow_set_fld(seg, src_port, ICE_FLOW_FLD_OFF_INVAL,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1134
ice_flow_set_fld(seg, dst_port, ICE_FLOW_FLD_OFF_INVAL,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1156
ice_set_fdir_ip6_usr_seg(struct ice_flow_seg_info *seg,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1177
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_IPV6);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1181
ice_flow_set_fld(seg, ICE_FLOW_FIELD_IDX_IPV6_SA,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1192
ice_flow_set_fld(seg, ICE_FLOW_FIELD_IDX_IPV6_DA,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1239
struct ice_flow_seg_info *seg,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1242
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_ETH);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1252
ice_flow_set_fld(seg, ICE_FLOW_FIELD_IDX_ETH_TYPE,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1263
ice_flow_set_fld(seg, ICE_FLOW_FIELD_IDX_ETH_SA,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1271
ice_flow_set_fld(seg, ICE_FLOW_FIELD_IDX_ETH_DA,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1292
ice_set_fdir_vlan_seg(struct ice_flow_seg_info *seg,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1295
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_VLAN);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1301
ice_flow_set_fld(seg, ICE_FLOW_FIELD_IDX_S_VLAN,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1311
ice_flow_set_fld(seg, ICE_FLOW_FIELD_IDX_C_VLAN,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1332
struct ice_flow_seg_info *seg, *tun_seg;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1339
seg = devm_kzalloc(dev, sizeof(*seg), GFP_KERNEL);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1340
if (!seg)
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1346
devm_kfree(dev, seg);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1352
ret = ice_set_fdir_ip4_seg(seg, &fsp->m_u.tcp_ip4_spec,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1357
ret = ice_set_fdir_ip4_seg(seg, &fsp->m_u.tcp_ip4_spec,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1362
ret = ice_set_fdir_ip4_seg(seg, &fsp->m_u.tcp_ip4_spec,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1367
ret = ice_set_fdir_ip4_usr_seg(seg, &fsp->m_u.usr_ip4_spec,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1371
ret = ice_set_fdir_ip6_seg(seg, &fsp->m_u.tcp_ip6_spec,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1376
ret = ice_set_fdir_ip6_seg(seg, &fsp->m_u.tcp_ip6_spec,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1381
ret = ice_set_fdir_ip6_seg(seg, &fsp->m_u.tcp_ip6_spec,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1386
ret = ice_set_fdir_ip6_usr_seg(seg, &fsp->m_u.usr_ip6_spec,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1390
ret = ice_set_ether_flow_seg(dev, seg, &fsp->m_u.ether_spec);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1396
ret = ice_set_fdir_vlan_seg(seg, &fsp->m_ext);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1406
memcpy(&tun_seg[1], seg, sizeof(*seg));
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1410
ice_flow_add_fld_raw(seg, user->flex_offset,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1425
ret = ice_fdir_set_hw_fltr_rule(pf, seg, fltr_idx,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1437
memcpy(&tun_seg[1], seg, sizeof(*seg));
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
1453
devm_kfree(dev, seg);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
632
ice_fdir_set_hw_fltr_rule(struct ice_pf *pf, struct ice_flow_seg_info *seg,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
666
if (!memcmp(old_seg, seg, sizeof(*seg)))
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
691
err = ice_flow_add_prof(hw, ICE_BLK_FD, ICE_FLOW_RX, seg,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
697
seg, &entry1_h);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
702
seg, &entry2_h);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
706
hw_prof->fdir_seg[tun] = seg;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
726
ICE_FLOW_PRIO_NORMAL, seg,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
791
ice_set_init_fdir_seg(struct ice_flow_seg_info *seg,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
797
if (!seg)
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
820
ICE_FLOW_SET_HDRS(seg, l3_proto | l4_proto);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
823
ice_flow_set_fld(seg, src_addr, ICE_FLOW_FLD_OFF_INVAL,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
827
ice_flow_set_fld(seg, dst_addr, ICE_FLOW_FLD_OFF_INVAL,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
831
ice_flow_set_fld(seg, src_port, ICE_FLOW_FLD_OFF_INVAL,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
835
ice_flow_set_fld(seg, dst_port, ICE_FLOW_FLD_OFF_INVAL,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
851
struct ice_flow_seg_info *seg, *tun_seg;
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
861
seg = devm_kzalloc(dev, sizeof(*seg), GFP_KERNEL);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
862
if (!seg)
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
868
devm_kfree(dev, seg);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
873
ret = ice_set_init_fdir_seg(seg, ICE_FLOW_SEG_HDR_IPV4,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
876
ret = ice_set_init_fdir_seg(seg, ICE_FLOW_SEG_HDR_IPV4,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
879
ret = ice_set_init_fdir_seg(seg, ICE_FLOW_SEG_HDR_IPV6,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
882
ret = ice_set_init_fdir_seg(seg, ICE_FLOW_SEG_HDR_IPV6,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
890
ret = ice_fdir_set_hw_fltr_rule(pf, seg, flow, ICE_FD_HW_SEG_NON_TUN);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
896
memcpy(&tun_seg[1], seg, sizeof(*seg));
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
908
devm_kfree(dev, seg);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
925
ice_set_fdir_ip4_seg(struct ice_flow_seg_info *seg,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
954
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_IPV4 | l4_proto);
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
958
ice_flow_set_fld(seg, ICE_FLOW_FIELD_IDX_IPV4_SA,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
968
ice_flow_set_fld(seg, ICE_FLOW_FIELD_IDX_IPV4_DA,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
978
ice_flow_set_fld(seg, src_port, ICE_FLOW_FLD_OFF_INVAL,
drivers/net/ethernet/intel/ice/ice_ethtool_fdir.c
988
ice_flow_set_fld(seg, dst_port, ICE_FLOW_FLD_OFF_INVAL,
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1899
if (!hw->seg) {
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
1906
sect = ice_pkg_enum_section(hw->seg, &state, sid);
drivers/net/ethernet/intel/ice/ice_flow.c
1006
u8 seg, enum ice_flow_field fld, u64 match)
drivers/net/ethernet/intel/ice/ice_flow.c
1017
flds = params->prof->segs[seg].fields;
drivers/net/ethernet/intel/ice/ice_flow.c
1024
prot_id = seg == 0 ? ICE_PROT_MAC_OF_OR_S : ICE_PROT_MAC_IL;
drivers/net/ethernet/intel/ice/ice_flow.c
1027
prot_id = seg == 0 ? ICE_PROT_ETYPE_OL : ICE_PROT_ETYPE_IL;
drivers/net/ethernet/intel/ice/ice_flow.c
1030
prot_id = seg == 0 ? ICE_PROT_IPV4_OF_OR_S : ICE_PROT_IPV4_IL;
drivers/net/ethernet/intel/ice/ice_flow.c
1033
prot_id = seg == 0 ? ICE_PROT_IPV6_OF_OR_S : ICE_PROT_IPV6_IL;
drivers/net/ethernet/intel/ice/ice_flow.c
1037
prot_id = seg == 0 ? ICE_PROT_IPV4_OF_OR_S : ICE_PROT_IPV4_IL;
drivers/net/ethernet/intel/ice/ice_flow.c
1056
prot_id = seg == 0 ? ICE_PROT_IPV6_OF_OR_S : ICE_PROT_IPV6_IL;
drivers/net/ethernet/intel/ice/ice_flow.c
1075
prot_id = seg == 0 ? ICE_PROT_IPV4_OF_OR_S : ICE_PROT_IPV4_IL;
drivers/net/ethernet/intel/ice/ice_flow.c
1079
prot_id = seg == 0 ? ICE_PROT_IPV6_OF_OR_S : ICE_PROT_IPV6_IL;
drivers/net/ethernet/intel/ice/ice_flow.c
1131
prot_id = (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_IPV4) ?
drivers/net/ethernet/intel/ice/ice_flow.c
1207
u8 seg)
drivers/net/ethernet/intel/ice/ice_flow.c
1213
if (!params->prof->segs[seg].raws_cnt)
drivers/net/ethernet/intel/ice/ice_flow.c
1216
if (params->prof->segs[seg].raws_cnt >
drivers/net/ethernet/intel/ice/ice_flow.c
1217
ARRAY_SIZE(params->prof->segs[seg].raws))
drivers/net/ethernet/intel/ice/ice_flow.c
1221
hdrs_sz = ice_flow_calc_seg_sz(params, seg);
drivers/net/ethernet/intel/ice/ice_flow.c
1227
for (i = 0; i < params->prof->segs[seg].raws_cnt; i++) {
drivers/net/ethernet/intel/ice/ice_flow.c
1231
raw = &params->prof->segs[seg].raws[i];
drivers/net/ethernet/intel/ice/ice_flow.c
1913
ice_flow_set_fld_ext(struct ice_flow_seg_info *seg, enum ice_flow_field fld,
drivers/net/ethernet/intel/ice/ice_flow.c
1919
seg->match |= bit;
drivers/net/ethernet/intel/ice/ice_flow.c
1921
seg->range |= bit;
drivers/net/ethernet/intel/ice/ice_flow.c
1923
seg->fields[fld].type = field_type;
drivers/net/ethernet/intel/ice/ice_flow.c
1924
seg->fields[fld].src.val = val_loc;
drivers/net/ethernet/intel/ice/ice_flow.c
1925
seg->fields[fld].src.mask = mask_loc;
drivers/net/ethernet/intel/ice/ice_flow.c
1926
seg->fields[fld].src.last = last_loc;
drivers/net/ethernet/intel/ice/ice_flow.c
1928
ICE_FLOW_SET_HDRS(seg, ice_flds_info[fld].hdr);
drivers/net/ethernet/intel/ice/ice_flow.c
1952
ice_flow_set_fld(struct ice_flow_seg_info *seg, enum ice_flow_field fld,
drivers/net/ethernet/intel/ice/ice_flow.c
1958
ice_flow_set_fld_ext(seg, fld, t, val_loc, mask_loc, last_loc);
drivers/net/ethernet/intel/ice/ice_flow.c
1979
ice_flow_add_fld_raw(struct ice_flow_seg_info *seg, u16 off, u8 len,
drivers/net/ethernet/intel/ice/ice_flow.c
1982
if (seg->raws_cnt < ICE_FLOW_SEG_RAW_FLD_MAX) {
drivers/net/ethernet/intel/ice/ice_flow.c
1983
seg->raws[seg->raws_cnt].off = off;
drivers/net/ethernet/intel/ice/ice_flow.c
1984
seg->raws[seg->raws_cnt].info.type = ICE_FLOW_FLD_TYPE_SIZE;
drivers/net/ethernet/intel/ice/ice_flow.c
1985
seg->raws[seg->raws_cnt].info.src.val = val_loc;
drivers/net/ethernet/intel/ice/ice_flow.c
1986
seg->raws[seg->raws_cnt].info.src.mask = mask_loc;
drivers/net/ethernet/intel/ice/ice_flow.c
1988
seg->raws[seg->raws_cnt].info.src.last = len;
drivers/net/ethernet/intel/ice/ice_flow.c
1994
seg->raws_cnt++;
drivers/net/ethernet/intel/ice/ice_flow.c
2076
struct ice_flow_seg_info *seg;
drivers/net/ethernet/intel/ice/ice_flow.c
2081
seg = &segs[seg_cnt - 1];
drivers/net/ethernet/intel/ice/ice_flow.c
2085
ice_flow_set_fld(seg, (enum ice_flow_field)i,
drivers/net/ethernet/intel/ice/ice_flow.c
2089
ICE_FLOW_SET_HDRS(seg, cfg->addl_hdrs);
drivers/net/ethernet/intel/ice/ice_flow.c
2099
if (seg->hdrs & ~ICE_FLOW_RSS_SEG_HDR_VAL_MASKS &
drivers/net/ethernet/intel/ice/ice_flow.c
2103
val = (u64)(seg->hdrs & ICE_FLOW_RSS_SEG_HDR_L3_MASKS);
drivers/net/ethernet/intel/ice/ice_flow.c
2107
val = (u64)(seg->hdrs & ICE_FLOW_RSS_SEG_HDR_L4_MASKS);
drivers/net/ethernet/intel/ice/ice_flow.c
2364
struct ice_flow_seg_info *seg;
drivers/net/ethernet/intel/ice/ice_flow.c
2366
seg = &prof->segs[prof->segs_cnt - 1];
drivers/net/ethernet/intel/ice/ice_flow.c
2368
ipv4_src = &seg->fields[ICE_FLOW_FIELD_IDX_IPV4_SA].xtrct;
drivers/net/ethernet/intel/ice/ice_flow.c
2369
ipv4_dst = &seg->fields[ICE_FLOW_FIELD_IDX_IPV4_DA].xtrct;
drivers/net/ethernet/intel/ice/ice_flow.c
2371
ipv6_src = &seg->fields[ICE_FLOW_FIELD_IDX_IPV6_SA].xtrct;
drivers/net/ethernet/intel/ice/ice_flow.c
2372
ipv6_dst = &seg->fields[ICE_FLOW_FIELD_IDX_IPV6_DA].xtrct;
drivers/net/ethernet/intel/ice/ice_flow.c
2374
tcp_src = &seg->fields[ICE_FLOW_FIELD_IDX_TCP_SRC_PORT].xtrct;
drivers/net/ethernet/intel/ice/ice_flow.c
2375
tcp_dst = &seg->fields[ICE_FLOW_FIELD_IDX_TCP_DST_PORT].xtrct;
drivers/net/ethernet/intel/ice/ice_flow.c
2377
udp_src = &seg->fields[ICE_FLOW_FIELD_IDX_UDP_SRC_PORT].xtrct;
drivers/net/ethernet/intel/ice/ice_flow.c
2378
udp_dst = &seg->fields[ICE_FLOW_FIELD_IDX_UDP_DST_PORT].xtrct;
drivers/net/ethernet/intel/ice/ice_flow.c
2380
sctp_src = &seg->fields[ICE_FLOW_FIELD_IDX_SCTP_SRC_PORT].xtrct;
drivers/net/ethernet/intel/ice/ice_flow.c
2381
sctp_dst = &seg->fields[ICE_FLOW_FIELD_IDX_SCTP_DST_PORT].xtrct;
drivers/net/ethernet/intel/ice/ice_flow.c
768
static u16 ice_flow_calc_seg_sz(struct ice_flow_prof_params *params, u8 seg)
drivers/net/ethernet/intel/ice/ice_flow.c
773
sz = (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_VLAN) ?
drivers/net/ethernet/intel/ice/ice_flow.c
777
if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_IPV4)
drivers/net/ethernet/intel/ice/ice_flow.c
779
else if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_IPV6)
drivers/net/ethernet/intel/ice/ice_flow.c
781
else if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_ARP)
drivers/net/ethernet/intel/ice/ice_flow.c
783
else if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDRS_L4_MASK)
drivers/net/ethernet/intel/ice/ice_flow.c
788
if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_ICMP)
drivers/net/ethernet/intel/ice/ice_flow.c
790
else if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_TCP)
drivers/net/ethernet/intel/ice/ice_flow.c
792
else if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_UDP)
drivers/net/ethernet/intel/ice/ice_flow.c
794
else if (params->prof->segs[seg].hdrs & ICE_FLOW_SEG_HDR_SCTP)
drivers/net/ethernet/intel/ice/ice_flow.h
404
#define ICE_FLOW_SET_HDRS(seg, val) ((seg)->hdrs |= (u32)(val))
drivers/net/ethernet/intel/ice/ice_flow.h
514
ice_flow_set_fld(struct ice_flow_seg_info *seg, enum ice_flow_field fld,
drivers/net/ethernet/intel/ice/ice_flow.h
517
ice_flow_add_fld_raw(struct ice_flow_seg_info *seg, u16 off, u8 len,
drivers/net/ethernet/intel/ice/ice_parser.c
102
if (!seg)
drivers/net/ethernet/intel/ice/ice_parser.c
110
data = ice_pkg_enum_entry(seg, &state, sect_type, NULL,
drivers/net/ethernet/intel/ice/ice_parser.c
112
seg = NULL;
drivers/net/ethernet/intel/ice/ice_parser.c
1891
struct ice_seg *seg = hw->seg;
drivers/net/ethernet/intel/ice/ice_parser.c
1895
if (!seg)
drivers/net/ethernet/intel/ice/ice_parser.c
1902
data = ice_pkg_enum_section(seg, &state, sect_type);
drivers/net/ethernet/intel/ice/ice_parser.c
98
struct ice_seg *seg = hw->seg;
drivers/net/ethernet/intel/ice/ice_type.h
999
struct ice_seg *seg;
drivers/net/ethernet/intel/ice/virt/fdir.c
290
struct ice_flow_seg_info *seg)
drivers/net/ethernet/intel/ice/virt/fdir.c
315
ice_flow_set_fld(seg, fld[i],
drivers/net/ethernet/intel/ice/virt/fdir.c
334
struct ice_flow_seg_info *seg)
drivers/net/ethernet/intel/ice/virt/fdir.c
342
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_ETH_NON_IP);
drivers/net/ethernet/intel/ice/virt/fdir.c
345
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_L2TPV3 |
drivers/net/ethernet/intel/ice/virt/fdir.c
350
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_ESP |
drivers/net/ethernet/intel/ice/virt/fdir.c
355
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_AH |
drivers/net/ethernet/intel/ice/virt/fdir.c
360
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_NAT_T_ESP |
drivers/net/ethernet/intel/ice/virt/fdir.c
365
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_PFCP_NODE |
drivers/net/ethernet/intel/ice/virt/fdir.c
370
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_PFCP_SESSION |
drivers/net/ethernet/intel/ice/virt/fdir.c
375
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_IPV4 |
drivers/net/ethernet/intel/ice/virt/fdir.c
379
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_TCP |
drivers/net/ethernet/intel/ice/virt/fdir.c
384
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_UDP |
drivers/net/ethernet/intel/ice/virt/fdir.c
393
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_GTPU_IP |
drivers/net/ethernet/intel/ice/virt/fdir.c
397
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_GTPU_EH |
drivers/net/ethernet/intel/ice/virt/fdir.c
408
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_SCTP |
drivers/net/ethernet/intel/ice/virt/fdir.c
413
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_L2TPV3 |
drivers/net/ethernet/intel/ice/virt/fdir.c
418
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_ESP |
drivers/net/ethernet/intel/ice/virt/fdir.c
423
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_AH |
drivers/net/ethernet/intel/ice/virt/fdir.c
428
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_NAT_T_ESP |
drivers/net/ethernet/intel/ice/virt/fdir.c
433
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_PFCP_NODE |
drivers/net/ethernet/intel/ice/virt/fdir.c
438
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_PFCP_SESSION |
drivers/net/ethernet/intel/ice/virt/fdir.c
443
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_IPV6 |
drivers/net/ethernet/intel/ice/virt/fdir.c
447
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_TCP |
drivers/net/ethernet/intel/ice/virt/fdir.c
452
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_UDP |
drivers/net/ethernet/intel/ice/virt/fdir.c
457
ICE_FLOW_SET_HDRS(seg, ICE_FLOW_SEG_HDR_SCTP |
drivers/net/ethernet/intel/ice/virt/fdir.c
636
struct ice_flow_seg_info *seg, int tun)
drivers/net/ethernet/intel/ice/virt/fdir.c
664
if (!memcmp(old_seg, seg, sizeof(*seg))) {
drivers/net/ethernet/intel/ice/virt/fdir.c
681
ret = ice_flow_add_prof(hw, ICE_BLK_FD, ICE_FLOW_RX, seg,
drivers/net/ethernet/intel/ice/virt/fdir.c
691
seg, &entry1_h);
drivers/net/ethernet/intel/ice/virt/fdir.c
700
seg, &entry2_h);
drivers/net/ethernet/intel/ice/virt/fdir.c
708
vf_prof->fdir_seg[tun] = seg;
drivers/net/ethernet/intel/ice/virt/fdir.c
753
struct ice_flow_seg_info *seg;
drivers/net/ethernet/intel/ice/virt/fdir.c
771
seg = devm_kzalloc(dev, sizeof(*seg), GFP_KERNEL);
drivers/net/ethernet/intel/ice/virt/fdir.c
772
if (!seg)
drivers/net/ethernet/intel/ice/virt/fdir.c
775
ret = ice_vc_fdir_set_flow_fld(vf, fltr, conf, seg);
drivers/net/ethernet/intel/ice/virt/fdir.c
781
ret = ice_vc_fdir_set_flow_hdr(vf, conf, seg);
drivers/net/ethernet/intel/ice/virt/fdir.c
787
ret = ice_vc_fdir_write_flow_prof(vf, flow, seg, tun);
drivers/net/ethernet/intel/ice/virt/fdir.c
789
devm_kfree(dev, seg);
drivers/net/ethernet/intel/ice/virt/fdir.c
799
devm_kfree(dev, seg);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
874
int seg, len;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
879
for (seg = 0; seg < num_segs; seg++) {
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
880
if ((seg % MAX_SEGS_PER_SG) == 0) {
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
890
if ((num_segs - seg) >= (MAX_SEGS_PER_SG - 1))
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
895
cpt_sg += (seg / MAX_SEGS_PER_SG) * 4;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
898
dma_addr = otx2_dma_map_skb_frag(pfvf, skb, seg, &len);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
902
sg_lens[seg % MAX_SEGS_PER_SG] = len;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
908
sq->sg[sq->head].dma_addr[seg] = dma_addr;
drivers/net/ethernet/marvell/octeontx2/nic/cn10k_ipsec.c
909
sq->sg[sq->head].size[seg] = len;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
2116
struct sk_buff *skb, int seg, int *len)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
2130
if (!seg) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
2135
frag = &skb_shinfo(skb)->frags[seg - 1];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
2147
int seg;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
2153
for (seg = 0; seg < sg->num_segs; seg++) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
2154
otx2_dma_unmap_page(pfvf, sg->dma_addr[seg],
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
2155
sg->size[seg], dir);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
1221
struct sk_buff *skb, int seg, int *len);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
252
int seg;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
259
for (seg = 0; seg < sg->segs; seg++, seg_addr++)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
343
int seg;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
368
for (seg = 0; seg < sg->segs; seg++, seg_addr++) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
370
seg_size[seg], parse, cq->cq_idx))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
668
int seg, len;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
672
for (seg = 0; seg < num_segs; seg++) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
673
if ((seg % MAX_SEGS_PER_SG) == 0) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
683
if ((num_segs - seg) >= (MAX_SEGS_PER_SG - 1))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
688
dma_addr = otx2_dma_map_skb_frag(pfvf, skb, seg, &len);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
692
sg_lens[frag_num(seg % MAX_SEGS_PER_SG)] = len;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
697
sq->sg[sq->head].dma_addr[seg] = dma_addr;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
698
sq->sg[sq->head].size[seg] = len;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
853
int seg, len;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
860
for (seg = 0; seg < num_segs; seg++) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
862
if (!seg && !len)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
864
dma_addr = otx2_dma_map_skb_frag(pfvf, skb, seg, &len);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
880
struct sk_buff *skb, int seg,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
887
if (seg < 0)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
890
frag = &skb_shinfo(skb)->frags[seg];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
893
seg++;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
894
return sg->dma_addr[seg] + offset;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
903
int seg;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
906
for (seg = 0; seg < list->num_segs; seg++) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
907
if ((seg % MAX_SEGS_PER_SG) == 0) {
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
917
if ((list->num_segs - seg) >= (MAX_SEGS_PER_SG - 1))
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
922
sg_lens[frag_num(seg % MAX_SEGS_PER_SG)] = list->size[seg];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
923
*iova++ = list->dma_addr[seg];
drivers/net/ethernet/mellanox/mlx4/mr.c
153
u32 seg;
drivers/net/ethernet/mellanox/mlx4/mr.c
159
seg = mlx4_buddy_alloc(&mr_table->mtt_buddy, seg_order);
drivers/net/ethernet/mellanox/mlx4/mr.c
160
if (seg == -1)
drivers/net/ethernet/mellanox/mlx4/mr.c
163
offset = seg * (1 << log_mtts_per_seg);
drivers/net/ethernet/mellanox/mlx4/mr.c
167
mlx4_buddy_free(&mr_table->mtt_buddy, seg, seg_order);
drivers/net/ethernet/mellanox/mlx4/mr.c
50
u32 seg;
drivers/net/ethernet/mellanox/mlx4/mr.c
57
seg = find_first_bit(buddy->bits[o], m);
drivers/net/ethernet/mellanox/mlx4/mr.c
58
if (seg < m)
drivers/net/ethernet/mellanox/mlx4/mr.c
66
clear_bit(seg, buddy->bits[o]);
drivers/net/ethernet/mellanox/mlx4/mr.c
71
seg <<= 1;
drivers/net/ethernet/mellanox/mlx4/mr.c
72
set_bit(seg ^ 1, buddy->bits[o]);
drivers/net/ethernet/mellanox/mlx4/mr.c
78
seg <<= order;
drivers/net/ethernet/mellanox/mlx4/mr.c
80
return seg;
drivers/net/ethernet/mellanox/mlx4/mr.c
83
static void mlx4_buddy_free(struct mlx4_buddy *buddy, u32 seg, int order)
drivers/net/ethernet/mellanox/mlx4/mr.c
85
seg >>= order;
drivers/net/ethernet/mellanox/mlx4/mr.c
89
while (test_bit(seg ^ 1, buddy->bits[order])) {
drivers/net/ethernet/mellanox/mlx4/mr.c
90
clear_bit(seg ^ 1, buddy->bits[order]);
drivers/net/ethernet/mellanox/mlx4/mr.c
92
seg >>= 1;
drivers/net/ethernet/mellanox/mlx4/mr.c
96
set_bit(seg, buddy->bits[order]);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/buddy.c
108
*segment = seg;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/buddy.c
115
u32 seg, order_iter, err;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/buddy.c
117
err = hws_buddy_find_free_seg(buddy, order, &seg, &order_iter);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/buddy.c
121
bitmap_clear(buddy->bitmap[order_iter], seg, 1);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/buddy.c
126
seg <<= 1;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/buddy.c
127
bitmap_set(buddy->bitmap[order_iter], seg ^ 1, 1);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/buddy.c
131
seg <<= order;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/buddy.c
133
return seg;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/buddy.c
136
void mlx5hws_buddy_free_mem(struct mlx5hws_buddy_mem *buddy, u32 seg, u32 order)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/buddy.c
138
seg >>= order;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/buddy.c
140
while (test_bit(seg ^ 1, buddy->bitmap[order])) {
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/buddy.c
141
bitmap_clear(buddy->bitmap[order], seg ^ 1, 1);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/buddy.c
143
seg >>= 1;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/buddy.c
147
bitmap_set(buddy->bitmap[order], seg, 1);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/buddy.c
87
unsigned int seg, order_iter, m;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/buddy.c
95
seg = find_first_bit(buddy->bitmap[order_iter], m);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/buddy.c
97
if (WARN(seg >= m,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/buddy.h
19
void mlx5hws_buddy_free_mem(struct mlx5hws_buddy_mem *buddy, u32 seg, u32 order);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_buddy.c
119
unsigned int seg, order_iter;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_buddy.c
122
err = dr_buddy_find_free_seg(buddy, order, &seg, &order_iter);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_buddy.c
126
bitmap_clear(buddy->bitmap[order_iter], seg, 1);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_buddy.c
135
seg <<= 1;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_buddy.c
136
bitmap_set(buddy->bitmap[order_iter], seg ^ 1, 1);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_buddy.c
140
seg <<= order;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_buddy.c
141
*segment = seg;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_buddy.c
147
unsigned int seg, unsigned int order)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_buddy.c
149
seg >>= order;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_buddy.c
154
while (test_bit(seg ^ 1, buddy->bitmap[order])) {
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_buddy.c
155
bitmap_clear(buddy->bitmap[order], seg ^ 1, 1);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_buddy.c
157
seg >>= 1;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_buddy.c
160
bitmap_set(buddy->bitmap[order], seg, 1);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_buddy.c
73
unsigned int seg, order_iter, m;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_buddy.c
81
seg = find_first_bit(buddy->bitmap[order_iter], m);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_buddy.c
83
if (WARN(seg >= m,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_buddy.c
94
*segment = seg;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
13
unsigned int seg;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
324
unsigned int seg)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
328
chunk->seg = seg;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
333
offset = mlx5dr_icm_pool_dm_type_to_entry_size(pool->icm_type) * seg;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
354
hot_chunk->seg, ilog2(num_entries));
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
388
unsigned int *seg)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
398
chunk_size, seg);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
439
unsigned int seg;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
447
ret = dr_icm_handle_buddies_get_mem(pool, chunk_size, &buddy, &seg);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
455
dr_icm_chunk_init(chunk, pool, chunk_size, buddy, seg);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
460
mlx5dr_buddy_free_mem(buddy, seg, chunk_size);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
482
hot_chunk->seg = chunk->seg;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
84
return (u64)offset * chunk->seg;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
96
return (u64)chunk->buddy_mem->icm_mr->icm_start_addr + size * chunk->seg;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
1178
unsigned int seg;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/mlx5dr.h
192
unsigned int seg, unsigned int order);
drivers/net/ethernet/qlogic/qed/qed_cxt.c
2126
u32 proto, seg, total_lines, i, shadow_line;
drivers/net/ethernet/qlogic/qed/qed_cxt.c
2135
seg = QED_CXT_FCOE_TID_SEG;
drivers/net/ethernet/qlogic/qed/qed_cxt.c
2140
seg = QED_CXT_TCP_ULP_TID_SEG;
drivers/net/ethernet/qlogic/qed/qed_cxt.c
2150
p_seg_info = &p_mngr->conn_cfg[proto].tid_seg[seg];
drivers/net/ethernet/qlogic/qed/qed_cxt.c
2154
p_fl_seg = &p_cli->pf_blks[CDUT_FL_SEG_BLK(seg, PF)];
drivers/net/ethernet/qlogic/qed/qed_cxt.c
2456
u32 proto, seg;
drivers/net/ethernet/qlogic/qed/qed_cxt.c
2462
seg = QED_CXT_FCOE_TID_SEG;
drivers/net/ethernet/qlogic/qed/qed_cxt.c
2467
seg = QED_CXT_TCP_ULP_TID_SEG;
drivers/net/ethernet/qlogic/qed/qed_cxt.c
2477
p_seg_info = &p_mngr->conn_cfg[proto].tid_seg[seg];
drivers/net/ethernet/qlogic/qed/qed_cxt.c
2480
p_seg = &p_cli->pf_blks[CDUT_SEG_BLK(seg)];
drivers/net/ethernet/qlogic/qed/qed_cxt.c
2484
p_seg = &p_cli->pf_blks[CDUT_FL_SEG_BLK(seg, PF)];
drivers/net/ethernet/qlogic/qed/qed_cxt.c
257
u32 seg)
drivers/net/ethernet/qlogic/qed/qed_cxt.c
266
if (p_cfg->conn_cfg[i].tid_seg[seg].count)
drivers/net/ethernet/qlogic/qed/qed_cxt.c
267
return &p_cfg->conn_cfg[i].tid_seg[seg];
drivers/net/ethernet/qlogic/qed/qed_cxt.c
357
u8 seg,
drivers/net/ethernet/qlogic/qed/qed_cxt.c
361
struct qed_tid_seg *p_seg = &p_mngr->conn_cfg[proto].tid_seg[seg];
drivers/net/ethernet/qlogic/qla3xxx.c
2311
int seg_cnt, seg = 0;
drivers/net/ethernet/qlogic/qla3xxx.c
2332
dma_unmap_addr_set(&tx_cb->map[seg], mapaddr, map);
drivers/net/ethernet/qlogic/qla3xxx.c
2333
dma_unmap_len_set(&tx_cb->map[seg], maplen, len);
drivers/net/ethernet/qlogic/qla3xxx.c
2334
seg++;
drivers/net/ethernet/qlogic/qla3xxx.c
2344
completed_segs++, seg++) {
drivers/net/ethernet/qlogic/qla3xxx.c
2352
if ((seg == 2 && seg_cnt > 3) ||
drivers/net/ethernet/qlogic/qla3xxx.c
2353
(seg == 7 && seg_cnt > 8) ||
drivers/net/ethernet/qlogic/qla3xxx.c
2354
(seg == 12 && seg_cnt > 13) ||
drivers/net/ethernet/qlogic/qla3xxx.c
2355
(seg == 17 && seg_cnt > 18)) {
drivers/net/ethernet/qlogic/qla3xxx.c
2372
dma_unmap_addr_set(&tx_cb->map[seg], mapaddr, map);
drivers/net/ethernet/qlogic/qla3xxx.c
2373
dma_unmap_len_set(&tx_cb->map[seg], maplen,
drivers/net/ethernet/qlogic/qla3xxx.c
2377
seg++;
drivers/net/ethernet/qlogic/qla3xxx.c
2394
dma_unmap_addr_set(&tx_cb->map[seg], mapaddr, map);
drivers/net/ethernet/qlogic/qla3xxx.c
2395
dma_unmap_len_set(&tx_cb->map[seg], maplen, skb_frag_size(frag));
drivers/net/ethernet/qlogic/qla3xxx.c
2407
seg = 1;
drivers/net/ethernet/qlogic/qla3xxx.c
2410
for (i = 0; i < completed_segs; i++, seg++) {
drivers/net/ethernet/qlogic/qla3xxx.c
2418
if ((seg == 2 && seg_cnt > 3) ||
drivers/net/ethernet/qlogic/qla3xxx.c
2419
(seg == 7 && seg_cnt > 8) ||
drivers/net/ethernet/qlogic/qla3xxx.c
2420
(seg == 12 && seg_cnt > 13) ||
drivers/net/ethernet/qlogic/qla3xxx.c
2421
(seg == 17 && seg_cnt > 18)) {
drivers/net/ethernet/qlogic/qla3xxx.c
2423
dma_unmap_addr(&tx_cb->map[seg], mapaddr),
drivers/net/ethernet/qlogic/qla3xxx.c
2424
dma_unmap_len(&tx_cb->map[seg], maplen),
drivers/net/ethernet/qlogic/qla3xxx.c
2427
seg++;
drivers/net/ethernet/qlogic/qla3xxx.c
2431
dma_unmap_addr(&tx_cb->map[seg], mapaddr),
drivers/net/ethernet/qlogic/qla3xxx.c
2432
dma_unmap_len(&tx_cb->map[seg], maplen),
drivers/net/usb/r8152.c
2216
struct sk_buff *segs, *seg, *next;
drivers/net/usb/r8152.c
2226
skb_list_walk_safe(segs, seg, next) {
drivers/net/usb/r8152.c
2227
skb_mark_not_on_list(seg);
drivers/net/usb/r8152.c
2228
__skb_queue_tail(&seg_list, seg);
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
3359
struct mt7925_clc_segment *seg;
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
3390
pos = clc->data + sizeof(*seg) * clc->t0.nr_seg;
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
3400
seg = (struct mt7925_clc_segment *)clc->data
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
3406
req.size = cpu_to_le16(seg->len);
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
3413
skb_put_data(skb, clc->data + seg->offset, seg->len);
drivers/net/wireless/rsi/rsi_91x_usb.c
48
u8 *seg = dev->tx_buffer;
drivers/net/wireless/rsi/rsi_91x_usb.c
52
memset(seg, 0, len + RSI_USB_TX_HEAD_ROOM);
drivers/net/wireless/rsi/rsi_91x_usb.c
53
memcpy(seg + RSI_USB_TX_HEAD_ROOM, buf, len);
drivers/net/wireless/rsi/rsi_91x_usb.c
58
(void *)seg,
drivers/nvme/target/pci-epf.c
1040
struct nvme_sgl_desc seg = cmd->common.dptr.sgl;
drivers/nvme/target/pci-epf.c
1049
if (seg.type != (NVME_SGL_FMT_SEG_DESC << 4) &&
drivers/nvme/target/pci-epf.c
1050
seg.type != (NVME_SGL_FMT_LAST_SEG_DESC << 4)) {
drivers/nvme/target/pci-epf.c
1055
while (seg.length) {
drivers/nvme/target/pci-epf.c
1056
sgls = nvmet_pci_epf_get_sgl_segment(ctrl, &seg, &nr_sgls);
drivers/nvme/target/pci-epf.c
1117
struct nvmet_pci_epf_segment *seg;
drivers/nvme/target/pci-epf.c
1151
seg = &iod->data_segs[i];
drivers/nvme/target/pci-epf.c
1152
seg->buf = kmalloc(seg->length, GFP_KERNEL);
drivers/nvme/target/pci-epf.c
1153
if (!seg->buf)
drivers/nvme/target/pci-epf.c
1155
sg_set_buf(sg, seg->buf, seg->length);
drivers/nvme/target/pci-epf.c
361
struct nvmet_pci_epf_segment *seg, enum dma_data_direction dir)
drivers/nvme/target/pci-epf.c
379
sconf.src_addr = seg->pci_addr;
drivers/nvme/target/pci-epf.c
385
sconf.dst_addr = seg->pci_addr;
drivers/nvme/target/pci-epf.c
394
dma_addr = dma_map_single(dma_dev, seg->buf, seg->length, dir);
drivers/nvme/target/pci-epf.c
405
desc = dmaengine_prep_slave_single(chan, dma_addr, seg->length,
drivers/nvme/target/pci-epf.c
428
dma_unmap_single(dma_dev, dma_addr, seg->length, dir);
drivers/nvme/target/pci-epf.c
437
struct nvmet_pci_epf_segment *seg, enum dma_data_direction dir)
drivers/nvme/target/pci-epf.c
439
u64 pci_addr = seg->pci_addr;
drivers/nvme/target/pci-epf.c
440
u32 length = seg->length;
drivers/nvme/target/pci-epf.c
441
void *buf = seg->buf;
drivers/nvme/target/pci-epf.c
482
struct nvmet_pci_epf_segment *seg, enum dma_data_direction dir)
drivers/nvme/target/pci-epf.c
485
return nvmet_pci_epf_dma_transfer(nvme_epf, seg, dir);
drivers/nvme/target/pci-epf.c
487
return nvmet_pci_epf_mmio_transfer(nvme_epf, seg, dir);
drivers/nvme/target/pci-epf.c
494
struct nvmet_pci_epf_segment seg = {
drivers/nvme/target/pci-epf.c
500
return nvmet_pci_epf_transfer_seg(ctrl->nvme_epf, &seg, dir);
drivers/nvme/target/pci-epf.c
738
struct nvmet_pci_epf_segment *seg = &iod->data_segs[0];
drivers/nvme/target/pci-epf.c
742
for (i = 0; i < iod->nr_data_segs; i++, seg++) {
drivers/nvme/target/pci-epf.c
743
ret = nvmet_pci_epf_transfer_seg(nvme_epf, seg, iod->dma_dir);
drivers/nvme/target/pci-epf.c
794
struct nvmet_pci_epf_segment *seg;
drivers/nvme/target/pci-epf.c
824
seg = &iod->data_segs[0];
drivers/nvme/target/pci-epf.c
825
seg->pci_addr = prp;
drivers/nvme/target/pci-epf.c
826
seg->length = nvmet_pci_epf_prp_size(ctrl, prp);
drivers/nvme/target/pci-epf.c
828
size = seg->length;
drivers/nvme/target/pci-epf.c
875
seg++;
drivers/nvme/target/pci-epf.c
876
seg->pci_addr = prp;
drivers/nvme/target/pci-epf.c
877
seg->length = 0;
drivers/nvme/target/pci-epf.c
882
seg->length += prp_size;
drivers/pci/controller/pci-thunder-pem.c
343
static void thunder_pem_reserve_range(struct device *dev, int seg,
drivers/pci/controller/pci-thunder-pem.c
350
regionid = kasprintf(GFP_KERNEL, "PEM RC:%d", seg);
drivers/pci/pci-acpi.c
1616
u16 seg = root->segment;
drivers/pci/pci-acpi.c
1625
dev_err(dev, "%04x:%pR ECAM region not found\n", seg, bus_res);
drivers/pci/pci-acpi.c
1639
dev_err(dev, "%04x:%pR error %ld mapping ECAM\n", seg, bus_res,
drivers/pci/pci.c
283
unsigned int seg, bus, slot, func;
drivers/pci/pci.c
323
ret = sscanf(wpath, "%x:%x:%x.%x%c", &seg, &bus, &slot,
drivers/pci/pci.c
326
seg = 0;
drivers/pci/pci.c
334
ret = (seg == pci_domain_nr(dev->bus) &&
drivers/pcmcia/omap_cf.c
201
unsigned seg;
drivers/pcmcia/omap_cf.c
208
seg = (int) pdev->dev.platform_data;
drivers/pcmcia/omap_cf.c
209
if (seg == 0 || seg > 3)
drivers/pcmcia/omap_cf.c
258
omap_writew(~(1 << seg), CF_CFG);
drivers/pcmcia/omap_cf.c
260
pr_info("%s: cs%d on irq %d\n", driver_name, seg, irq);
drivers/peci/cpu.c
118
int peci_ep_pci_local_read(struct peci_device *device, u8 seg,
drivers/peci/cpu.c
124
req = peci_xfer_ep_pci_cfg_local_readl(device, seg, bus, dev, func, reg);
drivers/peci/cpu.c
155
int peci_mmio_read(struct peci_device *device, u8 bar, u8 seg,
drivers/peci/cpu.c
161
req = peci_xfer_ep_mmio64_readl(device, bar, seg, bus, dev, func, address);
drivers/peci/internal.h
47
struct peci_request *peci_xfer_ep_pci_cfg_local_readb(struct peci_device *device, u8 seg,
drivers/peci/internal.h
49
struct peci_request *peci_xfer_ep_pci_cfg_local_readw(struct peci_device *device, u8 seg,
drivers/peci/internal.h
51
struct peci_request *peci_xfer_ep_pci_cfg_local_readl(struct peci_device *device, u8 seg,
drivers/peci/internal.h
54
struct peci_request *peci_xfer_ep_pci_cfg_readb(struct peci_device *device, u8 seg,
drivers/peci/internal.h
56
struct peci_request *peci_xfer_ep_pci_cfg_readw(struct peci_device *device, u8 seg,
drivers/peci/internal.h
58
struct peci_request *peci_xfer_ep_pci_cfg_readl(struct peci_device *device, u8 seg,
drivers/peci/internal.h
61
struct peci_request *peci_xfer_ep_mmio32_readl(struct peci_device *device, u8 bar, u8 seg,
drivers/peci/internal.h
64
struct peci_request *peci_xfer_ep_mmio64_readl(struct peci_device *device, u8 bar, u8 seg,
drivers/peci/request.c
327
__ep_pci_cfg_read(struct peci_device *device, u8 msg_type, u8 seg,
drivers/peci/request.c
348
req->tx.buf[7] = seg; /* PCI Segment */
drivers/peci/request.c
361
__ep_mmio_read(struct peci_device *device, u8 bar, u8 addr_type, u8 seg,
drivers/peci/request.c
378
req->tx.buf[7] = seg; /* PCI Segment */
drivers/peci/request.c
458
peci_xfer_ep_pci_cfg_##x(struct peci_device *device, u8 seg, u8 bus, u8 dev, u8 func, u16 reg) \
drivers/peci/request.c
460
return __ep_pci_cfg_read(device, msg_type, seg, bus, dev, func, reg, sizeof(type)); \
drivers/peci/request.c
472
struct peci_request *peci_xfer_ep_mmio##y##_##x(struct peci_device *device, u8 bar, u8 seg, \
drivers/peci/request.c
475
return __ep_mmio_read(device, bar, addr_type, seg, bus, dev, func, \
drivers/platform/mellanox/mlxbf-tmfifo.c
460
u32 len, idx, seg;
drivers/platform/mellanox/mlxbf-tmfifo.c
467
seg = CIRC_SPACE_TO_END(cons->tx_buf.head, cons->tx_buf.tail,
drivers/platform/mellanox/mlxbf-tmfifo.c
469
if (len <= seg) {
drivers/platform/mellanox/mlxbf-tmfifo.c
472
memcpy(cons->tx_buf.buf + cons->tx_buf.head, addr, seg);
drivers/platform/mellanox/mlxbf-tmfifo.c
473
addr += seg;
drivers/platform/mellanox/mlxbf-tmfifo.c
474
memcpy(cons->tx_buf.buf, addr, len - seg);
drivers/platform/mellanox/mlxbf-tmfifo.c
543
int size, seg;
drivers/platform/mellanox/mlxbf-tmfifo.c
576
seg = CIRC_CNT_TO_END(cons->tx_buf.head, cons->tx_buf.tail,
drivers/platform/mellanox/mlxbf-tmfifo.c
578
if (seg >= sizeof(u64)) {
drivers/platform/mellanox/mlxbf-tmfifo.c
581
memcpy(&data, addr, seg);
drivers/platform/mellanox/mlxbf-tmfifo.c
582
memcpy((u8 *)&data + seg, cons->tx_buf.buf,
drivers/platform/mellanox/mlxbf-tmfifo.c
583
sizeof(u64) - seg);
drivers/remoteproc/remoteproc_coredump.c
182
struct rproc_dump_segment *seg;
drivers/remoteproc/remoteproc_coredump.c
200
seg = rproc_coredump_find_segment(offset - header_sz,
drivers/remoteproc/remoteproc_coredump.c
204
if (!seg) {
drivers/remoteproc/remoteproc_coredump.c
212
rproc_copy_segment(rproc, buffer, seg, seg->size - seg_data,
drivers/scsi/megaraid.c
527
u32 seg;
drivers/scsi/megaraid.c
854
(u32 *)&mbox->m_out.xferaddr, &seg);
drivers/scsi/qla2xxx/qla_init.c
8774
uint32_t risc_addr, risc_size, fwclen, wlen, *seg;
drivers/scsi/qla2xxx/qla_init.c
8814
seg = blob->segs;
drivers/scsi/qla2xxx/qla_init.c
8815
while (*seg && rval == QLA_SUCCESS) {
drivers/scsi/qla2xxx/qla_init.c
8816
risc_addr = *seg;
drivers/scsi/qla2xxx/qla_init.c
8817
*srisc_addr = *srisc_addr == 0 ? *seg : *srisc_addr;
drivers/scsi/qla2xxx/qla_init.c
8857
seg++;
drivers/scsi/xen-scsifront.c
221
ring_req->seg[i] = shadow->seg[i];
drivers/scsi/xen-scsifront.c
482
struct scsiif_request_segment *seg;
drivers/scsi/xen-scsifront.c
502
seg = shadow->sg ? : shadow->seg;
drivers/scsi/xen-scsifront.c
514
page = virt_to_page(seg);
drivers/scsi/xen-scsifront.c
515
off = offset_in_page(seg);
drivers/scsi/xen-scsifront.c
527
shadow->seg[ref_cnt].gref = ref;
drivers/scsi/xen-scsifront.c
528
shadow->seg[ref_cnt].offset = (uint16_t)off;
drivers/scsi/xen-scsifront.c
529
shadow->seg[ref_cnt].length = (uint16_t)bytes;
drivers/scsi/xen-scsifront.c
563
seg->gref = ref;
drivers/scsi/xen-scsifront.c
564
seg->offset = (uint16_t)off;
drivers/scsi/xen-scsifront.c
565
seg->length = (uint16_t)bytes;
drivers/scsi/xen-scsifront.c
568
seg++;
drivers/scsi/xen-scsifront.c
87
struct scsiif_request_segment seg[VSCSIIF_SG_TABLESIZE];
drivers/ssb/pcmcia.c
166
int ssb_pcmcia_switch_segment(struct ssb_bus *bus, u8 seg)
drivers/ssb/pcmcia.c
172
WARN_ON((seg != 0) && (seg != 1));
drivers/ssb/pcmcia.c
174
err = ssb_pcmcia_cfg_write(bus, SSB_PCMCIA_MEMSEG, seg);
drivers/ssb/pcmcia.c
180
if (val == seg)
drivers/ssb/pcmcia.c
188
bus->mapped_pcmcia_seg = seg;
drivers/ssb/ssb_private.h
59
u8 seg);
drivers/ssb/ssb_private.h
75
u8 seg)
drivers/staging/media/meson/vdec/codec_vp9.c
495
static int segfeature_active(struct segmentation *seg, int segment_id,
drivers/staging/media/meson/vdec/codec_vp9.c
498
return seg->enabled &&
drivers/staging/media/meson/vdec/codec_vp9.c
499
(seg->feature_mask[segment_id] & (1 << feature_id));
drivers/staging/media/meson/vdec/codec_vp9.c
502
static int get_segdata(struct segmentation *seg, int segment_id,
drivers/staging/media/meson/vdec/codec_vp9.c
505
return seg->feature_data[segment_id][feature_id];
drivers/staging/media/meson/vdec/codec_vp9.c
574
vp9_loop_filter_frame_init(struct amvdec_core *core, struct segmentation *seg,
drivers/staging/media/meson/vdec/codec_vp9.c
611
if (segfeature_active(seg, seg_id, SEG_LVL_ALT_LF)) {
drivers/staging/media/meson/vdec/codec_vp9.c
612
const int data = get_segdata(seg, seg_id,
drivers/staging/media/meson/vdec/codec_vp9.c
615
seg->abs_delta == SEGMENT_ABSDATA ?
drivers/thermal/k3_j72xx_bandgap.c
132
int inc, i, seg;
drivers/thermal/k3_j72xx_bandgap.c
140
for (seg = 0; seg < 3; seg++)
drivers/thermal/k3_j72xx_bandgap.c
141
create_table_segments(err_vals, seg, ref_table);
drivers/thermal/k3_j72xx_bandgap.c
84
static void create_table_segments(struct err_values *err_vals, int seg,
drivers/thermal/k3_j72xx_bandgap.c
89
if (seg == 0)
drivers/thermal/k3_j72xx_bandgap.c
92
idx1 = err_vals->refs[seg];
drivers/thermal/k3_j72xx_bandgap.c
94
idx2 = err_vals->refs[seg + 1];
drivers/thermal/k3_j72xx_bandgap.c
95
err1 = err_vals->errs[seg];
drivers/thermal/k3_j72xx_bandgap.c
96
err2 = err_vals->errs[seg + 1];
drivers/thermal/k3_j72xx_bandgap.c
97
ref1 = err_vals->refs[seg];
drivers/thermal/k3_j72xx_bandgap.c
98
ref2 = err_vals->refs[seg + 1];
drivers/usb/cdns3/cdnsp-gadget.c
1401
struct cdnsp_segment *seg;
drivers/usb/cdns3/cdnsp-gadget.c
1407
seg = pdev->cmd_ring->first_seg;
drivers/usb/cdns3/cdnsp-gadget.c
1409
memset(seg->trbs, 0,
drivers/usb/cdns3/cdnsp-gadget.c
1411
seg = seg->next;
drivers/usb/cdns3/cdnsp-gadget.h
1532
dma_addr_t cdnsp_trb_virt_to_dma(struct cdnsp_segment *seg,
drivers/usb/cdns3/cdnsp-gadget.h
1534
bool cdnsp_last_trb_on_seg(struct cdnsp_segment *seg, union cdnsp_trb *trb);
drivers/usb/cdns3/cdnsp-gadget.h
1536
struct cdnsp_segment *seg,
drivers/usb/cdns3/cdnsp-mem.c
1029
struct cdnsp_segment *seg;
drivers/usb/cdns3/cdnsp-mem.c
1041
seg = evt_ring->first_seg;
drivers/usb/cdns3/cdnsp-mem.c
1044
entry->seg_addr = cpu_to_le64(seg->dma);
drivers/usb/cdns3/cdnsp-mem.c
1047
seg = seg->next;
drivers/usb/cdns3/cdnsp-mem.c
195
struct cdnsp_segment *seg,
drivers/usb/cdns3/cdnsp-mem.c
201
key = (unsigned long)(seg->dma >> TRB_SEGMENT_SHIFT);
drivers/usb/cdns3/cdnsp-mem.c
218
struct cdnsp_segment *seg)
drivers/usb/cdns3/cdnsp-mem.c
222
key = (unsigned long)(seg->dma >> TRB_SEGMENT_SHIFT);
drivers/usb/cdns3/cdnsp-mem.c
234
struct cdnsp_segment *seg;
drivers/usb/cdns3/cdnsp-mem.c
237
seg = first_seg;
drivers/usb/cdns3/cdnsp-mem.c
239
ret = cdnsp_insert_segment_mapping(trb_address_map, ring, seg,
drivers/usb/cdns3/cdnsp-mem.c
243
if (seg == last_seg)
drivers/usb/cdns3/cdnsp-mem.c
245
seg = seg->next;
drivers/usb/cdns3/cdnsp-mem.c
246
} while (seg != first_seg);
drivers/usb/cdns3/cdnsp-mem.c
251
failed_seg = seg;
drivers/usb/cdns3/cdnsp-mem.c
252
seg = first_seg;
drivers/usb/cdns3/cdnsp-mem.c
254
cdnsp_remove_segment_mapping(trb_address_map, seg);
drivers/usb/cdns3/cdnsp-mem.c
255
if (seg == failed_seg)
drivers/usb/cdns3/cdnsp-mem.c
257
seg = seg->next;
drivers/usb/cdns3/cdnsp-mem.c
258
} while (seg != first_seg);
drivers/usb/cdns3/cdnsp-mem.c
265
struct cdnsp_segment *seg;
drivers/usb/cdns3/cdnsp-mem.c
267
seg = ring->first_seg;
drivers/usb/cdns3/cdnsp-mem.c
269
cdnsp_remove_segment_mapping(ring->trb_address_map, seg);
drivers/usb/cdns3/cdnsp-mem.c
270
seg = seg->next;
drivers/usb/cdns3/cdnsp-mem.c
271
} while (seg != ring->first_seg);
drivers/usb/cdns3/cdnsp-mem.c
34
struct cdnsp_segment *seg;
drivers/usb/cdns3/cdnsp-mem.c
38
seg = kzalloc_obj(*seg, flags);
drivers/usb/cdns3/cdnsp-mem.c
39
if (!seg)
drivers/usb/cdns3/cdnsp-mem.c
42
seg->trbs = dma_pool_zalloc(pdev->segment_pool, flags, &dma);
drivers/usb/cdns3/cdnsp-mem.c
43
if (!seg->trbs) {
drivers/usb/cdns3/cdnsp-mem.c
44
kfree(seg);
drivers/usb/cdns3/cdnsp-mem.c
49
seg->bounce_buf = kzalloc(max_packet, flags | GFP_DMA);
drivers/usb/cdns3/cdnsp-mem.c
50
if (!seg->bounce_buf)
drivers/usb/cdns3/cdnsp-mem.c
57
seg->trbs[i].link.control |= cpu_to_le32(TRB_CYCLE);
drivers/usb/cdns3/cdnsp-mem.c
59
seg->dma = dma;
drivers/usb/cdns3/cdnsp-mem.c
60
seg->next = NULL;
drivers/usb/cdns3/cdnsp-mem.c
62
return seg;
drivers/usb/cdns3/cdnsp-mem.c
65
dma_pool_free(pdev->segment_pool, seg->trbs, dma);
drivers/usb/cdns3/cdnsp-mem.c
66
kfree(seg);
drivers/usb/cdns3/cdnsp-mem.c
72
struct cdnsp_segment *seg)
drivers/usb/cdns3/cdnsp-mem.c
74
if (seg->trbs)
drivers/usb/cdns3/cdnsp-mem.c
75
dma_pool_free(pdev->segment_pool, seg->trbs, seg->dma);
drivers/usb/cdns3/cdnsp-mem.c
77
kfree(seg->bounce_buf);
drivers/usb/cdns3/cdnsp-mem.c
78
kfree(seg);
drivers/usb/cdns3/cdnsp-mem.c
84
struct cdnsp_segment *seg;
drivers/usb/cdns3/cdnsp-mem.c
86
seg = first->next;
drivers/usb/cdns3/cdnsp-mem.c
88
while (seg != first) {
drivers/usb/cdns3/cdnsp-mem.c
89
struct cdnsp_segment *next = seg->next;
drivers/usb/cdns3/cdnsp-mem.c
91
cdnsp_segment_free(pdev, seg);
drivers/usb/cdns3/cdnsp-mem.c
92
seg = next;
drivers/usb/cdns3/cdnsp-ring.c
101
return cdnsp_last_trb_on_seg(seg, trb) && (seg->next == ring->first_seg);
drivers/usb/cdns3/cdnsp-ring.c
131
struct cdnsp_segment **seg,
drivers/usb/cdns3/cdnsp-ring.c
135
*seg = (*seg)->next;
drivers/usb/cdns3/cdnsp-ring.c
136
*trb = ((*seg)->trbs);
drivers/usb/cdns3/cdnsp-ring.c
1799
u32 *trb_buff_len, struct cdnsp_segment *seg)
drivers/usb/cdns3/cdnsp-ring.c
1835
seg->bounce_buf, new_buff_len, enqd_len);
drivers/usb/cdns3/cdnsp-ring.c
1836
seg->bounce_dma = dma_map_single(dev, seg->bounce_buf,
drivers/usb/cdns3/cdnsp-ring.c
1839
seg->bounce_dma = dma_map_single(dev, seg->bounce_buf,
drivers/usb/cdns3/cdnsp-ring.c
1843
if (dma_mapping_error(dev, seg->bounce_dma)) {
drivers/usb/cdns3/cdnsp-ring.c
1851
seg->bounce_len = new_buff_len;
drivers/usb/cdns3/cdnsp-ring.c
1852
seg->bounce_offs = enqd_len;
drivers/usb/cdns3/cdnsp-ring.c
1854
trace_cdnsp_bounce_map(preq, new_buff_len, enqd_len, seg->bounce_dma,
drivers/usb/cdns3/cdnsp-ring.c
509
struct cdnsp_segment *seg = td->start_seg;
drivers/usb/cdns3/cdnsp-ring.c
522
cdnsp_next_trb(pdev, ep_ring, &seg, &trb);
drivers/usb/cdns3/cdnsp-ring.c
602
struct cdnsp_segment *seg = td->bounce_seg;
drivers/usb/cdns3/cdnsp-ring.c
606
if (!seg)
drivers/usb/cdns3/cdnsp-ring.c
611
trace_cdnsp_bounce_unmap(td->preq, seg->bounce_len, seg->bounce_offs,
drivers/usb/cdns3/cdnsp-ring.c
612
seg->bounce_dma, 0);
drivers/usb/cdns3/cdnsp-ring.c
615
dma_unmap_single(pdev->dev, seg->bounce_dma,
drivers/usb/cdns3/cdnsp-ring.c
620
dma_unmap_single(pdev->dev, seg->bounce_dma, ring->bounce_buf_len,
drivers/usb/cdns3/cdnsp-ring.c
625
seg->bounce_buf, seg->bounce_len,
drivers/usb/cdns3/cdnsp-ring.c
626
seg->bounce_offs);
drivers/usb/cdns3/cdnsp-ring.c
627
if (len != seg->bounce_len)
drivers/usb/cdns3/cdnsp-ring.c
629
len, seg->bounce_len);
drivers/usb/cdns3/cdnsp-ring.c
631
seg->bounce_len = 0;
drivers/usb/cdns3/cdnsp-ring.c
632
seg->bounce_offs = 0;
drivers/usb/cdns3/cdnsp-ring.c
698
struct cdnsp_segment *seg;
drivers/usb/cdns3/cdnsp-ring.c
71
dma_addr_t cdnsp_trb_virt_to_dma(struct cdnsp_segment *seg,
drivers/usb/cdns3/cdnsp-ring.c
719
seg = cdnsp_trb_in_td(pdev, cur_td->start_seg, cur_td->first_trb,
drivers/usb/cdns3/cdnsp-ring.c
722
if (seg && (pep->ep_state & EP_ENABLED) &&
drivers/usb/cdns3/cdnsp-ring.c
74
unsigned long segment_offset = trb - seg->trbs;
drivers/usb/cdns3/cdnsp-ring.c
76
if (trb < seg->trbs || segment_offset >= TRBS_PER_SEGMENT)
drivers/usb/cdns3/cdnsp-ring.c
79
return seg->dma + (segment_offset * sizeof(*trb));
drivers/usb/cdns3/cdnsp-ring.c
92
bool cdnsp_last_trb_on_seg(struct cdnsp_segment *seg, union cdnsp_trb *trb)
drivers/usb/cdns3/cdnsp-ring.c
94
return trb == &seg->trbs[TRBS_PER_SEGMENT - 1];
drivers/usb/cdns3/cdnsp-ring.c
957
struct cdnsp_segment *seg = ring->deq_seg;
drivers/usb/cdns3/cdnsp-ring.c
961
for (sum = 0; trb != stop_trb; cdnsp_next_trb(pdev, ring, &seg, &trb)) {
drivers/usb/cdns3/cdnsp-ring.c
98
struct cdnsp_segment *seg,
drivers/usb/early/xhci-dbc.c
177
xdbc_alloc_ring(struct xdbc_segment *seg, struct xdbc_ring *ring)
drivers/usb/early/xhci-dbc.c
179
seg->trbs = xdbc_get_page(&seg->dma);
drivers/usb/early/xhci-dbc.c
180
if (!seg->trbs)
drivers/usb/early/xhci-dbc.c
183
ring->segment = seg;
drivers/usb/early/xhci-dbc.c
190
struct xdbc_segment *seg = ring->segment;
drivers/usb/early/xhci-dbc.c
192
if (!seg)
drivers/usb/early/xhci-dbc.c
195
memblock_phys_free(seg->dma, PAGE_SIZE);
drivers/usb/early/xhci-dbc.c
201
struct xdbc_segment *seg = ring->segment;
drivers/usb/early/xhci-dbc.c
204
memset(seg->trbs, 0, PAGE_SIZE);
drivers/usb/early/xhci-dbc.c
206
ring->enqueue = seg->trbs;
drivers/usb/early/xhci-dbc.c
207
ring->dequeue = seg->trbs;
drivers/usb/early/xhci-dbc.c
211
link_trb = &seg->trbs[XDBC_TRBS_PER_SEGMENT - 1];
drivers/usb/early/xhci-dbc.c
212
link_trb->field[0] = cpu_to_le32(lower_32_bits(seg->dma));
drivers/usb/early/xhci-dbc.c
213
link_trb->field[1] = cpu_to_le32(upper_32_bits(seg->dma));
drivers/usb/host/xen-hcd.c
587
struct xenusb_request_segment *seg,
drivers/usb/host/xen-hcd.c
606
seg[i].gref = ref;
drivers/usb/host/xen-hcd.c
607
seg[i].offset = (__u16)offset;
drivers/usb/host/xen-hcd.c
608
seg[i].length = (__u16)bytes;
drivers/usb/host/xen-hcd.c
673
&req->seg[0], nr_buff_pages,
drivers/usb/host/xen-hcd.c
694
&gref_head, &req->seg[nr_buff_pages],
drivers/usb/host/xen-hcd.c
734
if (!gnttab_try_end_foreign_access(shadow->req.seg[i].gref))
drivers/usb/host/xhci-dbgcap.c
465
struct xhci_segment *seg = ring->first_seg;
drivers/usb/host/xhci-dbgcap.c
468
memset(seg->trbs, 0, TRB_SEGMENT_SIZE);
drivers/usb/host/xhci-dbgcap.c
472
union xhci_trb *trb = &seg->trbs[TRBS_PER_SEGMENT - 1];
drivers/usb/host/xhci-dbgcap.c
503
struct xhci_segment *seg;
drivers/usb/host/xhci-dbgcap.c
513
seg = kzalloc_obj(*seg, flags);
drivers/usb/host/xhci-dbgcap.c
514
if (!seg)
drivers/usb/host/xhci-dbgcap.c
517
ring->first_seg = seg;
drivers/usb/host/xhci-dbgcap.c
518
ring->last_seg = seg;
drivers/usb/host/xhci-dbgcap.c
519
seg->next = seg;
drivers/usb/host/xhci-dbgcap.c
521
seg->trbs = dma_alloc_coherent(dev, TRB_SEGMENT_SIZE, &dma, flags);
drivers/usb/host/xhci-dbgcap.c
522
if (!seg->trbs)
drivers/usb/host/xhci-dbgcap.c
525
seg->dma = dma;
drivers/usb/host/xhci-dbgcap.c
533
kfree(seg);
drivers/usb/host/xhci-debugfs.c
197
struct xhci_segment *seg)
drivers/usb/host/xhci-debugfs.c
205
trb = &seg->trbs[i];
drivers/usb/host/xhci-debugfs.c
206
dma = seg->dma + i * sizeof(*trb);
drivers/usb/host/xhci-debugfs.c
207
seq_printf(s, "%2u %pad: %s\n", seg->num, &dma,
drivers/usb/host/xhci-debugfs.c
218
struct xhci_segment *seg = ring->first_seg;
drivers/usb/host/xhci-debugfs.c
220
xhci_for_each_ring_seg(ring->first_seg, seg)
drivers/usb/host/xhci-debugfs.c
221
xhci_ring_dump_segment(s, seg);
drivers/usb/host/xhci-mem.c
101
if (!seg || !seg->next)
drivers/usb/host/xhci-mem.c
104
trb = &seg->trbs[TRBS_PER_SEGMENT - 1];
drivers/usb/host/xhci-mem.c
113
trb->link.segment_ptr = cpu_to_le64(seg->next->dma);
drivers/usb/host/xhci-mem.c
118
struct xhci_segment *seg;
drivers/usb/host/xhci-mem.c
125
xhci_for_each_ring_seg(ring->first_seg, seg)
drivers/usb/host/xhci-mem.c
126
xhci_set_link_trb(seg, chain_links);
drivers/usb/host/xhci-mem.c
138
struct xhci_segment *seg;
drivers/usb/host/xhci-mem.c
146
xhci_for_each_ring_seg(src->first_seg, seg) {
drivers/usb/host/xhci-mem.c
148
seg->trbs[i].link.control |= cpu_to_le32(TRB_CYCLE);
drivers/usb/host/xhci-mem.c
171
for (seg = dst->enq_seg; seg != dst->last_seg; seg = seg->next)
drivers/usb/host/xhci-mem.c
172
seg->next->num = seg->num + 1;
drivers/usb/host/xhci-mem.c
1798
struct xhci_segment *seg;
drivers/usb/host/xhci-mem.c
1809
seg = evt_ring->first_seg;
drivers/usb/host/xhci-mem.c
1812
entry->seg_addr = cpu_to_le64(seg->dma);
drivers/usb/host/xhci-mem.c
1815
seg = seg->next;
drivers/usb/host/xhci-mem.c
208
struct xhci_segment *seg,
drivers/usb/host/xhci-mem.c
214
key = (unsigned long)(seg->dma >> TRB_SEGMENT_SHIFT);
drivers/usb/host/xhci-mem.c
229
struct xhci_segment *seg)
drivers/usb/host/xhci-mem.c
233
key = (unsigned long)(seg->dma >> TRB_SEGMENT_SHIFT);
drivers/usb/host/xhci-mem.c
244
struct xhci_segment *seg;
drivers/usb/host/xhci-mem.c
251
xhci_for_each_ring_seg(first_seg, seg) {
drivers/usb/host/xhci-mem.c
253
ring, seg, mem_flags);
drivers/usb/host/xhci-mem.c
261
failed_seg = seg;
drivers/usb/host/xhci-mem.c
262
xhci_for_each_ring_seg(first_seg, seg) {
drivers/usb/host/xhci-mem.c
263
xhci_remove_segment_mapping(trb_address_map, seg);
drivers/usb/host/xhci-mem.c
264
if (seg == failed_seg)
drivers/usb/host/xhci-mem.c
273
struct xhci_segment *seg;
drivers/usb/host/xhci-mem.c
278
xhci_for_each_ring_seg(ring->first_seg, seg)
drivers/usb/host/xhci-mem.c
279
xhci_remove_segment_mapping(ring->trb_address_map, seg);
drivers/usb/host/xhci-mem.c
34
struct xhci_segment *seg;
drivers/usb/host/xhci-mem.c
38
seg = kzalloc_node(sizeof(*seg), flags, dev_to_node(dev));
drivers/usb/host/xhci-mem.c
39
if (!seg)
drivers/usb/host/xhci-mem.c
42
seg->trbs = dma_pool_zalloc(xhci->segment_pool, flags, &dma);
drivers/usb/host/xhci-mem.c
43
if (!seg->trbs) {
drivers/usb/host/xhci-mem.c
44
kfree(seg);
drivers/usb/host/xhci-mem.c
49
seg->bounce_buf = kzalloc_node(max_packet, flags,
drivers/usb/host/xhci-mem.c
51
if (!seg->bounce_buf) {
drivers/usb/host/xhci-mem.c
52
dma_pool_free(xhci->segment_pool, seg->trbs, dma);
drivers/usb/host/xhci-mem.c
53
kfree(seg);
drivers/usb/host/xhci-mem.c
57
seg->num = num;
drivers/usb/host/xhci-mem.c
58
seg->dma = dma;
drivers/usb/host/xhci-mem.c
59
seg->next = NULL;
drivers/usb/host/xhci-mem.c
61
return seg;
drivers/usb/host/xhci-mem.c
64
static void xhci_segment_free(struct xhci_hcd *xhci, struct xhci_segment *seg)
drivers/usb/host/xhci-mem.c
66
if (seg->trbs) {
drivers/usb/host/xhci-mem.c
67
dma_pool_free(xhci->segment_pool, seg->trbs, seg->dma);
drivers/usb/host/xhci-mem.c
68
seg->trbs = NULL;
drivers/usb/host/xhci-mem.c
70
kfree(seg->bounce_buf);
drivers/usb/host/xhci-mem.c
71
kfree(seg);
drivers/usb/host/xhci-mem.c
76
struct xhci_segment *seg, *next;
drivers/usb/host/xhci-mem.c
79
seg = ring->first_seg;
drivers/usb/host/xhci-mem.c
81
while (seg) {
drivers/usb/host/xhci-mem.c
82
next = seg->next;
drivers/usb/host/xhci-mem.c
83
xhci_segment_free(xhci, seg);
drivers/usb/host/xhci-mem.c
84
seg = next;
drivers/usb/host/xhci-mem.c
96
static void xhci_set_link_trb(struct xhci_segment *seg, bool chain_links)
drivers/usb/host/xhci-ring.c
112
static bool last_trb_on_seg(struct xhci_segment *seg, union xhci_trb *trb)
drivers/usb/host/xhci-ring.c
114
return trb == &seg->trbs[TRBS_PER_SEGMENT - 1];
drivers/usb/host/xhci-ring.c
118
struct xhci_segment *seg, union xhci_trb *trb)
drivers/usb/host/xhci-ring.c
120
return last_trb_on_seg(seg, trb) && (seg->next == ring->first_seg);
drivers/usb/host/xhci-ring.c
163
static unsigned int trb_to_pos(struct xhci_segment *seg, union xhci_trb *trb)
drivers/usb/host/xhci-ring.c
165
return seg->num * TRBS_PER_SEGMENT + (trb - seg->trbs);
drivers/usb/host/xhci-ring.c
172
static void next_trb(struct xhci_segment **seg,
drivers/usb/host/xhci-ring.c
175
if (trb_is_link(*trb) || last_trb_on_seg(*seg, *trb)) {
drivers/usb/host/xhci-ring.c
176
*seg = (*seg)->next;
drivers/usb/host/xhci-ring.c
177
*trb = ((*seg)->trbs);
drivers/usb/host/xhci-ring.c
2281
struct xhci_segment *seg = td->start_seg;
drivers/usb/host/xhci-ring.c
2283
for (sum = 0; trb != stop_trb; next_trb(&seg, &trb)) {
drivers/usb/host/xhci-ring.c
3542
u32 *trb_buff_len, struct xhci_segment *seg)
drivers/usb/host/xhci-ring.c
3581
seg->bounce_buf, new_buff_len, enqd_len);
drivers/usb/host/xhci-ring.c
3586
memcpy(seg->bounce_buf, urb->transfer_buffer + enqd_len, new_buff_len);
drivers/usb/host/xhci-ring.c
3589
seg->bounce_dma = dma_map_single(dev, seg->bounce_buf,
drivers/usb/host/xhci-ring.c
3592
seg->bounce_dma = dma_map_single(dev, seg->bounce_buf,
drivers/usb/host/xhci-ring.c
3596
if (dma_mapping_error(dev, seg->bounce_dma)) {
drivers/usb/host/xhci-ring.c
3602
seg->bounce_len = new_buff_len;
drivers/usb/host/xhci-ring.c
3603
seg->bounce_offs = enqd_len;
drivers/usb/host/xhci-ring.c
382
struct xhci_segment *seg;
drivers/usb/host/xhci-ring.c
406
seg = ring->enq_seg;
drivers/usb/host/xhci-ring.c
409
seg = seg->next;
drivers/usb/host/xhci-ring.c
410
if (seg == ring->deq_seg) {
drivers/usb/host/xhci-ring.c
71
dma_addr_t xhci_trb_virt_to_dma(struct xhci_segment *seg,
drivers/usb/host/xhci-ring.c
76
if (!seg || !trb || trb < seg->trbs)
drivers/usb/host/xhci-ring.c
79
segment_offset = trb - seg->trbs;
drivers/usb/host/xhci-ring.c
805
struct xhci_segment *seg = td->start_seg;
drivers/usb/host/xhci-ring.c
82
return seg->dma + (segment_offset * sizeof(*trb));
drivers/usb/host/xhci-ring.c
821
next_trb(&seg, &trb);
drivers/usb/host/xhci-ring.c
849
struct xhci_segment *seg = td->bounce_seg;
drivers/usb/host/xhci-ring.c
853
if (!ring || !seg || !urb)
drivers/usb/host/xhci-ring.c
857
dma_unmap_single(dev, seg->bounce_dma, ring->bounce_buf_len,
drivers/usb/host/xhci-ring.c
862
dma_unmap_single(dev, seg->bounce_dma, ring->bounce_buf_len,
drivers/usb/host/xhci-ring.c
866
len = sg_pcopy_from_buffer(urb->sg, urb->num_sgs, seg->bounce_buf,
drivers/usb/host/xhci-ring.c
867
seg->bounce_len, seg->bounce_offs);
drivers/usb/host/xhci-ring.c
868
if (len != seg->bounce_len)
drivers/usb/host/xhci-ring.c
870
len, seg->bounce_len);
drivers/usb/host/xhci-ring.c
872
memcpy(urb->transfer_buffer + seg->bounce_offs, seg->bounce_buf,
drivers/usb/host/xhci-ring.c
873
seg->bounce_len);
drivers/usb/host/xhci-ring.c
875
seg->bounce_len = 0;
drivers/usb/host/xhci-ring.c
876
seg->bounce_offs = 0;
drivers/usb/host/xhci-ring.c
89
struct xhci_segment *seg;
drivers/usb/host/xhci-ring.c
91
xhci_for_each_ring_seg(start_seg, seg) {
drivers/usb/host/xhci-ring.c
92
if (in_range(dma, seg->dma, TRB_SEGMENT_SIZE)) {
drivers/usb/host/xhci-ring.c
94
*match_seg = seg;
drivers/usb/host/xhci-ring.c
95
return &seg->trbs[(dma - seg->dma) / sizeof(union xhci_trb)];
drivers/usb/host/xhci-sideband.c
20
struct xhci_segment *seg;
drivers/usb/host/xhci-sideband.c
41
seg = ring->first_seg;
drivers/usb/host/xhci-sideband.c
42
if (!seg)
drivers/usb/host/xhci-sideband.c
51
dma_get_sgtable(dev, sgt, seg->trbs, seg->dma,
drivers/usb/host/xhci-sideband.c
55
seg = seg->next;
drivers/usb/host/xhci.c
59
struct xhci_segment *seg;
drivers/usb/host/xhci.c
64
xhci_for_each_ring_seg(ring->first_seg, seg) {
drivers/usb/host/xhci.c
65
if (seg == td->start_seg)
drivers/usb/host/xhci.c
865
struct xhci_segment *seg;
drivers/usb/host/xhci.c
868
xhci_for_each_ring_seg(ring->first_seg, seg) {
drivers/usb/host/xhci.c
870
memset(seg->trbs, 0, sizeof(union xhci_trb) * (TRBS_PER_SEGMENT - 1));
drivers/usb/host/xhci.c
872
seg->trbs[TRBS_PER_SEGMENT - 1].link.control &= cpu_to_le32(~TRB_CYCLE);
drivers/usb/host/xhci.h
1278
#define xhci_for_each_ring_seg(head, seg) \
drivers/usb/host/xhci.h
1279
for (seg = head; seg != NULL; seg = (seg->next != head ? seg->next : NULL))
drivers/usb/host/xhci.h
1910
dma_addr_t xhci_trb_virt_to_dma(struct xhci_segment *seg, union xhci_trb *trb);
drivers/vhost/net.c
1057
int seg = 0;
drivers/vhost/net.c
1067
if (unlikely(seg >= UIO_MAXIOV)) {
drivers/vhost/net.c
1071
r = vhost_get_vq_desc_n(vq, vq->iov + seg,
drivers/vhost/net.c
1072
ARRAY_SIZE(vq->iov) - seg, &out,
drivers/vhost/net.c
1092
len = iov_length(vq->iov + seg, in);
drivers/vhost/net.c
1099
seg += in;
drivers/vhost/net.c
1103
*iovcount = seg;
drivers/xen/dbgp.c
26
dbgp.u.pci.seg = pci_domain_nr(pdev->bus);
drivers/xen/gntdev.c
863
struct gntdev_grant_copy_segment *seg,
drivers/xen/gntdev.c
873
if (!(seg->flags & (GNTCOPY_source_gref | GNTCOPY_dest_gref)))
drivers/xen/gntdev.c
877
if (seg->flags & GNTCOPY_source_gref) {
drivers/xen/gntdev.c
878
if (seg->source.foreign.offset + seg->len > XEN_PAGE_SIZE)
drivers/xen/gntdev.c
881
if (seg->flags & GNTCOPY_dest_gref) {
drivers/xen/gntdev.c
882
if (seg->dest.foreign.offset + seg->len > XEN_PAGE_SIZE)
drivers/xen/gntdev.c
889
while (copied < seg->len) {
drivers/xen/gntdev.c
902
len = seg->len - copied;
drivers/xen/gntdev.c
907
if (seg->flags & GNTCOPY_source_gref) {
drivers/xen/gntdev.c
908
op->source.u.ref = seg->source.foreign.ref;
drivers/xen/gntdev.c
909
op->source.domid = seg->source.foreign.domid;
drivers/xen/gntdev.c
910
op->source.offset = seg->source.foreign.offset + copied;
drivers/xen/gntdev.c
913
virt = seg->source.virt + copied;
drivers/xen/gntdev.c
927
if (seg->flags & GNTCOPY_dest_gref) {
drivers/xen/gntdev.c
928
op->dest.u.ref = seg->dest.foreign.ref;
drivers/xen/gntdev.c
929
op->dest.domid = seg->dest.foreign.domid;
drivers/xen/gntdev.c
930
op->dest.offset = seg->dest.foreign.offset + copied;
drivers/xen/gntdev.c
933
virt = seg->dest.virt + copied;
drivers/xen/gntdev.c
982
struct gntdev_grant_copy_segment seg;
drivers/xen/gntdev.c
984
if (copy_from_user(&seg, &copy.segments[i], sizeof(seg))) {
drivers/xen/gntdev.c
990
ret = gntdev_grant_copy_seg(batch, &seg, &copy.segments[i].status);
drivers/xen/pci.c
176
.seg = pci_domain_nr(pci_dev->bus),
drivers/xen/pci.c
201
.dev.seg = pci_domain_nr(dev->bus),
drivers/xen/pci.c
61
add->seg = pci_domain_nr(pci_dev->bus);
drivers/xen/xen-pciback/pci_stub.c
138
.seg = pci_domain_nr(dev->bus),
drivers/xen/xen-pciback/pci_stub.c
430
.seg = pci_domain_nr(dev->bus),
drivers/xen/xen-scsiback.c
213
static unsigned long vaddr(struct vscsibk_pend *req, int seg)
drivers/xen/xen-scsiback.c
215
return vaddr_page(req->pages[seg]);
drivers/xen/xen-scsiback.c
463
struct scsiif_request_segment *seg, struct page **pg,
drivers/xen/xen-scsiback.c
477
flags, seg[i].gref, info->domid);
drivers/xen/xen-scsiback.c
500
struct scsiif_request_segment *seg;
drivers/xen/xen-scsiback.c
522
err = scsiback_gnttab_data_map_list(pending_req, ring_req->seg,
drivers/xen/xen-scsiback.c
530
n_segs = ring_req->seg[i].length /
drivers/xen/xen-scsiback.c
532
if ((unsigned)ring_req->seg[i].offset +
drivers/xen/xen-scsiback.c
533
(unsigned)ring_req->seg[i].length > PAGE_SIZE ||
drivers/xen/xen-scsiback.c
535
ring_req->seg[i].length)
drivers/xen/xen-scsiback.c
560
seg = ring_req->seg;
drivers/xen/xen-scsiback.c
561
err = scsiback_gnttab_data_map_list(pending_req, seg,
drivers/xen/xen-scsiback.c
567
seg = (struct scsiif_request_segment *)(
drivers/xen/xen-scsiback.c
568
vaddr(pending_req, i) + ring_req->seg[i].offset);
drivers/xen/xen-scsiback.c
569
n_segs = ring_req->seg[i].length /
drivers/xen/xen-scsiback.c
571
err = scsiback_gnttab_data_map_list(pending_req, seg,
drivers/xen/xen-scsiback.c
578
end_seg = vaddr(pending_req, 0) + ring_req->seg[0].offset;
drivers/xen/xen-scsiback.c
579
seg = (struct scsiif_request_segment *)end_seg;
drivers/xen/xen-scsiback.c
580
end_seg += ring_req->seg[0].length;
drivers/xen/xen-scsiback.c
585
sg_set_page(sg, pg[i], seg->length, seg->offset);
drivers/xen/xen-scsiback.c
586
pending_req->data_len += seg->length;
drivers/xen/xen-scsiback.c
587
seg++;
drivers/xen/xen-scsiback.c
588
if (nr_sgl && (unsigned long)seg >= end_seg) {
drivers/xen/xen-scsiback.c
591
ring_req->seg[i_seg].offset;
drivers/xen/xen-scsiback.c
592
seg = (struct scsiif_request_segment *)end_seg;
drivers/xen/xen-scsiback.c
593
end_seg += ring_req->seg[i_seg].length;
fs/binfmt_elf_fdpic.c
1000
mm->end_data = seg->addr + phdr->p_memsz;
fs/binfmt_elf_fdpic.c
1004
seg++;
fs/binfmt_elf_fdpic.c
1019
struct elf_fdpic_loadseg *seg;
fs/binfmt_elf_fdpic.c
1028
seg = params->loadmap->segs;
fs/binfmt_elf_fdpic.c
1106
seg->addr = maddr + disp;
fs/binfmt_elf_fdpic.c
1107
seg->p_vaddr = phdr->p_vaddr;
fs/binfmt_elf_fdpic.c
1108
seg->p_memsz = phdr->p_memsz;
fs/binfmt_elf_fdpic.c
1112
params->elfhdr_addr = seg->addr;
fs/binfmt_elf_fdpic.c
1180
seg++;
fs/binfmt_elf_fdpic.c
755
struct elf_fdpic_loadseg *seg;
fs/binfmt_elf_fdpic.c
798
seg = loadmap->segs;
fs/binfmt_elf_fdpic.c
799
for (loop = loadmap->nsegs; loop > 0; loop--, seg++) {
fs/binfmt_elf_fdpic.c
800
if (params->hdr.e_entry >= seg->p_vaddr &&
fs/binfmt_elf_fdpic.c
801
params->hdr.e_entry < seg->p_vaddr + seg->p_memsz) {
fs/binfmt_elf_fdpic.c
803
(params->hdr.e_entry - seg->p_vaddr) +
fs/binfmt_elf_fdpic.c
804
seg->addr;
fs/binfmt_elf_fdpic.c
823
seg = loadmap->segs;
fs/binfmt_elf_fdpic.c
824
for (loop = loadmap->nsegs; loop > 0; loop--, seg++) {
fs/binfmt_elf_fdpic.c
825
if (phdr->p_vaddr >= seg->p_vaddr &&
fs/binfmt_elf_fdpic.c
827
seg->p_vaddr + seg->p_memsz) {
fs/binfmt_elf_fdpic.c
829
(phdr->p_vaddr - seg->p_vaddr) +
fs/binfmt_elf_fdpic.c
830
seg->addr +
fs/binfmt_elf_fdpic.c
844
seg = loadmap->segs;
fs/binfmt_elf_fdpic.c
845
for (loop = loadmap->nsegs; loop > 0; loop--, seg++) {
fs/binfmt_elf_fdpic.c
846
if (phdr->p_vaddr >= seg->p_vaddr &&
fs/binfmt_elf_fdpic.c
848
seg->p_vaddr + seg->p_memsz) {
fs/binfmt_elf_fdpic.c
853
(phdr->p_vaddr - seg->p_vaddr) +
fs/binfmt_elf_fdpic.c
854
seg->addr;
fs/binfmt_elf_fdpic.c
881
seg = mseg + 1;
fs/binfmt_elf_fdpic.c
884
if (seg->p_vaddr - mseg->p_vaddr == seg->addr - mseg->addr) {
fs/binfmt_elf_fdpic.c
886
if (load_addr == (seg->addr & PAGE_MASK)) {
fs/binfmt_elf_fdpic.c
890
mseg->p_memsz += seg->addr & ~PAGE_MASK;
fs/binfmt_elf_fdpic.c
891
mseg->p_memsz += seg->p_memsz;
fs/binfmt_elf_fdpic.c
898
if (mseg != seg)
fs/binfmt_elf_fdpic.c
899
*mseg = *seg;
fs/binfmt_elf_fdpic.c
908
seg = loadmap->segs;
fs/binfmt_elf_fdpic.c
909
for (loop = 0; loop < loadmap->nsegs; loop++, seg++)
fs/binfmt_elf_fdpic.c
912
(unsigned long long) seg->addr,
fs/binfmt_elf_fdpic.c
913
(unsigned long long) seg->addr + seg->p_memsz - 1,
fs/binfmt_elf_fdpic.c
914
(unsigned long long) seg->p_vaddr,
fs/binfmt_elf_fdpic.c
915
(unsigned long long) seg->p_memsz);
fs/binfmt_elf_fdpic.c
935
struct elf_fdpic_loadseg *seg;
fs/binfmt_elf_fdpic.c
941
seg = params->loadmap->segs;
fs/binfmt_elf_fdpic.c
971
seg->addr = maddr + (phdr->p_vaddr - base);
fs/binfmt_elf_fdpic.c
972
seg->p_vaddr = phdr->p_vaddr;
fs/binfmt_elf_fdpic.c
973
seg->p_memsz = phdr->p_memsz;
fs/binfmt_elf_fdpic.c
975
ret = read_code(file, seg->addr, phdr->p_offset,
fs/binfmt_elf_fdpic.c
982
params->elfhdr_addr = seg->addr;
fs/binfmt_elf_fdpic.c
986
if (clear_user((void *) (seg->addr + phdr->p_filesz),
fs/binfmt_elf_fdpic.c
994
mm->start_code = seg->addr;
fs/binfmt_elf_fdpic.c
995
mm->end_code = seg->addr +
fs/binfmt_elf_fdpic.c
999
mm->start_data = seg->addr;
fs/btrfs/direct-io.c
1005
int i, seg;
fs/btrfs/direct-io.c
1014
for (seg = 0; seg < iter->nr_segs; seg++) {
fs/btrfs/direct-io.c
1015
for (i = seg + 1; i < iter->nr_segs; i++) {
fs/btrfs/direct-io.c
1016
const struct iovec *iov1 = iter_iov(iter) + seg;
fs/f2fs/segment.c
3032
struct curseg_info *seg)
fs/f2fs/segment.c
3034
return __next_free_blkoff(sbi, seg->segno, seg->next_blkoff + 1);
fs/f2fs/segment.c
3762
struct curseg_info *seg)
fs/f2fs/segment.c
3765
if (--seg->fragment_remained_chunk > 0)
fs/f2fs/segment.c
3768
seg->fragment_remained_chunk =
fs/f2fs/segment.c
3770
seg->next_blkoff +=
fs/nfs/nfs42xdr.c
1073
static inline uint64_t read_plus_segment_length(struct read_plus_segment *seg)
fs/nfs/nfs42xdr.c
1075
return seg->type == NFS4_CONTENT_DATA ? seg->data.length : seg->hole.length;
fs/nfs/nfs42xdr.c
1079
struct read_plus_segment *seg)
fs/nfs/nfs42xdr.c
1086
seg->type = be32_to_cpup(p++);
fs/nfs/nfs42xdr.c
1088
p = xdr_inline_decode(xdr, seg->type == NFS4_CONTENT_DATA ? 12 : 16);
fs/nfs/nfs42xdr.c
1091
p = xdr_decode_hyper(p, &seg->offset);
fs/nfs/nfs42xdr.c
1093
if (seg->type == NFS4_CONTENT_DATA) {
fs/nfs/nfs42xdr.c
1097
seg->data.length = len;
fs/nfs/nfs42xdr.c
1098
seg->data.from = xdr_stream_pos(xdr);
fs/nfs/nfs42xdr.c
1102
} else if (seg->type == NFS4_CONTENT_HOLE) {
fs/nfs/nfs42xdr.c
1103
xdr_decode_hyper(p, &seg->hole.length);
fs/nfs/nfs42xdr.c
1112
struct read_plus_segment *seg)
fs/nfs/nfs42xdr.c
1114
unsigned long offset = seg->offset;
fs/nfs/nfs42xdr.c
1115
unsigned long length = read_plus_segment_length(seg);
fs/nfs/nfs42xdr.c
1132
if (seg->type == NFS4_CONTENT_HOLE)
fs/nfs/nfs42xdr.c
1135
return xdr_stream_move_subsegment(xdr, seg->data.from, bufpos, length);
fs/nfsd/blocklayout.c
102
if (seg->offset & (block_size - 1)) {
fs/nfsd/blocklayout.c
145
seg->iomode, args->lg_minlength, bex);
fs/nfsd/blocklayout.c
163
length = last_bex->foff + last_bex->len - seg->offset;
fs/nfsd/blocklayout.c
169
seg->offset = first_bex->foff;
fs/nfsd/blocklayout.c
170
seg->length = last_bex->foff - first_bex->foff + last_bex->len;
fs/nfsd/blocklayout.c
174
seg->length = 0;
fs/nfsd/blocklayout.c
91
struct nfsd4_layout_seg *seg = &args->lg_seg;
fs/nfsd/blocklayout.c
94
u64 offset = seg->offset, length = seg->length;
fs/nfsd/flexfilelayout.c
26
struct nfsd4_layout_seg *seg = &args->lg_seg;
fs/nfsd/flexfilelayout.c
53
if (seg->iomode == IOMODE_READ) {
fs/nfsd/flexfilelayout.c
68
seg->offset = 0;
fs/nfsd/flexfilelayout.c
69
seg->length = NFS4_MAX_UINT64;
fs/nfsd/flexfilelayout.c
71
dprintk("GET: 0x%llx:0x%llx %d\n", seg->offset, seg->length,
fs/nfsd/flexfilelayout.c
72
seg->iomode);
fs/nfsd/flexfilelayout.c
76
seg->length = 0;
fs/nfsd/nfs4layouts.c
355
layout_end(struct nfsd4_layout_seg *seg)
fs/nfsd/nfs4layouts.c
357
u64 end = seg->offset + seg->length;
fs/nfsd/nfs4layouts.c
358
return end >= seg->offset ? end : NFS4_MAX_UINT64;
fs/nfsd/nfs4layouts.c
419
struct nfsd4_layout_seg *seg = &lgp->lg_seg;
fs/nfsd/nfs4layouts.c
430
if (layouts_try_merge(&lp->lo_seg, seg))
fs/nfsd/nfs4layouts.c
439
memcpy(&new->lo_seg, seg, sizeof(new->lo_seg));
fs/nfsd/nfs4layouts.c
448
if (layouts_try_merge(&lp->lo_seg, seg))
fs/nfsd/nfs4layouts.c
479
nfsd4_return_file_layout(struct nfs4_layout *lp, struct nfsd4_layout_seg *seg,
fs/nfsd/nfs4layouts.c
485
if (seg->offset <= lo->offset) {
fs/nfsd/nfs4layouts.c
486
if (layout_end(seg) >= end) {
fs/nfsd/nfs4layouts.c
490
lo->offset = layout_end(seg);
fs/nfsd/nfs4layouts.c
493
if (layout_end(seg) < end) {
fs/nfsd/nfs4layouts.c
497
end = seg->offset;
fs/nfsd/nfs4proc.c
2717
const struct nfsd4_layout_seg *seg = &lcp->lc_seg;
fs/nfsd/nfs4proc.c
2740
if (new_size <= seg->offset)
fs/nfsd/nfs4proc.c
2742
if (new_size > seg->offset + seg->length)
fs/nilfs2/sufile.c
176
__u64 *seg;
fs/nilfs2/sufile.c
184
for (seg = segnumv; seg < segnumv + nsegs; seg++) {
fs/nilfs2/sufile.c
185
if (unlikely(*seg >= nilfs_sufile_get_nsegments(sufile))) {
fs/nilfs2/sufile.c
188
__func__, (unsigned long long)*seg);
fs/nilfs2/sufile.c
201
seg = segnumv;
fs/nilfs2/sufile.c
202
blkoff = nilfs_sufile_get_blkoff(sufile, *seg);
fs/nilfs2/sufile.c
208
dofunc(sufile, *seg, header_bh, bh);
fs/nilfs2/sufile.c
210
if (++seg >= segnumv + nsegs)
fs/nilfs2/sufile.c
213
blkoff = nilfs_sufile_get_blkoff(sufile, *seg);
fs/nilfs2/sufile.c
226
n = seg - segnumv;
fs/smb/client/smbdirect.c
3202
size_t off, len, seg;
fs/smb/client/smbdirect.c
3217
seg = min_t(size_t, len, PAGE_SIZE - off);
fs/smb/client/smbdirect.c
3224
if (!smb_set_sge(rdma, page, off, seg))
fs/smb/client/smbdirect.c
3227
ret += seg;
fs/smb/client/smbdirect.c
3228
len -= seg;
fs/splice.c
848
size_t seg;
fs/splice.c
855
seg = min_t(size_t, remain, buf->len);
fs/splice.c
864
bvec_set_page(&bvec[bc++], buf->page, seg, buf->offset);
fs/splice.c
865
remain -= seg;
fs/splice.c
893
size_t seg = min_t(size_t, ret, buf->len);
fs/splice.c
895
buf->offset += seg;
fs/splice.c
896
buf->len -= seg;
fs/splice.c
897
ret -= seg;
include/linux/dmar.h
62
u16 seg;
include/linux/mmzone.h
515
u8 seg;
include/linux/peci-cpu.h
58
int peci_ep_pci_local_read(struct peci_device *device, u8 seg,
include/linux/peci-cpu.h
61
int peci_mmio_read(struct peci_device *device, u8 bar, u8 seg,
include/linux/uio.h
177
unsigned long seg;
include/linux/uio.h
180
for (seg = 0; seg < nr_segs; seg++)
include/linux/uio.h
181
ret += iov[seg].iov_len;
include/uapi/linux/ublk_cmd.h
742
struct ublk_param_segment seg;
include/uapi/linux/v4l2-controls.h
2795
struct v4l2_vp9_segmentation seg;
include/xen/interface/io/blkif.h
202
struct blkif_request_segment seg[BLKIF_MAX_SEGMENTS_PER_REQUEST];
include/xen/interface/io/usbif.h
364
struct xenusb_request_segment seg[XENUSB_MAX_SEGMENTS_PER_REQUEST];
include/xen/interface/io/vscsiif.h
267
struct scsiif_request_segment seg[VSCSIIF_SG_TABLESIZE];
include/xen/interface/physdev.h
236
uint16_t seg;
include/xen/interface/physdev.h
268
uint16_t seg;
ipc/msgutil.c
107
for (seg = msg->next; seg != NULL; seg = seg->next) {
ipc/msgutil.c
111
if (copy_from_user(seg + 1, src, alen))
ipc/msgutil.c
161
struct msg_msgseg *seg;
ipc/msgutil.c
167
for (seg = msg->next; seg != NULL; seg = seg->next) {
ipc/msgutil.c
171
if (copy_to_user(dest, seg + 1, alen))
ipc/msgutil.c
179
struct msg_msgseg *seg;
ipc/msgutil.c
183
seg = msg->next;
ipc/msgutil.c
185
while (seg != NULL) {
ipc/msgutil.c
186
struct msg_msgseg *tmp = seg->next;
ipc/msgutil.c
189
kfree(seg);
ipc/msgutil.c
190
seg = tmp;
ipc/msgutil.c
71
struct msg_msgseg *seg;
ipc/msgutil.c
76
seg = kmalloc(sizeof(*seg) + alen, GFP_KERNEL_ACCOUNT);
ipc/msgutil.c
77
if (seg == NULL)
ipc/msgutil.c
79
*pseg = seg;
ipc/msgutil.c
80
seg->next = NULL;
ipc/msgutil.c
81
pseg = &seg->next;
ipc/msgutil.c
95
struct msg_msgseg *seg;
kernel/rcu/rcu_segcblist.c
110
static void rcu_segcblist_set_seglen(struct rcu_segcblist *rsclp, int seg, long v)
kernel/rcu/rcu_segcblist.c
112
WRITE_ONCE(rsclp->seglen[seg], v);
kernel/rcu/rcu_segcblist.c
116
static void rcu_segcblist_add_seglen(struct rcu_segcblist *rsclp, int seg, long v)
kernel/rcu/rcu_segcblist.c
118
WRITE_ONCE(rsclp->seglen[seg], rsclp->seglen[seg] + v);
kernel/rcu/rcu_segcblist.c
138
static void rcu_segcblist_inc_seglen(struct rcu_segcblist *rsclp, int seg)
kernel/rcu/rcu_segcblist.c
140
rcu_segcblist_add_seglen(rsclp, seg, 1);
kernel/rcu/rcu_segcblist.c
92
long rcu_segcblist_get_seglen(struct rcu_segcblist *rsclp, int seg)
kernel/rcu/rcu_segcblist.c
94
return READ_ONCE(rsclp->seglen[seg]);
kernel/rcu/rcu_segcblist.h
103
static inline bool rcu_segcblist_restempty(struct rcu_segcblist *rsclp, int seg)
kernel/rcu/rcu_segcblist.h
105
return !READ_ONCE(*READ_ONCE(rsclp->tails[seg]));
kernel/rcu/rcu_segcblist.h
112
static inline bool rcu_segcblist_segempty(struct rcu_segcblist *rsclp, int seg)
kernel/rcu/rcu_segcblist.h
114
if (seg == RCU_DONE_TAIL)
kernel/rcu/rcu_segcblist.h
116
return rsclp->tails[seg - 1] == rsclp->tails[seg];
kernel/rcu/rcu_segcblist.h
18
long rcu_segcblist_get_seglen(struct rcu_segcblist *rsclp, int seg);
lib/iov_iter.c
1369
unsigned long seg;
lib/iov_iter.c
1389
for (seg = 0; seg < nr_segs; seg++) {
lib/iov_iter.c
1390
ssize_t len = (ssize_t)iov[seg].iov_len;
lib/iov_iter.c
1392
if (!access_ok(iov[seg].iov_base, len)) {
lib/iov_iter.c
1401
iov[seg].iov_len = len;
lib/iov_iter.c
1709
size_t seg = min_t(size_t, len, PAGE_SIZE);
lib/iov_iter.c
1717
len -= seg;
lib/scatterlist.c
1144
size_t seg = min_t(size_t, PAGE_SIZE - off, len);
lib/scatterlist.c
1147
sg_set_page(sg, page, seg, off);
lib/scatterlist.c
1150
len -= seg;
lib/scatterlist.c
1228
size_t off, len, seg;
lib/scatterlist.c
1244
seg = min_t(size_t, len, PAGE_SIZE - off);
lib/scatterlist.c
1255
len -= seg;
mm/vmscan.c
4312
int seg;
mm/vmscan.c
4322
seg = 0;
mm/vmscan.c
4327
seg = MEMCG_LRU_HEAD;
mm/vmscan.c
4329
seg = MEMCG_LRU_TAIL;
mm/vmscan.c
4337
WRITE_ONCE(lruvec->lrugen.seg, seg);
mm/vmscan.c
4425
if (READ_ONCE(lruvec->lrugen.seg) != MEMCG_LRU_HEAD)
mm/vmscan.c
4920
if (READ_ONCE(lruvec->lrugen.seg) != MEMCG_LRU_TAIL)
mm/vmscan.c
4943
return READ_ONCE(lruvec->lrugen.seg) != MEMCG_LRU_TAIL ?
net/caif/cfrfml.c
216
u8 seg;
net/caif/cfrfml.c
241
seg = 1;
net/caif/cfrfml.c
244
if (cfpkt_add_head(frontpkt, &seg, 1) < 0)
net/caif/cfrfml.c
273
seg = 0;
net/caif/cfrfml.c
276
if (cfpkt_add_head(frontpkt, &seg, 1) < 0)
net/core/skbuff.c
3201
int seg;
net/core/skbuff.c
3222
for (seg = 0; seg < skb_shinfo(skb)->nr_frags; seg++) {
net/core/skbuff.c
3223
const skb_frag_t *f = &skb_shinfo(skb)->frags[seg];
net/ipv4/tcp_offload.c
34
static void __tcpv4_gso_segment_csum(struct sk_buff *seg,
net/ipv4/tcp_offload.c
44
th = tcp_hdr(seg);
net/ipv4/tcp_offload.c
45
iph = ip_hdr(seg);
net/ipv4/tcp_offload.c
47
inet_proto_csum_replace4(&th->check, seg, *oldip, newip, true);
net/ipv4/tcp_offload.c
48
inet_proto_csum_replace2(&th->check, seg, *oldport, newport, false);
net/ipv4/tcp_offload.c
59
struct sk_buff *seg;
net/ipv4/tcp_offload.c
63
seg = segs;
net/ipv4/tcp_offload.c
64
th = tcp_hdr(seg);
net/ipv4/tcp_offload.c
65
iph = ip_hdr(seg);
net/ipv4/tcp_offload.c
66
th2 = tcp_hdr(seg->next);
net/ipv4/tcp_offload.c
67
iph2 = ip_hdr(seg->next);
net/ipv4/tcp_offload.c
73
while ((seg = seg->next)) {
net/ipv4/tcp_offload.c
74
th2 = tcp_hdr(seg);
net/ipv4/tcp_offload.c
75
iph2 = ip_hdr(seg);
net/ipv4/tcp_offload.c
77
__tcpv4_gso_segment_csum(seg,
net/ipv4/tcp_offload.c
80
__tcpv4_gso_segment_csum(seg,
net/ipv4/udp_offload.c
348
static void __udpv4_gso_segment_csum(struct sk_buff *seg,
net/ipv4/udp_offload.c
358
uh = udp_hdr(seg);
net/ipv4/udp_offload.c
359
iph = ip_hdr(seg);
net/ipv4/udp_offload.c
362
inet_proto_csum_replace4(&uh->check, seg, *oldip, *newip,
net/ipv4/udp_offload.c
364
inet_proto_csum_replace2(&uh->check, seg, *oldport, *newport,
net/ipv4/udp_offload.c
377
struct sk_buff *seg;
net/ipv4/udp_offload.c
381
seg = segs;
net/ipv4/udp_offload.c
382
uh = udp_hdr(seg);
net/ipv4/udp_offload.c
383
iph = ip_hdr(seg);
net/ipv4/udp_offload.c
385
if ((udp_hdr(seg)->dest == udp_hdr(seg->next)->dest) &&
net/ipv4/udp_offload.c
386
(udp_hdr(seg)->source == udp_hdr(seg->next)->source) &&
net/ipv4/udp_offload.c
387
(ip_hdr(seg)->daddr == ip_hdr(seg->next)->daddr) &&
net/ipv4/udp_offload.c
388
(ip_hdr(seg)->saddr == ip_hdr(seg->next)->saddr))
net/ipv4/udp_offload.c
391
while ((seg = seg->next)) {
net/ipv4/udp_offload.c
392
uh2 = udp_hdr(seg);
net/ipv4/udp_offload.c
393
iph2 = ip_hdr(seg);
net/ipv4/udp_offload.c
395
__udpv4_gso_segment_csum(seg,
net/ipv4/udp_offload.c
398
__udpv4_gso_segment_csum(seg,
net/ipv4/udp_offload.c
406
static void __udpv6_gso_segment_csum(struct sk_buff *seg,
net/ipv4/udp_offload.c
411
struct udphdr *uh = udp_hdr(seg);
net/ipv4/udp_offload.c
417
inet_proto_csum_replace16(&uh->check, seg, oldip->s6_addr32,
net/ipv4/udp_offload.c
420
inet_proto_csum_replace2(&uh->check, seg, *oldport, newport,
net/ipv4/udp_offload.c
435
struct sk_buff *seg;
net/ipv4/udp_offload.c
438
seg = segs;
net/ipv4/udp_offload.c
439
uh = udp_hdr(seg);
net/ipv4/udp_offload.c
440
iph = ipv6_hdr(seg);
net/ipv4/udp_offload.c
441
uh2 = udp_hdr(seg->next);
net/ipv4/udp_offload.c
442
iph2 = ipv6_hdr(seg->next);
net/ipv4/udp_offload.c
449
while ((seg = seg->next)) {
net/ipv4/udp_offload.c
450
uh2 = udp_hdr(seg);
net/ipv4/udp_offload.c
451
iph2 = ipv6_hdr(seg);
net/ipv4/udp_offload.c
453
__udpv6_gso_segment_csum(seg, &iph2->saddr, &iph->saddr,
net/ipv4/udp_offload.c
455
__udpv6_gso_segment_csum(seg, &iph2->daddr, &iph->daddr,
net/ipv4/udp_offload.c
485
struct sk_buff *segs, *seg;
net/ipv4/udp_offload.c
568
seg = segs;
net/ipv4/udp_offload.c
569
uh = udp_hdr(seg);
net/ipv4/udp_offload.c
572
skb_shinfo(seg)->tskey = skb_shinfo(gso_skb)->tskey;
net/ipv4/udp_offload.c
573
skb_shinfo(seg)->tx_flags |=
net/ipv4/udp_offload.c
582
seg->destructor = sock_wfree;
net/ipv4/udp_offload.c
583
seg->sk = sk;
net/ipv4/udp_offload.c
584
sum_truesize += seg->truesize;
net/ipv4/udp_offload.c
587
if (!seg->next)
net/ipv4/udp_offload.c
593
if (seg->ip_summed == CHECKSUM_PARTIAL)
net/ipv4/udp_offload.c
594
gso_reset_checksum(seg, ~check);
net/ipv4/udp_offload.c
596
uh->check = gso_make_checksum(seg, ~check) ? :
net/ipv4/udp_offload.c
599
seg = seg->next;
net/ipv4/udp_offload.c
600
uh = udp_hdr(seg);
net/ipv4/udp_offload.c
604
newlen = htons(skb_tail_pointer(seg) - skb_transport_header(seg) +
net/ipv4/udp_offload.c
605
seg->data_len);
net/ipv4/udp_offload.c
611
if (seg->ip_summed == CHECKSUM_PARTIAL)
net/ipv4/udp_offload.c
612
gso_reset_checksum(seg, ~check);
net/ipv4/udp_offload.c
614
uh->check = gso_make_checksum(seg, ~check) ? : CSUM_MANGLED_0;
net/ipv6/tcpv6_offload.c
101
inet_proto_csum_replace16(&th->check, seg,
net/ipv6/tcpv6_offload.c
111
inet_proto_csum_replace2(&th->check, seg, *oldport, newport, false);
net/ipv6/tcpv6_offload.c
119
struct sk_buff *seg;
net/ipv6/tcpv6_offload.c
123
seg = segs;
net/ipv6/tcpv6_offload.c
124
th = tcp_hdr(seg);
net/ipv6/tcpv6_offload.c
125
iph = ipv6_hdr(seg);
net/ipv6/tcpv6_offload.c
126
th2 = tcp_hdr(seg->next);
net/ipv6/tcpv6_offload.c
127
iph2 = ipv6_hdr(seg->next);
net/ipv6/tcpv6_offload.c
134
while ((seg = seg->next)) {
net/ipv6/tcpv6_offload.c
135
th2 = tcp_hdr(seg);
net/ipv6/tcpv6_offload.c
136
iph2 = ipv6_hdr(seg);
net/ipv6/tcpv6_offload.c
138
__tcpv6_gso_segment_csum(seg, &iph2->saddr, &iph->saddr,
net/ipv6/tcpv6_offload.c
140
__tcpv6_gso_segment_csum(seg, &iph2->daddr, &iph->daddr,
net/ipv6/tcpv6_offload.c
93
static void __tcpv6_gso_segment_csum(struct sk_buff *seg,
net/ipv6/tcpv6_offload.c
98
struct tcphdr *th = tcp_hdr(seg);
net/mac80211/tx.c
4619
struct sk_buff *seg, *next;
net/mac80211/tx.c
4671
skb_list_walk_safe(skb, seg, next) {
net/mac80211/tx.c
4673
len += seg->len;
net/mac80211/tx.c
4674
if (seg != skb)
net/mac80211/tx.c
4675
memcpy(IEEE80211_SKB_CB(seg), info, sizeof(*info));
net/openvswitch/flow_table.c
793
int seg;
net/openvswitch/flow_table.c
816
for (seg = 0; seg < MC_HASH_SEGS; seg++) {
net/sunrpc/xprtrdma/frwr_ops.c
286
struct rpcrdma_mr_seg *seg,
net/sunrpc/xprtrdma/frwr_ops.c
299
sg_set_page(&mr->mr_sg[i], seg->mr_page,
net/sunrpc/xprtrdma/frwr_ops.c
300
seg->mr_len, seg->mr_offset);
net/sunrpc/xprtrdma/frwr_ops.c
302
++seg;
net/sunrpc/xprtrdma/frwr_ops.c
306
if ((i < nsegs && seg->mr_offset) ||
net/sunrpc/xprtrdma/frwr_ops.c
307
offset_in_page((seg-1)->mr_offset + (seg-1)->mr_len))
net/sunrpc/xprtrdma/frwr_ops.c
341
return seg;
net/sunrpc/xprtrdma/frwr_ops.c
673
struct rpcrdma_mr_seg seg;
net/sunrpc/xprtrdma/frwr_ops.c
682
seg.mr_len = XDR_UNIT;
net/sunrpc/xprtrdma/frwr_ops.c
683
seg.mr_page = virt_to_page(ep->re_write_pad);
net/sunrpc/xprtrdma/frwr_ops.c
684
seg.mr_offset = offset_in_page(ep->re_write_pad);
net/sunrpc/xprtrdma/frwr_ops.c
685
if (IS_ERR(frwr_map(r_xprt, &seg, 1, true, xdr_zero, mr)))
net/sunrpc/xprtrdma/rpc_rdma.c
209
rpcrdma_convert_kvec(struct kvec *vec, struct rpcrdma_mr_seg *seg,
net/sunrpc/xprtrdma/rpc_rdma.c
212
seg->mr_page = virt_to_page(vec->iov_base);
net/sunrpc/xprtrdma/rpc_rdma.c
213
seg->mr_offset = offset_in_page(vec->iov_base);
net/sunrpc/xprtrdma/rpc_rdma.c
214
seg->mr_len = vec->iov_len;
net/sunrpc/xprtrdma/rpc_rdma.c
215
++seg;
net/sunrpc/xprtrdma/rpc_rdma.c
217
return seg;
net/sunrpc/xprtrdma/rpc_rdma.c
230
struct rpcrdma_mr_seg *seg)
net/sunrpc/xprtrdma/rpc_rdma.c
238
seg = rpcrdma_convert_kvec(&xdrbuf->head[0], seg, &n);
net/sunrpc/xprtrdma/rpc_rdma.c
244
seg->mr_page = *ppages;
net/sunrpc/xprtrdma/rpc_rdma.c
245
seg->mr_offset = page_base;
net/sunrpc/xprtrdma/rpc_rdma.c
246
seg->mr_len = min_t(u32, PAGE_SIZE - page_base, len);
net/sunrpc/xprtrdma/rpc_rdma.c
247
len -= seg->mr_len;
net/sunrpc/xprtrdma/rpc_rdma.c
249
++seg;
net/sunrpc/xprtrdma/rpc_rdma.c
258
rpcrdma_convert_kvec(&xdrbuf->tail[0], seg, &n);
net/sunrpc/xprtrdma/rpc_rdma.c
297
struct rpcrdma_mr_seg *seg,
net/sunrpc/xprtrdma/rpc_rdma.c
310
return frwr_map(r_xprt, seg, nsegs, writing, req->rl_slot.rq_xid, *mr);
net/sunrpc/xprtrdma/rpc_rdma.c
339
struct rpcrdma_mr_seg *seg;
net/sunrpc/xprtrdma/rpc_rdma.c
350
seg = req->rl_segments;
net/sunrpc/xprtrdma/rpc_rdma.c
352
rtype, seg);
net/sunrpc/xprtrdma/rpc_rdma.c
357
seg = rpcrdma_mr_prepare(r_xprt, req, seg, nsegs, false, &mr);
net/sunrpc/xprtrdma/rpc_rdma.c
358
if (IS_ERR(seg))
net/sunrpc/xprtrdma/rpc_rdma.c
359
return PTR_ERR(seg);
net/sunrpc/xprtrdma/rpc_rdma.c
397
struct rpcrdma_mr_seg *seg;
net/sunrpc/xprtrdma/rpc_rdma.c
405
seg = req->rl_segments;
net/sunrpc/xprtrdma/rpc_rdma.c
408
wtype, seg);
net/sunrpc/xprtrdma/rpc_rdma.c
421
seg = rpcrdma_mr_prepare(r_xprt, req, seg, nsegs, true, &mr);
net/sunrpc/xprtrdma/rpc_rdma.c
422
if (IS_ERR(seg))
net/sunrpc/xprtrdma/rpc_rdma.c
423
return PTR_ERR(seg);
net/sunrpc/xprtrdma/rpc_rdma.c
474
struct rpcrdma_mr_seg *seg;
net/sunrpc/xprtrdma/rpc_rdma.c
485
seg = req->rl_segments;
net/sunrpc/xprtrdma/rpc_rdma.c
486
nsegs = rpcrdma_convert_iovs(r_xprt, &rqst->rq_rcv_buf, 0, wtype, seg);
net/sunrpc/xprtrdma/rpc_rdma.c
499
seg = rpcrdma_mr_prepare(r_xprt, req, seg, nsegs, true, &mr);
net/sunrpc/xprtrdma/rpc_rdma.c
500
if (IS_ERR(seg))
net/sunrpc/xprtrdma/rpc_rdma.c
501
return PTR_ERR(seg);
net/sunrpc/xprtrdma/svc_rdma_rw.c
492
const struct svc_rdma_segment *seg;
net/sunrpc/xprtrdma/svc_rdma_rw.c
503
seg = &info->wi_chunk->ch_segments[info->wi_seg_no];
net/sunrpc/xprtrdma/svc_rdma_rw.c
504
write_len = min(remaining, seg->rs_length - info->wi_seg_off);
net/sunrpc/xprtrdma/svc_rdma_rw.c
513
offset = seg->rs_offset + info->wi_seg_off;
net/sunrpc/xprtrdma/svc_rdma_rw.c
514
ret = svc_rdma_rw_ctx_init(rdma, ctxt, offset, seg->rs_handle,
net/sunrpc/xprtrdma/svc_rdma_rw.c
522
if (write_len == seg->rs_length - info->wi_seg_off) {
net/sunrpc/xprtrdma/xprt_rdma.h
539
struct rpcrdma_mr_seg *seg,
sound/core/oss/mulaw.c
72
int seg;
sound/core/oss/mulaw.c
87
seg = val_seg(pcm_val);
sound/core/oss/mulaw.c
93
uval = (seg << 4) | ((pcm_val >> (seg + 3)) & 0xF);
sound/pci/cs46xx/dsp_spos.c
1903
struct dsp_segment_desc *seg;
sound/pci/cs46xx/dsp_spos.c
1906
seg = get_segment_desc(module, SEGTYPE_SP_PARAMETER);
sound/pci/cs46xx/dsp_spos.c
1907
err = dsp_load_parameter(chip, seg);
sound/pci/cs46xx/dsp_spos.c
1911
seg = get_segment_desc(module, SEGTYPE_SP_SAMPLE);
sound/pci/cs46xx/dsp_spos.c
1912
err = dsp_load_sample(chip, seg);
sound/pci/cs46xx/dsp_spos.c
1916
seg = get_segment_desc(module, SEGTYPE_SP_PROGRAM);
sound/pci/cs46xx/dsp_spos.c
1917
if (!seg)
sound/pci/cs46xx/dsp_spos.c
1920
doffset = seg->offset * 4 + module->load_address * 4
sound/pci/cs46xx/dsp_spos.c
1922
dsize = seg->size * 4;
tools/bpf/bpftool/gen.c
2408
const struct btf_ext_info *seg;
tools/bpf/bpftool/gen.c
2442
seg = &btf_ext->core_relo_info;
tools/bpf/bpftool/gen.c
2443
for_each_btf_ext_sec(seg, sec) {
tools/bpf/bpftool/gen.c
2444
for_each_btf_ext_rec(seg, sec, relo_idx, relo) {
tools/lib/bpf/btf.c
5553
const struct btf_ext_info *seg;
tools/lib/bpf/btf.c
5557
seg = &btf_ext->func_info;
tools/lib/bpf/btf.c
5558
for_each_btf_ext_sec(seg, sec) {
tools/lib/bpf/btf.c
5561
for_each_btf_ext_rec(seg, sec, i, rec) {
tools/lib/bpf/btf.c
5568
seg = &btf_ext->core_relo_info;
tools/lib/bpf/btf.c
5569
for_each_btf_ext_sec(seg, sec) {
tools/lib/bpf/btf.c
5572
for_each_btf_ext_rec(seg, sec, i, rec) {
tools/lib/bpf/btf.c
5584
const struct btf_ext_info *seg;
tools/lib/bpf/btf.c
5588
seg = &btf_ext->func_info;
tools/lib/bpf/btf.c
5589
for_each_btf_ext_sec(seg, sec) {
tools/lib/bpf/btf.c
5595
seg = &btf_ext->line_info;
tools/lib/bpf/btf.c
5596
for_each_btf_ext_sec(seg, sec) {
tools/lib/bpf/btf.c
5603
for_each_btf_ext_rec(seg, sec, i, rec) {
tools/lib/bpf/btf.c
5613
seg = &btf_ext->core_relo_info;
tools/lib/bpf/btf.c
5614
for_each_btf_ext_sec(seg, sec) {
tools/lib/bpf/btf.c
5621
for_each_btf_ext_rec(seg, sec, i, rec) {
tools/lib/bpf/libbpf.c
3307
struct btf_ext_info *seg = ext_segs[seg_num];
tools/lib/bpf/libbpf.c
3312
if (seg->sec_cnt == 0)
tools/lib/bpf/libbpf.c
3315
seg->sec_idxs = calloc(seg->sec_cnt, sizeof(*seg->sec_idxs));
tools/lib/bpf/libbpf.c
3316
if (!seg->sec_idxs) {
tools/lib/bpf/libbpf.c
3322
for_each_btf_ext_sec(seg, sec) {
tools/lib/bpf/libbpf.c
3335
seg->sec_idxs[sec_num - 1] = elf_ndxscn(scn);
tools/lib/bpf/libbpf.c
6032
const struct btf_ext_info *seg;
tools/lib/bpf/libbpf.c
6058
seg = &obj->btf_ext->core_relo_info;
tools/lib/bpf/libbpf.c
6060
for_each_btf_ext_sec(seg, sec) {
tools/lib/bpf/libbpf.c
6061
sec_idx = seg->sec_idxs[sec_num];
tools/lib/bpf/libbpf.c
6072
for_each_btf_ext_rec(seg, sec, i, rec) {
tools/lib/bpf/libbpf_internal.h
449
#define for_each_btf_ext_sec(seg, sec) \
tools/lib/bpf/libbpf_internal.h
450
for (sec = (seg)->info; \
tools/lib/bpf/libbpf_internal.h
451
(void *)sec < (seg)->info + (seg)->len; \
tools/lib/bpf/libbpf_internal.h
453
(seg)->rec_size * sec->num_info)
tools/lib/bpf/libbpf_internal.h
455
#define for_each_btf_ext_rec(seg, sec, i, rec) \
tools/lib/bpf/libbpf_internal.h
458
i++, rec = (void *)rec + (seg)->rec_size)
tools/lib/bpf/usdt.c
416
struct elf_seg *seg;
tools/lib/bpf/usdt.c
443
seg = *segs + *seg_cnt;
tools/lib/bpf/usdt.c
446
seg->start = phdr.p_vaddr;
tools/lib/bpf/usdt.c
447
seg->end = phdr.p_vaddr + phdr.p_memsz;
tools/lib/bpf/usdt.c
448
seg->offset = phdr.p_offset;
tools/lib/bpf/usdt.c
449
seg->is_exec = phdr.p_flags & PF_X;
tools/lib/bpf/usdt.c
465
struct elf_seg *seg;
tools/lib/bpf/usdt.c
528
seg = *segs + *seg_cnt;
tools/lib/bpf/usdt.c
531
seg->start = seg_start;
tools/lib/bpf/usdt.c
532
seg->end = seg_end;
tools/lib/bpf/usdt.c
533
seg->offset = seg_off;
tools/lib/bpf/usdt.c
534
seg->is_exec = true;
tools/lib/bpf/usdt.c
553
struct elf_seg *seg;
tools/lib/bpf/usdt.c
560
for (i = 0, seg = segs; i < seg_cnt; i++, seg++) {
tools/lib/bpf/usdt.c
561
if (seg->start <= virtaddr && virtaddr < seg->end)
tools/lib/bpf/usdt.c
562
return seg;
tools/lib/bpf/usdt.c
569
struct elf_seg *seg;
tools/lib/bpf/usdt.c
576
for (i = 0, seg = segs; i < seg_cnt; i++, seg++) {
tools/lib/bpf/usdt.c
577
if (seg->offset <= offset && offset < seg->offset + (seg->end - seg->start))
tools/lib/bpf/usdt.c
578
return seg;
tools/lib/bpf/usdt.c
635
struct elf_seg *seg = NULL;
tools/lib/bpf/usdt.c
681
seg = find_elf_seg(segs, seg_cnt, usdt_abs_ip);
tools/lib/bpf/usdt.c
682
if (!seg) {
tools/lib/bpf/usdt.c
688
if (!seg->is_exec) {
tools/lib/bpf/usdt.c
691
path, seg->start, seg->end, usdt_provider, usdt_name,
tools/lib/bpf/usdt.c
696
usdt_rel_ip = usdt_abs_ip - seg->start + seg->offset;
tools/lib/bpf/usdt.c
728
seg = find_vma_seg(vma_segs, vma_seg_cnt, usdt_rel_ip);
tools/lib/bpf/usdt.c
729
if (!seg) {
tools/lib/bpf/usdt.c
736
usdt_abs_ip = seg->start - seg->offset + usdt_rel_ip;
tools/lib/bpf/usdt.c
742
seg ? seg->start : 0, seg ? seg->end : 0, seg ? seg->offset : 0);
tools/lib/bpf/usdt.c
753
seg = find_elf_seg(segs, seg_cnt, note.sema_addr);
tools/lib/bpf/usdt.c
754
if (!seg) {
tools/lib/bpf/usdt.c
760
if (seg->is_exec) {
tools/lib/bpf/usdt.c
763
path, seg->start, seg->end, usdt_provider, usdt_name,
tools/lib/bpf/usdt.c
768
usdt_sema_off = note.sema_addr - seg->start + seg->offset;
tools/lib/bpf/usdt.c
773
seg->start, seg->end, seg->offset);
tools/testing/selftests/bpf/progs/test_lwt_seg6local.c
291
struct ip6_addr_t *seg;
tools/testing/selftests/bpf/progs/test_lwt_seg6local.c
305
seg = (struct ip6_addr_t *)((char *)srh + sizeof(*srh));
tools/testing/selftests/bpf/progs/test_lwt_seg6local.c
309
seg->lo = bpf_cpu_to_be64(4 - lo);
tools/testing/selftests/bpf/progs/test_lwt_seg6local.c
310
seg->hi = bpf_cpu_to_be64(hi);
tools/testing/selftests/bpf/progs/test_lwt_seg6local.c
311
seg = (struct ip6_addr_t *)((char *)seg + sizeof(*seg));
tools/testing/selftests/kvm/lib/x86/processor.c
744
struct kvm_segment seg;
tools/testing/selftests/kvm/lib/x86/processor.c
758
kvm_seg_set_kernel_code_64bit(&seg);
tools/testing/selftests/kvm/lib/x86/processor.c
759
kvm_seg_fill_gdt_64bit(vm, &seg);
tools/testing/selftests/kvm/lib/x86/processor.c
761
kvm_seg_set_kernel_data_64bit(&seg);
tools/testing/selftests/kvm/lib/x86/processor.c
762
kvm_seg_fill_gdt_64bit(vm, &seg);
tools/testing/selftests/kvm/lib/x86/processor.c
764
kvm_seg_set_tss_64bit(vm->arch.tss, &seg);
tools/testing/selftests/kvm/lib/x86/processor.c
765
kvm_seg_fill_gdt_64bit(vm, &seg);
tools/testing/selftests/kvm/lib/x86/svm.c
56
static void vmcb_set_seg(struct vmcb_seg *seg, u16 selector,
tools/testing/selftests/kvm/lib/x86/svm.c
59
seg->selector = selector;
tools/testing/selftests/kvm/lib/x86/svm.c
60
seg->attrib = attr;
tools/testing/selftests/kvm/lib/x86/svm.c
61
seg->limit = limit;
tools/testing/selftests/kvm/lib/x86/svm.c
62
seg->base = base;
tools/testing/selftests/sgx/load.c
106
static bool encl_ioc_add_pages(struct encl *encl, struct encl_segment *seg)
tools/testing/selftests/sgx/load.c
113
secinfo.flags = seg->flags;
tools/testing/selftests/sgx/load.c
115
ioc.src = (uint64_t)seg->src;
tools/testing/selftests/sgx/load.c
116
ioc.offset = seg->offset;
tools/testing/selftests/sgx/load.c
117
ioc.length = seg->size;
tools/testing/selftests/sgx/load.c
119
if (seg->measure)
tools/testing/selftests/sgx/load.c
180
struct encl_segment *seg;
tools/testing/selftests/sgx/load.c
253
seg = &encl->segment_tbl[j];
tools/testing/selftests/sgx/load.c
273
seg->prot = PROT_READ | PROT_WRITE;
tools/testing/selftests/sgx/load.c
274
seg->flags = SGX_PAGE_TYPE_TCS << 8;
tools/testing/selftests/sgx/load.c
276
seg->prot = (phdr->p_flags & PF_R) ? PROT_READ : 0;
tools/testing/selftests/sgx/load.c
277
seg->prot |= (phdr->p_flags & PF_W) ? PROT_WRITE : 0;
tools/testing/selftests/sgx/load.c
278
seg->prot |= (phdr->p_flags & PF_X) ? PROT_EXEC : 0;
tools/testing/selftests/sgx/load.c
279
seg->flags = (SGX_PAGE_TYPE_REG << 8) | seg->prot;
tools/testing/selftests/sgx/load.c
282
seg->offset = (phdr->p_offset & PAGE_MASK) - src_offset;
tools/testing/selftests/sgx/load.c
283
seg->size = (phdr->p_filesz + PAGE_SIZE - 1) & PAGE_MASK;
tools/testing/selftests/sgx/load.c
284
seg->src = encl->src + seg->offset;
tools/testing/selftests/sgx/load.c
285
seg->measure = true;
tools/testing/selftests/sgx/load.c
292
seg = &encl->segment_tbl[j];
tools/testing/selftests/sgx/load.c
293
seg->offset = encl->segment_tbl[j - 1].offset + encl->segment_tbl[j - 1].size;
tools/testing/selftests/sgx/load.c
294
seg->size = heap_size;
tools/testing/selftests/sgx/load.c
295
seg->src = mmap(NULL, heap_size, PROT_READ | PROT_WRITE,
tools/testing/selftests/sgx/load.c
297
seg->prot = PROT_READ | PROT_WRITE;
tools/testing/selftests/sgx/load.c
298
seg->flags = (SGX_PAGE_TYPE_REG << 8) | seg->prot;
tools/testing/selftests/sgx/load.c
299
seg->measure = false;
tools/testing/selftests/sgx/load.c
301
if (seg->src == MAP_FAILED)
tools/testing/selftests/sgx/load.c
356
struct encl_segment *seg = &encl->segment_tbl[i];
tools/testing/selftests/sgx/load.c
358
if (!encl_ioc_add_pages(encl, seg))
tools/testing/selftests/sgx/main.c
1125
struct encl_segment *seg = &self->encl.segment_tbl[i];
tools/testing/selftests/sgx/main.c
1127
total_size += seg->size;
tools/testing/selftests/sgx/main.c
1274
struct encl_segment *seg = &self->encl.segment_tbl[i];
tools/testing/selftests/sgx/main.c
1276
total_size += seg->size;
tools/testing/selftests/sgx/main.c
142
struct encl_segment *seg = &encl->segment_tbl[i];
tools/testing/selftests/sgx/main.c
144
if (i == 0 && seg->prot == (PROT_READ | PROT_WRITE))
tools/testing/selftests/sgx/main.c
145
return seg->offset;
tools/testing/selftests/sgx/main.c
161
struct encl_segment *seg = &encl->segment_tbl[i];
tools/testing/selftests/sgx/main.c
163
if (seg->prot == (PROT_READ | PROT_WRITE))
tools/testing/selftests/sgx/main.c
164
return seg->offset;
tools/testing/selftests/sgx/main.c
180
struct encl_segment *seg;
tools/testing/selftests/sgx/main.c
202
struct encl_segment *seg = &encl->segment_tbl[i];
tools/testing/selftests/sgx/main.c
204
addr = mmap((void *)encl->encl_base + seg->offset, seg->size,
tools/testing/selftests/sgx/main.c
205
seg->prot, MAP_SHARED | MAP_FIXED, encl->fd, 0);
tools/testing/selftests/sgx/main.c
229
seg = &encl->segment_tbl[i];
tools/testing/selftests/sgx/main.c
231
TH_LOG("0x%016lx 0x%016lx 0x%02x", seg->offset, seg->size, seg->prot);
tools/testing/selftests/sgx/main.c
995
struct encl_segment *seg = &self->encl.segment_tbl[i];
tools/testing/selftests/sgx/main.c
997
total_size += seg->size;
tools/testing/selftests/sgx/sigstruct.c
296
struct encl_segment *seg)
tools/testing/selftests/sgx/sigstruct.c
298
uint64_t end = seg->size;
tools/testing/selftests/sgx/sigstruct.c
302
if (!mrenclave_eadd(ctx, seg->offset + offset, seg->flags))
tools/testing/selftests/sgx/sigstruct.c
305
if (seg->measure) {
tools/testing/selftests/sgx/sigstruct.c
306
if (!mrenclave_eextend(ctx, seg->offset + offset, seg->src + offset))
tools/testing/selftests/sgx/sigstruct.c
356
struct encl_segment *seg = &encl->segment_tbl[i];
tools/testing/selftests/sgx/sigstruct.c
358
if (!mrenclave_segment(ctx, encl, seg))
tools/testing/selftests/ublk/null.c
33
.seg = {
virt/kvm/kvm_main.c
3227
int seg;
virt/kvm/kvm_main.c
3231
while ((seg = next_segment(len, offset)) != 0) {
virt/kvm/kvm_main.c
3232
ret = kvm_read_guest_page(kvm, gfn, data, offset, seg);
virt/kvm/kvm_main.c
3236
len -= seg;
virt/kvm/kvm_main.c
3237
data += seg;
virt/kvm/kvm_main.c
3247
int seg;
virt/kvm/kvm_main.c
3251
while ((seg = next_segment(len, offset)) != 0) {
virt/kvm/kvm_main.c
3252
ret = kvm_vcpu_read_guest_page(vcpu, gfn, data, offset, seg);
virt/kvm/kvm_main.c
3256
len -= seg;
virt/kvm/kvm_main.c
3257
data += seg;
virt/kvm/kvm_main.c
3338
int seg;
virt/kvm/kvm_main.c
3342
while ((seg = next_segment(len, offset)) != 0) {
virt/kvm/kvm_main.c
3343
ret = kvm_write_guest_page(kvm, gfn, data, offset, seg);
virt/kvm/kvm_main.c
3347
len -= seg;
virt/kvm/kvm_main.c
3348
data += seg;
virt/kvm/kvm_main.c
3359
int seg;
virt/kvm/kvm_main.c
3363
while ((seg = next_segment(len, offset)) != 0) {
virt/kvm/kvm_main.c
3364
ret = kvm_vcpu_write_guest_page(vcpu, gfn, data, offset, seg);
virt/kvm/kvm_main.c
3368
len -= seg;
virt/kvm/kvm_main.c
3369
data += seg;
virt/kvm/kvm_main.c
3504
int seg;
virt/kvm/kvm_main.c
3508
while ((seg = next_segment(len, offset)) != 0) {
virt/kvm/kvm_main.c
3509
ret = kvm_write_guest_page(kvm, gfn, zero_page, offset, seg);
virt/kvm/kvm_main.c
3513
len -= seg;