arch/alpha/kernel/perf_event.c
340
static int collect_events(struct perf_event *group, int max_count,
arch/alpha/kernel/perf_event.c
347
if (!is_software_event(group)) {
arch/alpha/kernel/perf_event.c
350
event[n] = group;
arch/alpha/kernel/perf_event.c
351
evtype[n] = group->hw.event_base;
arch/alpha/kernel/perf_event.c
354
for_each_sibling_event(pe, group) {
arch/arm/kernel/module.c
223
group = 0;
arch/arm/kernel/module.c
238
shift = get_group_rem(group, &offset);
arch/arm/kernel/module.c
60
static u32 get_group_rem(u32 group, u32 *offset)
arch/arm/kernel/module.c
70
} while (group--);
arch/arm/kernel/module.c
90
u32 shift, group = 1;
arch/arm/mach-s3c/gpio-core.h
73
int group;
arch/arm/mach-s3c/irqs-s3c64xx.h
158
#define IRQ_EINT_GROUP(group, no) (IRQ_EINT_GROUP##group##_BASE + (no))
arch/arm64/kvm/arm.c
1694
switch (attr->group) {
arch/arm64/kvm/arm.c
1708
switch (attr->group) {
arch/arm64/kvm/arm.c
1722
switch (attr->group) {
arch/arm64/kvm/arm.c
1918
switch (attr->group) {
arch/arm64/kvm/arm.c
1928
switch (attr->group) {
arch/arm64/kvm/guest.c
909
switch (attr->group) {
arch/arm64/kvm/guest.c
934
switch (attr->group) {
arch/arm64/kvm/guest.c
957
switch (attr->group) {
arch/arm64/kvm/vgic/vgic-debug.c
226
irq->group,
arch/arm64/kvm/vgic/vgic-init.c
226
irq->group = 0;
arch/arm64/kvm/vgic/vgic-init.c
230
irq->group = 1;
arch/arm64/kvm/vgic/vgic-init.c
303
irq->group = 1;
arch/arm64/kvm/vgic/vgic-init.c
307
irq->group = 0;
arch/arm64/kvm/vgic/vgic-its.c
105
irq->group = 1;
arch/arm64/kvm/vgic/vgic-its.c
2641
switch (attr->group) {
arch/arm64/kvm/vgic/vgic-its.c
2729
switch (attr->group) {
arch/arm64/kvm/vgic/vgic-its.c
2771
switch (attr->group) {
arch/arm64/kvm/vgic/vgic-kvm-device.c
215
switch (attr->group) {
arch/arm64/kvm/vgic/vgic-kvm-device.c
296
switch (attr->group) {
arch/arm64/kvm/vgic/vgic-kvm-device.c
399
switch (attr->group) {
arch/arm64/kvm/vgic/vgic-kvm-device.c
425
switch (attr->group) {
arch/arm64/kvm/vgic/vgic-kvm-device.c
437
switch (attr->group) {
arch/arm64/kvm/vgic/vgic-kvm-device.c
449
switch (attr->group) {
arch/arm64/kvm/vgic/vgic-kvm-device.c
489
if (attr->group != KVM_DEV_ARM_VGIC_GRP_DIST_REGS) {
arch/arm64/kvm/vgic/vgic-kvm-device.c
513
if (attr->group != KVM_DEV_ARM_VGIC_GRP_DIST_REGS)
arch/arm64/kvm/vgic/vgic-kvm-device.c
550
switch (attr->group) {
arch/arm64/kvm/vgic/vgic-kvm-device.c
579
switch (attr->group) {
arch/arm64/kvm/vgic/vgic-kvm-device.c
625
switch (attr->group) {
arch/arm64/kvm/vgic/vgic-kvm-device.c
656
switch (attr->group) {
arch/arm64/kvm/vgic/vgic-kvm-device.c
676
switch (attr->group) {
arch/arm64/kvm/vgic/vgic-mmio-v2.c
537
switch (attr->group) {
arch/arm64/kvm/vgic/vgic-mmio-v3.c
1015
switch (attr->group) {
arch/arm64/kvm/vgic/vgic-mmio-v3.c
1065
if (!irq->group || allow_group1) {
arch/arm64/kvm/vgic/vgic-mmio.c
55
if (irq->group)
arch/arm64/kvm/vgic/vgic-mmio.c
66
WARN_ON(its_prop_update_vsgi(irq->host_irq, irq->priority, irq->group));
arch/arm64/kvm/vgic/vgic-mmio.c
80
irq->group = !!(val & BIT(i));
arch/arm64/kvm/vgic/vgic-v2.c
241
if (irq->group)
arch/arm64/kvm/vgic/vgic-v3.c
358
if (irq->group)
arch/arm64/kvm/vgic/vgic-v4.c
111
vpe->sgi_config[irq->intid].group = irq->group;
arch/arm64/kvm/vgic/vgic.c
319
ret = (int)(irqb->group ? info->vmcr.grpen1 : info->vmcr.grpen0);
arch/arm64/kvm/vgic/vgic.c
320
ret -= (int)(irqa->group ? info->vmcr.grpen1 : info->vmcr.grpen0);
arch/loongarch/kvm/intc/eiointc.c
567
switch (attr->group) {
arch/loongarch/kvm/intc/eiointc.c
596
switch (attr->group) {
arch/loongarch/kvm/intc/ipi.c
373
switch (attr->group) {
arch/loongarch/kvm/intc/ipi.c
377
kvm_err("%s: unknown group (%d)\n", __func__, attr->group);
arch/loongarch/kvm/intc/ipi.c
385
switch (attr->group) {
arch/loongarch/kvm/intc/ipi.c
389
kvm_err("%s: unknown group (%d)\n", __func__, attr->group);
arch/loongarch/kvm/intc/pch_pic.c
360
switch (attr->group) {
arch/loongarch/kvm/intc/pch_pic.c
374
switch (attr->group) {
arch/loongarch/kvm/intc/pch_pic.c
388
kvm_err("%s: unknown group (%d) attr (%lld)\n", __func__, attr->group,
arch/loongarch/kvm/vcpu.c
1078
switch (attr->group) {
arch/loongarch/kvm/vcpu.c
1139
switch (attr->group) {
arch/loongarch/kvm/vcpu.c
1222
switch (attr->group) {
arch/loongarch/kvm/vm.c
171
switch (attr->group) {
arch/mips/include/asm/octeon/cvmx-pow.h
1643
uint64_t group)
arch/mips/include/asm/octeon/cvmx-pow.h
1680
tag_req.s.grp = group;
arch/mips/include/asm/octeon/cvmx-pow.h
1717
uint64_t group)
arch/mips/include/asm/octeon/cvmx-pow.h
1728
cvmx_pow_tag_sw_full_nocheck(wqp, tag, tag_type, group);
arch/mips/include/asm/octeon/cvmx-pow.h
1955
uint64_t group,
arch/mips/include/asm/octeon/cvmx-pow.h
1980
tag_req.s.grp = group;
arch/mips/include/asm/octeon/cvmx-pow.h
2037
uint64_t group, uint64_t no_sched)
arch/mips/include/asm/octeon/cvmx-pow.h
2050
cvmx_pow_tag_sw_desched_nocheck(tag, tag_type, group, no_sched);
arch/mips/rb532/irq.c
114
unsigned int group, intr_bit, irq_nr = d->irq;
arch/mips/rb532/irq.c
121
group = ip >> 5;
arch/mips/rb532/irq.c
126
enable_local_irq(group_to_ip(group));
arch/mips/rb532/irq.c
128
addr = intr_group[group].base_addr;
arch/mips/rb532/irq.c
135
unsigned int group, intr_bit, mask, irq_nr = d->irq;
arch/mips/rb532/irq.c
142
group = ip >> 5;
arch/mips/rb532/irq.c
146
addr = intr_group[group].base_addr;
arch/mips/rb532/irq.c
152
if (group == GPIO_MAPPED_IRQ_GROUP && irq_nr <= (GROUP4_IRQ_BASE + 13))
arch/mips/rb532/irq.c
159
if (mask == intr_group[group].mask)
arch/mips/rb532/irq.c
160
disable_local_irq(group_to_ip(group));
arch/mips/rb532/irq.c
173
int group = irq_to_group(d->irq);
arch/mips/rb532/irq.c
175
if (group != GPIO_MAPPED_IRQ_GROUP || d->irq > (GROUP4_IRQ_BASE + 13))
arch/mips/rb532/irq.c
215
unsigned int ip, pend, group;
arch/mips/rb532/irq.c
224
group = 21 + (fls(ip) - 32);
arch/mips/rb532/irq.c
226
addr = intr_group[group].base_addr;
arch/mips/rb532/irq.c
231
do_IRQ((group << 5) + pend);
arch/mips/rb532/irq.c
86
static inline int group_to_ip(unsigned int group)
arch/mips/rb532/irq.c
88
return group + 2;
arch/powerpc/include/asm/eeh.h
306
struct eeh_pe *eeh_iommu_group_to_pe(struct iommu_group *group);
arch/powerpc/include/asm/iommu.h
205
struct iommu_group *group;
arch/powerpc/kernel/eeh.c
1277
struct eeh_pe *eeh_iommu_group_to_pe(struct iommu_group *group)
arch/powerpc/kernel/eeh.c
1284
if (!group)
arch/powerpc/kernel/eeh.c
1287
ret = iommu_group_for_each_dev(group, &pdev, dev_has_iommu_table);
arch/powerpc/kernel/iommu.c
1016
table_group->group = NULL;
arch/powerpc/kernel/iommu.c
1031
table_group->group = grp;
arch/powerpc/kernel/iommu.c
1141
__func__, dev_name(dev), iommu_group_id(table_group->group));
arch/powerpc/kvm/book3s_xics.c
1298
switch (attr->group) {
arch/powerpc/kvm/book3s_xics.c
1309
switch (attr->group) {
arch/powerpc/kvm/book3s_xics.c
1318
switch (attr->group) {
arch/powerpc/kvm/book3s_xive.c
2566
switch (attr->group) {
arch/powerpc/kvm/book3s_xive.c
2583
switch (attr->group) {
arch/powerpc/kvm/book3s_xive.c
2593
switch (attr->group) {
arch/powerpc/kvm/book3s_xive_native.c
946
switch (attr->group) {
arch/powerpc/kvm/book3s_xive_native.c
978
switch (attr->group) {
arch/powerpc/kvm/book3s_xive_native.c
989
switch (attr->group) {
arch/powerpc/kvm/mpic.c
1522
switch (attr->group) {
arch/powerpc/kvm/mpic.c
1563
switch (attr->group) {
arch/powerpc/kvm/mpic.c
1609
switch (attr->group) {
arch/powerpc/perf/core-book3s.c
1590
static int collect_events(struct perf_event *group, int max_count,
arch/powerpc/perf/core-book3s.c
1597
if (group->pmu->task_ctx_nr == perf_hw_context) {
arch/powerpc/perf/core-book3s.c
1600
ctrs[n] = group;
arch/powerpc/perf/core-book3s.c
1601
flags[n] = group->hw.event_base;
arch/powerpc/perf/core-book3s.c
1602
events[n++] = group->hw.config;
arch/powerpc/perf/core-book3s.c
1604
for_each_sibling_event(event, group) {
arch/powerpc/perf/core-fsl-emb.c
251
static int collect_events(struct perf_event *group, int max_count,
arch/powerpc/perf/core-fsl-emb.c
257
if (!is_software_event(group)) {
arch/powerpc/perf/core-fsl-emb.c
260
ctrs[n] = group;
arch/powerpc/perf/core-fsl-emb.c
263
for_each_sibling_event(event, group) {
arch/powerpc/platforms/powernv/pci-ioda.c
1043
if (pe->table_group.group)
arch/powerpc/platforms/powernv/pci-ioda.c
2301
if (pe->table_group.group) {
arch/powerpc/platforms/powernv/pci-ioda.c
2302
iommu_group_put(pe->table_group.group);
arch/powerpc/platforms/powernv/pci-ioda.c
2303
WARN_ON(pe->table_group.group);
arch/powerpc/platforms/powernv/pci-ioda.c
2466
if (!pe->table_group.group)
arch/powerpc/platforms/powernv/pci-ioda.c
2469
return iommu_group_ref_get(pe->table_group.group);
arch/powerpc/platforms/pseries/iommu.c
104
if (table_group->group) {
arch/powerpc/platforms/pseries/iommu.c
105
iommu_group_put(table_group->group);
arch/powerpc/platforms/pseries/iommu.c
106
BUG_ON(table_group->group);
arch/powerpc/platforms/pseries/iommu.c
1945
static struct pci_dev *iommu_group_get_first_pci_dev(struct iommu_group *group)
arch/powerpc/platforms/pseries/iommu.c
1951
if (!group)
arch/powerpc/platforms/pseries/iommu.c
1954
ret = iommu_group_for_each_dev(group, &pdev, dev_has_iommu_table);
arch/powerpc/platforms/pseries/iommu.c
2060
struct pci_dev *pdev = iommu_group_get_first_pci_dev(table_group->group);
arch/powerpc/platforms/pseries/iommu.c
2275
struct pci_dev *pdev = iommu_group_get_first_pci_dev(table_group->group);
arch/powerpc/platforms/pseries/iommu.c
2568
grp = pci->table_group->group;
arch/riscv/kvm/aia_device.c
214
u32 hart = 0, group = 0;
arch/riscv/kvm/aia_device.c
220
group = (addr >> aia->nr_group_shift) &
arch/riscv/kvm/aia_device.c
223
return (group << aia->nr_hart_bits) | hart;
arch/riscv/kvm/aia_device.c
308
switch (attr->group) {
arch/riscv/kvm/aia_device.c
375
switch (attr->group) {
arch/riscv/kvm/aia_device.c
446
switch (attr->group) {
arch/s390/include/uapi/asm/dasd.h
280
__u8 group:2;
arch/s390/kvm/interrupt.c
2309
switch (attr->group) {
arch/s390/kvm/interrupt.c
2625
switch (attr->group) {
arch/s390/kvm/interrupt.c
2678
switch (attr->group) {
arch/s390/kvm/kvm-s390.c
1958
switch (attr->group) {
arch/s390/kvm/kvm-s390.c
1989
switch (attr->group) {
arch/s390/kvm/kvm-s390.c
2017
switch (attr->group) {
arch/s390/tools/gen_opcode_table.c
263
struct insn_group *group;
arch/s390/tools/gen_opcode_table.c
265
group = desc->group ? &desc->group[desc->nr_groups - 1] : NULL;
arch/s390/tools/gen_opcode_table.c
266
if (group && (!strncmp(group->opcode, insn->opcode, 2) || group->type->byte == 0)) {
arch/s390/tools/gen_opcode_table.c
267
group->count++;
arch/s390/tools/gen_opcode_table.c
271
desc->group = realloc(desc->group, desc->nr_groups * sizeof(*desc->group));
arch/s390/tools/gen_opcode_table.c
272
if (!desc->group)
arch/s390/tools/gen_opcode_table.c
274
group = &desc->group[desc->nr_groups - 1];
arch/s390/tools/gen_opcode_table.c
275
memcpy(group->opcode, insn->opcode, 2);
arch/s390/tools/gen_opcode_table.c
276
group->type = insn->type;
arch/s390/tools/gen_opcode_table.c
277
group->offset = offset;
arch/s390/tools/gen_opcode_table.c
278
group->count = 1;
arch/s390/tools/gen_opcode_table.c
321
struct insn_group *group;
arch/s390/tools/gen_opcode_table.c
326
group = &desc->group[i];
arch/s390/tools/gen_opcode_table.c
328
group->opcode, group->type->mask, group->type->byte, group->offset, group->count);
arch/s390/tools/gen_opcode_table.c
46
struct insn_group *group;
arch/sparc/include/asm/hypervisor.h
3510
unsigned long sun4v_get_version(unsigned long group,
arch/sparc/include/asm/hypervisor.h
3513
unsigned long sun4v_set_version(unsigned long group,
arch/sparc/include/asm/hypervisor.h
3518
int sun4v_hvapi_register(unsigned long group, unsigned long major,
arch/sparc/include/asm/hypervisor.h
3520
void sun4v_hvapi_unregister(unsigned long group);
arch/sparc/include/asm/hypervisor.h
3521
int sun4v_hvapi_get(unsigned long group,
arch/sparc/kernel/hvapi.c
104
p = __get_info(group);
arch/sparc/kernel/hvapi.c
117
hv_ret = sun4v_set_version(group, major, *minor,
arch/sparc/kernel/hvapi.c
147
void sun4v_hvapi_unregister(unsigned long group)
arch/sparc/kernel/hvapi.c
153
p = __get_info(group);
arch/sparc/kernel/hvapi.c
160
int sun4v_hvapi_get(unsigned long group,
arch/sparc/kernel/hvapi.c
170
p = __get_info(group);
arch/sparc/kernel/hvapi.c
184
unsigned long group, major, minor;
arch/sparc/kernel/hvapi.c
186
group = HV_GRP_SUN4V;
arch/sparc/kernel/hvapi.c
189
if (sun4v_hvapi_register(group, major, &minor))
arch/sparc/kernel/hvapi.c
19
unsigned long group;
arch/sparc/kernel/hvapi.c
192
group = HV_GRP_CORE;
arch/sparc/kernel/hvapi.c
195
if (sun4v_hvapi_register(group, major, &minor))
arch/sparc/kernel/hvapi.c
203
group, major, minor);
arch/sparc/kernel/hvapi.c
28
{ .group = HV_GRP_SUN4V, .flags = FLAG_PRE_API },
arch/sparc/kernel/hvapi.c
29
{ .group = HV_GRP_CORE, .flags = FLAG_PRE_API },
arch/sparc/kernel/hvapi.c
30
{ .group = HV_GRP_INTR, },
arch/sparc/kernel/hvapi.c
31
{ .group = HV_GRP_SOFT_STATE, },
arch/sparc/kernel/hvapi.c
32
{ .group = HV_GRP_TM, },
arch/sparc/kernel/hvapi.c
33
{ .group = HV_GRP_PCI, .flags = FLAG_PRE_API },
arch/sparc/kernel/hvapi.c
34
{ .group = HV_GRP_LDOM, },
arch/sparc/kernel/hvapi.c
35
{ .group = HV_GRP_SVC_CHAN, .flags = FLAG_PRE_API },
arch/sparc/kernel/hvapi.c
36
{ .group = HV_GRP_NCS, .flags = FLAG_PRE_API },
arch/sparc/kernel/hvapi.c
37
{ .group = HV_GRP_RNG, },
arch/sparc/kernel/hvapi.c
38
{ .group = HV_GRP_PBOOT, },
arch/sparc/kernel/hvapi.c
39
{ .group = HV_GRP_TPM, },
arch/sparc/kernel/hvapi.c
40
{ .group = HV_GRP_SDIO, },
arch/sparc/kernel/hvapi.c
41
{ .group = HV_GRP_SDIO_ERR, },
arch/sparc/kernel/hvapi.c
42
{ .group = HV_GRP_REBOOT_DATA, },
arch/sparc/kernel/hvapi.c
43
{ .group = HV_GRP_ATU, .flags = FLAG_PRE_API },
arch/sparc/kernel/hvapi.c
44
{ .group = HV_GRP_DAX, },
arch/sparc/kernel/hvapi.c
45
{ .group = HV_GRP_NIAG_PERF, .flags = FLAG_PRE_API },
arch/sparc/kernel/hvapi.c
46
{ .group = HV_GRP_FIRE_PERF, },
arch/sparc/kernel/hvapi.c
47
{ .group = HV_GRP_N2_CPU, },
arch/sparc/kernel/hvapi.c
48
{ .group = HV_GRP_NIU, },
arch/sparc/kernel/hvapi.c
49
{ .group = HV_GRP_VF_CPU, },
arch/sparc/kernel/hvapi.c
50
{ .group = HV_GRP_KT_CPU, },
arch/sparc/kernel/hvapi.c
51
{ .group = HV_GRP_VT_CPU, },
arch/sparc/kernel/hvapi.c
52
{ .group = HV_GRP_T5_CPU, },
arch/sparc/kernel/hvapi.c
53
{ .group = HV_GRP_DIAG, .flags = FLAG_PRE_API },
arch/sparc/kernel/hvapi.c
54
{ .group = HV_GRP_M7_PERF, },
arch/sparc/kernel/hvapi.c
59
static struct api_info *__get_info(unsigned long group)
arch/sparc/kernel/hvapi.c
64
if (api_table[i].group == group)
arch/sparc/kernel/hvapi.c
80
sun4v_set_version(p->group, 0, 0, &ignore);
arch/sparc/kernel/hvapi.c
96
int sun4v_hvapi_register(unsigned long group, unsigned long major,
arch/sparc/kernel/perf_event.c
1345
static int collect_events(struct perf_event *group, int max_count,
arch/sparc/kernel/perf_event.c
1352
if (!is_software_event(group)) {
arch/sparc/kernel/perf_event.c
1355
evts[n] = group;
arch/sparc/kernel/perf_event.c
1356
events[n] = group->hw.event_base;
arch/sparc/kernel/perf_event.c
1359
for_each_sibling_event(event, group) {
arch/um/drivers/vector_user.c
462
char *group = uml_vector_fetch_arg(ifspec,"group");
arch/um/drivers/vector_user.c
475
if (group != NULL) {
arch/um/drivers/vector_user.c
477
argv[argc++] = group;
arch/um/drivers/vfio_kern.c
103
list_for_each_entry(group, ¨_vfio_groups, list) {
arch/um/drivers/vfio_kern.c
104
if (group->id == group_id) {
arch/um/drivers/vfio_kern.c
105
group->users++;
arch/um/drivers/vfio_kern.c
106
return group->fd;
arch/um/drivers/vfio_kern.c
110
group = kzalloc_obj(*group);
arch/um/drivers/vfio_kern.c
111
if (!group)
arch/um/drivers/vfio_kern.c
114
group->fd = uml_vfio_user_open_group(group_id);
arch/um/drivers/vfio_kern.c
115
if (group->fd < 0) {
arch/um/drivers/vfio_kern.c
116
err = group->fd;
arch/um/drivers/vfio_kern.c
120
err = uml_vfio_set_container(group->fd);
arch/um/drivers/vfio_kern.c
124
group->id = group_id;
arch/um/drivers/vfio_kern.c
125
group->users = 1;
arch/um/drivers/vfio_kern.c
127
list_add(&group->list, ¨_vfio_groups);
arch/um/drivers/vfio_kern.c
129
return group->fd;
arch/um/drivers/vfio_kern.c
132
os_close_file(group->fd);
arch/um/drivers/vfio_kern.c
134
kfree(group);
arch/um/drivers/vfio_kern.c
140
struct uml_vfio_group *group;
arch/um/drivers/vfio_kern.c
144
list_for_each_entry(group, ¨_vfio_groups, list) {
arch/um/drivers/vfio_kern.c
145
if (group->fd == group_fd) {
arch/um/drivers/vfio_kern.c
146
group->users--;
arch/um/drivers/vfio_kern.c
147
if (group->users == 0) {
arch/um/drivers/vfio_kern.c
150
list_del(&group->list);
arch/um/drivers/vfio_kern.c
151
kfree(group);
arch/um/drivers/vfio_kern.c
32
int group;
arch/um/drivers/vfio_kern.c
496
dev->group = uml_vfio_open_group(group_id);
arch/um/drivers/vfio_kern.c
497
if (dev->group < 0) {
arch/um/drivers/vfio_kern.c
499
group_id, dev->name, dev->group);
arch/um/drivers/vfio_kern.c
503
err = uml_vfio_user_setup_device(&dev->udev, dev->group, dev->name);
arch/um/drivers/vfio_kern.c
549
uml_vfio_release_group(dev->group);
arch/um/drivers/vfio_kern.c
568
uml_vfio_release_group(dev->group);
arch/um/drivers/vfio_kern.c
98
struct uml_vfio_group *group;
arch/um/drivers/vfio_user.c
136
int uml_vfio_user_set_container(int container, int group)
arch/um/drivers/vfio_user.c
138
if (ioctl(group, VFIO_GROUP_SET_CONTAINER, &container) < 0)
arch/um/drivers/vfio_user.c
143
int uml_vfio_user_unset_container(int container, int group)
arch/um/drivers/vfio_user.c
145
if (ioctl(group, VFIO_GROUP_UNSET_CONTAINER, &container) < 0)
arch/um/drivers/vfio_user.c
178
int group, const char *device)
arch/um/drivers/vfio_user.c
184
dev->device = ioctl(group, VFIO_GROUP_GET_DEVICE_FD, device);
arch/um/drivers/vfio_user.h
23
int uml_vfio_user_set_container(int container, int group);
arch/um/drivers/vfio_user.h
24
int uml_vfio_user_unset_container(int container, int group);
arch/um/drivers/vfio_user.h
27
int group, const char *device);
arch/x86/events/amd/uncore.c
1052
pmu->group = gid;
arch/x86/events/amd/uncore.c
517
if (gid != pmu->group)
arch/x86/events/amd/uncore.c
53
int group;
arch/x86/events/amd/uncore.c
744
pmu->group = amd_uncore_ctx_gid(uncore, cpu);
arch/x86/events/amd/uncore.c
878
pmu->group = amd_uncore_ctx_gid(uncore, cpu);
arch/x86/events/intel/uncore.c
1014
struct attribute_group group;
arch/x86/events/intel/uncore.c
1023
attr_group->group.name = "events";
arch/x86/events/intel/uncore.c
1024
attr_group->group.attrs = attr_group->attrs;
arch/x86/events/intel/uncore.c
1029
type->events_group = &attr_group->group;
arch/x86/include/asm/kvm_host.h
1931
int (*dev_get_attr)(u32 group, u64 attr, u64 *val);
arch/x86/kvm/emulate.c
196
const struct opcode *group;
arch/x86/kvm/emulate.c
3941
#define EXT(_f, _e) { .flags = ((_f) | RMExt), .u.group = (_e) }
arch/x86/kvm/emulate.c
3942
#define G(_f, _g) { .flags = ((_f) | Group | ModRM), .u.group = (_g) }
arch/x86/kvm/emulate.c
5004
opcode = opcode.u.group[goffset];
arch/x86/kvm/emulate.c
5015
opcode = opcode.u.group[goffset];
arch/x86/kvm/svm/sev.c
2147
int sev_dev_get_attr(u32 group, u64 attr, u64 *val)
arch/x86/kvm/svm/sev.c
2149
if (group != KVM_X86_GRP_SEV)
arch/x86/kvm/svm/svm.h
894
int sev_dev_get_attr(u32 group, u64 attr, u64 *val);
arch/x86/kvm/svm/svm.h
920
static inline int sev_dev_get_attr(u32 group, u64 attr, u64 *val) { return -ENXIO; }
arch/x86/kvm/x86.c
5004
if (attr->group) {
arch/x86/kvm/x86.c
5006
return kvm_x86_call(dev_get_attr)(attr->group, attr->attr, val);
arch/x86/kvm/x86.c
5989
if (attr.group != KVM_VCPU_TSC_CTRL)
arch/x86/platform/geode/geode-common.c
100
const struct software_node *group[MAX_LEDS + 2] = { 0 };
arch/x86/platform/geode/geode-common.c
138
group[0] = &geode_gpio_leds_node;
arch/x86/platform/geode/geode-common.c
159
group[i + 1] = &swnodes[i];
arch/x86/platform/geode/geode-common.c
162
err = software_node_register_node_group(group);
arch/x86/platform/geode/geode-common.c
180
software_node_unregister_node_group(group);
block/genhd.c
115
int group;
block/genhd.c
117
for (group = 0; group < NR_STAT_GROUPS; group++) {
block/genhd.c
118
stat->nsecs[group] += ptr->nsecs[group];
block/genhd.c
119
stat->sectors[group] += ptr->sectors[group];
block/genhd.c
120
stat->ios[group] += ptr->ios[group];
block/genhd.c
121
stat->merges[group] += ptr->merges[group];
block/partitions/ldm.c
1220
int rec, num, group;
block/partitions/ldm.c
1229
group = get_unaligned_be32(data + 0x08);
block/partitions/ldm.c
1243
if (f->group == group)
block/partitions/ldm.c
1253
f->group = group;
block/partitions/ldm.c
1319
f->group, f->map);
block/partitions/ldm.h
160
u32 group;
block/partitions/ldm.h
88
u32 group;
drivers/accessibility/speakup/i18n.c
582
struct msg_group_t *group = NULL;
drivers/accessibility/speakup/i18n.c
587
group = &all_groups[i];
drivers/accessibility/speakup/i18n.c
591
return group;
drivers/accessibility/speakup/i18n.c
594
void spk_reset_msg_group(struct msg_group_t *group)
drivers/accessibility/speakup/i18n.c
601
for (i = group->start; i <= group->end; i++) {
drivers/accessibility/speakup/i18n.h
244
void spk_reset_msg_group(struct msg_group_t *group);
drivers/accessibility/speakup/kobjects.c
749
struct msg_group_t *group)
drivers/accessibility/speakup/kobjects.c
763
enum msg_index_t firstmessage = group->start;
drivers/accessibility/speakup/kobjects.c
764
enum msg_index_t lastmessage = group->end;
drivers/accessibility/speakup/kobjects.c
830
spk_reset_msg_group(group);
drivers/accessibility/speakup/kobjects.c
832
report_msg_status(reset, received, used, rejected, group->name);
drivers/accessibility/speakup/kobjects.c
840
struct msg_group_t *group = spk_find_msg_group(attr->attr.name);
drivers/accessibility/speakup/kobjects.c
843
if (WARN_ON(!group))
drivers/accessibility/speakup/kobjects.c
847
retval = message_show_helper(buf, group->start, group->end);
drivers/accessibility/speakup/kobjects.c
855
struct msg_group_t *group = spk_find_msg_group(attr->attr.name);
drivers/accessibility/speakup/kobjects.c
857
if (WARN_ON(!group))
drivers/accessibility/speakup/kobjects.c
860
return message_store_helper(buf, count, group);
drivers/acpi/acpi_configfs.c
208
static struct config_item *acpi_table_make_item(struct config_group *group,
drivers/acpi/acpi_configfs.c
221
static void acpi_table_drop_item(struct config_group *group,
drivers/ata/sata_rcar.c
167
u32 val, int group)
drivers/ata/sata_rcar.c
179
if (group)
drivers/atm/he.c
1432
for (group = 1; group < HE_NUM_GROUPS; ++group) {
drivers/atm/he.c
1433
he_writel(he_dev, 0x0, G0_RBPS_S + (group * 32));
drivers/atm/he.c
1434
he_writel(he_dev, 0x0, G0_RBPS_T + (group * 32));
drivers/atm/he.c
1435
he_writel(he_dev, 0x0, G0_RBPS_QI + (group * 32));
drivers/atm/he.c
1437
G0_RBPS_BS + (group * 32));
drivers/atm/he.c
1439
he_writel(he_dev, 0x0, G0_RBPL_S + (group * 32));
drivers/atm/he.c
1440
he_writel(he_dev, 0x0, G0_RBPL_T + (group * 32));
drivers/atm/he.c
1442
G0_RBPL_QI + (group * 32));
drivers/atm/he.c
1443
he_writel(he_dev, 0x0, G0_RBPL_BS + (group * 32));
drivers/atm/he.c
1445
he_writel(he_dev, 0x0, G0_RBRQ_ST + (group * 16));
drivers/atm/he.c
1446
he_writel(he_dev, 0x0, G0_RBRQ_H + (group * 16));
drivers/atm/he.c
1448
G0_RBRQ_Q + (group * 16));
drivers/atm/he.c
1449
he_writel(he_dev, 0x0, G0_RBRQ_I + (group * 16));
drivers/atm/he.c
1451
he_writel(he_dev, 0x0, G0_TBRQ_B_T + (group * 16));
drivers/atm/he.c
1452
he_writel(he_dev, 0x0, G0_TBRQ_H + (group * 16));
drivers/atm/he.c
1454
G0_TBRQ_THRESH + (group * 16));
drivers/atm/he.c
1455
he_writel(he_dev, 0x0, G0_TBRQ_S + (group * 16));
drivers/atm/he.c
1637
he_service_rbrq(struct he_dev *he_dev, int group)
drivers/atm/he.c
1641
he_dev->hsp->group[group].rbrq_tail);
drivers/atm/he.c
1656
he_dev->rbrq_head, group,
drivers/atm/he.c
1790
G0_RBRQ_H + (group * 16));
drivers/atm/he.c
1797
he_service_tbrq(struct he_dev *he_dev, int group)
drivers/atm/he.c
1801
he_dev->hsp->group[group].tbrq_tail);
drivers/atm/he.c
1812
group,
drivers/atm/he.c
1871
G0_TBRQ_H + (group * 16));
drivers/atm/he.c
1876
he_service_rbpl(struct he_dev *he_dev, int group)
drivers/atm/he.c
1927
int group, type;
drivers/atm/he.c
1937
group = ITYPE_GROUP(he_dev->irq_head->isw);
drivers/atm/he.c
1941
HPRINTK("rbrq%d threshold\n", group);
drivers/atm/he.c
1944
if (he_service_rbrq(he_dev, group))
drivers/atm/he.c
1945
he_service_rbpl(he_dev, group);
drivers/atm/he.c
1948
HPRINTK("tbrq%d threshold\n", group);
drivers/atm/he.c
1951
he_service_tbrq(he_dev, group);
drivers/atm/he.c
1954
he_service_rbpl(he_dev, group);
drivers/atm/he.c
1969
switch (type|group) {
drivers/atm/he.c
769
static int he_init_group(struct he_dev *he_dev, int group)
drivers/atm/he.c
775
he_writel(he_dev, 0x0, G0_RBPS_S + (group * 32));
drivers/atm/he.c
776
he_writel(he_dev, 0x0, G0_RBPS_T + (group * 32));
drivers/atm/he.c
777
he_writel(he_dev, 0x0, G0_RBPS_QI + (group * 32));
drivers/atm/he.c
779
G0_RBPS_BS + (group * 32));
drivers/atm/he.c
829
he_writel(he_dev, he_dev->rbpl_phys, G0_RBPL_S + (group * 32));
drivers/atm/he.c
831
G0_RBPL_T + (group * 32));
drivers/atm/he.c
833
G0_RBPL_BS + (group * 32));
drivers/atm/he.c
838
G0_RBPL_QI + (group * 32));
drivers/atm/he.c
851
he_writel(he_dev, he_dev->rbrq_phys, G0_RBRQ_ST + (group * 16));
drivers/atm/he.c
852
he_writel(he_dev, 0, G0_RBRQ_H + (group * 16));
drivers/atm/he.c
855
G0_RBRQ_Q + (group * 16));
drivers/atm/he.c
859
G0_RBRQ_I + (group * 16));
drivers/atm/he.c
862
G0_RBRQ_I + (group * 16));
drivers/atm/he.c
876
he_writel(he_dev, he_dev->tbrq_phys, G0_TBRQ_B_T + (group * 16));
drivers/atm/he.c
877
he_writel(he_dev, 0, G0_TBRQ_H + (group * 16));
drivers/atm/he.c
878
he_writel(he_dev, CONFIG_TBRQ_SIZE - 1, G0_TBRQ_S + (group * 16));
drivers/atm/he.c
879
he_writel(he_dev, CONFIG_TBRQ_THRESH, G0_TBRQ_THRESH + (group * 16));
drivers/atm/he.c
979
int i, group;
drivers/atm/he.h
198
} group[HE_NUM_GROUPS];
drivers/atm/he.h
273
struct he_group group[HE_NUM_GROUPS];
drivers/base/core.c
2849
const struct attribute_group *group;
drivers/base/core.c
2856
const struct attribute_group *group = devres->group;
drivers/base/core.c
2858
dev_dbg(dev, "%s: removing group %p\n", __func__, group);
drivers/base/core.c
2859
sysfs_remove_group(&dev->kobj, group);
drivers/base/core.c
2888
devres->group = grp;
drivers/base/memory.c
1072
static int memory_group_register(struct memory_group group)
drivers/base/memory.c
1078
if (!node_possible(group.nid))
drivers/base/memory.c
1081
new_group = kzalloc_obj(group);
drivers/base/memory.c
1084
*new_group = group;
drivers/base/memory.c
1092
} else if (group.is_dynamic) {
drivers/base/memory.c
1115
struct memory_group group = {
drivers/base/memory.c
1124
return memory_group_register(group);
drivers/base/memory.c
1145
struct memory_group group = {
drivers/base/memory.c
1156
return memory_group_register(group);
drivers/base/memory.c
1173
struct memory_group *group;
drivers/base/memory.c
1178
group = xa_load(&memory_groups, mgid);
drivers/base/memory.c
1179
if (!group)
drivers/base/memory.c
1181
if (!list_empty(&group->memory_blocks))
drivers/base/memory.c
1184
kfree(group);
drivers/base/memory.c
1208
struct memory_group *group;
drivers/base/memory.c
1212
xa_for_each_marked(&memory_groups, index, group,
drivers/base/memory.c
1214
if (group == excluded)
drivers/base/memory.c
1217
if (nid != NUMA_NO_NODE && group->nid != nid)
drivers/base/memory.c
1220
ret = func(group, arg);
drivers/base/memory.c
235
zone = zone_for_pfn_range(mem->online_type, mem->nid, mem->group,
drivers/base/memory.c
256
nr_pages - nr_vmemmap_pages, zone, mem->group);
drivers/base/memory.c
268
adjust_present_page_count(pfn_to_page(start_pfn), mem->group,
drivers/base/memory.c
299
adjust_present_page_count(pfn_to_page(start_pfn), mem->group,
drivers/base/memory.c
303
nr_pages - nr_vmemmap_pages, mem->zone, mem->group);
drivers/base/memory.c
308
mem->group, nr_vmemmap_pages);
drivers/base/memory.c
453
struct memory_group *group,
drivers/base/memory.c
459
zone = zone_for_pfn_range(online_type, nid, group, start_pfn, nr_pages);
drivers/base/memory.c
472
struct memory_group *group = mem->group;
drivers/base/memory.c
490
default_zone = zone_for_pfn_range(MMOP_ONLINE, nid, group,
drivers/base/memory.c
494
len += print_allowed_zone(buf, len, nid, group, start_pfn, nr_pages,
drivers/base/memory.c
496
len += print_allowed_zone(buf, len, nid, group, start_pfn, nr_pages,
drivers/base/memory.c
793
struct memory_group *group)
drivers/base/memory.c
828
if (group) {
drivers/base/memory.c
829
mem->group = group;
drivers/base/memory.c
830
list_add(&mem->group_next, &group->memory_blocks);
drivers/base/memory.c
843
if (memory->group) {
drivers/base/memory.c
845
memory->group = NULL;
drivers/base/memory.c
862
struct memory_group *group)
drivers/base/memory.c
875
ret = add_memory_block(block_id, nid, MEM_OFFLINE, altmap, group);
drivers/base/test/property-entry-test.c
410
static const struct software_node *group[] = { &node1, &node2, NULL };
drivers/base/test/property-entry-test.c
428
error = software_node_register_node_group(group);
drivers/base/test/property-entry-test.c
489
software_node_unregister_node_group(group);
drivers/block/mtip32xx/mtip32xx.c
210
int group = tag >> 5;
drivers/block/mtip32xx/mtip32xx.c
213
spin_lock(&port->cmd_issue_lock[group]);
drivers/block/mtip32xx/mtip32xx.c
218
spin_unlock(&port->cmd_issue_lock[group]);
drivers/block/mtip32xx/mtip32xx.c
463
int group, tagmap_len = 0;
drivers/block/mtip32xx/mtip32xx.c
466
for (group = SLOTBITS_IN_LONGS; group > 0; group--)
drivers/block/mtip32xx/mtip32xx.c
468
tagbits[group-1]);
drivers/block/mtip32xx/mtip32xx.c
497
int group, tag, bit, reissue, rv;
drivers/block/mtip32xx/mtip32xx.c
523
for (group = 0; group < dd->slot_groups; group++) {
drivers/block/mtip32xx/mtip32xx.c
524
completed = readl(port->completed[group]);
drivers/block/mtip32xx/mtip32xx.c
526
dev_warn(&dd->pdev->dev, "g=%u, comp=%x\n", group, completed);
drivers/block/mtip32xx/mtip32xx.c
529
writel(completed, port->completed[group]);
drivers/block/mtip32xx/mtip32xx.c
535
tag = (group << 5) + bit;
drivers/block/mtip32xx/mtip32xx.c
591
for (group = 0; group < dd->slot_groups; group++) {
drivers/block/mtip32xx/mtip32xx.c
594
tag = (group << 5) + bit;
drivers/block/mtip32xx/mtip32xx.c
646
static inline void mtip_workq_sdbfx(struct mtip_port *port, int group,
drivers/block/mtip32xx/mtip32xx.c
658
writel(completed, port->completed[group]);
drivers/block/mtip32xx/mtip32xx.c
663
tag = (group << 5) | bit;
drivers/block/mtip32xx/mtip32xx.c
689
int group = MTIP_TAG_INDEX(MTIP_TAG_INTERNAL);
drivers/block/mtip32xx/mtip32xx.c
690
int status = readl(port->cmd_issue[group]);
drivers/block/mtip32xx/mtip32xx.h
177
#define DEFINE_HANDLER(group) \
drivers/block/mtip32xx/mtip32xx.h
178
void mtip_workq_sdbf##group(struct work_struct *work) \
drivers/block/mtip32xx/mtip32xx.h
181
mtip_workq_sdbfx(w->port, group, w->completed); \
drivers/block/null_blk/main.c
2029
if (config_item_name(&dev->group.cg_item)) {
drivers/block/null_blk/main.c
2032
"%s", config_item_name(&dev->group.cg_item));
drivers/block/null_blk/main.c
282
return item ? container_of(to_config_group(item), struct nullb_device, group) : NULL;
drivers/block/null_blk/main.c
663
configfs_add_default_group(&dev->timeout_config.group, &dev->group);
drivers/block/null_blk/main.c
664
configfs_add_default_group(&dev->requeue_config.group, &dev->group);
drivers/block/null_blk/main.c
665
configfs_add_default_group(&dev->init_hctx_fault_config.group, &dev->group);
drivers/block/null_blk/main.c
670
config_item_put(&dev->init_hctx_fault_config.group.cg_item);
drivers/block/null_blk/main.c
671
config_item_put(&dev->requeue_config.group.cg_item);
drivers/block/null_blk/main.c
672
config_item_put(&dev->timeout_config.group.cg_item);
drivers/block/null_blk/main.c
687
config_group *nullb_group_make_group(struct config_group *group, const char *name)
drivers/block/null_blk/main.c
698
config_group_init_type_name(&dev->group, name, &nullb_device_type);
drivers/block/null_blk/main.c
701
return &dev->group;
drivers/block/null_blk/main.c
705
nullb_group_drop_item(struct config_group *group, struct config_item *item)
drivers/block/null_blk/null_blk.h
55
struct config_group group;
drivers/clk/clk-si5341.c
970
unsigned int group = clkspec->args[0];
drivers/clk/clk-si5341.c
972
switch (group) {
drivers/clk/clk-si5341.c
995
dev_err(&data->i2c_client->dev, "invalid group %u\n", group);
drivers/clk/renesas/clk-mstp.c
100
spin_unlock_irqrestore(&group->lock, flags);
drivers/clk/renesas/clk-mstp.c
102
if (!enable || !group->mstpsr)
drivers/clk/renesas/clk-mstp.c
106
ret = readl_poll_timeout_atomic(group->mstpsr, value,
drivers/clk/renesas/clk-mstp.c
110
group->smstpcr, clock->bit_index);
drivers/clk/renesas/clk-mstp.c
128
struct mstp_clock_group *group = clock->group;
drivers/clk/renesas/clk-mstp.c
131
if (group->mstpsr)
drivers/clk/renesas/clk-mstp.c
132
value = cpg_mstp_read(group, group->mstpsr);
drivers/clk/renesas/clk-mstp.c
134
value = cpg_mstp_read(group, group->smstpcr);
drivers/clk/renesas/clk-mstp.c
147
struct mstp_clock_group *group)
drivers/clk/renesas/clk-mstp.c
169
clock->group = group;
drivers/clk/renesas/clk-mstp.c
182
struct mstp_clock_group *group;
drivers/clk/renesas/clk-mstp.c
187
group = kzalloc_flex(*group, clks, MSTP_MAX_CLOCKS);
drivers/clk/renesas/clk-mstp.c
188
if (!group)
drivers/clk/renesas/clk-mstp.c
191
clks = group->clks;
drivers/clk/renesas/clk-mstp.c
192
spin_lock_init(&group->lock);
drivers/clk/renesas/clk-mstp.c
193
group->data.clks = clks;
drivers/clk/renesas/clk-mstp.c
195
group->smstpcr = of_iomap(np, 0);
drivers/clk/renesas/clk-mstp.c
196
group->mstpsr = of_iomap(np, 1);
drivers/clk/renesas/clk-mstp.c
198
if (group->smstpcr == NULL) {
drivers/clk/renesas/clk-mstp.c
200
kfree(group);
drivers/clk/renesas/clk-mstp.c
205
group->width_8bit = true;
drivers/clk/renesas/clk-mstp.c
239
clkidx, group);
drivers/clk/renesas/clk-mstp.c
241
group->data.clk_num = max(group->data.clk_num,
drivers/clk/renesas/clk-mstp.c
248
of_clk_add_provider(np, of_clk_src_onecell_get, &group->data);
drivers/clk/renesas/clk-mstp.c
59
struct mstp_clock_group *group;
drivers/clk/renesas/clk-mstp.c
64
static inline u32 cpg_mstp_read(struct mstp_clock_group *group,
drivers/clk/renesas/clk-mstp.c
67
return group->width_8bit ? readb(reg) : readl(reg);
drivers/clk/renesas/clk-mstp.c
70
static inline void cpg_mstp_write(struct mstp_clock_group *group, u32 val,
drivers/clk/renesas/clk-mstp.c
73
group->width_8bit ? writeb(val, reg) : writel(val, reg);
drivers/clk/renesas/clk-mstp.c
79
struct mstp_clock_group *group = clock->group;
drivers/clk/renesas/clk-mstp.c
85
spin_lock_irqsave(&group->lock, flags);
drivers/clk/renesas/clk-mstp.c
87
value = cpg_mstp_read(group, group->smstpcr);
drivers/clk/renesas/clk-mstp.c
92
cpg_mstp_write(group, value, group->smstpcr);
drivers/clk/renesas/clk-mstp.c
94
if (!group->mstpsr) {
drivers/clk/renesas/clk-mstp.c
96
cpg_mstp_read(group, group->smstpcr);
drivers/clk/renesas/clk-mstp.c
97
barrier_data(group->smstpcr);
drivers/clk/renesas/r9a06g032-clocks.c
1376
uart_group_sel[d->dual.group] = d->dual.sel;
drivers/clk/renesas/r9a06g032-clocks.c
1382
uart_group_sel[d->dual.group]);
drivers/clk/renesas/r9a06g032-clocks.c
155
uint16_t group:1;
drivers/clk/renesas/r9a06g032-clocks.c
220
.group = _g, \
drivers/clk/renesas/r9a06g032-clocks.c
649
.dual.group = 0,
drivers/clk/renesas/r9a06g032-clocks.c
658
.dual.group = 1,
drivers/comedi/drivers/ni_at_ao.c
113
static void atao_select_reg_group(struct comedi_device *dev, int group)
drivers/comedi/drivers/ni_at_ao.c
117
if (group)
drivers/comedi/drivers/s626.c
1015
unsigned int group = chan / 16;
drivers/comedi/drivers/s626.c
1016
unsigned int mask = 1 << (chan - (16 * group));
drivers/comedi/drivers/s626.c
1020
status = s626_debi_read(dev, S626_LP_RDEDGSEL(group));
drivers/comedi/drivers/s626.c
1021
s626_debi_write(dev, S626_LP_WREDGSEL(group), mask | status);
drivers/comedi/drivers/s626.c
1024
status = s626_debi_read(dev, S626_LP_RDINTSEL(group));
drivers/comedi/drivers/s626.c
1025
s626_debi_write(dev, S626_LP_WRINTSEL(group), mask | status);
drivers/comedi/drivers/s626.c
1031
status = s626_debi_read(dev, S626_LP_RDCAPSEL(group));
drivers/comedi/drivers/s626.c
1032
s626_debi_write(dev, S626_LP_WRCAPSEL(group), mask | status);
drivers/comedi/drivers/s626.c
1037
static int s626_dio_reset_irq(struct comedi_device *dev, unsigned int group,
drivers/comedi/drivers/s626.c
1044
s626_debi_write(dev, S626_LP_WRCAPSEL(group), mask);
drivers/comedi/drivers/s626.c
1051
unsigned int group;
drivers/comedi/drivers/s626.c
1057
for (group = 0; group < S626_DIO_BANKS; group++)
drivers/comedi/drivers/s626.c
1058
s626_debi_write(dev, S626_LP_WRCAPSEL(group), 0xffff);
drivers/comedi/drivers/s626.c
1064
u16 irqbit, u8 group)
drivers/comedi/drivers/s626.c
1070
s626_dio_reset_irq(dev, group, irqbit);
drivers/comedi/drivers/s626.c
1074
if ((irqbit >> (cmd->start_arg - (16 * group))) == 1 &&
drivers/comedi/drivers/s626.c
1082
if ((irqbit >> (cmd->scan_begin_arg - (16 * group))) == 1 &&
drivers/comedi/drivers/s626.c
1098
if ((irqbit >> (cmd->convert_arg - (16 * group))) == 1 &&
drivers/comedi/drivers/s626.c
1113
u8 group;
drivers/comedi/drivers/s626.c
1115
for (group = 0; group < S626_DIO_BANKS; group++) {
drivers/comedi/drivers/s626.c
1117
irqbit = s626_debi_read(dev, S626_LP_RDCAPFLG(group));
drivers/comedi/drivers/s626.c
1121
s626_handle_dio_interrupt(dev, irqbit, group);
drivers/comedi/drivers/s626.c
1953
u16 group;
drivers/comedi/drivers/s626.c
1959
for (group = 0; group < S626_DIO_BANKS; group++) {
drivers/comedi/drivers/s626.c
1961
s626_debi_write(dev, S626_LP_WRINTSEL(group), 0);
drivers/comedi/drivers/s626.c
1963
s626_debi_write(dev, S626_LP_WRCAPSEL(group), 0xffff);
drivers/comedi/drivers/s626.c
1965
s626_debi_write(dev, S626_LP_WREDGSEL(group), 0);
drivers/comedi/drivers/s626.c
1967
s626_debi_write(dev, S626_LP_WRDOUT(group), 0);
drivers/comedi/drivers/s626.c
1976
unsigned long group = (unsigned long)s->private;
drivers/comedi/drivers/s626.c
1979
s626_debi_write(dev, S626_LP_WRDOUT(group), s->state);
drivers/comedi/drivers/s626.c
1981
data[1] = s626_debi_read(dev, S626_LP_RDDIN(group));
drivers/comedi/drivers/s626.c
1991
unsigned long group = (unsigned long)s->private;
drivers/comedi/drivers/s626.c
1998
s626_debi_write(dev, S626_LP_WRDOUT(group), s->io_bits);
drivers/connector/cn_queue.c
88
cbq->group = cbq->id.id.idx;
drivers/connector/connector.c
111
NETLINK_CB(skb).dst_group = group;
drivers/connector/connector.c
113
if (group)
drivers/connector/connector.c
114
return netlink_broadcast_filtered(dev->nls, skb, portid, group,
drivers/connector/connector.c
172
static int cn_bind(struct net *net, int group)
drivers/connector/connector.c
174
unsigned long groups = (unsigned long) group;
drivers/connector/connector.c
71
u32 group = 0;
drivers/connector/connector.c
75
group = __group;
drivers/connector/connector.c
82
group = __cbq->group;
drivers/connector/connector.c
92
if (!portid && !netlink_has_listeners(dev->nls, group))
drivers/counter/counter-sysfs.c
533
struct counter_attribute_group *const group,
drivers/counter/counter-sysfs.c
561
list_add(&counter_attr->l, &group->attr_list);
drivers/counter/counter-sysfs.c
562
group->num_attr++;
drivers/counter/counter-sysfs.c
568
struct counter_attribute_group *const group,
drivers/counter/counter-sysfs.c
659
list_add(&counter_attr->l, &group->attr_list);
drivers/counter/counter-sysfs.c
660
group->num_attr++;
drivers/counter/counter-sysfs.c
668
return counter_avail_attr_create(dev, group, comp, parent);
drivers/counter/counter-sysfs.c
681
struct counter_attribute_group *const group,
drivers/counter/counter-sysfs.c
700
list_add(&counter_attr->l, &group->attr_list);
drivers/counter/counter-sysfs.c
701
group->num_attr++;
drivers/counter/counter-sysfs.c
715
struct counter_attribute_group *const group,
drivers/counter/counter-sysfs.c
740
list_add(&counter_attr->l, &group->attr_list);
drivers/counter/counter-sysfs.c
741
group->num_attr++;
drivers/counter/counter-sysfs.c
747
struct counter_attribute_group *const group,
drivers/counter/counter-sysfs.c
755
err = counter_attr_create(dev, group, ext, scope, parent);
drivers/counter/counter-sysfs.c
760
return counter_comp_id_attr_create(dev, group, ext->name, id);
drivers/counter/counter-sysfs.c
764
struct counter_attribute_group *const group,
drivers/counter/counter-sysfs.c
793
err = counter_ext_attrs_create(dev, group, &ext, scope, parent,
drivers/counter/counter-sysfs.c
803
struct counter_attribute_group *const group,
drivers/counter/counter-sysfs.c
819
err = counter_array_attrs_create(dev, group, ext, scope,
drivers/counter/counter-sysfs.c
824
err = counter_ext_attrs_create(dev, group, ext, scope,
drivers/counter/counter-sysfs.c
891
struct counter_attribute_group *const group,
drivers/counter/counter-sysfs.c
918
err = counter_attr_create(dev, group, &comp,
drivers/counter/counter-sysfs.c
924
err = counter_comp_id_attr_create(dev, group, comp.name, i);
drivers/crypto/cavium/cpt/cptpf.h
25
u8 group;
drivers/crypto/cavium/cpt/cptpf_main.c
182
mcode->group = cpt->next_group;
drivers/crypto/cavium/cpt/cptpf_main.c
186
mcode->group);
drivers/crypto/cavium/cpt/cptpf_main.c
196
cpt_configure_group(cpt, mcode->group, mcode->core_mask,
drivers/crypto/cavium/cpt/cptpf_main.c
211
mcode->group = cpt->next_group;
drivers/crypto/cavium/cpt/cptpf_main.c
215
mcode->group);
drivers/crypto/cavium/cpt/cptpf_main.c
225
cpt_configure_group(cpt, mcode->group, mcode->core_mask,
drivers/crypto/cavium/cpt/cptpf_mbox.c
78
pf_qx_ctl.s.grp = mcode[grp].group;
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
416
u8 group = 0;
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
430
group = ctrl->s.grp;
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
472
vq_cmd.cptr.s.grp = group;
drivers/crypto/marvell/octeontx/otx_cptvf.h
97
int otx_cptvf_send_vf_to_grp_msg(struct otx_cptvf *cptvf, int group);
drivers/crypto/marvell/octeontx/otx_cptvf_mbox.c
185
int otx_cptvf_send_vf_to_grp_msg(struct otx_cptvf *cptvf, int group)
drivers/crypto/marvell/octeontx/otx_cptvf_mbox.c
192
mbx.data = group;
drivers/crypto/marvell/octeontx/otx_cptvf_mbox.c
196
cptvf->vfgrp = group;
drivers/crypto/nx/nx-842.c
151
if (s > be16_to_cpu(hdr->group[0].padding)) {
drivers/crypto/nx/nx-842.c
311
ret = compress(ctx, &p, &hdr->group[n], &c, &ignore, h);
drivers/crypto/nx/nx-842.c
507
ret = decompress(ctx, &p, &hdr->group[n], &c, ignore);
drivers/crypto/nx/nx-842.h
167
struct nx842_crypto_header_group group[];
drivers/crypto/nx/nx-842.h
169
static_assert(offsetof(struct nx842_crypto_header, group) == sizeof(struct nx842_crypto_header_hdr),
drivers/crypto/nx/nx-842.h
181
struct nx842_crypto_header_group group[NX842_CRYPTO_GROUP_MAX];
drivers/cxl/core/port.c
1123
void *group = devres_open_group(host, dport, GFP_KERNEL);
drivers/cxl/core/port.c
1125
if (!group) {
drivers/cxl/core/port.c
1132
devres_release_group(host, group);
drivers/cxl/core/port.c
1136
return group;
drivers/cxl/core/port.c
1139
static void cxl_dport_close_dr_group(struct cxl_dport *dport, void *group)
drivers/cxl/core/port.c
1141
devres_close_group(dport_to_host(dport), group);
drivers/dma/idxd/defaults.c
34
group = idxd->groups[0];
drivers/dma/idxd/defaults.c
35
wq->group = group;
drivers/dma/idxd/defaults.c
36
group->num_wqs++;
drivers/dma/idxd/defaults.c
47
engine->group = idxd->groups[0];
drivers/dma/idxd/defaults.c
48
engine->group->num_engines++;
drivers/dma/idxd/defaults.c
9
struct idxd_group *group;
drivers/dma/idxd/device.c
1039
struct idxd_group *group = idxd->groups[i];
drivers/dma/idxd/device.c
1041
if (group->tc_a == -1)
drivers/dma/idxd/device.c
1042
group->tc_a = group->grpcfg.flags.tc_a = 0;
drivers/dma/idxd/device.c
1044
group->grpcfg.flags.tc_a = group->tc_a;
drivers/dma/idxd/device.c
1045
if (group->tc_b == -1)
drivers/dma/idxd/device.c
1046
group->tc_b = group->grpcfg.flags.tc_b = 1;
drivers/dma/idxd/device.c
1048
group->grpcfg.flags.tc_b = group->tc_b;
drivers/dma/idxd/device.c
1049
group->grpcfg.flags.use_rdbuf_limit = group->use_rdbuf_limit;
drivers/dma/idxd/device.c
1050
group->grpcfg.flags.rdbufs_reserved = group->rdbufs_reserved;
drivers/dma/idxd/device.c
1051
group->grpcfg.flags.rdbufs_allowed = group->rdbufs_allowed;
drivers/dma/idxd/device.c
1052
group->grpcfg.flags.desc_progress_limit = group->desc_progress_limit;
drivers/dma/idxd/device.c
1053
group->grpcfg.flags.batch_progress_limit = group->batch_progress_limit;
drivers/dma/idxd/device.c
1061
struct idxd_group *group;
drivers/dma/idxd/device.c
1064
group = idxd->groups[i];
drivers/dma/idxd/device.c
1065
group->grpcfg.engines = 0;
drivers/dma/idxd/device.c
1070
group = eng->group;
drivers/dma/idxd/device.c
1072
if (!group)
drivers/dma/idxd/device.c
1075
group->grpcfg.engines |= BIT(eng->id);
drivers/dma/idxd/device.c
1088
struct idxd_group *group;
drivers/dma/idxd/device.c
1093
group = idxd->groups[i];
drivers/dma/idxd/device.c
1095
group->grpcfg.wqs[j] = 0;
drivers/dma/idxd/device.c
1100
group = wq->group;
drivers/dma/idxd/device.c
1102
if (!wq->group)
drivers/dma/idxd/device.c
1111
group->grpcfg.wqs[wq->id / 64] |= BIT(wq->id % 64);
drivers/dma/idxd/device.c
1187
static void idxd_group_load_config(struct idxd_group *group)
drivers/dma/idxd/device.c
1189
struct idxd_device *idxd = group->idxd;
drivers/dma/idxd/device.c
1200
grpcfg_offset = GRPWQCFG_OFFSET(idxd, group->id, i);
drivers/dma/idxd/device.c
1201
group->grpcfg.wqs[i] = ioread64(idxd->reg_base + grpcfg_offset);
drivers/dma/idxd/device.c
1203
group->id, i, grpcfg_offset, group->grpcfg.wqs[i]);
drivers/dma/idxd/device.c
1217
if (group->grpcfg.wqs[i] & BIT(j)) {
drivers/dma/idxd/device.c
1219
wq->group = group;
drivers/dma/idxd/device.c
1224
grpcfg_offset = GRPENGCFG_OFFSET(idxd, group->id);
drivers/dma/idxd/device.c
1225
group->grpcfg.engines = ioread64(idxd->reg_base + grpcfg_offset);
drivers/dma/idxd/device.c
1226
dev_dbg(dev, "GRPCFG engs[%d: %#x]: %#llx\n", group->id,
drivers/dma/idxd/device.c
1227
grpcfg_offset, group->grpcfg.engines);
drivers/dma/idxd/device.c
1234
if (group->grpcfg.engines & BIT(i)) {
drivers/dma/idxd/device.c
1237
engine->group = group;
drivers/dma/idxd/device.c
1241
grpcfg_offset = GRPFLGCFG_OFFSET(idxd, group->id);
drivers/dma/idxd/device.c
1242
group->grpcfg.flags.bits = ioread64(idxd->reg_base + grpcfg_offset);
drivers/dma/idxd/device.c
1244
group->id, grpcfg_offset, group->grpcfg.flags.bits);
drivers/dma/idxd/device.c
1256
struct idxd_group *group = idxd->groups[i];
drivers/dma/idxd/device.c
1258
idxd_group_load_config(group);
drivers/dma/idxd/device.c
1430
if (!wq->group) {
drivers/dma/idxd/device.c
403
wq->group = NULL;
drivers/dma/idxd/device.c
691
engine->group = NULL;
drivers/dma/idxd/device.c
697
struct idxd_group *group;
drivers/dma/idxd/device.c
702
group = idxd->groups[i];
drivers/dma/idxd/device.c
703
memset(&group->grpcfg, 0, sizeof(group->grpcfg));
drivers/dma/idxd/device.c
704
group->num_engines = 0;
drivers/dma/idxd/device.c
705
group->num_wqs = 0;
drivers/dma/idxd/device.c
706
group->use_rdbuf_limit = false;
drivers/dma/idxd/device.c
711
group->rdbufs_allowed = idxd->max_rdbufs;
drivers/dma/idxd/device.c
712
group->rdbufs_reserved = 0;
drivers/dma/idxd/device.c
714
group->tc_a = 1;
drivers/dma/idxd/device.c
715
group->tc_b = 1;
drivers/dma/idxd/device.c
717
group->tc_a = -1;
drivers/dma/idxd/device.c
718
group->tc_b = -1;
drivers/dma/idxd/device.c
720
group->desc_progress_limit = 0;
drivers/dma/idxd/device.c
721
group->batch_progress_limit = 0;
drivers/dma/idxd/device.c
859
static void idxd_group_config_write(struct idxd_group *group)
drivers/dma/idxd/device.c
861
struct idxd_device *idxd = group->idxd;
drivers/dma/idxd/device.c
866
dev_dbg(dev, "Writing group %d cfg registers\n", group->id);
drivers/dma/idxd/device.c
870
grpcfg_offset = GRPWQCFG_OFFSET(idxd, group->id, i);
drivers/dma/idxd/device.c
871
iowrite64(group->grpcfg.wqs[i], idxd->reg_base + grpcfg_offset);
drivers/dma/idxd/device.c
873
group->id, i, grpcfg_offset,
drivers/dma/idxd/device.c
878
grpcfg_offset = GRPENGCFG_OFFSET(idxd, group->id);
drivers/dma/idxd/device.c
879
iowrite64(group->grpcfg.engines, idxd->reg_base + grpcfg_offset);
drivers/dma/idxd/device.c
880
dev_dbg(dev, "GRPCFG engs[%d: %#x]: %#llx\n", group->id,
drivers/dma/idxd/device.c
884
grpcfg_offset = GRPFLGCFG_OFFSET(idxd, group->id);
drivers/dma/idxd/device.c
885
iowrite64(group->grpcfg.flags.bits, idxd->reg_base + grpcfg_offset);
drivers/dma/idxd/device.c
887
group->id, grpcfg_offset,
drivers/dma/idxd/device.c
909
struct idxd_group *group = idxd->groups[i];
drivers/dma/idxd/device.c
911
idxd_group_config_write(group);
drivers/dma/idxd/device.c
933
if (!wq->group)
drivers/dma/idxd/idxd.h
205
struct idxd_group *group;
drivers/dma/idxd/idxd.h
242
struct idxd_group *group;
drivers/dma/idxd/idxd.h
446
#define group_confdev(group) &group->idxd_dev.conf_dev
drivers/dma/idxd/init.c
1001
engine->group = saved_engine->group;
drivers/dma/idxd/init.c
1023
wq->group = saved_wq->group;
drivers/dma/idxd/init.c
339
struct idxd_group *group;
drivers/dma/idxd/init.c
343
group = idxd->groups[i];
drivers/dma/idxd/init.c
344
put_device(group_confdev(group));
drivers/dma/idxd/init.c
345
kfree(group);
drivers/dma/idxd/init.c
354
struct idxd_group *group;
drivers/dma/idxd/init.c
363
group = kzalloc_node(sizeof(*group), GFP_KERNEL, dev_to_node(dev));
drivers/dma/idxd/init.c
364
if (!group) {
drivers/dma/idxd/init.c
369
idxd_dev_set_type(&group->idxd_dev, IDXD_DEV_GROUP);
drivers/dma/idxd/init.c
370
conf_dev = group_confdev(group);
drivers/dma/idxd/init.c
371
group->id = i;
drivers/dma/idxd/init.c
372
group->idxd = idxd;
drivers/dma/idxd/init.c
377
rc = dev_set_name(conf_dev, "group%d.%d", idxd->id, group->id);
drivers/dma/idxd/init.c
380
kfree(group);
drivers/dma/idxd/init.c
384
idxd->groups[i] = group;
drivers/dma/idxd/init.c
386
group->tc_a = 1;
drivers/dma/idxd/init.c
387
group->tc_b = 1;
drivers/dma/idxd/init.c
389
group->tc_a = -1;
drivers/dma/idxd/init.c
390
group->tc_b = -1;
drivers/dma/idxd/init.c
396
group->rdbufs_allowed = idxd->max_rdbufs;
drivers/dma/idxd/init.c
403
group = idxd->groups[i];
drivers/dma/idxd/init.c
404
put_device(group_confdev(group));
drivers/dma/idxd/init.c
405
kfree(group);
drivers/dma/idxd/init.c
980
struct idxd_group *saved_group, *group;
drivers/dma/idxd/init.c
983
group = idxd->groups[i];
drivers/dma/idxd/init.c
985
group->rdbufs_allowed = saved_group->rdbufs_allowed;
drivers/dma/idxd/init.c
986
group->rdbufs_reserved = saved_group->rdbufs_reserved;
drivers/dma/idxd/init.c
987
group->tc_a = saved_group->tc_a;
drivers/dma/idxd/init.c
988
group->tc_b = saved_group->tc_b;
drivers/dma/idxd/init.c
989
group->use_rdbuf_limit = saved_group->use_rdbuf_limit;
drivers/dma/idxd/sysfs.c
119
struct idxd_group *group = confdev_to_group(dev);
drivers/dma/idxd/sysfs.c
121
return sysfs_emit(buf, "%u\n", group->rdbufs_reserved);
drivers/dma/idxd/sysfs.c
136
struct idxd_group *group = confdev_to_group(dev);
drivers/dma/idxd/sysfs.c
137
struct idxd_device *idxd = group->idxd;
drivers/dma/idxd/sysfs.c
157
if (val > idxd->nr_rdbufs + group->rdbufs_reserved)
drivers/dma/idxd/sysfs.c
160
group->rdbufs_reserved = val;
drivers/dma/idxd/sysfs.c
185
struct idxd_group *group = confdev_to_group(dev);
drivers/dma/idxd/sysfs.c
187
return sysfs_emit(buf, "%u\n", group->rdbufs_allowed);
drivers/dma/idxd/sysfs.c
1893
struct idxd_group *group;
drivers/dma/idxd/sysfs.c
1897
group = idxd->groups[i];
drivers/dma/idxd/sysfs.c
1898
rc = device_add(group_confdev(group));
drivers/dma/idxd/sysfs.c
1908
group = idxd->groups[i];
drivers/dma/idxd/sysfs.c
1909
put_device(group_confdev(group));
drivers/dma/idxd/sysfs.c
1913
group = idxd->groups[j];
drivers/dma/idxd/sysfs.c
1914
device_unregister(group_confdev(group));
drivers/dma/idxd/sysfs.c
2004
struct idxd_group *group = idxd->groups[i];
drivers/dma/idxd/sysfs.c
2006
device_unregister(group_confdev(group));
drivers/dma/idxd/sysfs.c
202
struct idxd_group *group = confdev_to_group(dev);
drivers/dma/idxd/sysfs.c
203
struct idxd_device *idxd = group->idxd;
drivers/dma/idxd/sysfs.c
220
if (val < 4 * group->num_engines ||
drivers/dma/idxd/sysfs.c
221
val > group->rdbufs_reserved + idxd->nr_rdbufs)
drivers/dma/idxd/sysfs.c
224
group->rdbufs_allowed = val;
drivers/dma/idxd/sysfs.c
248
struct idxd_group *group = confdev_to_group(dev);
drivers/dma/idxd/sysfs.c
25
if (engine->group)
drivers/dma/idxd/sysfs.c
250
return sysfs_emit(buf, "%u\n", group->use_rdbuf_limit);
drivers/dma/idxd/sysfs.c
26
return sysfs_emit(buf, "%d\n", engine->group->id);
drivers/dma/idxd/sysfs.c
265
struct idxd_group *group = confdev_to_group(dev);
drivers/dma/idxd/sysfs.c
266
struct idxd_device *idxd = group->idxd;
drivers/dma/idxd/sysfs.c
286
group->use_rdbuf_limit = !!val;
drivers/dma/idxd/sysfs.c
309
struct idxd_group *group = confdev_to_group(dev);
drivers/dma/idxd/sysfs.c
311
struct idxd_device *idxd = group->idxd;
drivers/dma/idxd/sysfs.c
316
if (!engine->group)
drivers/dma/idxd/sysfs.c
319
if (engine->group->id == group->id)
drivers/dma/idxd/sysfs.c
337
struct idxd_group *group = confdev_to_group(dev);
drivers/dma/idxd/sysfs.c
339
struct idxd_device *idxd = group->idxd;
drivers/dma/idxd/sysfs.c
344
if (!wq->group)
drivers/dma/idxd/sysfs.c
347
if (wq->group->id == group->id)
drivers/dma/idxd/sysfs.c
366
struct idxd_group *group = confdev_to_group(dev);
drivers/dma/idxd/sysfs.c
368
return sysfs_emit(buf, "%d\n", group->tc_a);
drivers/dma/idxd/sysfs.c
375
struct idxd_group *group = confdev_to_group(dev);
drivers/dma/idxd/sysfs.c
376
struct idxd_device *idxd = group->idxd;
drivers/dma/idxd/sysfs.c
396
group->tc_a = val;
drivers/dma/idxd/sysfs.c
408
struct idxd_group *group = confdev_to_group(dev);
drivers/dma/idxd/sysfs.c
410
return sysfs_emit(buf, "%d\n", group->tc_b);
drivers/dma/idxd/sysfs.c
417
struct idxd_group *group = confdev_to_group(dev);
drivers/dma/idxd/sysfs.c
418
struct idxd_device *idxd = group->idxd;
drivers/dma/idxd/sysfs.c
438
group->tc_b = val;
drivers/dma/idxd/sysfs.c
450
struct idxd_group *group = confdev_to_group(dev);
drivers/dma/idxd/sysfs.c
452
return sysfs_emit(buf, "%d\n", group->desc_progress_limit);
drivers/dma/idxd/sysfs.c
459
struct idxd_group *group = confdev_to_group(dev);
drivers/dma/idxd/sysfs.c
469
group->desc_progress_limit = val;
drivers/dma/idxd/sysfs.c
481
struct idxd_group *group = confdev_to_group(dev);
drivers/dma/idxd/sysfs.c
483
return sysfs_emit(buf, "%d\n", group->batch_progress_limit);
drivers/dma/idxd/sysfs.c
490
struct idxd_group *group = confdev_to_group(dev);
drivers/dma/idxd/sysfs.c
500
group->batch_progress_limit = val;
drivers/dma/idxd/sysfs.c
52
if (engine->group) {
drivers/dma/idxd/sysfs.c
53
engine->group->num_engines--;
drivers/dma/idxd/sysfs.c
54
engine->group = NULL;
drivers/dma/idxd/sysfs.c
551
struct idxd_group *group = confdev_to_group(dev);
drivers/dma/idxd/sysfs.c
552
struct idxd_device *idxd = group->idxd;
drivers/dma/idxd/sysfs.c
575
struct idxd_group *group = confdev_to_group(dev);
drivers/dma/idxd/sysfs.c
577
kfree(group);
drivers/dma/idxd/sysfs.c
59
prevg = engine->group;
drivers/dma/idxd/sysfs.c
621
if (wq->group)
drivers/dma/idxd/sysfs.c
622
return sysfs_emit(buf, "%u\n", wq->group->id);
drivers/dma/idxd/sysfs.c
63
engine->group = idxd->groups[id];
drivers/dma/idxd/sysfs.c
635
struct idxd_group *prevg, *group;
drivers/dma/idxd/sysfs.c
64
engine->group->num_engines++;
drivers/dma/idxd/sysfs.c
651
if (wq->group) {
drivers/dma/idxd/sysfs.c
652
wq->group->num_wqs--;
drivers/dma/idxd/sysfs.c
653
wq->group = NULL;
drivers/dma/idxd/sysfs.c
658
group = idxd->groups[id];
drivers/dma/idxd/sysfs.c
659
prevg = wq->group;
drivers/dma/idxd/sysfs.c
663
wq->group = group;
drivers/dma/idxd/sysfs.c
664
group->num_wqs++;
drivers/dma/ste_dma40.c
2314
u32 group = D40_TYPE_TO_GROUP(dev_type);
drivers/dma/ste_dma40.c
2337
writel(bit, d40c->base->virtbase + prioreg + group * 4);
drivers/dma/ste_dma40.c
2338
writel(bit, d40c->base->virtbase + rtreg + group * 4);
drivers/edac/ecs.c
165
struct attribute_group *group = &fru_ctx->group;
drivers/edac/ecs.c
179
group->name = fru_ctx->name;
drivers/edac/ecs.c
180
group->attrs = fru_ctx->ecs_attrs;
drivers/edac/ecs.c
181
group->is_visible = ecs_attr_visible;
drivers/edac/ecs.c
183
attr_groups[fru] = group;
drivers/edac/ecs.c
31
struct attribute_group group;
drivers/edac/mem_repair.c
315
struct attribute_group *group;
drivers/edac/mem_repair.c
330
group = &ctx->group;
drivers/edac/mem_repair.c
331
group->name = ctx->name;
drivers/edac/mem_repair.c
332
group->attrs = ctx->mem_repair_attrs;
drivers/edac/mem_repair.c
333
group->is_visible = mem_repair_attr_visible;
drivers/edac/mem_repair.c
334
attr_groups[0] = group;
drivers/edac/mem_repair.c
45
struct attribute_group group;
drivers/edac/scrub.c
161
struct attribute_group *group;
drivers/edac/scrub.c
176
group = &scrub_ctx->group;
drivers/edac/scrub.c
183
group->name = scrub_ctx->name;
drivers/edac/scrub.c
184
group->attrs = scrub_ctx->scrub_attrs;
drivers/edac/scrub.c
185
group->is_visible = scrub_attr_visible;
drivers/edac/scrub.c
187
attr_groups[0] = group;
drivers/edac/scrub.c
32
struct attribute_group group;
drivers/edac/versal_edac.c
165
u32 group:2;
drivers/edac/versal_edac.c
388
err_addr |= (pinf.group & BIT(0)) << priv->grp_bit[index];
drivers/edac/versal_edac.c
389
pinf.group >>= 1;
drivers/edac/versalnet_edac.c
365
pinf.group >>= MC5_EACHBIT;
drivers/edac/versalnet_edac.c
367
pinf.group >>= MC5_EACHBIT;
drivers/edac/versalnet_edac.c
369
pinf.group >>= MC5_EACHBIT;
drivers/edac/versalnet_edac.c
91
u32 group:3;
drivers/extcon/extcon-max14577.c
255
enum max14577_muic_cable_group group, bool *attached)
drivers/extcon/extcon-max14577.c
261
switch (group) {
drivers/extcon/extcon-max14577.c
313
dev_err(info->dev, "Unknown cable group (%d)\n", group);
drivers/extcon/extcon-max77693.c
310
enum max77693_muic_cable_group group, bool *attached)
drivers/extcon/extcon-max77693.c
319
switch (group) {
drivers/extcon/extcon-max77693.c
442
dev_err(info->dev, "Unknown cable group (%d)\n", group);
drivers/extcon/extcon-max77843.c
261
enum max77843_muic_cable_group group, bool *attached)
drivers/extcon/extcon-max77843.c
269
switch (group) {
drivers/extcon/extcon-max77843.c
360
dev_err(info->dev, "Unknown cable group (%d)\n", group);
drivers/extcon/extcon-max8997.c
247
enum max8997_muic_cable_group group, bool *attached)
drivers/extcon/extcon-max8997.c
253
switch (group) {
drivers/extcon/extcon-max8997.c
304
dev_err(info->dev, "Unknown cable group (%d)\n", group);
drivers/firewire/core-device.c
410
struct fw_attribute_group *group)
drivers/firewire/core-device.c
416
group->attrs[j] = &attrs[j].attr;
drivers/firewire/core-device.c
422
group->attrs[j++] = &attr->attr;
drivers/firewire/core-device.c
425
group->attrs[j] = NULL;
drivers/firewire/core-device.c
426
group->groups[0] = &group->group;
drivers/firewire/core-device.c
427
group->groups[1] = NULL;
drivers/firewire/core-device.c
428
group->group.attrs = group->attrs;
drivers/firewire/core-device.c
429
dev->groups = (const struct attribute_group **) group->groups;
drivers/firmware/arm_scmi/pinctrl.c
523
u32 group,
drivers/firmware/arm_scmi/pinctrl.c
532
ret = scmi_pinctrl_validate_id(ph, group, type);
drivers/firmware/arm_scmi/pinctrl.c
542
tx->identifier = cpu_to_le32(group);
drivers/firmware/arm_scmi/pinctrl.c
601
struct scmi_group_info *group;
drivers/firmware/arm_scmi/pinctrl.c
607
group = &pi->groups[selector];
drivers/firmware/arm_scmi/pinctrl.c
608
if (group->present)
drivers/firmware/arm_scmi/pinctrl.c
611
ret = scmi_pinctrl_attributes(ph, GROUP_TYPE, selector, group->name,
drivers/firmware/arm_scmi/pinctrl.c
612
&group->nr_pins);
drivers/firmware/arm_scmi/pinctrl.c
616
if (!group->nr_pins) {
drivers/firmware/arm_scmi/pinctrl.c
621
group->group_pins = kmalloc_array(group->nr_pins,
drivers/firmware/arm_scmi/pinctrl.c
622
sizeof(*group->group_pins),
drivers/firmware/arm_scmi/pinctrl.c
624
if (!group->group_pins)
drivers/firmware/arm_scmi/pinctrl.c
628
group->nr_pins, group->group_pins);
drivers/firmware/arm_scmi/pinctrl.c
630
kfree(group->group_pins);
drivers/firmware/arm_scmi/pinctrl.c
634
group->present = true;
drivers/firmware/arm_scmi/pinctrl.c
755
u32 selector, u32 group)
drivers/firmware/arm_scmi/pinctrl.c
757
return scmi_pinctrl_function_select(ph, group, GROUP_TYPE, selector);
drivers/firmware/imx/imx-scu-irq.c
117
int imx_scu_irq_get_status(u8 group, u32 *irq_status)
drivers/firmware/imx/imx-scu-irq.c
129
msg.data.req.group = group;
drivers/firmware/imx/imx-scu-irq.c
142
int imx_scu_irq_group_enable(u8 group, u32 mask, u8 enable)
drivers/firmware/imx/imx-scu-irq.c
157
msg.group = group;
drivers/firmware/imx/imx-scu-irq.c
164
group, mask, ret);
drivers/firmware/imx/imx-scu-irq.c
167
scu_irq_wakeup[group].mask |= mask;
drivers/firmware/imx/imx-scu-irq.c
169
scu_irq_wakeup[group].mask &= ~mask;
drivers/firmware/imx/imx-scu-irq.c
29
u8 group;
drivers/firmware/imx/imx-scu-irq.c
42
u8 group;
drivers/firmware/imx/imx-scu-irq.c
78
static int imx_scu_irq_notifier_call_chain(unsigned long status, u8 *group)
drivers/firmware/imx/imx-scu-irq.c
81
status, (void *)group);
drivers/gpio/gpio-aggregator.c
1004
subsys, &line->group.cg_item);
drivers/gpio/gpio-aggregator.c
1007
&line->group.cg_item);
drivers/gpio/gpio-aggregator.c
1254
gpio_aggregator_device_make_group(struct config_group *group, const char *name)
drivers/gpio/gpio-aggregator.c
1256
struct gpio_aggregator *aggr = to_gpio_aggregator(&group->cg_item);
drivers/gpio/gpio-aggregator.c
1287
config_group_init_type_name(&line->group, name, &gpio_aggregator_line_type);
drivers/gpio/gpio-aggregator.c
1291
return &line->group;
drivers/gpio/gpio-aggregator.c
1306
gpio_aggregator_make_group(struct config_group *group, const char *name)
drivers/gpio/gpio-aggregator.c
1324
config_group_init_type_name(&aggr->group, name, &gpio_aggregator_device_type);
drivers/gpio/gpio-aggregator.c
1327
return &aggr->group;
drivers/gpio/gpio-aggregator.c
1378
config_group_init_type_name(&line->group, name,
drivers/gpio/gpio-aggregator.c
1380
error = configfs_register_group(&aggr->group,
drivers/gpio/gpio-aggregator.c
1381
&line->group);
drivers/gpio/gpio-aggregator.c
1409
config_group_init_type_name(&line->group, name,
drivers/gpio/gpio-aggregator.c
1411
error = configfs_register_group(&aggr->group,
drivers/gpio/gpio-aggregator.c
1412
&line->group);
drivers/gpio/gpio-aggregator.c
1472
config_group_init_type_name(&aggr->group, name, &gpio_aggregator_device_type);
drivers/gpio/gpio-aggregator.c
1482
&aggr->group);
drivers/gpio/gpio-aggregator.c
1505
configfs_unregister_group(&aggr->group);
drivers/gpio/gpio-aggregator.c
1528
configfs_unregister_group(&aggr->group);
drivers/gpio/gpio-aggregator.c
212
configfs_unregister_group(&line->group);
drivers/gpio/gpio-aggregator.c
46
struct config_group group;
drivers/gpio/gpio-aggregator.c
60
struct config_group group;
drivers/gpio/gpio-aggregator.c
865
struct config_group *group = to_config_group(item);
drivers/gpio/gpio-aggregator.c
867
return container_of(group, struct gpio_aggregator, group);
drivers/gpio/gpio-aggregator.c
873
struct config_group *group = to_config_group(item);
drivers/gpio/gpio-aggregator.c
875
return container_of(group, struct gpio_aggregator_line, group);
drivers/gpio/gpio-aggregator.c
993
struct configfs_subsystem *subsys = aggr->group.cg_subsys;
drivers/gpio/gpio-cs5535.c
148
int cs5535_gpio_set_irq(unsigned group, unsigned irq)
drivers/gpio/gpio-cs5535.c
152
if (group > 7 || irq > 15)
drivers/gpio/gpio-cs5535.c
157
lo &= ~(0xF << (group * 4));
drivers/gpio/gpio-cs5535.c
158
lo |= (irq & 0xF) << (group * 4);
drivers/gpio/gpio-it87.c
149
u8 mask, group;
drivers/gpio/gpio-it87.c
154
group = (gpio_num / 8);
drivers/gpio/gpio-it87.c
165
if (group < it87_gpio->simple_size)
drivers/gpio/gpio-it87.c
166
superio_set_mask(mask, group + it87_gpio->simple_base);
drivers/gpio/gpio-it87.c
171
superio_clear_mask(mask, group + it87_gpio->output_base);
drivers/gpio/gpio-it87.c
194
u8 mask, group;
drivers/gpio/gpio-it87.c
199
group = (gpio_num / 8);
drivers/gpio/gpio-it87.c
208
superio_clear_mask(mask, group + it87_gpio->output_base);
drivers/gpio/gpio-it87.c
238
u8 mask, group;
drivers/gpio/gpio-it87.c
243
group = (gpio_num / 8);
drivers/gpio/gpio-it87.c
252
superio_set_mask(mask, group + it87_gpio->output_base);
drivers/gpio/gpio-lpc32xx.c
168
static inline u32 gpreg_read(struct lpc32xx_gpio_chip *group, unsigned long offset)
drivers/gpio/gpio-lpc32xx.c
170
return __raw_readl(group->reg_base + offset);
drivers/gpio/gpio-lpc32xx.c
173
static inline void gpreg_write(struct lpc32xx_gpio_chip *group, u32 val, unsigned long offset)
drivers/gpio/gpio-lpc32xx.c
175
__raw_writel(val, group->reg_base + offset);
drivers/gpio/gpio-lpc32xx.c
178
static void __set_gpio_dir_p012(struct lpc32xx_gpio_chip *group,
drivers/gpio/gpio-lpc32xx.c
182
gpreg_write(group, GPIO012_PIN_TO_BIT(pin),
drivers/gpio/gpio-lpc32xx.c
183
group->gpio_grp->dir_clr);
drivers/gpio/gpio-lpc32xx.c
185
gpreg_write(group, GPIO012_PIN_TO_BIT(pin),
drivers/gpio/gpio-lpc32xx.c
186
group->gpio_grp->dir_set);
drivers/gpio/gpio-lpc32xx.c
189
static void __set_gpio_dir_p3(struct lpc32xx_gpio_chip *group,
drivers/gpio/gpio-lpc32xx.c
195
gpreg_write(group, u, group->gpio_grp->dir_clr);
drivers/gpio/gpio-lpc32xx.c
197
gpreg_write(group, u, group->gpio_grp->dir_set);
drivers/gpio/gpio-lpc32xx.c
200
static void __set_gpio_level_p012(struct lpc32xx_gpio_chip *group,
drivers/gpio/gpio-lpc32xx.c
204
gpreg_write(group, GPIO012_PIN_TO_BIT(pin),
drivers/gpio/gpio-lpc32xx.c
205
group->gpio_grp->outp_set);
drivers/gpio/gpio-lpc32xx.c
207
gpreg_write(group, GPIO012_PIN_TO_BIT(pin),
drivers/gpio/gpio-lpc32xx.c
208
group->gpio_grp->outp_clr);
drivers/gpio/gpio-lpc32xx.c
211
static void __set_gpio_level_p3(struct lpc32xx_gpio_chip *group,
drivers/gpio/gpio-lpc32xx.c
217
gpreg_write(group, u, group->gpio_grp->outp_set);
drivers/gpio/gpio-lpc32xx.c
219
gpreg_write(group, u, group->gpio_grp->outp_clr);
drivers/gpio/gpio-lpc32xx.c
222
static void __set_gpo_level_p3(struct lpc32xx_gpio_chip *group,
drivers/gpio/gpio-lpc32xx.c
226
gpreg_write(group, GPO3_PIN_TO_BIT(pin), group->gpio_grp->outp_set);
drivers/gpio/gpio-lpc32xx.c
228
gpreg_write(group, GPO3_PIN_TO_BIT(pin), group->gpio_grp->outp_clr);
drivers/gpio/gpio-lpc32xx.c
231
static int __get_gpio_state_p012(struct lpc32xx_gpio_chip *group,
drivers/gpio/gpio-lpc32xx.c
234
return GPIO012_PIN_IN_SEL(gpreg_read(group, group->gpio_grp->inp_state),
drivers/gpio/gpio-lpc32xx.c
238
static int __get_gpio_state_p3(struct lpc32xx_gpio_chip *group,
drivers/gpio/gpio-lpc32xx.c
241
int state = gpreg_read(group, group->gpio_grp->inp_state);
drivers/gpio/gpio-lpc32xx.c
250
static int __get_gpi_state_p3(struct lpc32xx_gpio_chip *group,
drivers/gpio/gpio-lpc32xx.c
253
return GPI3_PIN_IN_SEL(gpreg_read(group, group->gpio_grp->inp_state), pin);
drivers/gpio/gpio-lpc32xx.c
256
static int __get_gpo_state_p3(struct lpc32xx_gpio_chip *group,
drivers/gpio/gpio-lpc32xx.c
259
return GPO3_PIN_IN_SEL(gpreg_read(group, group->gpio_grp->outp_state), pin);
drivers/gpio/gpio-lpc32xx.c
268
struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip);
drivers/gpio/gpio-lpc32xx.c
270
__set_gpio_dir_p012(group, pin, 1);
drivers/gpio/gpio-lpc32xx.c
278
struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip);
drivers/gpio/gpio-lpc32xx.c
280
__set_gpio_dir_p3(group, pin, 1);
drivers/gpio/gpio-lpc32xx.c
293
struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip);
drivers/gpio/gpio-lpc32xx.c
295
return !!__get_gpio_state_p012(group, pin);
drivers/gpio/gpio-lpc32xx.c
300
struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip);
drivers/gpio/gpio-lpc32xx.c
302
return !!__get_gpio_state_p3(group, pin);
drivers/gpio/gpio-lpc32xx.c
307
struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip);
drivers/gpio/gpio-lpc32xx.c
309
return !!__get_gpi_state_p3(group, pin);
drivers/gpio/gpio-lpc32xx.c
315
struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip);
drivers/gpio/gpio-lpc32xx.c
317
__set_gpio_level_p012(group, pin, value);
drivers/gpio/gpio-lpc32xx.c
318
__set_gpio_dir_p012(group, pin, 0);
drivers/gpio/gpio-lpc32xx.c
326
struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip);
drivers/gpio/gpio-lpc32xx.c
328
__set_gpio_level_p3(group, pin, value);
drivers/gpio/gpio-lpc32xx.c
329
__set_gpio_dir_p3(group, pin, 0);
drivers/gpio/gpio-lpc32xx.c
337
struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip);
drivers/gpio/gpio-lpc32xx.c
339
__set_gpo_level_p3(group, pin, value);
drivers/gpio/gpio-lpc32xx.c
346
struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip);
drivers/gpio/gpio-lpc32xx.c
348
__set_gpio_level_p012(group, pin, value);
drivers/gpio/gpio-lpc32xx.c
356
struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip);
drivers/gpio/gpio-lpc32xx.c
358
__set_gpio_level_p3(group, pin, value);
drivers/gpio/gpio-lpc32xx.c
366
struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip);
drivers/gpio/gpio-lpc32xx.c
368
__set_gpo_level_p3(group, pin, value);
drivers/gpio/gpio-lpc32xx.c
375
struct lpc32xx_gpio_chip *group = gpiochip_get_data(chip);
drivers/gpio/gpio-lpc32xx.c
377
return !!__get_gpo_state_p3(group, pin);
drivers/gpio/gpio-nct6694.c
117
cmd_hd.offset = cpu_to_le16(NCT6694_GPO_DATA + data->group);
drivers/gpio/gpio-nct6694.c
135
.offset = cpu_to_le16(NCT6694_GPO_DIR + data->group),
drivers/gpio/gpio-nct6694.c
147
cmd_hd.offset = cpu_to_le16(NCT6694_GPO_DATA + data->group);
drivers/gpio/gpio-nct6694.c
155
cmd_hd.offset = cpu_to_le16(NCT6694_GPI_DATA + data->group);
drivers/gpio/gpio-nct6694.c
169
.offset = cpu_to_le16(NCT6694_GPO_DATA + data->group),
drivers/gpio/gpio-nct6694.c
194
.offset = cpu_to_le16(NCT6694_GPO_TYPE + data->group),
drivers/gpio/gpio-nct6694.c
226
.offset = cpu_to_le16(NCT6694_GPIO_VALID + data->group),
drivers/gpio/gpio-nct6694.c
247
.offset = cpu_to_le16(NCT6694_GPI_STS + data->group),
drivers/gpio/gpio-nct6694.c
267
cmd_hd.offset = cpu_to_le16(NCT6694_GPI_CLR + data->group);
drivers/gpio/gpio-nct6694.c
278
.offset = cpu_to_le16(NCT6694_GPI_FALLING + data->group),
drivers/gpio/gpio-nct6694.c
289
cmd_hd.offset = cpu_to_le16(NCT6694_GPI_RISING + data->group);
drivers/gpio/gpio-nct6694.c
352
.offset = cpu_to_le16(NCT6694_GPI_FALLING + data->group),
drivers/gpio/gpio-nct6694.c
359
cmd_hd.offset = cpu_to_le16(NCT6694_GPI_RISING + data->group);
drivers/gpio/gpio-nct6694.c
389
ida_free(&nct6694->gpio_ida, data->group);
drivers/gpio/gpio-nct6694.c
410
data->group = ret;
drivers/gpio/gpio-nct6694.c
423
data->group, i);
drivers/gpio/gpio-nct6694.c
429
NCT6694_IRQ_GPIO0 + data->group);
drivers/gpio/gpio-nct6694.c
49
unsigned char group;
drivers/gpio/gpio-nct6694.c
58
.offset = cpu_to_le16(NCT6694_GPO_DIR + data->group),
drivers/gpio/gpio-nct6694.c
77
.offset = cpu_to_le16(NCT6694_GPO_DIR + data->group),
drivers/gpio/gpio-nct6694.c
99
.offset = cpu_to_le16(NCT6694_GPO_DIR + data->group),
drivers/gpio/gpio-sim.c
1017
struct configfs_subsystem *subsys = dev->group.cg_subsys;
drivers/gpio/gpio-sim.c
1030
: &line->group.cg_item;
drivers/gpio/gpio-sim.c
1398
gpio_sim_line_config_make_hog_item(struct config_group *group, const char *name)
drivers/gpio/gpio-sim.c
1400
struct gpio_sim_line *line = to_gpio_sim_line(&group->cg_item);
drivers/gpio/gpio-sim.c
1452
gpio_sim_bank_config_make_line_group(struct config_group *group,
drivers/gpio/gpio-sim.c
1455
struct gpio_sim_bank *bank = to_gpio_sim_bank(&group->cg_item);
drivers/gpio/gpio-sim.c
1474
config_group_init_type_name(&line->group, name,
drivers/gpio/gpio-sim.c
1482
return &line->group;
drivers/gpio/gpio-sim.c
1513
gpio_sim_device_config_make_bank_group(struct config_group *group,
drivers/gpio/gpio-sim.c
1516
struct gpio_sim_device *dev = to_gpio_sim_device(&group->cg_item);
drivers/gpio/gpio-sim.c
1528
config_group_init_type_name(&bank->group, name,
drivers/gpio/gpio-sim.c
1535
return &bank->group;
drivers/gpio/gpio-sim.c
1568
gpio_sim_config_make_device_group(struct config_group *group, const char *name)
drivers/gpio/gpio-sim.c
1580
config_group_init_type_name(&dev->group, name,
drivers/gpio/gpio-sim.c
1588
return &no_free_ptr(dev)->group;
drivers/gpio/gpio-sim.c
549
struct config_group group;
drivers/gpio/gpio-sim.c
571
struct config_group *group = to_config_group(item);
drivers/gpio/gpio-sim.c
573
return container_of(group, struct gpio_sim_device, group);
drivers/gpio/gpio-sim.c
577
struct config_group group;
drivers/gpio/gpio-sim.c
605
struct config_group *group = to_config_group(item);
drivers/gpio/gpio-sim.c
607
return container_of(group, struct gpio_sim_bank, group);
drivers/gpio/gpio-sim.c
624
struct config_group group;
drivers/gpio/gpio-sim.c
639
struct config_group *group = to_config_group(item);
drivers/gpio/gpio-sim.c
641
return container_of(group, struct gpio_sim_line, group);
drivers/gpio/gpio-tegra186.c
106
const char *group;
drivers/gpio/gpio-tegra186.c
472
const char *group = gpio->soc->pin_ranges[i].group;
drivers/gpio/gpio-tegra186.c
479
port, group);
drivers/gpio/gpio-tegra186.c
486
err = gpiochip_add_pingroup_range(chip, pctldev, pin, group);
drivers/gpio/gpio-virtuser.c
1009
struct config_group group;
drivers/gpio/gpio-virtuser.c
1022
struct config_group *group = to_config_group(item);
drivers/gpio/gpio-virtuser.c
1024
return container_of(group, struct gpio_virtuser_lookup, group);
drivers/gpio/gpio-virtuser.c
1028
struct config_group group;
drivers/gpio/gpio-virtuser.c
1042
struct config_group *group = to_config_group(item);
drivers/gpio/gpio-virtuser.c
1044
return container_of(group, struct gpio_virtuser_lookup_entry, group);
drivers/gpio/gpio-virtuser.c
1504
struct configfs_subsystem *subsys = dev->group.cg_subsys;
drivers/gpio/gpio-virtuser.c
1517
subsys, &entry->group.cg_item));
drivers/gpio/gpio-virtuser.c
1520
&entry->group.cg_item);
drivers/gpio/gpio-virtuser.c
1595
gpio_virtuser_make_lookup_entry_group(struct config_group *group,
drivers/gpio/gpio-virtuser.c
1599
to_gpio_virtuser_lookup(&group->cg_item);
drivers/gpio/gpio-virtuser.c
1612
config_group_init_type_name(&entry->group, name,
drivers/gpio/gpio-virtuser.c
1618
return &no_free_ptr(entry)->group;
drivers/gpio/gpio-virtuser.c
1650
gpio_virtuser_make_lookup_group(struct config_group *group, const char *name)
drivers/gpio/gpio-virtuser.c
1653
to_gpio_virtuser_device(&group->cg_item);
drivers/gpio/gpio-virtuser.c
1672
config_group_init_type_name(&lookup->group, name,
drivers/gpio/gpio-virtuser.c
1678
return &no_free_ptr(lookup)->group;
drivers/gpio/gpio-virtuser.c
1711
gpio_virtuser_config_make_device_group(struct config_group *group,
drivers/gpio/gpio-virtuser.c
1722
config_group_init_type_name(&dev->group, name,
drivers/gpio/gpio-virtuser.c
1728
return &no_free_ptr(dev)->group;
drivers/gpio/gpio-virtuser.c
982
struct config_group group;
drivers/gpio/gpio-virtuser.c
995
struct config_group *group = to_config_group(item);
drivers/gpio/gpio-virtuser.c
997
return container_of(group, struct gpio_virtuser_device, group);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2061
struct attribute_group group = {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2067
sysfs_remove_group(&adev->dev->kobj, &group);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2300
struct attribute_group group = {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2316
group.attrs = attrs;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2339
group.bin_attrs = bin_attrs;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2342
r = sysfs_create_group(&adev->dev->kobj, &group);
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2807
uint32_t group, start_group, end_group;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2825
for (group = start_group; group < end_group; group++) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2827
cmd_input.group_index = group;
drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c
2835
pos_in_group = pos - group * RAS_CMD_MAX_BAD_PAGES_PER_GROUP;
drivers/gpu/drm/amd/include/kgd_pp_interface.h
347
#define PP_CG_MSG_ID(group, block, support, state) \
drivers/gpu/drm/amd/include/kgd_pp_interface.h
348
((group) << PP_GROUP_SHIFT | (block) << PP_BLOCK_SHIFT | \
drivers/gpu/drm/display/drm_dp_tunnel.c
1075
struct drm_dp_tunnel_mgr *mgr = tunnel->group->mgr;
drivers/gpu/drm/display/drm_dp_tunnel.c
1142
DPTUN_BW_ARG(group_allocated_bw(tunnel->group)),
drivers/gpu/drm/display/drm_dp_tunnel.c
1143
DPTUN_BW_ARG(tunnel->group->available_bw));
drivers/gpu/drm/display/drm_dp_tunnel.c
1277
DPTUN_BW_ARG(group_allocated_bw(tunnel->group)),
drivers/gpu/drm/display/drm_dp_tunnel.c
1278
DPTUN_BW_ARG(tunnel->group->available_bw));
drivers/gpu/drm/display/drm_dp_tunnel.c
130
struct drm_dp_tunnel_group *group;
drivers/gpu/drm/display/drm_dp_tunnel.c
1367
return tunnel->group->available_bw;
drivers/gpu/drm/display/drm_dp_tunnel.c
1377
&tunnel->group->base);
drivers/gpu/drm/display/drm_dp_tunnel.c
1551
if (to_group(old_group_state->base.obj) == tunnel->group)
drivers/gpu/drm/display/drm_dp_tunnel.c
1575
if (to_group(new_group_state->base.obj) == tunnel->group)
drivers/gpu/drm/display/drm_dp_tunnel.c
1582
static bool init_group(struct drm_dp_tunnel_mgr *mgr, struct drm_dp_tunnel_group *group)
drivers/gpu/drm/display/drm_dp_tunnel.c
1592
group->mgr = mgr;
drivers/gpu/drm/display/drm_dp_tunnel.c
1593
group->available_bw = -1;
drivers/gpu/drm/display/drm_dp_tunnel.c
1594
INIT_LIST_HEAD(&group->tunnels);
drivers/gpu/drm/display/drm_dp_tunnel.c
1596
drm_atomic_private_obj_init(mgr->dev, &group->base, &group_state->base,
drivers/gpu/drm/display/drm_dp_tunnel.c
1602
static void cleanup_group(struct drm_dp_tunnel_group *group)
drivers/gpu/drm/display/drm_dp_tunnel.c
1604
drm_atomic_private_obj_fini(&group->base);
drivers/gpu/drm/display/drm_dp_tunnel.c
1710
if (drm_WARN_ON(tunnel->group->mgr->dev,
drivers/gpu/drm/display/drm_dp_tunnel.c
1802
struct drm_dp_tunnel_group *group = to_group(new_group_state->base.obj);
drivers/gpu/drm/display/drm_dp_tunnel.c
1828
tun_grp_dbg(group,
drivers/gpu/drm/display/drm_dp_tunnel.c
1830
group_bw > group->available_bw ? "Not enough BW: " : "",
drivers/gpu/drm/display/drm_dp_tunnel.c
1832
DPTUN_BW_ARG(group->available_bw));
drivers/gpu/drm/display/drm_dp_tunnel.c
1834
if (group_bw > group->available_bw) {
drivers/gpu/drm/display/drm_dp_tunnel.c
308
static const char *drm_dp_tunnel_group_name(const struct drm_dp_tunnel_group *group)
drivers/gpu/drm/display/drm_dp_tunnel.c
310
return group->name;
drivers/gpu/drm/display/drm_dp_tunnel.c
316
struct drm_dp_tunnel_group *group = NULL;
drivers/gpu/drm/display/drm_dp_tunnel.c
328
if (!group && !mgr->groups[i].active)
drivers/gpu/drm/display/drm_dp_tunnel.c
329
group = &mgr->groups[i];
drivers/gpu/drm/display/drm_dp_tunnel.c
332
if (!group) {
drivers/gpu/drm/display/drm_dp_tunnel.c
338
group->drv_group_id = drv_group_id;
drivers/gpu/drm/display/drm_dp_tunnel.c
339
group->active = true;
drivers/gpu/drm/display/drm_dp_tunnel.c
345
snprintf(group->name, sizeof(group->name), "%d:%d:*",
drivers/gpu/drm/display/drm_dp_tunnel.c
349
return group;
drivers/gpu/drm/display/drm_dp_tunnel.c
352
static void free_group(struct drm_dp_tunnel_group *group)
drivers/gpu/drm/display/drm_dp_tunnel.c
354
struct drm_dp_tunnel_mgr *mgr = group->mgr;
drivers/gpu/drm/display/drm_dp_tunnel.c
356
if (drm_WARN_ON(mgr->dev, !list_empty(&group->tunnels)))
drivers/gpu/drm/display/drm_dp_tunnel.c
359
group->drv_group_id = 0;
drivers/gpu/drm/display/drm_dp_tunnel.c
360
group->available_bw = -1;
drivers/gpu/drm/display/drm_dp_tunnel.c
361
group->active = false;
drivers/gpu/drm/display/drm_dp_tunnel.c
375
struct drm_dp_tunnel_group *group = tunnel->group;
drivers/gpu/drm/display/drm_dp_tunnel.c
378
if (list_empty(&group->tunnels))
drivers/gpu/drm/display/drm_dp_tunnel.c
379
free_group(group);
drivers/gpu/drm/display/drm_dp_tunnel.c
393
ref_tracker_alloc(&tunnel->group->mgr->ref_tracker,
drivers/gpu/drm/display/drm_dp_tunnel.c
400
ref_tracker_free(&tunnel->group->mgr->ref_tracker,
drivers/gpu/drm/display/drm_dp_tunnel.c
459
struct drm_dp_tunnel_group *group;
drivers/gpu/drm/display/drm_dp_tunnel.c
461
group = lookup_or_alloc_group(mgr, drv_group_id);
drivers/gpu/drm/display/drm_dp_tunnel.c
462
if (!group)
drivers/gpu/drm/display/drm_dp_tunnel.c
465
tunnel->group = group;
drivers/gpu/drm/display/drm_dp_tunnel.c
466
list_add(&tunnel->node, &group->tunnels);
drivers/gpu/drm/display/drm_dp_tunnel.c
61
drm_##__level##__type((__tunnel)->group->mgr->dev, \
drivers/gpu/drm/display/drm_dp_tunnel.c
637
if (tunnel->group->drv_group_id != new_drv_group_id) {
drivers/gpu/drm/display/drm_dp_tunnel.c
640
tunnel_group_drv_id(tunnel->group->drv_group_id),
drivers/gpu/drm/display/drm_dp_tunnel.c
641
tunnel_group_id(tunnel->group->drv_group_id),
drivers/gpu/drm/display/drm_dp_tunnel.c
693
if (!tunnel_regs_are_valid(tunnel->group->mgr, regs, flags))
drivers/gpu/drm/display/drm_dp_tunnel.c
743
return min(get_max_dprx_bw(tunnel), tunnel->group->available_bw);
drivers/gpu/drm/display/drm_dp_tunnel.c
815
if (drm_WARN_ON(tunnel->group->mgr->dev, tunnel->destroyed))
drivers/gpu/drm/display/drm_dp_tunnel.c
838
static int group_allocated_bw(struct drm_dp_tunnel_group *group)
drivers/gpu/drm/display/drm_dp_tunnel.c
843
for_each_tunnel_in_group(group, tunnel) {
drivers/gpu/drm/display/drm_dp_tunnel.c
864
return group_allocated_bw(tunnel->group) +
drivers/gpu/drm/display/drm_dp_tunnel.c
877
if (calc_group_available_bw(tunnel) == tunnel->group->available_bw)
drivers/gpu/drm/display/drm_dp_tunnel.c
880
for_each_tunnel_in_group(tunnel->group, tunnel_iter) {
drivers/gpu/drm/display/drm_dp_tunnel.c
902
DPTUN_BW_ARG(tunnel->group->available_bw),
drivers/gpu/drm/display/drm_dp_tunnel.c
905
changed = tunnel->group->available_bw != group_available_bw;
drivers/gpu/drm/display/drm_dp_tunnel.c
907
tunnel->group->available_bw = group_available_bw;
drivers/gpu/drm/display/drm_dp_tunnel.c
958
if (!tunnel_group_id(tunnel->group->drv_group_id))
drivers/gpu/drm/display/drm_dp_tunnel.c
983
if (tunnel->group->available_bw == -1) {
drivers/gpu/drm/display/drm_dp_tunnel.c
992
DPTUN_BW_ARG(group_allocated_bw(tunnel->group)),
drivers/gpu/drm/display/drm_dp_tunnel.c
993
DPTUN_BW_ARG(tunnel->group->available_bw));
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
209
int group, int instance, u32 value)
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
225
REG_FIELD_PREP(MTL_MCR_GROUPID, group) |
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
230
mcr_ss = GEN11_MCR_SLICE(group) | GEN11_MCR_SUBSLICE(instance);
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
255
mcr_ss = GEN8_MCR_SLICE(group) | GEN8_MCR_SUBSLICE(instance);
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
286
int group, int instance,
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
304
val = rw_with_mcr_steering_fw(gt, reg, rw_flag, group, instance, value);
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
438
int group, int instance)
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
440
return rw_with_mcr_steering(gt, reg, FW_REG_READ, group, instance, 0);
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
457
int group, int instance)
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
459
rw_with_mcr_steering(gt, reg, FW_REG_WRITE, group, instance, value);
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
595
u8 *group, u8 *instance)
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
601
*group = 0; /* unused */
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
606
*group = __ffs(gt->info.mslice_mask);
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
615
*group = __ffs(gt->info.mslice_mask) << 1;
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
619
*group = IS_DG2(gt->i915) ? 1 : 0;
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
624
*group = dss / GEN_DSS_PER_GSLICE;
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
632
*group = 0;
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
637
*group = 0;
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
639
*group = 1;
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
644
*group = 0;
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
664
u8 *group, u8 *instance)
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
670
get_nonterminated_steering(gt, type, group, instance);
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
675
*group = gt->default_steering.groupid;
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
697
u8 group, instance;
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
703
get_nonterminated_steering(gt, type, &group, &instance);
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
706
group, instance, 0);
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
728
u8 group, instance;
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
732
get_nonterminated_steering(gt, type, &group, &instance);
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
735
group, instance, 0);
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
748
u8 group, instance;
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
758
get_nonterminated_steering(gt, type, &group, &instance);
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
760
intel_steering_types[type], group, instance);
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
802
unsigned int *group, unsigned int *instance)
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
805
*group = dss / GEN_DSS_PER_GSLICE;
drivers/gpu/drm/i915/gt/intel_gt_mcr.c
808
*group = dss / GEN_MAX_SS_PER_HSW_SLICE;
drivers/gpu/drm/i915/gt/intel_gt_mcr.h
18
int group, int instance);
drivers/gpu/drm/i915/gt/intel_gt_mcr.h
24
int group, int instance);
drivers/gpu/drm/i915/gt/intel_gt_mcr.h
35
u8 *group, u8 *instance);
drivers/gpu/drm/i915/gt/intel_gt_mcr.h
41
unsigned int *group, unsigned int *instance);
drivers/gpu/drm/i915/gt/intel_sseu.c
674
if (gt->perf.group && gt->perf.group[PERF_GROUP_OAG].exclusive_stream)
drivers/gpu/drm/i915/gt/uc/intel_guc_ads.c
343
#define GUC_REGSET_STEERING(group, instance) ( \
drivers/gpu/drm/i915/gt/uc/intel_guc_ads.c
344
FIELD_PREP(GUC_REGSET_STEERING_GROUP, (group)) | \
drivers/gpu/drm/i915/gt/uc/intel_guc_ads.c
353
u8 group, inst;
drivers/gpu/drm/i915/gt/uc/intel_guc_ads.c
363
intel_gt_mcr_get_nonterminated_steering(gt, reg, &group, &inst);
drivers/gpu/drm/i915/gt/uc/intel_guc_ads.c
364
flags |= GUC_REGSET_STEERING(group, inst);
drivers/gpu/drm/i915/gvt/interrupt.c
363
if (info->group != map->down_irq_group)
drivers/gpu/drm/i915/gvt/interrupt.c
383
if (up_irq_info->group == INTEL_GVT_IRQ_INFO_MASTER) {
drivers/gpu/drm/i915/gvt/interrupt.c
416
up_info->group, up_bit,
drivers/gpu/drm/i915/gvt/interrupt.c
417
down_info->group, map->down_irq_bitmask);
drivers/gpu/drm/i915/gvt/interrupt.c
47
int group;
drivers/gpu/drm/i915/gvt/interrupt.c
568
(i)->group = g; \
drivers/gpu/drm/i915/i915_perf.c
4890
struct i915_perf_group *g = >->perf.group[i];
drivers/gpu/drm/i915/i915_perf.c
4928
gt->perf.group = g;
drivers/gpu/drm/i915/i915_perf.c
5205
kfree(gt->perf.group);
drivers/gpu/drm/i915/i915_perf_types.h
455
struct i915_perf_group *group;
drivers/gpu/drm/nouveau/include/nvkm/subdev/bios/M0203.h
25
u8 group;
drivers/gpu/drm/nouveau/nvkm/subdev/bios/M0203.c
94
info->group = (nvbios_rd08(bios, data + 0x01) & 0x0f) >> 0;
drivers/gpu/drm/nouveau/nvkm/subdev/bios/ramcfg.c
70
return M0203E.group;
drivers/gpu/drm/panthor/panthor_sched.c
1001
struct panthor_device *ptdev = group->ptdev;
drivers/gpu/drm/panthor/panthor_sched.c
1007
if (drm_WARN_ON(&ptdev->base, group->csg_id != -1 || csg_id >= MAX_CSGS ||
drivers/gpu/drm/panthor/panthor_sched.c
1008
ptdev->scheduler->csg_slots[csg_id].group))
drivers/gpu/drm/panthor/panthor_sched.c
1011
ret = panthor_vm_active(group->vm);
drivers/gpu/drm/panthor/panthor_sched.c
1016
group_get(group);
drivers/gpu/drm/panthor/panthor_sched.c
1017
group->csg_id = csg_id;
drivers/gpu/drm/panthor/panthor_sched.c
1026
for (u32 i = 0; i < group->queue_count; i++)
drivers/gpu/drm/panthor/panthor_sched.c
1027
group->queues[i]->doorbell_id = csg_id + 1;
drivers/gpu/drm/panthor/panthor_sched.c
1029
csg_slot->group = group;
drivers/gpu/drm/panthor/panthor_sched.c
1041
group_unbind_locked(struct panthor_group *group)
drivers/gpu/drm/panthor/panthor_sched.c
1043
struct panthor_device *ptdev = group->ptdev;
drivers/gpu/drm/panthor/panthor_sched.c
1048
if (drm_WARN_ON(&ptdev->base, group->csg_id < 0 || group->csg_id >= MAX_CSGS))
drivers/gpu/drm/panthor/panthor_sched.c
1051
if (drm_WARN_ON(&ptdev->base, group->state == PANTHOR_CS_GROUP_ACTIVE))
drivers/gpu/drm/panthor/panthor_sched.c
1054
slot = &ptdev->scheduler->csg_slots[group->csg_id];
drivers/gpu/drm/panthor/panthor_sched.c
1055
panthor_vm_idle(group->vm);
drivers/gpu/drm/panthor/panthor_sched.c
1056
group->csg_id = -1;
drivers/gpu/drm/panthor/panthor_sched.c
1059
atomic_set(&group->tiler_oom, 0);
drivers/gpu/drm/panthor/panthor_sched.c
1060
cancel_work(&group->tiler_oom_work);
drivers/gpu/drm/panthor/panthor_sched.c
1062
for (u32 i = 0; i < group->queue_count; i++)
drivers/gpu/drm/panthor/panthor_sched.c
1063
group->queues[i]->doorbell_id = -1;
drivers/gpu/drm/panthor/panthor_sched.c
1065
slot->group = NULL;
drivers/gpu/drm/panthor/panthor_sched.c
1067
group_put(group);
drivers/gpu/drm/panthor/panthor_sched.c
1072
group_is_idle(struct panthor_group *group)
drivers/gpu/drm/panthor/panthor_sched.c
1074
u32 inactive_queues = group->idle_queues | group->blocked_queues;
drivers/gpu/drm/panthor/panthor_sched.c
1076
return hweight32(inactive_queues) == group->queue_count;
drivers/gpu/drm/panthor/panthor_sched.c
108
struct panthor_group *group;
drivers/gpu/drm/panthor/panthor_sched.c
1080
group_can_run(struct panthor_group *group)
drivers/gpu/drm/panthor/panthor_sched.c
1082
return group->state != PANTHOR_CS_GROUP_TERMINATED &&
drivers/gpu/drm/panthor/panthor_sched.c
1083
group->state != PANTHOR_CS_GROUP_UNKNOWN_STATE &&
drivers/gpu/drm/panthor/panthor_sched.c
1084
!group->destroyed && group->fatal_queues == 0 &&
drivers/gpu/drm/panthor/panthor_sched.c
1085
!group->timedout;
drivers/gpu/drm/panthor/panthor_sched.c
1111
struct panthor_group *group;
drivers/gpu/drm/panthor/panthor_sched.c
1123
group = job ? job->group : NULL;
drivers/gpu/drm/panthor/panthor_sched.c
1130
if (group && group_can_run(group) &&
drivers/gpu/drm/panthor/panthor_sched.c
1131
(group->blocked_queues & BIT(job->queue_idx)) &&
drivers/gpu/drm/panthor/panthor_sched.c
1132
group_is_idle(group))
drivers/gpu/drm/panthor/panthor_sched.c
1189
struct panthor_queue *queue = ptdev->scheduler->csg_slots[csg_id].group->queues[cs_id];
drivers/gpu/drm/panthor/panthor_sched.c
1230
struct panthor_group *group = ptdev->scheduler->csg_slots[csg_id].group;
drivers/gpu/drm/panthor/panthor_sched.c
1231
struct panthor_queue *queue = group->queues[cs_id];
drivers/gpu/drm/panthor/panthor_sched.c
1278
struct panthor_group *group = ptdev->scheduler->csg_slots[csg_id].group;
drivers/gpu/drm/panthor/panthor_sched.c
1279
struct panthor_queue *queue = group->queues[cs_id];
drivers/gpu/drm/panthor/panthor_sched.c
1281
panthor_fw_get_cs_iface(group->ptdev, csg_id, cs_id);
drivers/gpu/drm/panthor/panthor_sched.c
1289
group->idle_queues |= BIT(cs_id);
drivers/gpu/drm/panthor/panthor_sched.c
1293
if (list_empty(&group->wait_node)) {
drivers/gpu/drm/panthor/panthor_sched.c
1294
list_move_tail(&group->wait_node,
drivers/gpu/drm/panthor/panthor_sched.c
1295
&group->ptdev->scheduler->groups.waiting);
drivers/gpu/drm/panthor/panthor_sched.c
1302
group->blocked_queues |= BIT(cs_id);
drivers/gpu/drm/panthor/panthor_sched.c
1330
struct panthor_group *group = csg_slot->group;
drivers/gpu/drm/panthor/panthor_sched.c
1335
group->idle_queues = 0;
drivers/gpu/drm/panthor/panthor_sched.c
1336
group->blocked_queues = 0;
drivers/gpu/drm/panthor/panthor_sched.c
1338
for (i = 0; i < group->queue_count; i++) {
drivers/gpu/drm/panthor/panthor_sched.c
1339
if (group->queues[i])
drivers/gpu/drm/panthor/panthor_sched.c
1349
struct panthor_group *group;
drivers/gpu/drm/panthor/panthor_sched.c
1356
group = csg_slot->group;
drivers/gpu/drm/panthor/panthor_sched.c
1358
if (!group)
drivers/gpu/drm/panthor/panthor_sched.c
1361
old_state = group->state;
drivers/gpu/drm/panthor/panthor_sched.c
1408
for (i = 0; i < group->queue_count; i++) {
drivers/gpu/drm/panthor/panthor_sched.c
1409
if (group->queues[i])
drivers/gpu/drm/panthor/panthor_sched.c
1414
group->state = new_state;
drivers/gpu/drm/panthor/panthor_sched.c
1422
struct panthor_group *group;
drivers/gpu/drm/panthor/panthor_sched.c
1435
group = csg_slot->group;
drivers/gpu/drm/panthor/panthor_sched.c
1436
if (!group || group->state == PANTHOR_CS_GROUP_ACTIVE)
drivers/gpu/drm/panthor/panthor_sched.c
1439
csg_iface = panthor_fw_get_csg_iface(group->ptdev, csg_id);
drivers/gpu/drm/panthor/panthor_sched.c
1441
for (i = 0; i < group->queue_count; i++) {
drivers/gpu/drm/panthor/panthor_sched.c
1442
if (group->queues[i]) {
drivers/gpu/drm/panthor/panthor_sched.c
1448
csg_iface->input->allow_compute = group->compute_core_mask;
drivers/gpu/drm/panthor/panthor_sched.c
1449
csg_iface->input->allow_fragment = group->fragment_core_mask;
drivers/gpu/drm/panthor/panthor_sched.c
1450
csg_iface->input->allow_other = group->tiler_core_mask;
drivers/gpu/drm/panthor/panthor_sched.c
1451
endpoint_req = CSG_EP_REQ_COMPUTE(group->max_compute_cores) |
drivers/gpu/drm/panthor/panthor_sched.c
1452
CSG_EP_REQ_FRAGMENT(group->max_fragment_cores) |
drivers/gpu/drm/panthor/panthor_sched.c
1453
CSG_EP_REQ_TILER(group->max_tiler_cores) |
drivers/gpu/drm/panthor/panthor_sched.c
1457
csg_iface->input->config = panthor_vm_as(group->vm);
drivers/gpu/drm/panthor/panthor_sched.c
1459
if (group->suspend_buf)
drivers/gpu/drm/panthor/panthor_sched.c
1460
csg_iface->input->suspend_buf = panthor_kernel_bo_gpuva(group->suspend_buf);
drivers/gpu/drm/panthor/panthor_sched.c
1464
if (group->protm_suspend_buf) {
drivers/gpu/drm/panthor/panthor_sched.c
1466
panthor_kernel_bo_gpuva(group->protm_suspend_buf);
drivers/gpu/drm/panthor/panthor_sched.c
1482
struct panthor_group *group = csg_slot->group;
drivers/gpu/drm/panthor/panthor_sched.c
1493
if (group) {
drivers/gpu/drm/panthor/panthor_sched.c
1495
group->task_info.pid, group->task_info.comm);
drivers/gpu/drm/panthor/panthor_sched.c
1497
group->fatal_queues |= BIT(cs_id);
drivers/gpu/drm/panthor/panthor_sched.c
1528
struct panthor_group *group = csg_slot->group;
drivers/gpu/drm/panthor/panthor_sched.c
1529
struct panthor_queue *queue = group && cs_id < group->queue_count ?
drivers/gpu/drm/panthor/panthor_sched.c
1530
group->queues[cs_id] : NULL;
drivers/gpu/drm/panthor/panthor_sched.c
1558
if (group) {
drivers/gpu/drm/panthor/panthor_sched.c
1560
group->task_info.pid, group->task_info.comm);
drivers/gpu/drm/panthor/panthor_sched.c
1575
static int group_process_tiler_oom(struct panthor_group *group, u32 cs_id)
drivers/gpu/drm/panthor/panthor_sched.c
1577
struct panthor_device *ptdev = group->ptdev;
drivers/gpu/drm/panthor/panthor_sched.c
1586
csg_id = group->csg_id;
drivers/gpu/drm/panthor/panthor_sched.c
1591
heaps = panthor_vm_get_heap_pool(group->vm, false);
drivers/gpu/drm/panthor/panthor_sched.c
1625
group->fatal_queues |= BIT(cs_id);
drivers/gpu/drm/panthor/panthor_sched.c
1631
csg_id = group->csg_id;
drivers/gpu/drm/panthor/panthor_sched.c
1663
struct panthor_group *group =
drivers/gpu/drm/panthor/panthor_sched.c
1665
u32 tiler_oom = atomic_xchg(&group->tiler_oom, 0);
drivers/gpu/drm/panthor/panthor_sched.c
1670
group_process_tiler_oom(group, cs_id);
drivers/gpu/drm/panthor/panthor_sched.c
1674
group_put(group);
drivers/gpu/drm/panthor/panthor_sched.c
1683
struct panthor_group *group = csg_slot->group;
drivers/gpu/drm/panthor/panthor_sched.c
1687
if (drm_WARN_ON(&ptdev->base, !group))
drivers/gpu/drm/panthor/panthor_sched.c
1690
atomic_or(BIT(cs_id), &group->tiler_oom);
drivers/gpu/drm/panthor/panthor_sched.c
1695
group_get(group);
drivers/gpu/drm/panthor/panthor_sched.c
1696
if (!queue_work(sched->heap_alloc_wq, &group->tiler_oom_work))
drivers/gpu/drm/panthor/panthor_sched.c
1697
group_put(group);
drivers/gpu/drm/panthor/panthor_sched.c
1749
struct panthor_group *group = csg_slot->group;
drivers/gpu/drm/panthor/panthor_sched.c
1753
if (group)
drivers/gpu/drm/panthor/panthor_sched.c
1754
group_queue_work(group, sync_upd);
drivers/gpu/drm/panthor/panthor_sched.c
1764
struct panthor_group *group = csg_slot->group;
drivers/gpu/drm/panthor/panthor_sched.c
1768
group = csg_slot->group;
drivers/gpu/drm/panthor/panthor_sched.c
1769
if (!drm_WARN_ON(&ptdev->base, !group)) {
drivers/gpu/drm/panthor/panthor_sched.c
1771
group->task_info.pid, group->task_info.comm);
drivers/gpu/drm/panthor/panthor_sched.c
1773
group->timedout = true;
drivers/gpu/drm/panthor/panthor_sched.c
2033
struct panthor_group *group, *tmp;
drivers/gpu/drm/panthor/panthor_sched.c
2038
list_for_each_entry_safe(group, tmp, queue, run_node) {
drivers/gpu/drm/panthor/panthor_sched.c
2041
if (!group_can_run(group))
drivers/gpu/drm/panthor/panthor_sched.c
2044
if (skip_idle_groups && group_is_idle(group))
drivers/gpu/drm/panthor/panthor_sched.c
2048
if (ctx->vms[i] == group->vm)
drivers/gpu/drm/panthor/panthor_sched.c
2056
group_get(group);
drivers/gpu/drm/panthor/panthor_sched.c
2063
if (group_is_idle(group))
drivers/gpu/drm/panthor/panthor_sched.c
2065
else if (!list_empty(&ctx->groups[group->priority]))
drivers/gpu/drm/panthor/panthor_sched.c
2068
list_move_tail(&group->run_node, &ctx->groups[group->priority]);
drivers/gpu/drm/panthor/panthor_sched.c
2071
ctx->vms[ctx->as_count++] = group->vm;
drivers/gpu/drm/panthor/panthor_sched.c
2081
struct panthor_group *group)
drivers/gpu/drm/panthor/panthor_sched.c
2083
struct panthor_csg_slot *csg_slot = &sched->csg_slots[group->csg_id];
drivers/gpu/drm/panthor/panthor_sched.c
2088
&ctx->old_groups[csg_slot->group->priority],
drivers/gpu/drm/panthor/panthor_sched.c
2096
list_add_tail(&group->run_node, &other_group->run_node);
drivers/gpu/drm/panthor/panthor_sched.c
2101
list_add_tail(&group->run_node, &ctx->old_groups[group->priority]);
drivers/gpu/drm/panthor/panthor_sched.c
2124
struct panthor_group *group = csg_slot->group;
drivers/gpu/drm/panthor/panthor_sched.c
2127
if (!group)
drivers/gpu/drm/panthor/panthor_sched.c
2131
group_get(group);
drivers/gpu/drm/panthor/panthor_sched.c
2136
if (panthor_vm_has_unhandled_faults(group->vm)) {
drivers/gpu/drm/panthor/panthor_sched.c
2140
if (!group->fatal_queues)
drivers/gpu/drm/panthor/panthor_sched.c
2141
group->fatal_queues |= GENMASK(group->queue_count - 1, 0);
drivers/gpu/drm/panthor/panthor_sched.c
2144
tick_ctx_insert_old_group(sched, ctx, group);
drivers/gpu/drm/panthor/panthor_sched.c
2158
group_term_post_processing(struct panthor_group *group)
drivers/gpu/drm/panthor/panthor_sched.c
2165
if (drm_WARN_ON(&group->ptdev->base, group_can_run(group)))
drivers/gpu/drm/panthor/panthor_sched.c
2169
for (i = 0; i < group->queue_count; i++) {
drivers/gpu/drm/panthor/panthor_sched.c
2170
struct panthor_queue *queue = group->queues[i];
drivers/gpu/drm/panthor/panthor_sched.c
2174
if (group->fatal_queues & BIT(i))
drivers/gpu/drm/panthor/panthor_sched.c
2176
else if (group->timedout)
drivers/gpu/drm/panthor/panthor_sched.c
2193
syncobj = group->syncobjs->kmap + (i * sizeof(*syncobj));
drivers/gpu/drm/panthor/panthor_sched.c
2196
sched_queue_work(group->ptdev->scheduler, sync_upd);
drivers/gpu/drm/panthor/panthor_sched.c
2208
struct panthor_group *group =
drivers/gpu/drm/panthor/panthor_sched.c
2211
group_term_post_processing(group);
drivers/gpu/drm/panthor/panthor_sched.c
2212
group_put(group);
drivers/gpu/drm/panthor/panthor_sched.c
2220
struct panthor_group *group, *tmp;
drivers/gpu/drm/panthor/panthor_sched.c
2224
list_for_each_entry_safe(group, tmp, &ctx->old_groups[i], run_node) {
drivers/gpu/drm/panthor/panthor_sched.c
2229
group_can_run(group));
drivers/gpu/drm/panthor/panthor_sched.c
2231
if (!group_can_run(group)) {
drivers/gpu/drm/panthor/panthor_sched.c
2232
list_del_init(&group->run_node);
drivers/gpu/drm/panthor/panthor_sched.c
2233
list_del_init(&group->wait_node);
drivers/gpu/drm/panthor/panthor_sched.c
2234
group_queue_work(group, term);
drivers/gpu/drm/panthor/panthor_sched.c
2235
} else if (group->csg_id >= 0) {
drivers/gpu/drm/panthor/panthor_sched.c
2236
list_del_init(&group->run_node);
drivers/gpu/drm/panthor/panthor_sched.c
2238
list_move(&group->run_node,
drivers/gpu/drm/panthor/panthor_sched.c
2239
group_is_idle(group) ?
drivers/gpu/drm/panthor/panthor_sched.c
2240
&sched->groups.idle[group->priority] :
drivers/gpu/drm/panthor/panthor_sched.c
2241
&sched->groups.runnable[group->priority]);
drivers/gpu/drm/panthor/panthor_sched.c
2243
group_put(group);
drivers/gpu/drm/panthor/panthor_sched.c
2254
list_for_each_entry_safe(group, tmp, &ctx->groups[i], run_node) {
drivers/gpu/drm/panthor/panthor_sched.c
2255
if (group->csg_id >= 0) {
drivers/gpu/drm/panthor/panthor_sched.c
2256
list_del_init(&group->run_node);
drivers/gpu/drm/panthor/panthor_sched.c
2258
list_move(&group->run_node,
drivers/gpu/drm/panthor/panthor_sched.c
2259
group_is_idle(group) ?
drivers/gpu/drm/panthor/panthor_sched.c
2260
&sched->groups.idle[group->priority] :
drivers/gpu/drm/panthor/panthor_sched.c
2261
&sched->groups.runnable[group->priority]);
drivers/gpu/drm/panthor/panthor_sched.c
2263
group_put(group);
drivers/gpu/drm/panthor/panthor_sched.c
2271
struct panthor_group *group, *tmp;
drivers/gpu/drm/panthor/panthor_sched.c
2283
list_for_each_entry(group, &ctx->old_groups[prio], run_node) {
drivers/gpu/drm/panthor/panthor_sched.c
2284
bool term = !group_can_run(group);
drivers/gpu/drm/panthor/panthor_sched.c
2285
int csg_id = group->csg_id;
drivers/gpu/drm/panthor/panthor_sched.c
2297
list_for_each_entry(group, &ctx->groups[prio], run_node) {
drivers/gpu/drm/panthor/panthor_sched.c
2299
int csg_id = group->csg_id;
drivers/gpu/drm/panthor/panthor_sched.c
2332
list_for_each_entry(group, &ctx->old_groups[prio], run_node) {
drivers/gpu/drm/panthor/panthor_sched.c
2337
if (group->csg_id >= 0)
drivers/gpu/drm/panthor/panthor_sched.c
2338
sched_process_csg_irq_locked(ptdev, group->csg_id);
drivers/gpu/drm/panthor/panthor_sched.c
2340
group_unbind_locked(group);
drivers/gpu/drm/panthor/panthor_sched.c
2345
if (!sched->csg_slots[i].group)
drivers/gpu/drm/panthor/panthor_sched.c
2354
list_for_each_entry(group, &ctx->groups[prio], run_node) {
drivers/gpu/drm/panthor/panthor_sched.c
2355
int csg_id = group->csg_id;
drivers/gpu/drm/panthor/panthor_sched.c
2369
group_bind_locked(group, csg_id);
drivers/gpu/drm/panthor/panthor_sched.c
2372
group->state == PANTHOR_CS_GROUP_SUSPENDED ?
drivers/gpu/drm/panthor/panthor_sched.c
2390
list_for_each_entry_safe(group, tmp, &ctx->groups[prio], run_node) {
drivers/gpu/drm/panthor/panthor_sched.c
2391
list_del_init(&group->run_node);
drivers/gpu/drm/panthor/panthor_sched.c
2398
if (group->destroyed)
drivers/gpu/drm/panthor/panthor_sched.c
2400
group_put(group);
drivers/gpu/drm/panthor/panthor_sched.c
2408
list_for_each_entry_safe(group, tmp, &ctx->old_groups[prio], run_node) {
drivers/gpu/drm/panthor/panthor_sched.c
2409
if (!group_can_run(group))
drivers/gpu/drm/panthor/panthor_sched.c
2412
if (group_is_idle(group))
drivers/gpu/drm/panthor/panthor_sched.c
2413
list_move_tail(&group->run_node, &sched->groups.idle[prio]);
drivers/gpu/drm/panthor/panthor_sched.c
2415
list_move_tail(&group->run_node, &sched->groups.runnable[prio]);
drivers/gpu/drm/panthor/panthor_sched.c
2416
group_put(group);
drivers/gpu/drm/panthor/panthor_sched.c
2577
static int panthor_queue_eval_syncwait(struct panthor_group *group, u8 queue_idx)
drivers/gpu/drm/panthor/panthor_sched.c
2579
struct panthor_queue *queue = group->queues[queue_idx];
drivers/gpu/drm/panthor/panthor_sched.c
2587
syncobj = panthor_queue_get_syncwait_obj(group, queue);
drivers/gpu/drm/panthor/panthor_sched.c
2611
struct panthor_group *group, *tmp;
drivers/gpu/drm/panthor/panthor_sched.c
2615
list_for_each_entry_safe(group, tmp, &sched->groups.waiting, wait_node) {
drivers/gpu/drm/panthor/panthor_sched.c
2616
u32 tested_queues = group->blocked_queues;
drivers/gpu/drm/panthor/panthor_sched.c
2623
ret = panthor_queue_eval_syncwait(group, cs_id);
drivers/gpu/drm/panthor/panthor_sched.c
2624
drm_WARN_ON(&group->ptdev->base, ret < 0);
drivers/gpu/drm/panthor/panthor_sched.c
2632
group->blocked_queues &= ~unblocked_queues;
drivers/gpu/drm/panthor/panthor_sched.c
2634
if (group->csg_id < 0) {
drivers/gpu/drm/panthor/panthor_sched.c
2635
list_move(&group->run_node,
drivers/gpu/drm/panthor/panthor_sched.c
2636
&sched->groups.runnable[group->priority]);
drivers/gpu/drm/panthor/panthor_sched.c
2637
if (group->priority == PANTHOR_CSG_PRIORITY_RT)
drivers/gpu/drm/panthor/panthor_sched.c
2642
if (!group->blocked_queues)
drivers/gpu/drm/panthor/panthor_sched.c
2643
list_del_init(&group->wait_node);
drivers/gpu/drm/panthor/panthor_sched.c
2672
static void group_schedule_locked(struct panthor_group *group, u32 queue_mask)
drivers/gpu/drm/panthor/panthor_sched.c
2674
struct panthor_device *ptdev = group->ptdev;
drivers/gpu/drm/panthor/panthor_sched.c
2676
struct list_head *queue = &sched->groups.runnable[group->priority];
drivers/gpu/drm/panthor/panthor_sched.c
2679
if (!group_can_run(group))
drivers/gpu/drm/panthor/panthor_sched.c
2683
if ((queue_mask & group->blocked_queues) == queue_mask)
drivers/gpu/drm/panthor/panthor_sched.c
2686
was_idle = group_is_idle(group);
drivers/gpu/drm/panthor/panthor_sched.c
2687
group->idle_queues &= ~queue_mask;
drivers/gpu/drm/panthor/panthor_sched.c
2693
if (was_idle && !group_is_idle(group))
drivers/gpu/drm/panthor/panthor_sched.c
2694
list_move_tail(&group->run_node, queue);
drivers/gpu/drm/panthor/panthor_sched.c
2697
if (group->priority == PANTHOR_CSG_PRIORITY_RT) {
drivers/gpu/drm/panthor/panthor_sched.c
2744
static void panthor_group_stop(struct panthor_group *group)
drivers/gpu/drm/panthor/panthor_sched.c
2746
struct panthor_scheduler *sched = group->ptdev->scheduler;
drivers/gpu/drm/panthor/panthor_sched.c
2750
for (u32 i = 0; i < group->queue_count; i++)
drivers/gpu/drm/panthor/panthor_sched.c
2751
queue_stop(group->queues[i], NULL);
drivers/gpu/drm/panthor/panthor_sched.c
2753
group_get(group);
drivers/gpu/drm/panthor/panthor_sched.c
2754
list_move_tail(&group->run_node, &sched->reset.stopped_groups);
drivers/gpu/drm/panthor/panthor_sched.c
2757
static void panthor_group_start(struct panthor_group *group)
drivers/gpu/drm/panthor/panthor_sched.c
2759
struct panthor_scheduler *sched = group->ptdev->scheduler;
drivers/gpu/drm/panthor/panthor_sched.c
2761
lockdep_assert_held(&group->ptdev->scheduler->reset.lock);
drivers/gpu/drm/panthor/panthor_sched.c
2763
for (u32 i = 0; i < group->queue_count; i++)
drivers/gpu/drm/panthor/panthor_sched.c
2764
queue_start(group->queues[i]);
drivers/gpu/drm/panthor/panthor_sched.c
2766
if (group_can_run(group)) {
drivers/gpu/drm/panthor/panthor_sched.c
2767
list_move_tail(&group->run_node,
drivers/gpu/drm/panthor/panthor_sched.c
2768
group_is_idle(group) ?
drivers/gpu/drm/panthor/panthor_sched.c
2769
&sched->groups.idle[group->priority] :
drivers/gpu/drm/panthor/panthor_sched.c
2770
&sched->groups.runnable[group->priority]);
drivers/gpu/drm/panthor/panthor_sched.c
2772
list_del_init(&group->run_node);
drivers/gpu/drm/panthor/panthor_sched.c
2773
list_del_init(&group->wait_node);
drivers/gpu/drm/panthor/panthor_sched.c
2774
group_queue_work(group, term);
drivers/gpu/drm/panthor/panthor_sched.c
2777
group_put(group);
drivers/gpu/drm/panthor/panthor_sched.c
2822
if (csg_slot->group) {
drivers/gpu/drm/panthor/panthor_sched.c
2824
group_can_run(csg_slot->group) ?
drivers/gpu/drm/panthor/panthor_sched.c
2847
if (group_can_run(csg_slot->group))
drivers/gpu/drm/panthor/panthor_sched.c
2848
csg_slot->group->innocent = true;
drivers/gpu/drm/panthor/panthor_sched.c
2853
csg_slot->group->timedout = true;
drivers/gpu/drm/panthor/panthor_sched.c
2867
struct panthor_group *group = csg_slot->group;
drivers/gpu/drm/panthor/panthor_sched.c
2873
if (group->state != PANTHOR_CS_GROUP_TERMINATED) {
drivers/gpu/drm/panthor/panthor_sched.c
2874
group->state = PANTHOR_CS_GROUP_TERMINATED;
drivers/gpu/drm/panthor/panthor_sched.c
2879
for (i = 0; i < group->queue_count; i++) {
drivers/gpu/drm/panthor/panthor_sched.c
2880
if (group->queues[i])
drivers/gpu/drm/panthor/panthor_sched.c
2903
csg_slot->group->state = PANTHOR_CS_GROUP_TERMINATED;
drivers/gpu/drm/panthor/panthor_sched.c
2913
struct panthor_group *group = csg_slot->group;
drivers/gpu/drm/panthor/panthor_sched.c
2915
if (!group)
drivers/gpu/drm/panthor/panthor_sched.c
2918
group_get(group);
drivers/gpu/drm/panthor/panthor_sched.c
2920
if (group->csg_id >= 0)
drivers/gpu/drm/panthor/panthor_sched.c
2921
sched_process_csg_irq_locked(ptdev, group->csg_id);
drivers/gpu/drm/panthor/panthor_sched.c
2923
group_unbind_locked(group);
drivers/gpu/drm/panthor/panthor_sched.c
2925
drm_WARN_ON(&group->ptdev->base, !list_empty(&group->run_node));
drivers/gpu/drm/panthor/panthor_sched.c
2927
if (group_can_run(group)) {
drivers/gpu/drm/panthor/panthor_sched.c
2928
list_add(&group->run_node,
drivers/gpu/drm/panthor/panthor_sched.c
2929
&sched->groups.idle[group->priority]);
drivers/gpu/drm/panthor/panthor_sched.c
2934
list_del_init(&group->wait_node);
drivers/gpu/drm/panthor/panthor_sched.c
2935
group_queue_work(group, term);
drivers/gpu/drm/panthor/panthor_sched.c
2937
group_put(group);
drivers/gpu/drm/panthor/panthor_sched.c
2945
struct panthor_group *group, *group_tmp;
drivers/gpu/drm/panthor/panthor_sched.c
2963
list_for_each_entry_safe(group, group_tmp, &sched->groups.runnable[i], run_node)
drivers/gpu/drm/panthor/panthor_sched.c
2964
panthor_group_stop(group);
drivers/gpu/drm/panthor/panthor_sched.c
2968
list_for_each_entry_safe(group, group_tmp, &sched->groups.idle[i], run_node)
drivers/gpu/drm/panthor/panthor_sched.c
2969
panthor_group_stop(group);
drivers/gpu/drm/panthor/panthor_sched.c
2978
struct panthor_group *group, *group_tmp;
drivers/gpu/drm/panthor/panthor_sched.c
2982
list_for_each_entry_safe(group, group_tmp, &sched->reset.stopped_groups, run_node) {
drivers/gpu/drm/panthor/panthor_sched.c
2987
group->state = PANTHOR_CS_GROUP_TERMINATED;
drivers/gpu/drm/panthor/panthor_sched.c
2989
panthor_group_start(group);
drivers/gpu/drm/panthor/panthor_sched.c
3007
struct panthor_group *group = job->group;
drivers/gpu/drm/panthor/panthor_sched.c
3008
struct panthor_queue *queue = group->queues[job->queue_idx];
drivers/gpu/drm/panthor/panthor_sched.c
3009
struct panthor_gpu_usage *fdinfo = &group->fdinfo.data;
drivers/gpu/drm/panthor/panthor_sched.c
3013
scoped_guard(spinlock, &group->fdinfo.lock) {
drivers/gpu/drm/panthor/panthor_sched.c
3024
struct panthor_group *group;
drivers/gpu/drm/panthor/panthor_sched.c
3031
xa_for_each_marked(&gpool->xa, i, group, GROUP_REGISTERED) {
drivers/gpu/drm/panthor/panthor_sched.c
3032
guard(spinlock)(&group->fdinfo.lock);
drivers/gpu/drm/panthor/panthor_sched.c
3033
pfile->stats.cycles += group->fdinfo.data.cycles;
drivers/gpu/drm/panthor/panthor_sched.c
3034
pfile->stats.time += group->fdinfo.data.time;
drivers/gpu/drm/panthor/panthor_sched.c
3035
group->fdinfo.data.cycles = 0;
drivers/gpu/drm/panthor/panthor_sched.c
3036
group->fdinfo.data.time = 0;
drivers/gpu/drm/panthor/panthor_sched.c
3052
struct panthor_group *group = job->group;
drivers/gpu/drm/panthor/panthor_sched.c
3054
syncobj = group->syncobjs->kmap +
drivers/gpu/drm/panthor/panthor_sched.c
3096
struct panthor_group *group =
drivers/gpu/drm/panthor/panthor_sched.c
3102
for (queue_idx = 0; queue_idx < group->queue_count; queue_idx++) {
drivers/gpu/drm/panthor/panthor_sched.c
3103
struct panthor_queue *queue = group->queues[queue_idx];
drivers/gpu/drm/panthor/panthor_sched.c
3112
group_put(group);
drivers/gpu/drm/panthor/panthor_sched.c
3170
struct panthor_group *group = job->group;
drivers/gpu/drm/panthor/panthor_sched.c
3171
struct panthor_queue *queue = group->queues[job->queue_idx];
drivers/gpu/drm/panthor/panthor_sched.c
3172
struct panthor_device *ptdev = group->ptdev;
drivers/gpu/drm/panthor/panthor_sched.c
3181
params->sync_addr = panthor_kernel_bo_gpuva(group->syncobjs) +
drivers/gpu/drm/panthor/panthor_sched.c
3299
struct panthor_group *group = job->group;
drivers/gpu/drm/panthor/panthor_sched.c
3300
struct panthor_queue *queue = group->queues[job->queue_idx];
drivers/gpu/drm/panthor/panthor_sched.c
3301
struct panthor_device *ptdev = group->ptdev;
drivers/gpu/drm/panthor/panthor_sched.c
3322
if (!group_can_run(group)) {
drivers/gpu/drm/panthor/panthor_sched.c
3358
if (group->csg_id < 0) {
drivers/gpu/drm/panthor/panthor_sched.c
3359
group_schedule_locked(group, BIT(job->queue_idx));
drivers/gpu/drm/panthor/panthor_sched.c
3362
bool resume_tick = group_is_idle(group) &&
drivers/gpu/drm/panthor/panthor_sched.c
3363
(group->idle_queues & queue_mask) &&
drivers/gpu/drm/panthor/panthor_sched.c
3364
!(group->blocked_queues & queue_mask) &&
drivers/gpu/drm/panthor/panthor_sched.c
3368
group->idle_queues &= ~queue_mask;
drivers/gpu/drm/panthor/panthor_sched.c
3375
!(group->blocked_queues & BIT(job->queue_idx))) {
drivers/gpu/drm/panthor/panthor_sched.c
3401
struct panthor_group *group = job->group;
drivers/gpu/drm/panthor/panthor_sched.c
3402
struct panthor_device *ptdev = group->ptdev;
drivers/gpu/drm/panthor/panthor_sched.c
3404
struct panthor_queue *queue = group->queues[job->queue_idx];
drivers/gpu/drm/panthor/panthor_sched.c
3407
group->task_info.pid, group->task_info.comm, job->done_fence->seqno);
drivers/gpu/drm/panthor/panthor_sched.c
3414
group->timedout = true;
drivers/gpu/drm/panthor/panthor_sched.c
3415
if (group->csg_id >= 0) {
drivers/gpu/drm/panthor/panthor_sched.c
3421
list_del_init(&group->run_node);
drivers/gpu/drm/panthor/panthor_sched.c
3422
list_del_init(&group->wait_node);
drivers/gpu/drm/panthor/panthor_sched.c
3424
group_queue_work(group, term);
drivers/gpu/drm/panthor/panthor_sched.c
3483
group_create_queue(struct panthor_group *group,
drivers/gpu/drm/panthor/panthor_sched.c
3489
.submit_wq = group->ptdev->scheduler->wq,
drivers/gpu/drm/panthor/panthor_sched.c
3499
.timeout_wq = group->ptdev->reset.wq,
drivers/gpu/drm/panthor/panthor_sched.c
3500
.dev = group->ptdev->base.dev,
drivers/gpu/drm/panthor/panthor_sched.c
3528
queue->ringbuf = panthor_kernel_bo_create(group->ptdev, group->vm,
drivers/gpu/drm/panthor/panthor_sched.c
3544
queue->iface.mem = panthor_fw_alloc_queue_iface_mem(group->ptdev,
drivers/gpu/drm/panthor/panthor_sched.c
3555
calc_profiling_ringbuf_num_slots(group->ptdev, args->ringbuf_size);
drivers/gpu/drm/panthor/panthor_sched.c
3558
panthor_kernel_bo_create(group->ptdev, group->vm,
drivers/gpu/drm/panthor/panthor_sched.c
3597
group_free_queue(group, queue);
drivers/gpu/drm/panthor/panthor_sched.c
3601
static void group_init_task_info(struct panthor_group *group)
drivers/gpu/drm/panthor/panthor_sched.c
3605
group->task_info.pid = task->pid;
drivers/gpu/drm/panthor/panthor_sched.c
3606
get_task_comm(group->task_info.comm, task);
drivers/gpu/drm/panthor/panthor_sched.c
3610
struct panthor_group *group)
drivers/gpu/drm/panthor/panthor_sched.c
3615
if (drm_WARN_ON(&ptdev->base, IS_ERR_OR_NULL(group)))
drivers/gpu/drm/panthor/panthor_sched.c
3617
if (drm_WARN_ON(&ptdev->base, ptdev != group->ptdev))
drivers/gpu/drm/panthor/panthor_sched.c
3620
group->fdinfo.kbo_sizes += group->suspend_buf->obj->size;
drivers/gpu/drm/panthor/panthor_sched.c
3621
group->fdinfo.kbo_sizes += group->protm_suspend_buf->obj->size;
drivers/gpu/drm/panthor/panthor_sched.c
3622
group->fdinfo.kbo_sizes += group->syncobjs->obj->size;
drivers/gpu/drm/panthor/panthor_sched.c
3624
for (i = 0; i < group->queue_count; i++) {
drivers/gpu/drm/panthor/panthor_sched.c
3625
queue = group->queues[i];
drivers/gpu/drm/panthor/panthor_sched.c
3626
group->fdinfo.kbo_sizes += queue->ringbuf->obj->size;
drivers/gpu/drm/panthor/panthor_sched.c
3627
group->fdinfo.kbo_sizes += queue->iface.mem->obj->size;
drivers/gpu/drm/panthor/panthor_sched.c
3628
group->fdinfo.kbo_sizes += queue->profiling.slots->obj->size;
drivers/gpu/drm/panthor/panthor_sched.c
3643
struct panthor_group *group = NULL;
drivers/gpu/drm/panthor/panthor_sched.c
3663
group = kzalloc_obj(*group);
drivers/gpu/drm/panthor/panthor_sched.c
3664
if (!group)
drivers/gpu/drm/panthor/panthor_sched.c
3667
spin_lock_init(&group->fatal_lock);
drivers/gpu/drm/panthor/panthor_sched.c
3668
kref_init(&group->refcount);
drivers/gpu/drm/panthor/panthor_sched.c
3669
group->state = PANTHOR_CS_GROUP_CREATED;
drivers/gpu/drm/panthor/panthor_sched.c
3670
group->csg_id = -1;
drivers/gpu/drm/panthor/panthor_sched.c
3672
group->ptdev = ptdev;
drivers/gpu/drm/panthor/panthor_sched.c
3673
group->max_compute_cores = group_args->max_compute_cores;
drivers/gpu/drm/panthor/panthor_sched.c
3674
group->compute_core_mask = group_args->compute_core_mask;
drivers/gpu/drm/panthor/panthor_sched.c
3675
group->max_fragment_cores = group_args->max_fragment_cores;
drivers/gpu/drm/panthor/panthor_sched.c
3676
group->fragment_core_mask = group_args->fragment_core_mask;
drivers/gpu/drm/panthor/panthor_sched.c
3677
group->max_tiler_cores = group_args->max_tiler_cores;
drivers/gpu/drm/panthor/panthor_sched.c
3678
group->tiler_core_mask = group_args->tiler_core_mask;
drivers/gpu/drm/panthor/panthor_sched.c
3679
group->priority = group_args->priority;
drivers/gpu/drm/panthor/panthor_sched.c
3681
INIT_LIST_HEAD(&group->wait_node);
drivers/gpu/drm/panthor/panthor_sched.c
3682
INIT_LIST_HEAD(&group->run_node);
drivers/gpu/drm/panthor/panthor_sched.c
3683
INIT_WORK(&group->term_work, group_term_work);
drivers/gpu/drm/panthor/panthor_sched.c
3684
INIT_WORK(&group->sync_upd_work, group_sync_upd_work);
drivers/gpu/drm/panthor/panthor_sched.c
3685
INIT_WORK(&group->tiler_oom_work, group_tiler_oom_work);
drivers/gpu/drm/panthor/panthor_sched.c
3686
INIT_WORK(&group->release_work, group_release_work);
drivers/gpu/drm/panthor/panthor_sched.c
3688
group->vm = panthor_vm_pool_get_vm(pfile->vms, group_args->vm_id);
drivers/gpu/drm/panthor/panthor_sched.c
3689
if (!group->vm) {
drivers/gpu/drm/panthor/panthor_sched.c
3695
group->suspend_buf = panthor_fw_alloc_suspend_buf_mem(ptdev, suspend_size);
drivers/gpu/drm/panthor/panthor_sched.c
3696
if (IS_ERR(group->suspend_buf)) {
drivers/gpu/drm/panthor/panthor_sched.c
3697
ret = PTR_ERR(group->suspend_buf);
drivers/gpu/drm/panthor/panthor_sched.c
3698
group->suspend_buf = NULL;
drivers/gpu/drm/panthor/panthor_sched.c
3703
group->protm_suspend_buf = panthor_fw_alloc_suspend_buf_mem(ptdev, suspend_size);
drivers/gpu/drm/panthor/panthor_sched.c
3704
if (IS_ERR(group->protm_suspend_buf)) {
drivers/gpu/drm/panthor/panthor_sched.c
3705
ret = PTR_ERR(group->protm_suspend_buf);
drivers/gpu/drm/panthor/panthor_sched.c
3706
group->protm_suspend_buf = NULL;
drivers/gpu/drm/panthor/panthor_sched.c
3710
group->syncobjs = panthor_kernel_bo_create(ptdev, group->vm,
drivers/gpu/drm/panthor/panthor_sched.c
3718
if (IS_ERR(group->syncobjs)) {
drivers/gpu/drm/panthor/panthor_sched.c
3719
ret = PTR_ERR(group->syncobjs);
drivers/gpu/drm/panthor/panthor_sched.c
3723
ret = panthor_kernel_bo_vmap(group->syncobjs);
drivers/gpu/drm/panthor/panthor_sched.c
3727
memset(group->syncobjs->kmap, 0,
drivers/gpu/drm/panthor/panthor_sched.c
3730
ret = xa_alloc(&gpool->xa, &gid, group, XA_LIMIT(1, MAX_GROUPS_PER_POOL), GFP_KERNEL);
drivers/gpu/drm/panthor/panthor_sched.c
3735
group->queues[i] = group_create_queue(group, &queue_args[i], drm_client_id, gid, i);
drivers/gpu/drm/panthor/panthor_sched.c
3736
if (IS_ERR(group->queues[i])) {
drivers/gpu/drm/panthor/panthor_sched.c
3737
ret = PTR_ERR(group->queues[i]);
drivers/gpu/drm/panthor/panthor_sched.c
3738
group->queues[i] = NULL;
drivers/gpu/drm/panthor/panthor_sched.c
3742
group->queue_count++;
drivers/gpu/drm/panthor/panthor_sched.c
3745
group->idle_queues = GENMASK(group->queue_count - 1, 0);
drivers/gpu/drm/panthor/panthor_sched.c
3749
panthor_group_stop(group);
drivers/gpu/drm/panthor/panthor_sched.c
3752
list_add_tail(&group->run_node,
drivers/gpu/drm/panthor/panthor_sched.c
3753
&sched->groups.idle[group->priority]);
drivers/gpu/drm/panthor/panthor_sched.c
3758
add_group_kbo_sizes(group->ptdev, group);
drivers/gpu/drm/panthor/panthor_sched.c
3759
spin_lock_init(&group->fdinfo.lock);
drivers/gpu/drm/panthor/panthor_sched.c
3761
group_init_task_info(group);
drivers/gpu/drm/panthor/panthor_sched.c
3771
group_put(group);
drivers/gpu/drm/panthor/panthor_sched.c
3780
struct panthor_group *group;
drivers/gpu/drm/panthor/panthor_sched.c
3785
group = xa_erase(&gpool->xa, group_handle);
drivers/gpu/drm/panthor/panthor_sched.c
3786
if (!group)
drivers/gpu/drm/panthor/panthor_sched.c
3791
group->destroyed = true;
drivers/gpu/drm/panthor/panthor_sched.c
3792
if (group->csg_id >= 0) {
drivers/gpu/drm/panthor/panthor_sched.c
3798
list_del_init(&group->run_node);
drivers/gpu/drm/panthor/panthor_sched.c
3799
list_del_init(&group->wait_node);
drivers/gpu/drm/panthor/panthor_sched.c
3800
group_queue_work(group, term);
drivers/gpu/drm/panthor/panthor_sched.c
3805
group_put(group);
drivers/gpu/drm/panthor/panthor_sched.c
3812
struct panthor_group *group;
drivers/gpu/drm/panthor/panthor_sched.c
3815
group = group_get(xa_find(&pool->xa, &group_handle, group_handle, GROUP_REGISTERED));
drivers/gpu/drm/panthor/panthor_sched.c
3818
return group;
drivers/gpu/drm/panthor/panthor_sched.c
3827
struct panthor_group *group;
drivers/gpu/drm/panthor/panthor_sched.c
3832
group = group_from_handle(gpool, get_state->group_handle);
drivers/gpu/drm/panthor/panthor_sched.c
3833
if (!group)
drivers/gpu/drm/panthor/panthor_sched.c
3839
if (group->timedout)
drivers/gpu/drm/panthor/panthor_sched.c
3841
if (group->fatal_queues) {
drivers/gpu/drm/panthor/panthor_sched.c
3843
get_state->fatal_queues = group->fatal_queues;
drivers/gpu/drm/panthor/panthor_sched.c
3845
if (group->innocent)
drivers/gpu/drm/panthor/panthor_sched.c
3849
group_put(group);
drivers/gpu/drm/panthor/panthor_sched.c
3869
struct panthor_group *group;
drivers/gpu/drm/panthor/panthor_sched.c
3875
xa_for_each(&gpool->xa, i, group)
drivers/gpu/drm/panthor/panthor_sched.c
3895
struct panthor_group *group;
drivers/gpu/drm/panthor/panthor_sched.c
3902
xa_for_each_marked(&gpool->xa, i, group, GROUP_REGISTERED) {
drivers/gpu/drm/panthor/panthor_sched.c
3903
stats->resident += group->fdinfo.kbo_sizes;
drivers/gpu/drm/panthor/panthor_sched.c
3904
if (group->csg_id >= 0)
drivers/gpu/drm/panthor/panthor_sched.c
3905
stats->active += group->fdinfo.kbo_sizes;
drivers/gpu/drm/panthor/panthor_sched.c
3914
drm_WARN_ON(&job->group->ptdev->base, !list_empty(&job->node));
drivers/gpu/drm/panthor/panthor_sched.c
3924
group_put(job->group);
drivers/gpu/drm/panthor/panthor_sched.c
3952
return job->group->vm;
drivers/gpu/drm/panthor/panthor_sched.c
3994
job->group = group_from_handle(gpool, group_handle);
drivers/gpu/drm/panthor/panthor_sched.c
3995
if (!job->group) {
drivers/gpu/drm/panthor/panthor_sched.c
4000
if (!group_can_run(job->group)) {
drivers/gpu/drm/panthor/panthor_sched.c
4005
if (job->queue_idx >= job->group->queue_count ||
drivers/gpu/drm/panthor/panthor_sched.c
4006
!job->group->queues[job->queue_idx]) {
drivers/gpu/drm/panthor/panthor_sched.c
4030
&job->group->queues[job->queue_idx]->entity,
drivers/gpu/drm/panthor/panthor_sched.c
4031
credits, job->group, drm_client_id);
drivers/gpu/drm/panthor/panthor_sched.c
4046
panthor_vm_update_resvs(job->group->vm, exec, &sched_job->s_fence->finished,
drivers/gpu/drm/panthor/panthor_sched.c
731
#define group_queue_work(group, wname) \
drivers/gpu/drm/panthor/panthor_sched.c
733
group_get(group); \
drivers/gpu/drm/panthor/panthor_sched.c
734
if (!queue_work((group)->ptdev->scheduler->wq, &(group)->wname ## _work)) \
drivers/gpu/drm/panthor/panthor_sched.c
735
group_put(group); \
drivers/gpu/drm/panthor/panthor_sched.c
801
struct panthor_group *group;
drivers/gpu/drm/panthor/panthor_sched.c
866
panthor_queue_get_syncwait_obj(struct panthor_group *group, struct panthor_queue *queue)
drivers/gpu/drm/panthor/panthor_sched.c
868
struct panthor_device *ptdev = group->ptdev;
drivers/gpu/drm/panthor/panthor_sched.c
879
bo = panthor_vm_get_bo_for_va(group->vm,
drivers/gpu/drm/panthor/panthor_sched.c
913
static void group_free_queue(struct panthor_group *group, struct panthor_queue *queue)
drivers/gpu/drm/panthor/panthor_sched.c
943
struct panthor_group *group = container_of(work,
drivers/gpu/drm/panthor/panthor_sched.c
951
for (i = 0; i < group->queue_count; i++)
drivers/gpu/drm/panthor/panthor_sched.c
952
group_free_queue(group, group->queues[i]);
drivers/gpu/drm/panthor/panthor_sched.c
954
panthor_kernel_bo_destroy(group->suspend_buf);
drivers/gpu/drm/panthor/panthor_sched.c
955
panthor_kernel_bo_destroy(group->protm_suspend_buf);
drivers/gpu/drm/panthor/panthor_sched.c
956
panthor_kernel_bo_destroy(group->syncobjs);
drivers/gpu/drm/panthor/panthor_sched.c
958
panthor_vm_put(group->vm);
drivers/gpu/drm/panthor/panthor_sched.c
959
kfree(group);
drivers/gpu/drm/panthor/panthor_sched.c
964
struct panthor_group *group = container_of(kref,
drivers/gpu/drm/panthor/panthor_sched.c
967
struct panthor_device *ptdev = group->ptdev;
drivers/gpu/drm/panthor/panthor_sched.c
969
drm_WARN_ON(&ptdev->base, group->csg_id >= 0);
drivers/gpu/drm/panthor/panthor_sched.c
970
drm_WARN_ON(&ptdev->base, !list_empty(&group->run_node));
drivers/gpu/drm/panthor/panthor_sched.c
971
drm_WARN_ON(&ptdev->base, !list_empty(&group->wait_node));
drivers/gpu/drm/panthor/panthor_sched.c
973
queue_work(panthor_cleanup_wq, &group->release_work);
drivers/gpu/drm/panthor/panthor_sched.c
976
static void group_put(struct panthor_group *group)
drivers/gpu/drm/panthor/panthor_sched.c
978
if (group)
drivers/gpu/drm/panthor/panthor_sched.c
979
kref_put(&group->refcount, group_release);
drivers/gpu/drm/panthor/panthor_sched.c
983
group_get(struct panthor_group *group)
drivers/gpu/drm/panthor/panthor_sched.c
985
if (group)
drivers/gpu/drm/panthor/panthor_sched.c
986
kref_get(&group->refcount);
drivers/gpu/drm/panthor/panthor_sched.c
988
return group;
drivers/gpu/drm/panthor/panthor_sched.c
999
group_bind_locked(struct panthor_group *group, u32 csg_id)
drivers/gpu/drm/qxl/qxl_dev.h
787
uint32_t group;
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
1282
rcrtc->group = rgrp;
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
242
rcar_du_group_write(rcrtc->group, DPLLCR, dpllcr);
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
304
if (rcrtc->group->cmms_mask & BIT(rcrtc->index % 2))
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
352
for (i = 0; i < rcrtc->group->num_planes; ++i) {
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
353
struct rcar_du_plane *plane = &rcrtc->group->planes[i];
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
409
mutex_lock(&rcrtc->group->lock);
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
411
dptsr_planes = rcrtc->index % 2 ? rcrtc->group->dptsr_planes | hwplanes
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
412
: rcrtc->group->dptsr_planes & ~hwplanes;
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
414
if (dptsr_planes != rcrtc->group->dptsr_planes) {
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
415
rcar_du_group_write(rcrtc->group, DPTSR,
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
417
rcrtc->group->dptsr_planes = dptsr_planes;
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
419
if (rcrtc->group->used_crtcs)
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
420
rcar_du_group_restart(rcrtc->group);
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
424
if (rcrtc->group->need_restart)
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
425
rcar_du_group_restart(rcrtc->group);
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
427
mutex_unlock(&rcrtc->group->lock);
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
429
rcar_du_group_write(rcrtc->group, rcrtc->index % 2 ? DS2PR : DS1PR,
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
537
rcar_du_group_set_routing(rcrtc->group);
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
540
rcar_du_group_write(rcrtc->group, rcrtc->index % 2 ? DS2PR : DS1PR, 0);
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
569
ret = rcar_du_group_get(rcrtc->group);
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
587
rcar_du_group_put(rcrtc->group);
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
609
rcar_du_group_start_stop(rcrtc->group, true);
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
629
rcar_du_group_write(rcrtc->group, rcrtc->index % 2 ? DS2PR : DS1PR, 0);
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
684
rcar_du_group_start_stop(rcrtc->group, false);
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
884
if (rcrtc->group->cmms_mask & BIT(rcrtc->index % 2))
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.h
67
struct rcar_du_group *group;
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
106
if (plane->group->index != 0)
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
112
fixed = plane->group->index == 0 ? 1 : 0;
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
155
plane->group->index, plane - plane->group->planes);
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
165
index = plane - plane->group->planes;
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
166
group_freed_planes[plane->group->index] |= 1 << index;
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
178
groups |= 1 << plane->group->index;
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
181
index = plane - plane->group->planes;
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
182
group_freed_planes[plane->group->index] |= 1 << index;
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
201
struct rcar_du_group *group = &rcdu->groups[index];
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
207
for (i = 0; i < group->num_planes; ++i) {
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
208
struct rcar_du_plane *plane = &group->planes[i];
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
227
__func__, plane->group->index,
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
228
plane - plane->group->planes);
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
237
__func__, plane->group->index,
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
238
plane - plane->group->planes,
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
266
plane->group->index, plane - plane->group->planes);
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
283
? plane->group->dptsr_planes
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
284
: ~plane->group->dptsr_planes;
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
285
free = group_free_planes[plane->group->index];
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
303
group_free_planes[plane->group->index] &=
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
307
__func__, plane->group->index,
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
308
group_free_planes[plane->group->index]);
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
675
rplane->group->need_restart = true;
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
741
struct rcar_du_device *rcdu = to_rcar_plane(plane)->group->dev;
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
757
struct rcar_du_device *rcdu = to_rcar_plane(plane)->group->dev;
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
812
plane->group = rgrp;
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.h
35
struct rcar_du_group *group;
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.h
83
return __rcar_du_plane_setup(plane->group, state);
drivers/gpu/drm/renesas/rcar-du/rcar_du_vsp.c
87
__rcar_du_plane_setup(crtc->group, &state);
drivers/gpu/drm/rockchip/rockchip_drm_vop.c
68
#define VOP_REG_SET(vop, group, name, v) \
drivers/gpu/drm/rockchip/rockchip_drm_vop.c
69
vop_reg_set(vop, &vop->data->group->name, 0, ~0, v, #name)
drivers/gpu/drm/rockchip/rockchip_drm_vop.c
71
#define VOP_HAS_REG(vop, group, name) \
drivers/gpu/drm/rockchip/rockchip_drm_vop.c
72
(!!(vop->data->group->name.mask))
drivers/gpu/drm/tegra/dpaux.c
386
unsigned int group)
drivers/gpu/drm/tegra/dpaux.c
388
return tegra_dpaux_groups[group];
drivers/gpu/drm/tegra/dpaux.c
392
unsigned group, const unsigned **pins,
drivers/gpu/drm/tegra/dpaux.c
432
unsigned int function, unsigned int group)
drivers/gpu/drm/tegra/drm.c
1005
iommu_detach_group(tegra->domain, client->group);
drivers/gpu/drm/tegra/drm.c
1007
iommu_group_put(client->group);
drivers/gpu/drm/tegra/drm.c
1008
client->group = NULL;
drivers/gpu/drm/tegra/drm.c
947
struct iommu_group *group = NULL;
drivers/gpu/drm/tegra/drm.c
971
group = iommu_group_get(client->dev);
drivers/gpu/drm/tegra/drm.c
972
if (!group)
drivers/gpu/drm/tegra/drm.c
976
err = iommu_attach_group(tegra->domain, group);
drivers/gpu/drm/tegra/drm.c
978
iommu_group_put(group);
drivers/gpu/drm/tegra/drm.c
986
client->group = group;
drivers/gpu/drm/tegra/drm.c
997
if (client->group) {
drivers/gpu/drm/tegra/nvdec.c
224
if (client->group) {
drivers/gpu/drm/tegra/nvdec.c
262
if (!client->group) {
drivers/gpu/drm/tegra/nvdec.c
284
if (client->group) {
drivers/gpu/drm/tegra/nvdec.c
299
if (!client->group)
drivers/gpu/drm/tegra/nvjpg.c
128
if (!client->group) {
drivers/gpu/drm/tegra/nvjpg.c
150
if (client->group) {
drivers/gpu/drm/tegra/nvjpg.c
165
if (!client->group)
drivers/gpu/drm/tegra/nvjpg.c
90
if (client->group) {
drivers/gpu/drm/tegra/plane.c
155
if (!dc->client.group) {
drivers/gpu/drm/tegra/vic.c
190
if (client->group) {
drivers/gpu/drm/tegra/vic.c
234
if (!client->group) {
drivers/gpu/drm/tegra/vic.c
260
if (client->group) {
drivers/gpu/drm/tegra/vic.c
296
if (!client->group)
drivers/gpu/drm/vkms/vkms_configfs.c
106
group)
drivers/gpu/drm/vkms/vkms_configfs.c
108
#define child_group_to_vkms_configfs_device(group) \
drivers/gpu/drm/vkms/vkms_configfs.c
109
device_item_to_vkms_configfs_device((&(group)->cg_item)->ci_parent)
drivers/gpu/drm/vkms/vkms_configfs.c
112
container_of(to_config_group((item)), struct vkms_configfs_plane, group)
drivers/gpu/drm/vkms/vkms_configfs.c
119
container_of(to_config_group((item)), struct vkms_configfs_crtc, group)
drivers/gpu/drm/vkms/vkms_configfs.c
123
group)
drivers/gpu/drm/vkms/vkms_configfs.c
131
group)
drivers/gpu/drm/vkms/vkms_configfs.c
202
static struct config_group *make_crtc_group(struct config_group *group,
drivers/gpu/drm/vkms/vkms_configfs.c
209
dev = child_group_to_vkms_configfs_device(group);
drivers/gpu/drm/vkms/vkms_configfs.c
228
config_group_init_type_name(&crtc->group, name, &crtc_item_type);
drivers/gpu/drm/vkms/vkms_configfs.c
231
return &crtc->group;
drivers/gpu/drm/vkms/vkms_configfs.c
30
struct config_group group;
drivers/gpu/drm/vkms/vkms_configfs.c
358
static struct config_group *make_plane_group(struct config_group *group,
drivers/gpu/drm/vkms/vkms_configfs.c
365
dev = child_group_to_vkms_configfs_device(group);
drivers/gpu/drm/vkms/vkms_configfs.c
384
config_group_init_type_name(&plane->group, name, &plane_item_type);
drivers/gpu/drm/vkms/vkms_configfs.c
390
&plane->group);
drivers/gpu/drm/vkms/vkms_configfs.c
393
return &plane->group;
drivers/gpu/drm/vkms/vkms_configfs.c
474
static struct config_group *make_encoder_group(struct config_group *group,
drivers/gpu/drm/vkms/vkms_configfs.c
481
dev = child_group_to_vkms_configfs_device(group);
drivers/gpu/drm/vkms/vkms_configfs.c
500
config_group_init_type_name(&encoder->group, name,
drivers/gpu/drm/vkms/vkms_configfs.c
507
&encoder->group);
drivers/gpu/drm/vkms/vkms_configfs.c
51
struct config_group group;
drivers/gpu/drm/vkms/vkms_configfs.c
510
return &encoder->group;
drivers/gpu/drm/vkms/vkms_configfs.c
641
static struct config_group *make_connector_group(struct config_group *group,
drivers/gpu/drm/vkms/vkms_configfs.c
648
dev = child_group_to_vkms_configfs_device(group);
drivers/gpu/drm/vkms/vkms_configfs.c
66
struct config_group group;
drivers/gpu/drm/vkms/vkms_configfs.c
667
config_group_init_type_name(&connector->group, name,
drivers/gpu/drm/vkms/vkms_configfs.c
674
&connector->group);
drivers/gpu/drm/vkms/vkms_configfs.c
677
return &connector->group;
drivers/gpu/drm/vkms/vkms_configfs.c
763
static struct config_group *make_device_group(struct config_group *group,
drivers/gpu/drm/vkms/vkms_configfs.c
783
config_group_init_type_name(&dev->group, name, &device_item_type);
drivers/gpu/drm/vkms/vkms_configfs.c
788
configfs_add_default_group(&dev->planes_group, &dev->group);
drivers/gpu/drm/vkms/vkms_configfs.c
792
configfs_add_default_group(&dev->crtcs_group, &dev->group);
drivers/gpu/drm/vkms/vkms_configfs.c
796
configfs_add_default_group(&dev->encoders_group, &dev->group);
drivers/gpu/drm/vkms/vkms_configfs.c
800
configfs_add_default_group(&dev->connectors_group, &dev->group);
drivers/gpu/drm/vkms/vkms_configfs.c
802
return &dev->group;
drivers/gpu/drm/vkms/vkms_configfs.c
81
struct config_group group;
drivers/gpu/drm/vkms/vkms_configfs.c
98
struct config_group group;
drivers/gpu/drm/xe/xe_configfs.c
1011
config_group_init_type_name(&dev->group, name, &xe_config_device_type);
drivers/gpu/drm/xe/xe_configfs.c
1014
configfs_add_default_group(&dev->sriov, &dev->group);
drivers/gpu/drm/xe/xe_configfs.c
1019
return &dev->group;
drivers/gpu/drm/xe/xe_configfs.c
1091
config_group_put(&dev->group);
drivers/gpu/drm/xe/xe_configfs.c
1109
config_group_put(&dev->group);
drivers/gpu/drm/xe/xe_configfs.c
1123
config_group_put(&dev->group);
drivers/gpu/drm/xe/xe_configfs.c
1167
config_group_put(&dev->group);
drivers/gpu/drm/xe/xe_configfs.c
1187
config_group_put(&dev->group);
drivers/gpu/drm/xe/xe_configfs.c
1214
config_group_put(&dev->group);
drivers/gpu/drm/xe/xe_configfs.c
1239
config_group_put(&dev->group);
drivers/gpu/drm/xe/xe_configfs.c
1266
config_group_put(&dev->group);
drivers/gpu/drm/xe/xe_configfs.c
254
struct config_group group;
drivers/gpu/drm/xe/xe_configfs.c
324
return container_of(to_config_group(item), struct xe_config_group_device, group);
drivers/gpu/drm/xe/xe_configfs.c
341
name = dev->group.cg_item.ci_name;
drivers/gpu/drm/xe/xe_configfs.c
947
static struct config_group *xe_config_make_device_group(struct config_group *group,
drivers/gpu/drm/xe/xe_eu_stall.c
410
u16 group, instance;
drivers/gpu/drm/xe/xe_eu_stall.c
414
for_each_dss_steering(xecore, gt, group, instance) {
drivers/gpu/drm/xe/xe_eu_stall.c
418
group, instance);
drivers/gpu/drm/xe/xe_eu_stall.c
436
static void clear_dropped_eviction_line_bit(struct xe_gt *gt, u16 group, u16 instance)
drivers/gpu/drm/xe/xe_eu_stall.c
449
xe_gt_mcr_unicast_write(gt, XEHPC_EUSTALL_REPORT, write_ptr_reg, group, instance);
drivers/gpu/drm/xe/xe_eu_stall.c
455
u16 group, u16 instance, unsigned int xecore)
drivers/gpu/drm/xe/xe_eu_stall.c
508
xe_gt_mcr_unicast_write(gt, XEHPC_EUSTALL_REPORT1, read_ptr_reg, group, instance);
drivers/gpu/drm/xe/xe_eu_stall.c
510
trace_xe_eu_stall_data_read(group, instance, read_ptr, write_ptr,
drivers/gpu/drm/xe/xe_eu_stall.c
518
clear_dropped_eviction_line_bit(gt, group, instance);
drivers/gpu/drm/xe/xe_eu_stall.c
542
u16 group, instance;
drivers/gpu/drm/xe/xe_eu_stall.c
558
for_each_dss_steering(xecore, gt, group, instance) {
drivers/gpu/drm/xe/xe_eu_stall.c
560
gt, group, instance, xecore);
drivers/gpu/drm/xe/xe_eu_stall.c
663
u16 group, instance;
drivers/gpu/drm/xe/xe_eu_stall.c
679
for_each_dss_steering(xecore, gt, group, instance) {
drivers/gpu/drm/xe/xe_eu_stall.c
680
write_ptr_reg = xe_gt_mcr_unicast_read(gt, XEHPC_EUSTALL_REPORT, group, instance);
drivers/gpu/drm/xe/xe_eu_stall.c
683
clear_dropped_eviction_line_bit(gt, group, instance);
drivers/gpu/drm/xe/xe_eu_stall.c
688
xe_gt_mcr_unicast_write(gt, XEHPC_EUSTALL_REPORT1, read_ptr_reg, group, instance);
drivers/gpu/drm/xe/xe_eu_stall.c
736
u16 group, instance;
drivers/gpu/drm/xe/xe_eu_stall.c
766
for_each_dss_steering(xecore, gt, group, instance) {
drivers/gpu/drm/xe/xe_exec.c
129
struct xe_hw_engine_group *group;
drivers/gpu/drm/xe/xe_exec.c
204
group = q->hwe->hw_engine_group;
drivers/gpu/drm/xe/xe_exec.c
208
err = xe_hw_engine_group_get_mode(group, mode, &previous_mode,
drivers/gpu/drm/xe/xe_exec.c
346
xe_hw_engine_group_resume_faulting_lr_jobs(group);
drivers/gpu/drm/xe/xe_exec.c
363
xe_hw_engine_group_put(group);
drivers/gpu/drm/xe/xe_exec_queue.c
115
struct xe_exec_queue_group *group = q->multi_queue.group;
drivers/gpu/drm/xe/xe_exec_queue.c
128
if (!group)
drivers/gpu/drm/xe/xe_exec_queue.c
132
xa_for_each(&group->xa, idx, lrc)
drivers/gpu/drm/xe/xe_exec_queue.c
135
xa_destroy(&group->xa);
drivers/gpu/drm/xe/xe_exec_queue.c
136
mutex_destroy(&group->list_lock);
drivers/gpu/drm/xe/xe_exec_queue.c
137
xe_bo_unpin_map_no_vm(group->cgp_bo);
drivers/gpu/drm/xe/xe_exec_queue.c
138
kfree(group);
drivers/gpu/drm/xe/xe_exec_queue.c
680
struct xe_exec_queue_group *group;
drivers/gpu/drm/xe/xe_exec_queue.c
683
group = kzalloc_obj(*group);
drivers/gpu/drm/xe/xe_exec_queue.c
684
if (!group)
drivers/gpu/drm/xe/xe_exec_queue.c
696
kfree(group);
drivers/gpu/drm/xe/xe_exec_queue.c
702
group->primary = q;
drivers/gpu/drm/xe/xe_exec_queue.c
703
group->cgp_bo = bo;
drivers/gpu/drm/xe/xe_exec_queue.c
704
INIT_LIST_HEAD(&group->list);
drivers/gpu/drm/xe/xe_exec_queue.c
705
xa_init_flags(&group->xa, XA_FLAGS_ALLOC1);
drivers/gpu/drm/xe/xe_exec_queue.c
706
mutex_init(&group->list_lock);
drivers/gpu/drm/xe/xe_exec_queue.c
707
q->multi_queue.group = group;
drivers/gpu/drm/xe/xe_exec_queue.c
712
might_lock(&group->list_lock);
drivers/gpu/drm/xe/xe_exec_queue.c
727
struct xe_exec_queue_group *group;
drivers/gpu/drm/xe/xe_exec_queue.c
746
group = primary->multi_queue.group;
drivers/gpu/drm/xe/xe_exec_queue.c
748
q->multi_queue.group = group;
drivers/gpu/drm/xe/xe_exec_queue.c
759
struct xe_exec_queue_group *group = q->multi_queue.group;
drivers/gpu/drm/xe/xe_exec_queue.c
766
err = xa_alloc(&group->xa, &pos, xe_lrc_get(q->lrc[0]),
drivers/gpu/drm/xe/xe_exec_queue.c
785
struct xe_exec_queue_group *group = q->multi_queue.group;
drivers/gpu/drm/xe/xe_exec_queue.c
790
lrc = xa_erase(&group->xa, q->multi_queue.pos);
drivers/gpu/drm/xe/xe_exec_queue.h
114
return xe_exec_queue_is_multi_queue(q) ? q->multi_queue.group->primary : q;
drivers/gpu/drm/xe/xe_exec_queue_types.h
159
struct xe_exec_queue_group *group;
drivers/gpu/drm/xe/xe_gt.c
310
u8 group, instance;
drivers/gpu/drm/xe/xe_gt.c
312
xe_gt_mcr_get_nonterminated_steering(gt, reg, &group, &instance);
drivers/gpu/drm/xe/xe_gt.c
316
REG_FIELD_PREP(SELECTIVE_READ_GROUP, group) |
drivers/gpu/drm/xe/xe_gt_mcr.c
412
void xe_gt_mcr_get_dss_steering(const struct xe_gt *gt, unsigned int dss, u16 *group, u16 *instance)
drivers/gpu/drm/xe/xe_gt_mcr.c
416
*group = dss / gt->steering_dss_per_grp;
drivers/gpu/drm/xe/xe_gt_mcr.c
428
u32 xe_gt_mcr_steering_info_to_dss_id(struct xe_gt *gt, u16 group, u16 instance)
drivers/gpu/drm/xe/xe_gt_mcr.c
430
return group * dss_per_group(gt) + instance;
drivers/gpu/drm/xe/xe_gt_mcr.c
630
u8 *group, u8 *instance)
drivers/gpu/drm/xe/xe_gt_mcr.c
645
*group = gt->steering[type].group_target;
drivers/gpu/drm/xe/xe_gt_mcr.c
665
*group = 0;
drivers/gpu/drm/xe/xe_gt_mcr.c
711
u8 rw_flag, int group, int instance, u32 value)
drivers/gpu/drm/xe/xe_gt_mcr.c
722
steer_val = REG_FIELD_PREP(MTL_MCR_GROUPID, group) |
drivers/gpu/drm/xe/xe_gt_mcr.c
726
steer_val = REG_FIELD_PREP(MCR_SLICE_MASK, group) |
drivers/gpu/drm/xe/xe_gt_mcr.c
781
u8 group, instance;
drivers/gpu/drm/xe/xe_gt_mcr.c
788
&group, &instance);
drivers/gpu/drm/xe/xe_gt_mcr.c
793
group, instance, 0);
drivers/gpu/drm/xe/xe_gt_mcr.c
814
int group, int instance)
drivers/gpu/drm/xe/xe_gt_mcr.c
821
val = rw_with_mcr_steering(gt, reg_mcr, MCR_OP_READ, group, instance, 0);
drivers/gpu/drm/xe/xe_gt_mcr.c
839
u32 value, int group, int instance)
drivers/gpu/drm/xe/xe_gt_mcr.c
844
rw_with_mcr_steering(gt, reg_mcr, MCR_OP_WRITE, group, instance, value);
drivers/gpu/drm/xe/xe_gt_mcr.h
21
int group, int instance);
drivers/gpu/drm/xe/xe_gt_mcr.h
25
u32 value, int group, int instance);
drivers/gpu/drm/xe/xe_gt_mcr.h
31
u8 *group, u8 *instance);
drivers/gpu/drm/xe/xe_gt_mcr.h
35
unsigned int dss, u16 *group, u16 *instance);
drivers/gpu/drm/xe/xe_gt_mcr.h
36
u32 xe_gt_mcr_steering_info_to_dss_id(struct xe_gt *gt, u16 group, u16 instance);
drivers/gpu/drm/xe/xe_gt_mcr.h
46
#define for_each_dss_steering(dss, gt, group, instance) \
drivers/gpu/drm/xe/xe_gt_mcr.h
48
for_each_if((xe_gt_mcr_get_dss_steering((gt), (dss), &(group), &(instance)), true))
drivers/gpu/drm/xe/xe_gt_mcr.h
58
#define for_each_geometry_dss(dss, gt, group, instance) \
drivers/gpu/drm/xe/xe_gt_mcr.h
59
for_each_dss_steering(dss, gt, group, instance) \
drivers/gpu/drm/xe/xe_gt_mcr.h
70
#define for_each_compute_dss(dss, gt, group, instance) \
drivers/gpu/drm/xe/xe_gt_mcr.h
71
for_each_dss_steering(dss, gt, group, instance) \
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
1023
static char *to_group_name(const char *what, u8 group, char *buf, size_t size)
drivers/gpu/drm/xe/xe_gt_sriov_pf_config.c
1025
snprintf(buf, size, "group%u%s%s", group, what ? " " : "", what ?: "");
drivers/gpu/drm/xe/xe_gt_sriov_pf_debugfs.c
279
u8 group;
drivers/gpu/drm/xe/xe_gt_sriov_pf_debugfs.c
283
for (group = 0; group < ARRAY_SIZE(values); group++) {
drivers/gpu/drm/xe/xe_gt_sriov_pf_debugfs.c
284
drm_printf(&p, "%s%u", first ? "" : ",", values[group]);
drivers/gpu/drm/xe/xe_gt_sriov_pf_debugfs.c
375
unsigned int group = (uintptr_t)extract_priv(dent);
drivers/gpu/drm/xe/xe_gt_sriov_pf_debugfs.c
382
if (group < num_groups) {
drivers/gpu/drm/xe/xe_gt_sriov_pf_debugfs.c
385
u32 mask = groups[group].engines[guc_class];
drivers/gpu/drm/xe/xe_gt_sriov_pf_debugfs.c
408
u8 group;
drivers/gpu/drm/xe/xe_gt_sriov_pf_debugfs.c
442
for (group = 0; group < GUC_MAX_SCHED_GROUPS; group++) {
drivers/gpu/drm/xe/xe_gt_sriov_pf_debugfs.c
445
snprintf(name, sizeof(name), "group%u", group);
drivers/gpu/drm/xe/xe_gt_sriov_pf_debugfs.c
446
debugfs_create_file(name, 0644, groups, (void *)(uintptr_t)group,
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
384
int group = 0;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
402
slice_to_group[slice] = group++;
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
414
if (group < 2)
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
422
if (group > gt->sriov.pf.policy.guc.sched_groups.max_groups) {
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
424
group, gt->sriov.pf.policy.guc.sched_groups.max_groups);
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
429
values = drmm_kcalloc(>_to_xe(gt)->drm, group, sizeof(struct guc_sched_group),
drivers/gpu/drm/xe/xe_gt_sriov_pf_policy.c
458
*num_groups = group;
drivers/gpu/drm/xe/xe_gt_throttle.c
125
const struct attribute_group *group;
drivers/gpu/drm/xe/xe_gt_throttle.c
134
group = get_platform_throttle_group(xe);
drivers/gpu/drm/xe/xe_gt_throttle.c
135
for (pother = group->attrs; *pother; pother++) {
drivers/gpu/drm/xe/xe_gt_throttle.c
251
const struct attribute_group *group = get_platform_throttle_group(xe);
drivers/gpu/drm/xe/xe_gt_throttle.c
253
sysfs_remove_group(gt->freq, group);
drivers/gpu/drm/xe/xe_gt_throttle.c
259
const struct attribute_group *group = get_platform_throttle_group(xe);
drivers/gpu/drm/xe/xe_gt_throttle.c
262
err = sysfs_create_group(gt->freq, group);
drivers/gpu/drm/xe/xe_gt_topology.c
362
u16 group, instance;
drivers/gpu/drm/xe/xe_gt_topology.c
364
for_each_dss_steering(xecore, gt, group, instance) {
drivers/gpu/drm/xe/xe_gt_topology.c
365
if (last_group != group) {
drivers/gpu/drm/xe/xe_gt_topology.c
366
if (group - last_group > 1)
drivers/gpu/drm/xe/xe_gt_topology.c
368
last_group = group;
drivers/gpu/drm/xe/xe_guc_ads.c
716
u8 group, instance;
drivers/gpu/drm/xe/xe_guc_ads.c
719
&group, &instance);
drivers/gpu/drm/xe/xe_guc_ads.c
722
entry.flags |= FIELD_PREP(GUC_REGSET_STEERING_GROUP, group);
drivers/gpu/drm/xe/xe_guc_capture.c
1575
int group, instance;
drivers/gpu/drm/xe/xe_guc_capture.c
1577
group = FIELD_GET(GUC_REGSET_STEERING_GROUP, desc.flags);
drivers/gpu/drm/xe/xe_guc_capture.c
1580
group, instance);
drivers/gpu/drm/xe/xe_guc_submit.c
1475
READ_ONCE(q->multi_queue.group->banned))
drivers/gpu/drm/xe/xe_guc_submit.c
1662
struct xe_exec_queue_group *group = q->multi_queue.group;
drivers/gpu/drm/xe/xe_guc_submit.c
1664
mutex_lock(&group->list_lock);
drivers/gpu/drm/xe/xe_guc_submit.c
1666
mutex_unlock(&group->list_lock);
drivers/gpu/drm/xe/xe_guc_submit.c
1859
struct xe_exec_queue_group *group = q->multi_queue.group;
drivers/gpu/drm/xe/xe_guc_submit.c
1864
action[len++] = group->primary->guc->id;
drivers/gpu/drm/xe/xe_guc_submit.c
1992
struct xe_exec_queue_group *group = q->multi_queue.group;
drivers/gpu/drm/xe/xe_guc_submit.c
1995
mutex_lock(&group->list_lock);
drivers/gpu/drm/xe/xe_guc_submit.c
1996
list_add_tail(&q->multi_queue.link, &group->list);
drivers/gpu/drm/xe/xe_guc_submit.c
1997
mutex_unlock(&group->list_lock);
drivers/gpu/drm/xe/xe_guc_submit.c
3096
WRITE_ONCE(q->multi_queue.group->sync_pending, false);
drivers/gpu/drm/xe/xe_guc_submit.c
581
struct xe_exec_queue_group *group = q->multi_queue.group;
drivers/gpu/drm/xe/xe_guc_submit.c
588
WRITE_ONCE(group->banned, true);
drivers/gpu/drm/xe/xe_guc_submit.c
591
mutex_lock(&group->list_lock);
drivers/gpu/drm/xe/xe_guc_submit.c
592
list_for_each_entry(eq, &group->list, multi_queue.link)
drivers/gpu/drm/xe/xe_guc_submit.c
594
mutex_unlock(&group->list_lock);
drivers/gpu/drm/xe/xe_guc_submit.c
601
struct xe_exec_queue_group *group = q->multi_queue.group;
drivers/gpu/drm/xe/xe_guc_submit.c
605
WRITE_ONCE(group->banned, true);
drivers/gpu/drm/xe/xe_guc_submit.c
611
mutex_lock(&group->list_lock);
drivers/gpu/drm/xe/xe_guc_submit.c
612
list_for_each_entry(eq, &group->list, multi_queue.link) {
drivers/gpu/drm/xe/xe_guc_submit.c
617
mutex_unlock(&group->list_lock);
drivers/gpu/drm/xe/xe_guc_submit.c
628
struct xe_exec_queue_group *group = q->multi_queue.group;
drivers/gpu/drm/xe/xe_guc_submit.c
636
mutex_lock(&group->list_lock);
drivers/gpu/drm/xe/xe_guc_submit.c
637
list_for_each_entry(eq, &group->list, multi_queue.link)
drivers/gpu/drm/xe/xe_guc_submit.c
639
mutex_unlock(&group->list_lock);
drivers/gpu/drm/xe/xe_guc_submit.c
728
struct xe_exec_queue_group *group = q->multi_queue.group;
drivers/gpu/drm/xe/xe_guc_submit.c
729
u32 guc_id = group->primary->guc->id;
drivers/gpu/drm/xe/xe_guc_submit.c
732
xe_map_wr(xe, &group->cgp_bo->vmap, 0, u32,
drivers/gpu/drm/xe/xe_guc_submit.c
735
xe_map_wr(xe, &group->cgp_bo->vmap,
drivers/gpu/drm/xe/xe_guc_submit.c
739
xe_map_wr(xe, &group->cgp_bo->vmap,
drivers/gpu/drm/xe/xe_guc_submit.c
744
xe_map_wr(xe, &group->cgp_bo->vmap, 17 * sizeof(u32),
drivers/gpu/drm/xe/xe_guc_submit.c
746
xe_map_wr(xe, &group->cgp_bo->vmap, 16 * sizeof(u32), u32, 0);
drivers/gpu/drm/xe/xe_guc_submit.c
748
xe_map_wr(xe, &group->cgp_bo->vmap, 16 * sizeof(u32),
drivers/gpu/drm/xe/xe_guc_submit.c
750
xe_map_wr(xe, &group->cgp_bo->vmap, 17 * sizeof(u32), u32, 0);
drivers/gpu/drm/xe/xe_guc_submit.c
758
struct xe_exec_queue_group *group = q->multi_queue.group;
drivers/gpu/drm/xe/xe_guc_submit.c
772
!READ_ONCE(group->sync_pending) ||
drivers/gpu/drm/xe/xe_guc_submit.c
786
WRITE_ONCE(group->sync_pending, true);
drivers/gpu/drm/xe/xe_guc_submit.c
828
action[len++] = q->multi_queue.group->primary->guc->id;
drivers/gpu/drm/xe/xe_guc_submit.c
937
struct xe_exec_queue_group *group = q->multi_queue.group;
drivers/gpu/drm/xe/xe_guc_submit.c
939
info.cgp_lo = xe_bo_ggtt_addr(group->cgp_bo);
drivers/gpu/drm/xe/xe_hw_engine_group.c
125
int xe_hw_engine_group_add_exec_queue(struct xe_hw_engine_group *group, struct xe_exec_queue *q)
drivers/gpu/drm/xe/xe_hw_engine_group.c
130
xe_assert(xe, group);
drivers/gpu/drm/xe/xe_hw_engine_group.c
137
err = down_write_killable(&group->mode_sem);
drivers/gpu/drm/xe/xe_hw_engine_group.c
141
if (xe_vm_in_fault_mode(q->vm) && group->cur_mode == EXEC_MODE_DMA_FENCE) {
drivers/gpu/drm/xe/xe_hw_engine_group.c
147
xe_hw_engine_group_resume_faulting_lr_jobs(group);
drivers/gpu/drm/xe/xe_hw_engine_group.c
150
list_add(&q->hw_engine_group_link, &group->exec_queue_list);
drivers/gpu/drm/xe/xe_hw_engine_group.c
151
up_write(&group->mode_sem);
drivers/gpu/drm/xe/xe_hw_engine_group.c
156
up_write(&group->mode_sem);
drivers/gpu/drm/xe/xe_hw_engine_group.c
166
void xe_hw_engine_group_del_exec_queue(struct xe_hw_engine_group *group, struct xe_exec_queue *q)
drivers/gpu/drm/xe/xe_hw_engine_group.c
170
xe_assert(xe, group);
drivers/gpu/drm/xe/xe_hw_engine_group.c
173
down_write(&group->mode_sem);
drivers/gpu/drm/xe/xe_hw_engine_group.c
178
up_write(&group->mode_sem);
drivers/gpu/drm/xe/xe_hw_engine_group.c
186
void xe_hw_engine_group_resume_faulting_lr_jobs(struct xe_hw_engine_group *group)
drivers/gpu/drm/xe/xe_hw_engine_group.c
188
queue_work(group->resume_wq, &group->resume_work);
drivers/gpu/drm/xe/xe_hw_engine_group.c
198
static int xe_hw_engine_group_suspend_faulting_lr_jobs(struct xe_hw_engine_group *group,
drivers/gpu/drm/xe/xe_hw_engine_group.c
207
lockdep_assert_held_write(&group->mode_sem);
drivers/gpu/drm/xe/xe_hw_engine_group.c
209
list_for_each_entry(q, &group->exec_queue_list, hw_engine_group_link) {
drivers/gpu/drm/xe/xe_hw_engine_group.c
21
struct xe_hw_engine_group *group = container_of(w, struct xe_hw_engine_group, resume_work);
drivers/gpu/drm/xe/xe_hw_engine_group.c
229
list_for_each_entry(q, &group->exec_queue_list, hw_engine_group_link) {
drivers/gpu/drm/xe/xe_hw_engine_group.c
245
xe_hw_engine_group_resume_faulting_lr_jobs(group);
drivers/gpu/drm/xe/xe_hw_engine_group.c
25
err = xe_hw_engine_group_get_mode(group, EXEC_MODE_LR, &previous_mode,
drivers/gpu/drm/xe/xe_hw_engine_group.c
260
static int xe_hw_engine_group_wait_for_dma_fence_jobs(struct xe_hw_engine_group *group)
drivers/gpu/drm/xe/xe_hw_engine_group.c
268
lockdep_assert_held_write(&group->mode_sem);
drivers/gpu/drm/xe/xe_hw_engine_group.c
270
list_for_each_entry(q, &group->exec_queue_list, hw_engine_group_link) {
drivers/gpu/drm/xe/xe_hw_engine_group.c
293
static int switch_mode(struct xe_hw_engine_group *group, bool has_deps)
drivers/gpu/drm/xe/xe_hw_engine_group.c
298
lockdep_assert_held_write(&group->mode_sem);
drivers/gpu/drm/xe/xe_hw_engine_group.c
300
switch (group->cur_mode) {
drivers/gpu/drm/xe/xe_hw_engine_group.c
303
err = xe_hw_engine_group_suspend_faulting_lr_jobs(group,
drivers/gpu/drm/xe/xe_hw_engine_group.c
308
err = xe_hw_engine_group_wait_for_dma_fence_jobs(group);
drivers/gpu/drm/xe/xe_hw_engine_group.c
315
group->cur_mode = new_mode;
drivers/gpu/drm/xe/xe_hw_engine_group.c
33
list_for_each_entry(q, &group->exec_queue_list, hw_engine_group_link) {
drivers/gpu/drm/xe/xe_hw_engine_group.c
343
int xe_hw_engine_group_get_mode(struct xe_hw_engine_group *group,
drivers/gpu/drm/xe/xe_hw_engine_group.c
347
__acquires(&group->mode_sem)
drivers/gpu/drm/xe/xe_hw_engine_group.c
350
int err = down_read_interruptible(&group->mode_sem);
drivers/gpu/drm/xe/xe_hw_engine_group.c
355
*previous_mode = group->cur_mode;
drivers/gpu/drm/xe/xe_hw_engine_group.c
357
if (new_mode != group->cur_mode) {
drivers/gpu/drm/xe/xe_hw_engine_group.c
358
up_read(&group->mode_sem);
drivers/gpu/drm/xe/xe_hw_engine_group.c
360
err = down_write_killable(&group->mode_sem);
drivers/gpu/drm/xe/xe_hw_engine_group.c
364
if (new_mode != group->cur_mode) {
drivers/gpu/drm/xe/xe_hw_engine_group.c
365
err = switch_mode(group, has_deps);
drivers/gpu/drm/xe/xe_hw_engine_group.c
367
up_write(&group->mode_sem);
drivers/gpu/drm/xe/xe_hw_engine_group.c
380
downgrade_write(&group->mode_sem);
drivers/gpu/drm/xe/xe_hw_engine_group.c
390
void xe_hw_engine_group_put(struct xe_hw_engine_group *group)
drivers/gpu/drm/xe/xe_hw_engine_group.c
391
__releases(&group->mode_sem)
drivers/gpu/drm/xe/xe_hw_engine_group.c
393
up_read(&group->mode_sem);
drivers/gpu/drm/xe/xe_hw_engine_group.c
41
xe_hw_engine_group_put(group);
drivers/gpu/drm/xe/xe_hw_engine_group.c
47
struct xe_hw_engine_group *group;
drivers/gpu/drm/xe/xe_hw_engine_group.c
50
group = drmm_kzalloc(&xe->drm, sizeof(*group), GFP_KERNEL);
drivers/gpu/drm/xe/xe_hw_engine_group.c
51
if (!group)
drivers/gpu/drm/xe/xe_hw_engine_group.c
54
group->resume_wq = alloc_workqueue("xe-resume-lr-jobs-wq", 0, 0);
drivers/gpu/drm/xe/xe_hw_engine_group.c
55
if (!group->resume_wq)
drivers/gpu/drm/xe/xe_hw_engine_group.c
58
err = drmm_add_action_or_reset(&xe->drm, __drmm_workqueue_release, group->resume_wq);
drivers/gpu/drm/xe/xe_hw_engine_group.c
62
init_rwsem(&group->mode_sem);
drivers/gpu/drm/xe/xe_hw_engine_group.c
63
INIT_WORK(&group->resume_work, hw_engine_group_resume_lr_jobs_func);
drivers/gpu/drm/xe/xe_hw_engine_group.c
64
INIT_LIST_HEAD(&group->exec_queue_list);
drivers/gpu/drm/xe/xe_hw_engine_group.c
66
return group;
drivers/gpu/drm/xe/xe_hw_engine_group.h
18
int xe_hw_engine_group_add_exec_queue(struct xe_hw_engine_group *group, struct xe_exec_queue *q);
drivers/gpu/drm/xe/xe_hw_engine_group.h
19
void xe_hw_engine_group_del_exec_queue(struct xe_hw_engine_group *group, struct xe_exec_queue *q);
drivers/gpu/drm/xe/xe_hw_engine_group.h
21
int xe_hw_engine_group_get_mode(struct xe_hw_engine_group *group,
drivers/gpu/drm/xe/xe_hw_engine_group.h
25
void xe_hw_engine_group_put(struct xe_hw_engine_group *group);
drivers/gpu/drm/xe/xe_hw_engine_group.h
29
void xe_hw_engine_group_resume_faulting_lr_jobs(struct xe_hw_engine_group *group);
drivers/gpu/host1x/dev.c
406
host->group = iommu_group_get(host->dev);
drivers/gpu/host1x/dev.c
407
if (host->group) {
drivers/gpu/host1x/dev.c
423
err = iommu_attach_group(host->domain, host->group);
drivers/gpu/host1x/dev.c
450
iommu_group_put(host->group);
drivers/gpu/host1x/dev.c
451
host->group = NULL;
drivers/gpu/host1x/dev.c
493
iommu_detach_group(host->domain, host->group);
drivers/gpu/host1x/dev.c
500
iommu_group_put(host->group);
drivers/gpu/host1x/dev.c
501
host->group = NULL;
drivers/gpu/host1x/dev.h
144
struct iommu_group *group;
drivers/hid/bpf/progs/hid_bpf_helpers.h
121
__uint(group, (g)); \
drivers/hid/hid-core.c
2177
(id->group == HID_GROUP_ANY || id->group == hdev->group) &&
drivers/hid/hid-core.c
2619
dynid->id.group = HID_GROUP_ANY;
drivers/hid/hid-core.c
2726
hdev->group = HID_GROUP_GENERIC;
drivers/hid/hid-core.c
2727
} else if (!hdev->group &&
drivers/hid/hid-core.c
2759
hdev->group = 0;
drivers/hid/hid-core.c
2853
hdev->bus, hdev->group, hdev->vendor, hdev->product);
drivers/hid/hid-core.c
2889
hdev->bus, hdev->group, hdev->vendor, hdev->product))
drivers/hid/hid-core.c
839
hid->group = HID_GROUP_MULTITOUCH;
drivers/hid/hid-core.c
861
hid->group = HID_GROUP_SENSOR_HUB;
drivers/hid/hid-core.c
865
hid->group == HID_GROUP_MULTITOUCH)
drivers/hid/hid-core.c
866
hid->group = HID_GROUP_GENERIC;
drivers/hid/hid-core.c
880
parser->device->group =
drivers/hid/hid-core.c
944
hid->group = HID_GROUP_GENERIC;
drivers/hid/hid-core.c
967
(hid->group == HID_GROUP_MULTITOUCH))
drivers/hid/hid-core.c
968
hid->group = HID_GROUP_MULTITOUCH_WIN_8;
drivers/hid/hid-core.c
975
hid->group = HID_GROUP_WACOM;
drivers/hid/hid-core.c
978
if (hid->group == HID_GROUP_GENERIC)
drivers/hid/hid-core.c
985
hid->group = HID_GROUP_RMI;
drivers/hid/hid-logitech-dj.c
846
dj_hiddev->group = HID_GROUP_LOGITECH_27MHZ_DEVICE;
drivers/hid/hid-logitech-dj.c
848
dj_hiddev->group = HID_GROUP_LOGITECH_DJ_DEVICE;
drivers/hid/hid-logitech-hidpp.c
3759
if (hdev->group == HID_GROUP_LOGITECH_27MHZ_DEVICE ||
drivers/hid/hid-logitech-hidpp.c
3879
if (hidpp->hid_dev->group == HID_GROUP_LOGITECH_27MHZ_DEVICE &&
drivers/hid/hid-logitech-hidpp.c
4409
if (id->group == HID_GROUP_LOGITECH_27MHZ_DEVICE &&
drivers/hid/hid-logitech-hidpp.c
4414
if (id->group == HID_GROUP_LOGITECH_27MHZ_DEVICE &&
drivers/hid/hid-logitech-hidpp.c
4462
if (id->group == HID_GROUP_LOGITECH_DJ_DEVICE)
drivers/hid/hid-multitouch.c
2003
if (id->group != HID_GROUP_MULTITOUCH_WIN_8)
drivers/hid/hid-multitouch.c
711
if (field->logical == HID_DG_FINGER || td->hdev->group != HID_GROUP_MULTITOUCH_WIN_8) {
drivers/hid/hid-quirks.c
1179
hdev->group = q_new->hid_bl_item.group = id->group;
drivers/hid/hid-steam.c
1203
client_hdev->group = HID_GROUP_STEAM;
drivers/hid/hid-steam.c
1224
if (hdev->group == HID_GROUP_STEAM)
drivers/hid/hid-steam.c
1317
if (!steam || hdev->group == HID_GROUP_STEAM) {
drivers/hid/i2c-hid/i2c-hid-core.c
574
if (ihid->hid->group != HID_GROUP_RMI)
drivers/hid/wacom.h
117
unsigned int group;
drivers/hid/wacom.h
148
struct attribute_group group;
drivers/hid/wacom.h
246
struct wacom_led *wacom_led_find(struct wacom *wacom, unsigned int group,
drivers/hid/wacom_sys.c
1103
struct wacom_group_leds *group = &wacom->led.groups[i];
drivers/hid/wacom_sys.c
1105
for (unsigned int j = 0; j < group->count; j++) {
drivers/hid/wacom_sys.c
1107
group->leds[j].llv = *dest;
drivers/hid/wacom_sys.c
1109
group->leds[j].hlv = *dest;
drivers/hid/wacom_sys.c
1246
const struct attribute_group *group;
drivers/hid/wacom_sys.c
1256
__func__, devres->group->name);
drivers/hid/wacom_sys.c
1257
sysfs_remove_group(kobj, devres->group);
drivers/hid/wacom_sys.c
1262
const struct attribute_group *group)
drivers/hid/wacom_sys.c
1273
devres->group = group;
drivers/hid/wacom_sys.c
1276
error = sysfs_create_group(devres->root, group);
drivers/hid/wacom_sys.c
1288
const struct attribute_group *group)
drivers/hid/wacom_sys.c
1291
group);
drivers/hid/wacom_sys.c
1346
if (wacom->led.groups[led->group].select != led->id)
drivers/hid/wacom_sys.c
1362
wacom->led.groups[led->group].select != led->id)) {
drivers/hid/wacom_sys.c
1370
wacom->led.groups[led->group].select = led->id;
drivers/hid/wacom_sys.c
1386
struct wacom_led *led, unsigned int group,
drivers/hid/wacom_sys.c
1395
group,
drivers/hid/wacom_sys.c
1400
led->group = group;
drivers/hid/wacom_sys.c
1418
if (id == wacom->led.groups[group].select)
drivers/hid/wacom_sys.c
1443
struct wacom_group_leds *group = data;
drivers/hid/wacom_sys.c
1445
devres_release_group(group->dev, group);
drivers/hid/wacom_sys.c
1507
struct wacom_group_leds *group;
drivers/hid/wacom_sys.c
1512
group = &wacom->led.groups[group_id];
drivers/hid/wacom_sys.c
1514
if (!group->leds)
drivers/hid/wacom_sys.c
1517
id %= group->count;
drivers/hid/wacom_sys.c
1519
return &group->leds[id];
drivers/hid/wacom_sys.c
1531
int group, next;
drivers/hid/wacom_sys.c
1536
group = cur->group;
drivers/hid/wacom_sys.c
1540
next_led = wacom_led_find(wacom, group, ++next);
drivers/hid/wacom_sys.c
1930
remote->remotes[index].group.name = devm_kasprintf(&wacom->hdev->dev,
drivers/hid/wacom_sys.c
1933
if (!remote->remotes[index].group.name)
drivers/hid/wacom_sys.c
1937
&remote->remotes[index].group);
drivers/hid/wacom_sys.c
1939
remote->remotes[index].group.name = NULL;
drivers/hid/wacom_sys.c
2043
remote->remotes[0].group = remote0_serial_group;
drivers/hid/wacom_sys.c
2044
remote->remotes[1].group = remote1_serial_group;
drivers/hid/wacom_sys.c
2045
remote->remotes[2].group = remote2_serial_group;
drivers/hid/wacom_sys.c
2046
remote->remotes[3].group = remote3_serial_group;
drivers/hid/wacom_sys.c
2047
remote->remotes[4].group = remote4_serial_group;
drivers/hid/wacom_sys.c
2622
if (remote->remotes[i].group.name)
drivers/hid/wacom_sys.c
2627
remote->remotes[i].group.name = NULL;
drivers/hid/wacom_sys.c
2665
remote->remotes[index].input->uniq = remote->remotes[index].group.name;
drivers/hid/wacom_wac.c
39
int group);
drivers/hid/wacom_wac.c
4114
static void wacom_24hd_update_leds(struct wacom *wacom, int mask, int group)
drivers/hid/wacom_wac.c
4125
if (group == 0)
drivers/hid/wacom_wac.c
4129
led = wacom_led_find(wacom, group, i);
drivers/hid/wacom_wac.c
4132
i, group);
drivers/hid/wacom_wac.c
4145
int mask, int group)
drivers/hid/wacom_wac.c
4155
group = 1 - group;
drivers/hid/wacom_wac.c
4157
group_button = group * (button_count/wacom->led.count);
drivers/hid/wacom_wac.c
4166
int group)
drivers/hid/wacom_wac.c
4173
return wacom_24hd_update_leds(wacom, mask, group);
drivers/hid/wacom_wac.c
4175
pressed = wacom_is_led_toggled(wacom, button_count, mask, group);
drivers/hid/wacom_wac.c
4176
cur = wacom->led.groups[group].select;
drivers/hid/wacom_wac.c
4178
led = wacom_led_find(wacom, group, cur);
drivers/hid/wacom_wac.c
4181
cur, group);
drivers/hid/wacom_wac.c
4196
group);
drivers/hwmon/dme1737.c
2053
const struct attribute_group *group,
drivers/hwmon/dme1737.c
2058
for (attr = group->attrs; *attr; attr++)
drivers/hwmon/fam15h_power.c
323
data->group.attrs = fam15h_power_attrs;
drivers/hwmon/fam15h_power.c
472
data->groups[0] = &data->group;
drivers/hwmon/fam15h_power.c
54
struct attribute_group group;
drivers/hwmon/hwmon.c
135
if (hwdev->group.attrs)
drivers/hwmon/hwmon.c
136
hwmon_free_attrs(hwdev->group.attrs);
drivers/hwmon/hwmon.c
42
struct attribute_group group;
drivers/hwmon/hwmon.c
948
hwdev->group.attrs = attrs;
drivers/hwmon/hwmon.c
950
hwdev->groups[ngroups++] = &hwdev->group;
drivers/hwmon/ibmpowernv.c
468
sensor_groups[type].group.attrs = devm_kcalloc(&pdev->dev,
drivers/hwmon/ibmpowernv.c
472
if (!sensor_groups[type].group.attrs)
drivers/hwmon/ibmpowernv.c
475
pgroups[type] = &sensor_groups[type].group;
drivers/hwmon/ibmpowernv.c
61
struct attribute_group group;
drivers/hwmon/max31760.c
50
struct attribute_group group;
drivers/hwmon/max31760.c
513
state->group.attrs = state->attrs;
drivers/hwmon/max31760.c
514
state->groups[0] = &state->group;
drivers/hwmon/nct6683.c
1199
struct attribute_group *group;
drivers/hwmon/nct6683.c
1262
group = nct6683_create_attr_group(dev,
drivers/hwmon/nct6683.c
1265
if (IS_ERR(group))
drivers/hwmon/nct6683.c
1266
return PTR_ERR(group);
drivers/hwmon/nct6683.c
1267
data->groups[groups++] = group;
drivers/hwmon/nct6683.c
1271
group = nct6683_create_attr_group(dev,
drivers/hwmon/nct6683.c
1274
if (IS_ERR(group))
drivers/hwmon/nct6683.c
1275
return PTR_ERR(group);
drivers/hwmon/nct6683.c
1276
data->groups[groups++] = group;
drivers/hwmon/nct6683.c
1280
group = nct6683_create_attr_group(dev,
drivers/hwmon/nct6683.c
1283
if (IS_ERR(group))
drivers/hwmon/nct6683.c
1284
return PTR_ERR(group);
drivers/hwmon/nct6683.c
1285
data->groups[groups++] = group;
drivers/hwmon/nct6683.c
1289
group = nct6683_create_attr_group(dev,
drivers/hwmon/nct6683.c
1292
if (IS_ERR(group))
drivers/hwmon/nct6683.c
1293
return PTR_ERR(group);
drivers/hwmon/nct6683.c
1294
data->groups[groups++] = group;
drivers/hwmon/nct6683.c
420
struct attribute_group *group;
drivers/hwmon/nct6683.c
434
group = devm_kzalloc(dev, sizeof(*group), GFP_KERNEL);
drivers/hwmon/nct6683.c
435
if (group == NULL)
drivers/hwmon/nct6683.c
448
group->attrs = attrs;
drivers/hwmon/nct6683.c
449
group->is_visible = tg->is_visible;
drivers/hwmon/nct6683.c
484
return group;
drivers/hwmon/nct6775-core.c
1080
struct attribute_group *group;
drivers/hwmon/nct6775-core.c
1098
group = devm_kzalloc(dev, sizeof(*group), GFP_KERNEL);
drivers/hwmon/nct6775-core.c
1099
if (group == NULL)
drivers/hwmon/nct6775-core.c
1112
group->attrs = attrs;
drivers/hwmon/nct6775-core.c
1113
group->is_visible = tg->is_visible;
drivers/hwmon/nct6775-core.c
1148
return nct6775_add_attr_group(data, group);
drivers/hwmon/nct6775.h
225
nct6775_add_attr_group(struct nct6775_data *data, const struct attribute_group *group)
drivers/hwmon/nct6775.h
231
data->groups[data->num_groups++] = group;
drivers/hwmon/occ/common.c
1042
occ->group.attrs[i] = &occ->attrs[i].sensor.dev_attr.attr;
drivers/hwmon/occ/common.c
1173
occ->groups[0] = &occ->group;
drivers/hwmon/occ/common.c
863
occ->group.attrs = devm_kcalloc(dev, num_attrs + 1,
drivers/hwmon/occ/common.c
864
sizeof(*occ->group.attrs),
drivers/hwmon/occ/common.c
866
if (!occ->group.attrs)
drivers/hwmon/occ/common.h
107
struct attribute_group group;
drivers/hwmon/pmbus/pmbus_core.c
109
struct attribute_group group;
drivers/hwmon/pmbus/pmbus_core.c
1280
void *new_attrs = devm_krealloc_array(data->dev, data->group.attrs,
drivers/hwmon/pmbus/pmbus_core.c
1285
data->group.attrs = new_attrs;
drivers/hwmon/pmbus/pmbus_core.c
1289
data->group.attrs[data->num_attributes++] = attr;
drivers/hwmon/pmbus/pmbus_core.c
1290
data->group.attrs[data->num_attributes] = NULL;
drivers/hwmon/pmbus/pmbus_core.c
2971
struct device_attribute *da = to_dev_attr(data->group.attrs[i]);
drivers/hwmon/pmbus/pmbus_core.c
3837
data->groups[0] = &data->group;
drivers/hwmon/scpi-hwmon.c
244
scpi_sensors->group.attrs = scpi_sensors->attrs;
drivers/hwmon/scpi-hwmon.c
245
scpi_sensors->groups[0] = &scpi_sensors->group;
drivers/hwmon/scpi-hwmon.c
37
struct attribute_group group;
drivers/hwtracing/coresight/coresight-cti-sysfs.c
1050
struct attribute_group *group = NULL;
drivers/hwtracing/coresight/coresight-cti-sysfs.c
1053
group = devm_kzalloc(dev, sizeof(struct attribute_group), GFP_KERNEL);
drivers/hwtracing/coresight/coresight-cti-sysfs.c
1054
if (!group)
drivers/hwtracing/coresight/coresight-cti-sysfs.c
1057
group->name = devm_kasprintf(dev, GFP_KERNEL, "triggers%d", con_idx);
drivers/hwtracing/coresight/coresight-cti-sysfs.c
1058
if (!group->name)
drivers/hwtracing/coresight/coresight-cti-sysfs.c
1062
ctidev->con_groups[grp_idx] = group;
drivers/hwtracing/coresight/coresight-cti-sysfs.c
1063
tc->attr_group = group;
drivers/hwtracing/coresight/coresight-cti-sysfs.c
1064
return group;
drivers/hwtracing/coresight/coresight-syscfg-configfs.c
103
struct cscfg_fs_config, group);
drivers/hwtracing/coresight/coresight-syscfg-configfs.c
122
struct cscfg_fs_config, group);
drivers/hwtracing/coresight/coresight-syscfg-configfs.c
131
struct cscfg_fs_config, group);
drivers/hwtracing/coresight/coresight-syscfg-configfs.c
198
config_group_init_type_name(&cfg_fs_preset->group, name,
drivers/hwtracing/coresight/coresight-syscfg-configfs.c
200
configfs_add_default_group(&cfg_fs_preset->group, &cfg_view->group);
drivers/hwtracing/coresight/coresight-syscfg-configfs.c
219
config_group_init_type_name(&cfg_view->group, config_desc->name, &cscfg_config_view_type);
drivers/hwtracing/coresight/coresight-syscfg-configfs.c
226
return &cfg_view->group;
drivers/hwtracing/coresight/coresight-syscfg-configfs.c
234
struct cscfg_fs_feature, group);
drivers/hwtracing/coresight/coresight-syscfg-configfs.c
243
struct cscfg_fs_feature, group);
drivers/hwtracing/coresight/coresight-syscfg-configfs.c
261
struct cscfg_fs_feature, group);
drivers/hwtracing/coresight/coresight-syscfg-configfs.c
283
struct cscfg_fs_param, group);
drivers/hwtracing/coresight/coresight-syscfg-configfs.c
293
struct cscfg_fs_param, group);
drivers/hwtracing/coresight/coresight-syscfg-configfs.c
30
struct cscfg_fs_config, group);
drivers/hwtracing/coresight/coresight-syscfg-configfs.c
336
config_group_init_type_name(¶m_item->group,
drivers/hwtracing/coresight/coresight-syscfg-configfs.c
339
configfs_add_default_group(¶m_item->group, params_group);
drivers/hwtracing/coresight/coresight-syscfg-configfs.c
370
config_group_init_type_name(&feat_view->group,
drivers/hwtracing/coresight/coresight-syscfg-configfs.c
375
configfs_add_default_group(params_group, &feat_view->group);
drivers/hwtracing/coresight/coresight-syscfg-configfs.c
380
return &feat_view->group;
drivers/hwtracing/coresight/coresight-syscfg-configfs.c
39
struct cscfg_fs_config, group);
drivers/hwtracing/coresight/coresight-syscfg-configfs.c
60
fs_preset = container_of(to_config_group(item), struct cscfg_fs_preset, group);
drivers/hwtracing/coresight/coresight-syscfg-configfs.c
94
struct cscfg_fs_config, group);
drivers/hwtracing/coresight/coresight-syscfg-configfs.h
17
struct config_group group;
drivers/hwtracing/coresight/coresight-syscfg-configfs.h
25
struct config_group group;
drivers/hwtracing/coresight/coresight-syscfg-configfs.h
32
struct config_group group;
drivers/hwtracing/coresight/coresight-syscfg-configfs.h
39
struct config_group group;
drivers/hwtracing/stm/policy.c
223
stp_policy_node_make(struct config_group *group, const char *name)
drivers/hwtracing/stm/policy.c
230
if (group->cg_item.ci_type == &stp_policy_type) {
drivers/hwtracing/stm/policy.c
231
policy = container_of(group, struct stp_policy, group);
drivers/hwtracing/stm/policy.c
233
parent_node = container_of(group, struct stp_policy_node,
drivers/hwtracing/stm/policy.c
234
group);
drivers/hwtracing/stm/policy.c
25
struct config_group group;
drivers/hwtracing/stm/policy.c
254
config_group_init_type_name(&policy_node->group, name, type);
drivers/hwtracing/stm/policy.c
264
return &policy_node->group;
drivers/hwtracing/stm/policy.c
268
stp_policy_node_drop(struct config_group *group, struct config_item *item)
drivers/hwtracing/stm/policy.c
30
struct config_group group;
drivers/hwtracing/stm/policy.c
383
stp_policy_make(struct config_group *group, const char *name)
drivers/hwtracing/stm/policy.c
446
config_group_init_type_name(&stm->policy->group, name,
drivers/hwtracing/stm/policy.c
452
ret = &stm->policy->group;
drivers/hwtracing/stm/policy.c
494
struct list_head *head = &policy->group.cg_children;
drivers/hwtracing/stm/policy.c
514
policy_node->group.cg_item.ci_name)) {
drivers/hwtracing/stm/policy.c
520
head = &policy_node->group.cg_children;
drivers/hwtracing/stm/policy.c
545
config_item_get(&policy_node->group.cg_item);
drivers/hwtracing/stm/policy.c
557
config_item_put(&policy_node->group.cg_item);
drivers/hwtracing/stm/policy.c
63
container_of(to_config_group(item), struct stp_policy, group) :
drivers/hwtracing/stm/policy.c
72
group) :
drivers/iio/adc/ti-tsc2046.c
343
unsigned int group,
drivers/iio/adc/ti-tsc2046.c
351
if (group)
drivers/iio/adc/ti-tsc2046.c
352
offset = priv->l[group - 1].offset + priv->l[group - 1].count;
drivers/iio/adc/ti-tsc2046.c
357
cur = &priv->l[group];
drivers/iio/adc/ti-tsc2046.c
366
unsigned int group, int ch_idx)
drivers/iio/adc/ti-tsc2046.c
368
struct tsc2046_adc_group_layout *l = &priv->l[group];
drivers/iio/adc/ti-tsc2046.c
385
static u16 tsc2046_adc_get_val(struct tsc2046_adc_priv *priv, int group)
drivers/iio/adc/ti-tsc2046.c
391
l = &priv->l[group];
drivers/iio/adc/ti-tsc2046.c
406
int group;
drivers/iio/adc/ti-tsc2046.c
415
for (group = 0; group < priv->groups; group++)
drivers/iio/adc/ti-tsc2046.c
416
priv->scan_buf.data[group] = tsc2046_adc_get_val(priv, group);
drivers/iio/adc/ti-tsc2046.c
482
unsigned int ch_idx, group = 0;
drivers/iio/adc/ti-tsc2046.c
489
size += tsc2046_adc_group_set_layout(priv, group, ch_idx);
drivers/iio/adc/ti-tsc2046.c
490
tsc2046_adc_group_set_cmd(priv, group, ch_idx);
drivers/iio/adc/ti-tsc2046.c
491
group++;
drivers/iio/adc/ti-tsc2046.c
494
priv->groups = group;
drivers/iio/iio_core.h
58
const struct attribute_group *group);
drivers/iio/industrialio-buffer.c
1508
struct attribute_group *group;
drivers/iio/industrialio-buffer.c
1518
group = &iio_dev_opaque->legacy_buffer_group;
drivers/iio/industrialio-buffer.c
1519
group->attrs = attrs;
drivers/iio/industrialio-buffer.c
1520
group->name = "buffer";
drivers/iio/industrialio-buffer.c
1522
ret = iio_device_register_sysfs_group(indio_dev, group);
drivers/iio/industrialio-buffer.c
1535
group = &iio_dev_opaque->legacy_scan_el_group;
drivers/iio/industrialio-buffer.c
1536
group->attrs = attrs;
drivers/iio/industrialio-buffer.c
1537
group->name = "scan_elements";
drivers/iio/industrialio-buffer.c
1539
ret = iio_device_register_sysfs_group(indio_dev, group);
drivers/iio/industrialio-core.c
1526
const struct attribute_group *group)
drivers/iio/industrialio-core.c
1536
new[iio_dev_opaque->groupcounter++] = group;
drivers/iio/industrialio-event.c
42
struct attribute_group group;
drivers/iio/industrialio-event.c
608
ev_int->group.name = iio_event_group_name;
drivers/iio/industrialio-event.c
609
ev_int->group.attrs = kzalloc_objs(ev_int->group.attrs[0],
drivers/iio/industrialio-event.c
611
if (ev_int->group.attrs == NULL) {
drivers/iio/industrialio-event.c
616
memcpy(ev_int->group.attrs,
drivers/iio/industrialio-event.c
618
sizeof(ev_int->group.attrs[0]) * attrcount_orig);
drivers/iio/industrialio-event.c
622
ev_int->group.attrs[attrn++] = &p->dev_attr.attr;
drivers/iio/industrialio-event.c
624
ret = iio_device_register_sysfs_group(indio_dev, &ev_int->group);
drivers/iio/industrialio-event.c
635
kfree(ev_int->group.attrs);
drivers/iio/industrialio-event.c
669
kfree(ev_int->group.attrs);
drivers/iio/industrialio-sw-device.c
128
static struct config_group *device_make_group(struct config_group *group,
drivers/iio/industrialio-sw-device.c
133
d = iio_sw_device_create(group->cg_item.ci_name, name);
drivers/iio/industrialio-sw-device.c
137
config_item_set_name(&d->group.cg_item, "%s", name);
drivers/iio/industrialio-sw-device.c
139
return &d->group;
drivers/iio/industrialio-sw-device.c
142
static void device_drop_group(struct config_group *group,
drivers/iio/industrialio-sw-device.c
59
d->group = configfs_register_default_group(iio_devices_group, d->name,
drivers/iio/industrialio-sw-device.c
61
if (IS_ERR(d->group))
drivers/iio/industrialio-sw-device.c
62
ret = PTR_ERR(d->group);
drivers/iio/industrialio-sw-device.c
78
configfs_unregister_default_group(dt->group);
drivers/iio/industrialio-sw-trigger.c
132
static struct config_group *trigger_make_group(struct config_group *group,
drivers/iio/industrialio-sw-trigger.c
137
t = iio_sw_trigger_create(group->cg_item.ci_name, name);
drivers/iio/industrialio-sw-trigger.c
141
config_item_set_name(&t->group.cg_item, "%s", name);
drivers/iio/industrialio-sw-trigger.c
143
return &t->group;
drivers/iio/industrialio-sw-trigger.c
146
static void trigger_drop_group(struct config_group *group,
drivers/iio/industrialio-sw-trigger.c
59
t->group = configfs_register_default_group(iio_triggers_group, t->name,
drivers/iio/industrialio-sw-trigger.c
61
if (IS_ERR(t->group)) {
drivers/iio/industrialio-sw-trigger.c
65
ret = PTR_ERR(t->group);
drivers/iio/industrialio-sw-trigger.c
82
configfs_unregister_default_group(t->group);
drivers/infiniband/core/cm.c
156
unsigned short group;
drivers/infiniband/core/cm.c
4280
->counters[cm_attr->group][cm_attr->index]));
drivers/infiniband/core/cm.c
4286
.group = _group, .index = _index \
drivers/infiniband/core/cma_configfs.c
103
struct cma_dev_port_group *group;
drivers/infiniband/core/cma_configfs.c
107
ret = cma_configfs_params_get(item, &cma_dev, &group);
drivers/infiniband/core/cma_configfs.c
111
gid_type = cma_get_default_gid_type(cma_dev, group->port_num);
drivers/infiniband/core/cma_configfs.c
124
struct cma_dev_port_group *group;
drivers/infiniband/core/cma_configfs.c
128
ret = cma_configfs_params_get(item, &cma_dev, &group);
drivers/infiniband/core/cma_configfs.c
138
ret = cma_set_default_gid_type(cma_dev, group->port_num, gid_type);
drivers/infiniband/core/cma_configfs.c
150
struct cma_dev_port_group *group;
drivers/infiniband/core/cma_configfs.c
154
ret = cma_configfs_params_get(item, &cma_dev, &group);
drivers/infiniband/core/cma_configfs.c
158
tos = cma_get_default_roce_tos(cma_dev, group->port_num);
drivers/infiniband/core/cma_configfs.c
168
struct cma_dev_port_group *group;
drivers/infiniband/core/cma_configfs.c
176
ret = cma_configfs_params_get(item, &cma_dev, &group);
drivers/infiniband/core/cma_configfs.c
180
ret = cma_set_default_roce_tos(cma_dev, group->port_num, tos);
drivers/infiniband/core/cma_configfs.c
224
config_group_init_type_name(&ports[i].group,
drivers/infiniband/core/cma_configfs.c
227
configfs_add_default_group(&ports[i].group,
drivers/infiniband/core/cma_configfs.c
237
struct config_group *group = container_of(item, struct config_group,
drivers/infiniband/core/cma_configfs.c
239
struct cma_dev_group *cma_dev_group = container_of(group,
drivers/infiniband/core/cma_configfs.c
248
struct config_group *group = container_of(item, struct config_group,
drivers/infiniband/core/cma_configfs.c
250
struct cma_dev_group *cma_dev_group = container_of(group,
drivers/infiniband/core/cma_configfs.c
276
static struct config_group *make_cma_dev(struct config_group *group,
drivers/infiniband/core/cma_configfs.c
320
struct config_group *group =
drivers/infiniband/core/cma_configfs.c
323
container_of(group, struct cma_dev_group, device_group);
drivers/infiniband/core/cma_configfs.c
47
struct config_group group;
drivers/infiniband/core/cma_configfs.c
59
struct config_group *group;
drivers/infiniband/core/cma_configfs.c
64
group = container_of(item, struct config_group, cg_item);
drivers/infiniband/core/cma_configfs.c
65
return container_of(group, struct cma_dev_port_group, group);
drivers/infiniband/core/cma_configfs.c
77
struct cma_dev_port_group *group = to_dev_port_group(item);
drivers/infiniband/core/cma_configfs.c
80
if (!group)
drivers/infiniband/core/cma_configfs.c
84
group->cma_dev_group->name);
drivers/infiniband/core/cma_configfs.c
89
*pgroup = group;
drivers/infiniband/core/multicast.c
117
struct mcast_group *group;
drivers/infiniband/core/multicast.c
133
struct mcast_group *group;
drivers/infiniband/core/multicast.c
137
group = rb_entry(node, struct mcast_group, node);
drivers/infiniband/core/multicast.c
138
ret = memcmp(mgid->raw, group->rec.mgid.raw, sizeof *mgid);
drivers/infiniband/core/multicast.c
140
return group;
drivers/infiniband/core/multicast.c
151
struct mcast_group *group,
drivers/infiniband/core/multicast.c
163
ret = memcmp(group->rec.mgid.raw, cur_group->rec.mgid.raw,
drivers/infiniband/core/multicast.c
164
sizeof group->rec.mgid);
drivers/infiniband/core/multicast.c
174
rb_link_node(&group->node, parent, link);
drivers/infiniband/core/multicast.c
175
rb_insert_color(&group->node, &port->table);
drivers/infiniband/core/multicast.c
185
static void release_group(struct mcast_group *group)
drivers/infiniband/core/multicast.c
187
struct mcast_port *port = group->port;
drivers/infiniband/core/multicast.c
191
if (atomic_dec_and_test(&group->refcount)) {
drivers/infiniband/core/multicast.c
192
rb_erase(&group->node, &port->table);
drivers/infiniband/core/multicast.c
194
kfree(group);
drivers/infiniband/core/multicast.c
208
struct mcast_group *group = member->group;
drivers/infiniband/core/multicast.c
211
spin_lock_irqsave(&group->lock, flags);
drivers/infiniband/core/multicast.c
212
list_add_tail(&member->list, &group->pending_list);
drivers/infiniband/core/multicast.c
213
if (group->state == MCAST_IDLE) {
drivers/infiniband/core/multicast.c
214
group->state = MCAST_BUSY;
drivers/infiniband/core/multicast.c
215
atomic_inc(&group->refcount);
drivers/infiniband/core/multicast.c
216
queue_work(mcast_wq, &group->work);
drivers/infiniband/core/multicast.c
218
spin_unlock_irqrestore(&group->lock, flags);
drivers/infiniband/core/multicast.c
228
static void adjust_membership(struct mcast_group *group, u8 join_state, int inc)
drivers/infiniband/core/multicast.c
234
group->members[i] += inc;
drivers/infiniband/core/multicast.c
243
static u8 get_leave_state(struct mcast_group *group)
drivers/infiniband/core/multicast.c
249
if (!group->members[i])
drivers/infiniband/core/multicast.c
252
return leave_state & group->rec.join_state;
drivers/infiniband/core/multicast.c
330
static int send_join(struct mcast_group *group, struct mcast_member *member)
drivers/infiniband/core/multicast.c
332
struct mcast_port *port = group->port;
drivers/infiniband/core/multicast.c
335
group->last_join = member;
drivers/infiniband/core/multicast.c
340
3000, GFP_KERNEL, join_handler, group,
drivers/infiniband/core/multicast.c
341
&group->query);
drivers/infiniband/core/multicast.c
345
static int send_leave(struct mcast_group *group, u8 leave_state)
drivers/infiniband/core/multicast.c
347
struct mcast_port *port = group->port;
drivers/infiniband/core/multicast.c
351
rec = group->rec;
drivers/infiniband/core/multicast.c
353
group->leave_state = leave_state;
drivers/infiniband/core/multicast.c
361
group, &group->query);
drivers/infiniband/core/multicast.c
365
static void join_group(struct mcast_group *group, struct mcast_member *member,
drivers/infiniband/core/multicast.c
369
adjust_membership(group, join_state, 1);
drivers/infiniband/core/multicast.c
370
group->rec.join_state |= join_state;
drivers/infiniband/core/multicast.c
371
member->multicast.rec = group->rec;
drivers/infiniband/core/multicast.c
373
list_move(&member->list, &group->active_list);
drivers/infiniband/core/multicast.c
376
static int fail_join(struct mcast_group *group, struct mcast_member *member,
drivers/infiniband/core/multicast.c
379
spin_lock_irq(&group->lock);
drivers/infiniband/core/multicast.c
381
spin_unlock_irq(&group->lock);
drivers/infiniband/core/multicast.c
385
static void process_group_error(struct mcast_group *group)
drivers/infiniband/core/multicast.c
391
if (group->state == MCAST_PKEY_EVENT)
drivers/infiniband/core/multicast.c
392
ret = ib_find_pkey(group->port->dev->device,
drivers/infiniband/core/multicast.c
393
group->port->port_num,
drivers/infiniband/core/multicast.c
394
be16_to_cpu(group->rec.pkey), &pkey_index);
drivers/infiniband/core/multicast.c
396
spin_lock_irq(&group->lock);
drivers/infiniband/core/multicast.c
397
if (group->state == MCAST_PKEY_EVENT && !ret &&
drivers/infiniband/core/multicast.c
398
group->pkey_index == pkey_index)
drivers/infiniband/core/multicast.c
401
while (!list_empty(&group->active_list)) {
drivers/infiniband/core/multicast.c
402
member = list_entry(group->active_list.next,
drivers/infiniband/core/multicast.c
406
adjust_membership(group, member->multicast.rec.join_state, -1);
drivers/infiniband/core/multicast.c
408
spin_unlock_irq(&group->lock);
drivers/infiniband/core/multicast.c
415
spin_lock_irq(&group->lock);
drivers/infiniband/core/multicast.c
418
group->rec.join_state = 0;
drivers/infiniband/core/multicast.c
420
group->state = MCAST_BUSY;
drivers/infiniband/core/multicast.c
421
spin_unlock_irq(&group->lock);
drivers/infiniband/core/multicast.c
426
struct mcast_group *group;
drivers/infiniband/core/multicast.c
432
group = container_of(work, typeof(*group), work);
drivers/infiniband/core/multicast.c
434
spin_lock_irq(&group->lock);
drivers/infiniband/core/multicast.c
435
while (!list_empty(&group->pending_list) ||
drivers/infiniband/core/multicast.c
436
(group->state != MCAST_BUSY)) {
drivers/infiniband/core/multicast.c
438
if (group->state != MCAST_BUSY) {
drivers/infiniband/core/multicast.c
439
spin_unlock_irq(&group->lock);
drivers/infiniband/core/multicast.c
440
process_group_error(group);
drivers/infiniband/core/multicast.c
444
member = list_entry(group->pending_list.next,
drivers/infiniband/core/multicast.c
450
if (join_state == (group->rec.join_state & join_state)) {
drivers/infiniband/core/multicast.c
451
status = cmp_rec(&group->rec, &multicast->rec,
drivers/infiniband/core/multicast.c
454
join_group(group, member, join_state);
drivers/infiniband/core/multicast.c
457
spin_unlock_irq(&group->lock);
drivers/infiniband/core/multicast.c
460
spin_unlock_irq(&group->lock);
drivers/infiniband/core/multicast.c
461
status = send_join(group, member);
drivers/infiniband/core/multicast.c
466
ret = fail_join(group, member, status);
drivers/infiniband/core/multicast.c
472
spin_lock_irq(&group->lock);
drivers/infiniband/core/multicast.c
475
join_state = get_leave_state(group);
drivers/infiniband/core/multicast.c
477
group->rec.join_state &= ~join_state;
drivers/infiniband/core/multicast.c
478
spin_unlock_irq(&group->lock);
drivers/infiniband/core/multicast.c
479
if (send_leave(group, join_state))
drivers/infiniband/core/multicast.c
482
group->state = MCAST_IDLE;
drivers/infiniband/core/multicast.c
483
spin_unlock_irq(&group->lock);
drivers/infiniband/core/multicast.c
484
release_group(group);
drivers/infiniband/core/multicast.c
491
static void process_join_error(struct mcast_group *group, int status)
drivers/infiniband/core/multicast.c
496
spin_lock_irq(&group->lock);
drivers/infiniband/core/multicast.c
497
member = list_entry(group->pending_list.next,
drivers/infiniband/core/multicast.c
499
if (group->last_join == member) {
drivers/infiniband/core/multicast.c
502
spin_unlock_irq(&group->lock);
drivers/infiniband/core/multicast.c
508
spin_unlock_irq(&group->lock);
drivers/infiniband/core/multicast.c
514
struct mcast_group *group = context;
drivers/infiniband/core/multicast.c
518
process_join_error(group, status);
drivers/infiniband/core/multicast.c
522
if (ib_find_pkey(group->port->dev->device,
drivers/infiniband/core/multicast.c
523
group->port->port_num, be16_to_cpu(rec->pkey),
drivers/infiniband/core/multicast.c
527
spin_lock_irq(&group->port->lock);
drivers/infiniband/core/multicast.c
528
if (group->state == MCAST_BUSY &&
drivers/infiniband/core/multicast.c
529
group->pkey_index == MCAST_INVALID_PKEY_INDEX)
drivers/infiniband/core/multicast.c
530
group->pkey_index = pkey_index;
drivers/infiniband/core/multicast.c
531
mgids_changed = memcmp(&rec->mgid, &group->rec.mgid,
drivers/infiniband/core/multicast.c
532
sizeof(group->rec.mgid));
drivers/infiniband/core/multicast.c
533
group->rec = *rec;
drivers/infiniband/core/multicast.c
535
rb_erase(&group->node, &group->port->table);
drivers/infiniband/core/multicast.c
536
is_mgid0 = !memcmp(&mgid0, &group->rec.mgid,
drivers/infiniband/core/multicast.c
538
mcast_insert(group->port, group, is_mgid0);
drivers/infiniband/core/multicast.c
540
spin_unlock_irq(&group->port->lock);
drivers/infiniband/core/multicast.c
542
mcast_work_handler(&group->work);
drivers/infiniband/core/multicast.c
548
struct mcast_group *group = context;
drivers/infiniband/core/multicast.c
550
if (status && group->retries > 0 &&
drivers/infiniband/core/multicast.c
551
!send_leave(group, group->leave_state))
drivers/infiniband/core/multicast.c
552
group->retries--;
drivers/infiniband/core/multicast.c
554
mcast_work_handler(&group->work);
drivers/infiniband/core/multicast.c
560
struct mcast_group *group, *cur_group;
drivers/infiniband/core/multicast.c
567
group = mcast_find(port, mgid);
drivers/infiniband/core/multicast.c
568
if (group)
drivers/infiniband/core/multicast.c
573
group = kzalloc_obj(*group, gfp_mask);
drivers/infiniband/core/multicast.c
574
if (!group)
drivers/infiniband/core/multicast.c
577
group->retries = 3;
drivers/infiniband/core/multicast.c
578
group->port = port;
drivers/infiniband/core/multicast.c
579
group->rec.mgid = *mgid;
drivers/infiniband/core/multicast.c
580
group->pkey_index = MCAST_INVALID_PKEY_INDEX;
drivers/infiniband/core/multicast.c
581
INIT_LIST_HEAD(&group->pending_list);
drivers/infiniband/core/multicast.c
582
INIT_LIST_HEAD(&group->active_list);
drivers/infiniband/core/multicast.c
583
INIT_WORK(&group->work, mcast_work_handler);
drivers/infiniband/core/multicast.c
584
spin_lock_init(&group->lock);
drivers/infiniband/core/multicast.c
587
cur_group = mcast_insert(port, group, is_mgid0);
drivers/infiniband/core/multicast.c
589
kfree(group);
drivers/infiniband/core/multicast.c
590
group = cur_group;
drivers/infiniband/core/multicast.c
594
atomic_inc(&group->refcount);
drivers/infiniband/core/multicast.c
596
return group;
drivers/infiniband/core/multicast.c
638
member->group = acquire_group(&dev->port[port_num - dev->start_port],
drivers/infiniband/core/multicast.c
640
if (!member->group) {
drivers/infiniband/core/multicast.c
665
struct mcast_group *group;
drivers/infiniband/core/multicast.c
668
group = member->group;
drivers/infiniband/core/multicast.c
670
spin_lock_irq(&group->lock);
drivers/infiniband/core/multicast.c
672
adjust_membership(group, multicast->rec.join_state, -1);
drivers/infiniband/core/multicast.c
676
if (group->state == MCAST_IDLE) {
drivers/infiniband/core/multicast.c
677
group->state = MCAST_BUSY;
drivers/infiniband/core/multicast.c
678
spin_unlock_irq(&group->lock);
drivers/infiniband/core/multicast.c
680
queue_work(mcast_wq, &group->work);
drivers/infiniband/core/multicast.c
682
spin_unlock_irq(&group->lock);
drivers/infiniband/core/multicast.c
683
release_group(group);
drivers/infiniband/core/multicast.c
698
struct mcast_group *group;
drivers/infiniband/core/multicast.c
708
group = mcast_find(port, mgid);
drivers/infiniband/core/multicast.c
709
if (group)
drivers/infiniband/core/multicast.c
710
*rec = group->rec;
drivers/infiniband/core/multicast.c
774
struct mcast_group *group;
drivers/infiniband/core/multicast.c
780
group = rb_entry(node, struct mcast_group, node);
drivers/infiniband/core/multicast.c
781
spin_lock(&group->lock);
drivers/infiniband/core/multicast.c
782
if (group->state == MCAST_IDLE) {
drivers/infiniband/core/multicast.c
783
atomic_inc(&group->refcount);
drivers/infiniband/core/multicast.c
784
queue_work(mcast_wq, &group->work);
drivers/infiniband/core/multicast.c
786
if (group->state != MCAST_GROUP_ERROR)
drivers/infiniband/core/multicast.c
787
group->state = state;
drivers/infiniband/core/multicast.c
788
spin_unlock(&group->lock);
drivers/infiniband/core/netlink.c
283
unsigned int group, gfp_t flags)
drivers/infiniband/core/netlink.c
287
return nlmsg_multicast(rnet->nl_sock, skb, 0, group, flags);
drivers/infiniband/core/netlink.c
53
bool rdma_nl_chk_listeners(unsigned int group)
drivers/infiniband/core/netlink.c
57
return netlink_has_listeners(rnet->nl_sock, group);
drivers/infiniband/core/sysfs.c
1018
group->attrs[pos] = &attr->attr.attr;
drivers/infiniband/core/sysfs.c
1030
group->attrs[pos] = &attr->attr.attr;
drivers/infiniband/core/sysfs.c
1046
alloc_port_table_group(const char *name, struct attribute_group *group,
drivers/infiniband/core/sysfs.c
1073
group->name = name;
drivers/infiniband/core/sysfs.c
1074
group->attrs = attr_list;
drivers/infiniband/core/sysfs.c
861
data->group.attrs = kzalloc_objs(*data->group.attrs,
drivers/infiniband/core/sysfs.c
863
if (!data->group.attrs)
drivers/infiniband/core/sysfs.c
866
data->group.name = "hw_counters";
drivers/infiniband/core/sysfs.c
879
kfree(data->group.attrs);
drivers/infiniband/core/sysfs.c
922
data->group.attrs[pos] = &attr->attr.attr;
drivers/infiniband/core/sysfs.c
934
data->group.attrs[pos] = &attr->attr.attr;
drivers/infiniband/core/sysfs.c
937
ibdev->groups[i] = &data->group;
drivers/infiniband/core/sysfs.c
946
alloc_hw_stats_port(struct ib_port *port, struct attribute_group *group)
drivers/infiniband/core/sysfs.c
95
struct attribute_group group;
drivers/infiniband/core/sysfs.c
967
group->attrs = kzalloc_objs(*group->attrs, stats->num_counters + 2);
drivers/infiniband/core/sysfs.c
968
if (!group->attrs)
drivers/infiniband/core/sysfs.c
971
group->name = "hw_counters";
drivers/infiniband/core/sysfs.c
983
struct attribute_group *group)
drivers/infiniband/core/sysfs.c
990
data = alloc_hw_stats_port(port, group);
drivers/infiniband/hw/efa/efa_admin_defs.h
97
u16 group;
drivers/infiniband/hw/efa/efa_com.c
826
u16 group)
drivers/infiniband/hw/efa/efa_com.c
830
if (group < EFA_MAX_HANDLERS && aenq_handlers->handlers[group])
drivers/infiniband/hw/efa/efa_com.c
831
return aenq_handlers->handlers[group];
drivers/infiniband/hw/efa/efa_com.c
869
aenq_common->group);
drivers/infiniband/hw/hfi1/exp_rcv.h
119
static inline void tid_group_move(struct tid_group *group,
drivers/infiniband/hw/hfi1/exp_rcv.h
123
tid_group_remove(group, s1);
drivers/infiniband/hw/hfi1/exp_rcv.h
124
tid_group_add_tail(group, s2);
drivers/infiniband/hw/hfi1/pio.c
1724
gc = group_context(hw_context, sc->group);
drivers/infiniband/hw/hfi1/pio.c
1725
gc_end = gc + group_size(sc->group);
drivers/infiniband/hw/hfi1/pio.c
2115
sc->flags, sc->sw_index, sc->hw_context, sc->group);
drivers/infiniband/hw/hfi1/pio.c
494
static inline u32 group_context(u32 context, u32 group)
drivers/infiniband/hw/hfi1/pio.c
496
return (context >> group) << group;
drivers/infiniband/hw/hfi1/pio.c
500
static inline u32 group_size(u32 group)
drivers/infiniband/hw/hfi1/pio.c
502
return 1 << group;
drivers/infiniband/hw/hfi1/pio.c
520
u32 gc = group_context(sc->hw_context, sc->group);
drivers/infiniband/hw/hfi1/pio.c
708
sc->group = 0;
drivers/infiniband/hw/hfi1/pio.c
826
sc->group,
drivers/infiniband/hw/hfi1/pio.h
70
u8 group; /* credit return group */
drivers/infiniband/hw/hfi1/tid_rdma.c
1242
struct tid_group *group = NULL, *used;
drivers/infiniband/hw/hfi1/tid_rdma.c
1251
list_for_each_entry(group, &rcd->tid_group_list.list, list) {
drivers/infiniband/hw/hfi1/tid_rdma.c
1252
kern_add_tid_node(flow, rcd, "complete groups", group,
drivers/infiniband/hw/hfi1/tid_rdma.c
1253
group->size);
drivers/infiniband/hw/hfi1/tid_rdma.c
1255
pageidx += group->size;
drivers/infiniband/hw/hfi1/tid_rdma.c
1280
if (group && &group->list == &rcd->tid_group_list.list)
drivers/infiniband/hw/hfi1/tid_rdma.c
1282
group = list_prepare_entry(group, &rcd->tid_group_list.list,
drivers/infiniband/hw/hfi1/tid_rdma.c
1284
if (list_is_last(&group->list, &rcd->tid_group_list.list))
drivers/infiniband/hw/hfi1/tid_rdma.c
1286
group = list_next_entry(group, list);
drivers/infiniband/hw/hfi1/tid_rdma.c
1287
use = min_t(u32, flow->npagesets - pageidx, group->size);
drivers/infiniband/hw/hfi1/tid_rdma.c
1288
kern_add_tid_node(flow, rcd, "complete continue", group, use);
drivers/infiniband/hw/mlx4/mcg.c
1002
if (group->state == MCAST_IDLE)
drivers/infiniband/hw/mlx4/mcg.c
1004
get_state_string(group->state));
drivers/infiniband/hw/mlx4/mcg.c
1007
get_state_string(group->state),
drivers/infiniband/hw/mlx4/mcg.c
1008
be64_to_cpu(group->last_req_tid));
drivers/infiniband/hw/mlx4/mcg.c
1010
if (list_empty(&group->pending_list)) {
drivers/infiniband/hw/mlx4/mcg.c
1013
req = list_first_entry(&group->pending_list, struct mcast_req,
drivers/infiniband/hw/mlx4/mcg.c
1020
group->rec.scope_join_state & 0xf,
drivers/infiniband/hw/mlx4/mcg.c
1021
group->members[2],
drivers/infiniband/hw/mlx4/mcg.c
1022
group->members[1],
drivers/infiniband/hw/mlx4/mcg.c
1023
group->members[0],
drivers/infiniband/hw/mlx4/mcg.c
1024
atomic_read(&group->refcount),
drivers/infiniband/hw/mlx4/mcg.c
1029
if (group->func[i].state == MCAST_MEMBER)
drivers/infiniband/hw/mlx4/mcg.c
1031
group->func[i].join_state);
drivers/infiniband/hw/mlx4/mcg.c
1034
hoplimit = be32_to_cpu(group->rec.sl_flowlabel_hoplimit);
drivers/infiniband/hw/mlx4/mcg.c
1037
be16_to_cpu(group->rec.pkey),
drivers/infiniband/hw/mlx4/mcg.c
1038
be32_to_cpu(group->rec.qkey),
drivers/infiniband/hw/mlx4/mcg.c
1039
(group->rec.mtusel_mtu & 0xc0) >> 6,
drivers/infiniband/hw/mlx4/mcg.c
1040
(group->rec.mtusel_mtu & 0x3f),
drivers/infiniband/hw/mlx4/mcg.c
1041
group->rec.tclass,
drivers/infiniband/hw/mlx4/mcg.c
1042
(group->rec.ratesel_rate & 0xc0) >> 6,
drivers/infiniband/hw/mlx4/mcg.c
1043
(group->rec.ratesel_rate & 0x3f),
drivers/infiniband/hw/mlx4/mcg.c
1047
group->rec.proxy_join);
drivers/infiniband/hw/mlx4/mcg.c
1070
static void force_clean_group(struct mcast_group *group)
drivers/infiniband/hw/mlx4/mcg.c
1074
list_for_each_entry_safe(req, tmp, &group->pending_list, group_list) {
drivers/infiniband/hw/mlx4/mcg.c
1078
del_sysfs_port_mcg_attr(group->demux->dev, group->demux->port, &group->dentry.attr);
drivers/infiniband/hw/mlx4/mcg.c
1079
rb_erase(&group->node, &group->demux->mcg_table);
drivers/infiniband/hw/mlx4/mcg.c
1080
kfree(group);
drivers/infiniband/hw/mlx4/mcg.c
1087
struct mcast_group *group;
drivers/infiniband/hw/mlx4/mcg.c
1113
group = rb_entry(p, struct mcast_group, node);
drivers/infiniband/hw/mlx4/mcg.c
1114
if (atomic_read(&group->refcount))
drivers/infiniband/hw/mlx4/mcg.c
1115
mcg_debug_group(group, "group refcount %d!!! (pointer %p)\n",
drivers/infiniband/hw/mlx4/mcg.c
1116
atomic_read(&group->refcount), group);
drivers/infiniband/hw/mlx4/mcg.c
1118
force_clean_group(group);
drivers/infiniband/hw/mlx4/mcg.c
1173
static void clear_pending_reqs(struct mcast_group *group, int vf)
drivers/infiniband/hw/mlx4/mcg.c
1179
if (!list_empty(&group->pending_list))
drivers/infiniband/hw/mlx4/mcg.c
1180
group_first = list_first_entry(&group->pending_list, struct mcast_req, group_list);
drivers/infiniband/hw/mlx4/mcg.c
1182
list_for_each_entry_safe(req, tmp, &group->func[vf].pending, func_list) {
drivers/infiniband/hw/mlx4/mcg.c
1185
(group->state == MCAST_JOIN_SENT ||
drivers/infiniband/hw/mlx4/mcg.c
1186
group->state == MCAST_LEAVE_SENT)) {
drivers/infiniband/hw/mlx4/mcg.c
1187
clear = cancel_delayed_work(&group->timeout_work);
drivers/infiniband/hw/mlx4/mcg.c
1189
group->state = MCAST_IDLE;
drivers/infiniband/hw/mlx4/mcg.c
1192
--group->func[vf].num_pend_reqs;
drivers/infiniband/hw/mlx4/mcg.c
1196
atomic_dec(&group->refcount);
drivers/infiniband/hw/mlx4/mcg.c
1200
if (!pend && (!list_empty(&group->func[vf].pending) || group->func[vf].num_pend_reqs)) {
drivers/infiniband/hw/mlx4/mcg.c
1201
mcg_warn_group(group, "DRIVER BUG: list_empty %d, num_pend_reqs %d\n",
drivers/infiniband/hw/mlx4/mcg.c
1202
list_empty(&group->func[vf].pending), group->func[vf].num_pend_reqs);
drivers/infiniband/hw/mlx4/mcg.c
1206
static int push_deleteing_req(struct mcast_group *group, int slave)
drivers/infiniband/hw/mlx4/mcg.c
1211
if (!group->func[slave].join_state)
drivers/infiniband/hw/mlx4/mcg.c
1218
if (!list_empty(&group->func[slave].pending)) {
drivers/infiniband/hw/mlx4/mcg.c
1219
pend_req = list_entry(group->func[slave].pending.prev, struct mcast_req, group_list);
drivers/infiniband/hw/mlx4/mcg.c
1228
req->group = group;
drivers/infiniband/hw/mlx4/mcg.c
1229
++group->func[slave].num_pend_reqs;
drivers/infiniband/hw/mlx4/mcg.c
1237
struct mcast_group *group;
drivers/infiniband/hw/mlx4/mcg.c
1242
group = rb_entry(p, struct mcast_group, node);
drivers/infiniband/hw/mlx4/mcg.c
1243
mutex_lock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
1244
if (atomic_read(&group->refcount)) {
drivers/infiniband/hw/mlx4/mcg.c
1246
clear_pending_reqs(group, slave);
drivers/infiniband/hw/mlx4/mcg.c
1247
push_deleteing_req(group, slave);
drivers/infiniband/hw/mlx4/mcg.c
1249
mutex_unlock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
136
struct mcast_group *group;
drivers/infiniband/hw/mlx4/mcg.c
144
mcg_warn_group(group, "did not expect to reach zero\n"); \
drivers/infiniband/hw/mlx4/mcg.c
166
struct mcast_group *group;
drivers/infiniband/hw/mlx4/mcg.c
170
group = rb_entry(node, struct mcast_group, node);
drivers/infiniband/hw/mlx4/mcg.c
171
ret = memcmp(mgid->raw, group->rec.mgid.raw, sizeof *mgid);
drivers/infiniband/hw/mlx4/mcg.c
173
return group;
drivers/infiniband/hw/mlx4/mcg.c
184
struct mcast_group *group)
drivers/infiniband/hw/mlx4/mcg.c
195
ret = memcmp(group->rec.mgid.raw, cur_group->rec.mgid.raw,
drivers/infiniband/hw/mlx4/mcg.c
196
sizeof group->rec.mgid);
drivers/infiniband/hw/mlx4/mcg.c
204
rb_link_node(&group->node, parent, link);
drivers/infiniband/hw/mlx4/mcg.c
205
rb_insert_color(&group->node, &ctx->mcg_table);
drivers/infiniband/hw/mlx4/mcg.c
252
static int send_join_to_wire(struct mcast_group *group, struct ib_sa_mad *sa_mad)
drivers/infiniband/hw/mlx4/mcg.c
262
sa_mad_data->port_gid.global.interface_id = group->demux->guid_cache[0];
drivers/infiniband/hw/mlx4/mcg.c
265
mad.mad_hdr.tid = mlx4_ib_get_new_demux_tid(group->demux);
drivers/infiniband/hw/mlx4/mcg.c
266
group->last_req_tid = mad.mad_hdr.tid; /* keep it for later validation */
drivers/infiniband/hw/mlx4/mcg.c
268
ret = send_mad_to_wire(group->demux, (struct ib_mad *)&mad);
drivers/infiniband/hw/mlx4/mcg.c
272
queue_delayed_work(group->demux->mcg_wq, &group->timeout_work,
drivers/infiniband/hw/mlx4/mcg.c
279
static int send_leave_to_wire(struct mcast_group *group, u8 join_state)
drivers/infiniband/hw/mlx4/mcg.c
292
mad.mad_hdr.tid = mlx4_ib_get_new_demux_tid(group->demux);
drivers/infiniband/hw/mlx4/mcg.c
293
group->last_req_tid = mad.mad_hdr.tid; /* keep it for later validation */
drivers/infiniband/hw/mlx4/mcg.c
301
*sa_data = group->rec;
drivers/infiniband/hw/mlx4/mcg.c
304
ret = send_mad_to_wire(group->demux, (struct ib_mad *)&mad);
drivers/infiniband/hw/mlx4/mcg.c
306
group->state = MCAST_IDLE;
drivers/infiniband/hw/mlx4/mcg.c
311
queue_delayed_work(group->demux->mcg_wq, &group->timeout_work,
drivers/infiniband/hw/mlx4/mcg.c
318
static int send_reply_to_slave(int slave, struct mcast_group *group,
drivers/infiniband/hw/mlx4/mcg.c
341
*sa_data = group->rec;
drivers/infiniband/hw/mlx4/mcg.c
345
sa_data->scope_join_state |= (group->func[slave].join_state & 0x0f);
drivers/infiniband/hw/mlx4/mcg.c
348
ret = send_mad_to_slave(slave, group->demux, (struct ib_mad *)&mad);
drivers/infiniband/hw/mlx4/mcg.c
437
static int release_group(struct mcast_group *group, int from_timeout_handler)
drivers/infiniband/hw/mlx4/mcg.c
439
struct mlx4_ib_demux_ctx *ctx = group->demux;
drivers/infiniband/hw/mlx4/mcg.c
443
mutex_lock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
444
if (atomic_dec_and_test(&group->refcount)) {
drivers/infiniband/hw/mlx4/mcg.c
446
if (group->state != MCAST_IDLE &&
drivers/infiniband/hw/mlx4/mcg.c
447
!cancel_delayed_work(&group->timeout_work)) {
drivers/infiniband/hw/mlx4/mcg.c
448
atomic_inc(&group->refcount);
drivers/infiniband/hw/mlx4/mcg.c
449
mutex_unlock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
455
nzgroup = memcmp(&group->rec.mgid, &mgid0, sizeof mgid0);
drivers/infiniband/hw/mlx4/mcg.c
457
del_sysfs_port_mcg_attr(ctx->dev, ctx->port, &group->dentry.attr);
drivers/infiniband/hw/mlx4/mcg.c
458
if (!list_empty(&group->pending_list))
drivers/infiniband/hw/mlx4/mcg.c
459
mcg_warn_group(group, "releasing a group with non empty pending list\n");
drivers/infiniband/hw/mlx4/mcg.c
461
rb_erase(&group->node, &ctx->mcg_table);
drivers/infiniband/hw/mlx4/mcg.c
462
list_del_init(&group->mgid0_list);
drivers/infiniband/hw/mlx4/mcg.c
463
mutex_unlock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
465
kfree(group);
drivers/infiniband/hw/mlx4/mcg.c
468
mutex_unlock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
474
static void adjust_membership(struct mcast_group *group, u8 join_state, int inc)
drivers/infiniband/hw/mlx4/mcg.c
480
group->members[i] += inc;
drivers/infiniband/hw/mlx4/mcg.c
483
static u8 get_leave_state(struct mcast_group *group)
drivers/infiniband/hw/mlx4/mcg.c
489
if (!group->members[i])
drivers/infiniband/hw/mlx4/mcg.c
492
return leave_state & (group->rec.scope_join_state & 0xf);
drivers/infiniband/hw/mlx4/mcg.c
495
static int join_group(struct mcast_group *group, int slave, u8 join_mask)
drivers/infiniband/hw/mlx4/mcg.c
50
#define mcg_warn_group(group, format, arg...) \
drivers/infiniband/hw/mlx4/mcg.c
501
join_state = join_mask & (~group->func[slave].join_state);
drivers/infiniband/hw/mlx4/mcg.c
502
adjust_membership(group, join_state, 1);
drivers/infiniband/hw/mlx4/mcg.c
503
group->func[slave].join_state |= join_state;
drivers/infiniband/hw/mlx4/mcg.c
504
if (group->func[slave].state != MCAST_MEMBER && join_state) {
drivers/infiniband/hw/mlx4/mcg.c
505
group->func[slave].state = MCAST_MEMBER;
drivers/infiniband/hw/mlx4/mcg.c
511
static int leave_group(struct mcast_group *group, int slave, u8 leave_state)
drivers/infiniband/hw/mlx4/mcg.c
515
adjust_membership(group, leave_state, -1);
drivers/infiniband/hw/mlx4/mcg.c
516
group->func[slave].join_state &= ~leave_state;
drivers/infiniband/hw/mlx4/mcg.c
517
if (!group->func[slave].join_state) {
drivers/infiniband/hw/mlx4/mcg.c
518
group->func[slave].state = MCAST_NOT_MEMBER;
drivers/infiniband/hw/mlx4/mcg.c
52
(group)->name, group->demux->port, ## arg)
drivers/infiniband/hw/mlx4/mcg.c
524
static int check_leave(struct mcast_group *group, int slave, u8 leave_mask)
drivers/infiniband/hw/mlx4/mcg.c
526
if (group->func[slave].state != MCAST_MEMBER)
drivers/infiniband/hw/mlx4/mcg.c
530
if (~group->func[slave].join_state & leave_mask)
drivers/infiniband/hw/mlx4/mcg.c
54
#define mcg_debug_group(group, format, arg...) \
drivers/infiniband/hw/mlx4/mcg.c
542
struct mcast_group *group;
drivers/infiniband/hw/mlx4/mcg.c
545
group = container_of(delay, typeof(*group), timeout_work);
drivers/infiniband/hw/mlx4/mcg.c
547
mutex_lock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
548
if (group->state == MCAST_JOIN_SENT) {
drivers/infiniband/hw/mlx4/mcg.c
549
if (!list_empty(&group->pending_list)) {
drivers/infiniband/hw/mlx4/mcg.c
550
req = list_first_entry(&group->pending_list, struct mcast_req, group_list);
drivers/infiniband/hw/mlx4/mcg.c
553
--group->func[req->func].num_pend_reqs;
drivers/infiniband/hw/mlx4/mcg.c
554
mutex_unlock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
556
if (memcmp(&group->rec.mgid, &mgid0, sizeof mgid0)) {
drivers/infiniband/hw/mlx4/mcg.c
557
if (release_group(group, 1))
drivers/infiniband/hw/mlx4/mcg.c
56
(group)->name, (group)->demux->port, ## arg)
drivers/infiniband/hw/mlx4/mcg.c
560
kfree(group);
drivers/infiniband/hw/mlx4/mcg.c
563
mutex_lock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
565
mcg_warn_group(group, "DRIVER BUG\n");
drivers/infiniband/hw/mlx4/mcg.c
566
} else if (group->state == MCAST_LEAVE_SENT) {
drivers/infiniband/hw/mlx4/mcg.c
567
if (group->rec.scope_join_state & 0xf)
drivers/infiniband/hw/mlx4/mcg.c
568
group->rec.scope_join_state &= 0xf0;
drivers/infiniband/hw/mlx4/mcg.c
569
group->state = MCAST_IDLE;
drivers/infiniband/hw/mlx4/mcg.c
570
mutex_unlock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
571
if (release_group(group, 1))
drivers/infiniband/hw/mlx4/mcg.c
573
mutex_lock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
575
mcg_warn_group(group, "invalid state %s\n", get_state_string(group->state));
drivers/infiniband/hw/mlx4/mcg.c
576
group->state = MCAST_IDLE;
drivers/infiniband/hw/mlx4/mcg.c
577
atomic_inc(&group->refcount);
drivers/infiniband/hw/mlx4/mcg.c
578
if (!queue_work(group->demux->mcg_wq, &group->work))
drivers/infiniband/hw/mlx4/mcg.c
579
safe_atomic_dec(&group->refcount);
drivers/infiniband/hw/mlx4/mcg.c
58
#define mcg_error_group(group, format, arg...) \
drivers/infiniband/hw/mlx4/mcg.c
581
mutex_unlock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
584
static int handle_leave_req(struct mcast_group *group, u8 leave_mask,
drivers/infiniband/hw/mlx4/mcg.c
59
pr_err(" %16s: " format, (group)->name, ## arg)
drivers/infiniband/hw/mlx4/mcg.c
590
leave_mask = group->func[req->func].join_state;
drivers/infiniband/hw/mlx4/mcg.c
592
status = check_leave(group, req->func, leave_mask);
drivers/infiniband/hw/mlx4/mcg.c
594
leave_group(group, req->func, leave_mask);
drivers/infiniband/hw/mlx4/mcg.c
597
send_reply_to_slave(req->func, group, &req->sa_mad, status);
drivers/infiniband/hw/mlx4/mcg.c
598
--group->func[req->func].num_pend_reqs;
drivers/infiniband/hw/mlx4/mcg.c
605
static int handle_join_req(struct mcast_group *group, u8 join_mask,
drivers/infiniband/hw/mlx4/mcg.c
608
u8 group_join_state = group->rec.scope_join_state & 0xf;
drivers/infiniband/hw/mlx4/mcg.c
615
status = cmp_rec(&group->rec, sa_data, req->sa_mad.sa_hdr.comp_mask);
drivers/infiniband/hw/mlx4/mcg.c
617
join_group(group, req->func, join_mask);
drivers/infiniband/hw/mlx4/mcg.c
619
--group->func[req->func].num_pend_reqs;
drivers/infiniband/hw/mlx4/mcg.c
620
send_reply_to_slave(req->func, group, &req->sa_mad, status);
drivers/infiniband/hw/mlx4/mcg.c
627
group->prev_state = group->state;
drivers/infiniband/hw/mlx4/mcg.c
628
if (send_join_to_wire(group, &req->sa_mad)) {
drivers/infiniband/hw/mlx4/mcg.c
629
--group->func[req->func].num_pend_reqs;
drivers/infiniband/hw/mlx4/mcg.c
634
group->state = group->prev_state;
drivers/infiniband/hw/mlx4/mcg.c
636
group->state = MCAST_JOIN_SENT;
drivers/infiniband/hw/mlx4/mcg.c
644
struct mcast_group *group;
drivers/infiniband/hw/mlx4/mcg.c
652
group = container_of(work, typeof(*group), work);
drivers/infiniband/hw/mlx4/mcg.c
654
mutex_lock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
660
if (group->state == MCAST_RESP_READY) {
drivers/infiniband/hw/mlx4/mcg.c
662
cancel_delayed_work(&group->timeout_work);
drivers/infiniband/hw/mlx4/mcg.c
663
status = be16_to_cpu(group->response_sa_mad.mad_hdr.status);
drivers/infiniband/hw/mlx4/mcg.c
664
method = group->response_sa_mad.mad_hdr.method;
drivers/infiniband/hw/mlx4/mcg.c
665
if (group->last_req_tid != group->response_sa_mad.mad_hdr.tid) {
drivers/infiniband/hw/mlx4/mcg.c
666
mcg_warn_group(group, "Got MAD response to existing MGID but wrong TID, dropping. Resp TID=%llx, group TID=%llx\n",
drivers/infiniband/hw/mlx4/mcg.c
667
be64_to_cpu(group->response_sa_mad.mad_hdr.tid),
drivers/infiniband/hw/mlx4/mcg.c
668
be64_to_cpu(group->last_req_tid));
drivers/infiniband/hw/mlx4/mcg.c
669
group->state = group->prev_state;
drivers/infiniband/hw/mlx4/mcg.c
673
if (!list_empty(&group->pending_list))
drivers/infiniband/hw/mlx4/mcg.c
674
req = list_first_entry(&group->pending_list,
drivers/infiniband/hw/mlx4/mcg.c
678
send_reply_to_slave(req->func, group, &req->sa_mad, status);
drivers/infiniband/hw/mlx4/mcg.c
679
--group->func[req->func].num_pend_reqs;
drivers/infiniband/hw/mlx4/mcg.c
685
mcg_warn_group(group, "no request for failed join\n");
drivers/infiniband/hw/mlx4/mcg.c
686
} else if (method == IB_SA_METHOD_DELETE_RESP && group->demux->flushing)
drivers/infiniband/hw/mlx4/mcg.c
693
group->response_sa_mad.data)->scope_join_state & 0xf;
drivers/infiniband/hw/mlx4/mcg.c
694
cur_join_state = group->rec.scope_join_state & 0xf;
drivers/infiniband/hw/mlx4/mcg.c
702
memcpy(&group->rec, group->response_sa_mad.data, sizeof group->rec);
drivers/infiniband/hw/mlx4/mcg.c
704
group->state = MCAST_IDLE;
drivers/infiniband/hw/mlx4/mcg.c
709
while (!list_empty(&group->pending_list) && group->state == MCAST_IDLE) {
drivers/infiniband/hw/mlx4/mcg.c
710
req = list_first_entry(&group->pending_list, struct mcast_req,
drivers/infiniband/hw/mlx4/mcg.c
719
rc += handle_leave_req(group, req_join_state, req);
drivers/infiniband/hw/mlx4/mcg.c
721
rc += handle_join_req(group, req_join_state, req);
drivers/infiniband/hw/mlx4/mcg.c
725
if (group->state == MCAST_IDLE) {
drivers/infiniband/hw/mlx4/mcg.c
726
req_join_state = get_leave_state(group);
drivers/infiniband/hw/mlx4/mcg.c
728
group->rec.scope_join_state &= ~req_join_state;
drivers/infiniband/hw/mlx4/mcg.c
729
group->prev_state = group->state;
drivers/infiniband/hw/mlx4/mcg.c
730
if (send_leave_to_wire(group, req_join_state)) {
drivers/infiniband/hw/mlx4/mcg.c
731
group->state = group->prev_state;
drivers/infiniband/hw/mlx4/mcg.c
734
group->state = MCAST_LEAVE_SENT;
drivers/infiniband/hw/mlx4/mcg.c
738
if (!list_empty(&group->pending_list) && group->state == MCAST_IDLE)
drivers/infiniband/hw/mlx4/mcg.c
740
mutex_unlock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
743
release_group(group, 0);
drivers/infiniband/hw/mlx4/mcg.c
750
struct mcast_group *group = NULL, *cur_group, *n;
drivers/infiniband/hw/mlx4/mcg.c
754
list_for_each_entry_safe(group, n, &ctx->mcg_mgid0_list, mgid0_list) {
drivers/infiniband/hw/mlx4/mcg.c
755
mutex_lock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
756
if (group->last_req_tid == tid) {
drivers/infiniband/hw/mlx4/mcg.c
758
group->rec.mgid = *new_mgid;
drivers/infiniband/hw/mlx4/mcg.c
759
sprintf(group->name, "%016llx%016llx",
drivers/infiniband/hw/mlx4/mcg.c
760
be64_to_cpu(group->rec.mgid.global.subnet_prefix),
drivers/infiniband/hw/mlx4/mcg.c
761
be64_to_cpu(group->rec.mgid.global.interface_id));
drivers/infiniband/hw/mlx4/mcg.c
762
list_del_init(&group->mgid0_list);
drivers/infiniband/hw/mlx4/mcg.c
763
cur_group = mcast_insert(ctx, group);
drivers/infiniband/hw/mlx4/mcg.c
766
req = list_first_entry(&group->pending_list,
drivers/infiniband/hw/mlx4/mcg.c
768
--group->func[req->func].num_pend_reqs;
drivers/infiniband/hw/mlx4/mcg.c
772
mutex_unlock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
774
release_group(group, 0);
drivers/infiniband/hw/mlx4/mcg.c
778
atomic_inc(&group->refcount);
drivers/infiniband/hw/mlx4/mcg.c
779
add_sysfs_port_mcg_attr(ctx->dev, ctx->port, &group->dentry.attr);
drivers/infiniband/hw/mlx4/mcg.c
780
mutex_unlock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
782
return group;
drivers/infiniband/hw/mlx4/mcg.c
786
list_del(&group->mgid0_list);
drivers/infiniband/hw/mlx4/mcg.c
787
if (!list_empty(&group->pending_list) && group->state != MCAST_IDLE)
drivers/infiniband/hw/mlx4/mcg.c
788
cancel_delayed_work_sync(&group->timeout_work);
drivers/infiniband/hw/mlx4/mcg.c
790
list_for_each_entry_safe(tmp1, tmp2, &group->pending_list, group_list) {
drivers/infiniband/hw/mlx4/mcg.c
794
mutex_unlock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
796
kfree(group);
drivers/infiniband/hw/mlx4/mcg.c
800
mutex_unlock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
813
struct mcast_group *group, *cur_group;
drivers/infiniband/hw/mlx4/mcg.c
819
group = mcast_find(ctx, mgid);
drivers/infiniband/hw/mlx4/mcg.c
820
if (group)
drivers/infiniband/hw/mlx4/mcg.c
827
group = kzalloc_obj(*group);
drivers/infiniband/hw/mlx4/mcg.c
828
if (!group)
drivers/infiniband/hw/mlx4/mcg.c
831
group->demux = ctx;
drivers/infiniband/hw/mlx4/mcg.c
832
group->rec.mgid = *mgid;
drivers/infiniband/hw/mlx4/mcg.c
833
INIT_LIST_HEAD(&group->pending_list);
drivers/infiniband/hw/mlx4/mcg.c
834
INIT_LIST_HEAD(&group->mgid0_list);
drivers/infiniband/hw/mlx4/mcg.c
836
INIT_LIST_HEAD(&group->func[i].pending);
drivers/infiniband/hw/mlx4/mcg.c
837
INIT_WORK(&group->work, mlx4_ib_mcg_work_handler);
drivers/infiniband/hw/mlx4/mcg.c
838
INIT_DELAYED_WORK(&group->timeout_work, mlx4_ib_mcg_timeout_handler);
drivers/infiniband/hw/mlx4/mcg.c
839
mutex_init(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
840
sprintf(group->name, "%016llx%016llx",
drivers/infiniband/hw/mlx4/mcg.c
841
be64_to_cpu(group->rec.mgid.global.subnet_prefix),
drivers/infiniband/hw/mlx4/mcg.c
842
be64_to_cpu(group->rec.mgid.global.interface_id));
drivers/infiniband/hw/mlx4/mcg.c
843
sysfs_attr_init(&group->dentry.attr);
drivers/infiniband/hw/mlx4/mcg.c
844
group->dentry.show = sysfs_show_group;
drivers/infiniband/hw/mlx4/mcg.c
845
group->dentry.store = NULL;
drivers/infiniband/hw/mlx4/mcg.c
846
group->dentry.attr.name = group->name;
drivers/infiniband/hw/mlx4/mcg.c
847
group->dentry.attr.mode = 0400;
drivers/infiniband/hw/mlx4/mcg.c
848
group->state = MCAST_IDLE;
drivers/infiniband/hw/mlx4/mcg.c
851
list_add(&group->mgid0_list, &ctx->mcg_mgid0_list);
drivers/infiniband/hw/mlx4/mcg.c
855
cur_group = mcast_insert(ctx, group);
drivers/infiniband/hw/mlx4/mcg.c
858
kfree(group);
drivers/infiniband/hw/mlx4/mcg.c
862
add_sysfs_port_mcg_attr(ctx->dev, ctx->port, &group->dentry.attr);
drivers/infiniband/hw/mlx4/mcg.c
865
atomic_inc(&group->refcount);
drivers/infiniband/hw/mlx4/mcg.c
866
return group;
drivers/infiniband/hw/mlx4/mcg.c
871
struct mcast_group *group = req->group;
drivers/infiniband/hw/mlx4/mcg.c
873
atomic_inc(&group->refcount); /* for the request */
drivers/infiniband/hw/mlx4/mcg.c
874
atomic_inc(&group->refcount); /* for scheduling the work */
drivers/infiniband/hw/mlx4/mcg.c
875
list_add_tail(&req->group_list, &group->pending_list);
drivers/infiniband/hw/mlx4/mcg.c
876
list_add_tail(&req->func_list, &group->func[req->func].pending);
drivers/infiniband/hw/mlx4/mcg.c
878
if (!queue_work(group->demux->mcg_wq, &group->work))
drivers/infiniband/hw/mlx4/mcg.c
879
safe_atomic_dec(&group->refcount);
drivers/infiniband/hw/mlx4/mcg.c
888
struct mcast_group *group;
drivers/infiniband/hw/mlx4/mcg.c
894
group = acquire_group(ctx, &rec->mgid, 0);
drivers/infiniband/hw/mlx4/mcg.c
896
if (IS_ERR(group)) {
drivers/infiniband/hw/mlx4/mcg.c
900
group = search_relocate_mgid0_group(ctx, tid, &rec->mgid);
drivers/infiniband/hw/mlx4/mcg.c
902
group = NULL;
drivers/infiniband/hw/mlx4/mcg.c
905
if (!group)
drivers/infiniband/hw/mlx4/mcg.c
908
mutex_lock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
909
group->response_sa_mad = *mad;
drivers/infiniband/hw/mlx4/mcg.c
910
group->prev_state = group->state;
drivers/infiniband/hw/mlx4/mcg.c
911
group->state = MCAST_RESP_READY;
drivers/infiniband/hw/mlx4/mcg.c
913
atomic_inc(&group->refcount);
drivers/infiniband/hw/mlx4/mcg.c
914
if (!queue_work(ctx->mcg_wq, &group->work))
drivers/infiniband/hw/mlx4/mcg.c
915
safe_atomic_dec(&group->refcount);
drivers/infiniband/hw/mlx4/mcg.c
916
mutex_unlock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
917
release_group(group, 0);
drivers/infiniband/hw/mlx4/mcg.c
937
struct mcast_group *group;
drivers/infiniband/hw/mlx4/mcg.c
957
group = acquire_group(ctx, &rec->mgid, may_create);
drivers/infiniband/hw/mlx4/mcg.c
959
if (IS_ERR(group)) {
drivers/infiniband/hw/mlx4/mcg.c
961
return PTR_ERR(group);
drivers/infiniband/hw/mlx4/mcg.c
963
mutex_lock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
964
if (group->func[slave].num_pend_reqs > MAX_PEND_REQS_PER_FUNC) {
drivers/infiniband/hw/mlx4/mcg.c
965
mutex_unlock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
966
mcg_debug_group(group, "Port %d, Func %d has too many pending requests (%d), dropping\n",
drivers/infiniband/hw/mlx4/mcg.c
968
release_group(group, 0);
drivers/infiniband/hw/mlx4/mcg.c
972
++group->func[slave].num_pend_reqs;
drivers/infiniband/hw/mlx4/mcg.c
973
req->group = group;
drivers/infiniband/hw/mlx4/mcg.c
975
mutex_unlock(&group->lock);
drivers/infiniband/hw/mlx4/mcg.c
976
release_group(group, 0);
drivers/infiniband/hw/mlx4/mcg.c
993
struct mcast_group *group =
drivers/infiniband/hw/mthca/mthca_memfree.c
572
int group;
drivers/infiniband/hw/mthca/mthca_memfree.c
583
group = 0;
drivers/infiniband/hw/mthca/mthca_memfree.c
592
group = 1;
drivers/infiniband/hw/mthca/mthca_memfree.c
622
if (group == 0)
drivers/infiniband/hw/mthca/mthca_memfree.c
652
if (group == 1)
drivers/infiniband/ulp/srpt/ib_srpt.c
3834
struct config_group *group,
drivers/input/keyboard/imx_sc_key.c
42
unsigned long event, void *group)
drivers/input/keyboard/imx_sc_key.c
49
if ((event & SC_IRQ_BUTTON) && (*(u8 *)group == SC_IRQ_GROUP_WAKE)) {
drivers/input/touchscreen/mms114.c
293
int group;
drivers/input/touchscreen/mms114.c
311
group = i2c_smbus_read_byte_data(data->client,
drivers/input/touchscreen/mms114.c
313
if (group < 0)
drivers/input/touchscreen/mms114.c
314
return group;
drivers/input/touchscreen/mms114.c
317
buf[0], buf[1], buf[2], group);
drivers/iommu/amd/amd_iommu_types.h
791
struct iommu_group *group;
drivers/iommu/amd/iommu.c
515
if ((devid == p->devid) && p->group)
drivers/iommu/amd/iommu.c
516
entry->group = p->group;
drivers/iommu/amd/iommu.c
519
if (!entry->group)
drivers/iommu/amd/iommu.c
520
entry->group = generic_device_group(dev);
drivers/iommu/amd/iommu.c
522
iommu_group_ref_get(entry->group);
drivers/iommu/amd/iommu.c
524
return entry->group;
drivers/iommu/apple-dart.c
897
struct iommu_group *group = NULL;
drivers/iommu/apple-dart.c
907
if (group && group != stream_group) {
drivers/iommu/apple-dart.c
912
group = stream_group;
drivers/iommu/apple-dart.c
916
if (group) {
drivers/iommu/apple-dart.c
917
res = iommu_group_ref_get(group);
drivers/iommu/apple-dart.c
923
group = pci_device_group(dev);
drivers/iommu/apple-dart.c
926
group = generic_device_group(dev);
drivers/iommu/apple-dart.c
929
if (!group)
drivers/iommu/apple-dart.c
932
group_master_cfg = iommu_group_get_iommudata(group);
drivers/iommu/apple-dart.c
939
iommu_group_put(group);
drivers/iommu/apple-dart.c
947
iommu_group_put(group);
drivers/iommu/apple-dart.c
951
iommu_group_set_iommudata(group, group_master_cfg,
drivers/iommu/apple-dart.c
957
stream_map->dart->sid2group[sid] = group;
drivers/iommu/apple-dart.c
959
res = group;
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
3681
struct iommu_group *group;
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
3689
group = pci_device_group(dev);
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
3691
group = generic_device_group(dev);
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
3693
return group;
drivers/iommu/arm/arm-smmu/arm-smmu.c
1539
struct iommu_group *group = NULL;
drivers/iommu/arm/arm-smmu/arm-smmu.c
1544
if (group && smmu->s2crs[idx].group &&
drivers/iommu/arm/arm-smmu/arm-smmu.c
1545
group != smmu->s2crs[idx].group) {
drivers/iommu/arm/arm-smmu/arm-smmu.c
1550
group = smmu->s2crs[idx].group;
drivers/iommu/arm/arm-smmu/arm-smmu.c
1553
if (group) {
drivers/iommu/arm/arm-smmu/arm-smmu.c
1555
return iommu_group_ref_get(group);
drivers/iommu/arm/arm-smmu/arm-smmu.c
1559
group = pci_device_group(dev);
drivers/iommu/arm/arm-smmu/arm-smmu.c
1561
group = fsl_mc_device_group(dev);
drivers/iommu/arm/arm-smmu/arm-smmu.c
1563
group = generic_device_group(dev);
drivers/iommu/arm/arm-smmu/arm-smmu.c
1566
if (!IS_ERR(group))
drivers/iommu/arm/arm-smmu/arm-smmu.c
1568
smmu->s2crs[idx].group = group;
drivers/iommu/arm/arm-smmu/arm-smmu.c
1571
return group;
drivers/iommu/arm/arm-smmu/arm-smmu.h
275
struct iommu_group *group;
drivers/iommu/fsl_pamu_domain.c
385
struct iommu_group *group;
drivers/iommu/fsl_pamu_domain.c
411
group = iommu_group_get(pci_bus_to_host(pdev->bus)->parent);
drivers/iommu/fsl_pamu_domain.c
412
if (WARN_ON(!group))
drivers/iommu/fsl_pamu_domain.c
414
return group;
drivers/iommu/io-pgfault.c
100
INIT_LIST_HEAD(&group->pending_node);
drivers/iommu/io-pgfault.c
101
list_add(&group->last_fault.list, &group->faults);
drivers/iommu/io-pgfault.c
108
list_move(&iopf->list, &group->faults);
drivers/iommu/io-pgfault.c
110
list_add(&group->pending_node, &iopf_param->faults);
drivers/iommu/io-pgfault.c
113
group->fault_count = list_count_nodes(&group->faults);
drivers/iommu/io-pgfault.c
115
return group;
drivers/iommu/io-pgfault.c
220
struct iopf_group *group;
drivers/iommu/io-pgfault.c
252
group = iopf_group_alloc(iopf_param, evt, &abort_group);
drivers/iommu/io-pgfault.c
253
if (group == &abort_group)
drivers/iommu/io-pgfault.c
256
group->attach_handle = attach_handle;
drivers/iommu/io-pgfault.c
262
if (group->attach_handle->domain->iopf_handler(group))
drivers/iommu/io-pgfault.c
270
iopf_group_response(group, IOMMU_PAGE_RESP_FAILURE);
drivers/iommu/io-pgfault.c
271
if (group == &abort_group)
drivers/iommu/io-pgfault.c
272
__iopf_free_group(group);
drivers/iommu/io-pgfault.c
274
iopf_free_group(group);
drivers/iommu/io-pgfault.c
322
void iopf_group_response(struct iopf_group *group,
drivers/iommu/io-pgfault.c
325
struct iommu_fault_param *fault_param = group->fault_param;
drivers/iommu/io-pgfault.c
326
struct iopf_fault *iopf = &group->last_fault;
drivers/iommu/io-pgfault.c
327
struct device *dev = group->fault_param->dev;
drivers/iommu/io-pgfault.c
337
if (!list_empty(&group->pending_node)) {
drivers/iommu/io-pgfault.c
338
ops->page_response(dev, &group->last_fault, &resp);
drivers/iommu/io-pgfault.c
339
list_del_init(&group->pending_node);
drivers/iommu/io-pgfault.c
42
static void __iopf_free_group(struct iopf_group *group)
drivers/iommu/io-pgfault.c
454
struct iopf_group *group, *temp;
drivers/iommu/io-pgfault.c
46
list_for_each_entry_safe(iopf, next, &group->faults, list) {
drivers/iommu/io-pgfault.c
471
list_for_each_entry_safe(group, temp, &fault_param->faults, pending_node) {
drivers/iommu/io-pgfault.c
472
struct iopf_fault *iopf = &group->last_fault;
drivers/iommu/io-pgfault.c
480
list_del_init(&group->pending_node);
drivers/iommu/io-pgfault.c
481
iopf_free_group(group);
drivers/iommu/io-pgfault.c
52
iopf_put_dev_fault_param(group->fault_param);
drivers/iommu/io-pgfault.c
55
void iopf_free_group(struct iopf_group *group)
drivers/iommu/io-pgfault.c
57
__iopf_free_group(group);
drivers/iommu/io-pgfault.c
58
kfree(group);
drivers/iommu/io-pgfault.c
86
struct iopf_group *group;
drivers/iommu/io-pgfault.c
88
group = kzalloc_obj(*group);
drivers/iommu/io-pgfault.c
89
if (!group) {
drivers/iommu/io-pgfault.c
94
group = abort_group;
drivers/iommu/io-pgfault.c
97
group->fault_param = iopf_param;
drivers/iommu/io-pgfault.c
98
group->last_fault.fault = evt->fault;
drivers/iommu/io-pgfault.c
99
INIT_LIST_HEAD(&group->faults);
drivers/iommu/iommu-priv.h
43
struct iommu_attach_handle *iommu_attach_handle_get(struct iommu_group *group,
drivers/iommu/iommu-priv.h
47
struct iommu_group *group,
drivers/iommu/iommu-priv.h
50
struct iommu_group *group);
drivers/iommu/iommu-priv.h
51
int iommu_replace_group_handle(struct iommu_group *group,
drivers/iommu/iommu-sva.c
278
struct iopf_group *group;
drivers/iommu/iommu-sva.c
281
group = container_of(work, struct iopf_group, work);
drivers/iommu/iommu-sva.c
282
list_for_each_entry(iopf, &group->faults, list) {
drivers/iommu/iommu-sva.c
291
group->attach_handle->domain->mm);
drivers/iommu/iommu-sva.c
294
iopf_group_response(group, status);
drivers/iommu/iommu-sva.c
295
iopf_free_group(group);
drivers/iommu/iommu-sva.c
298
static int iommu_sva_iopf_handler(struct iopf_group *group)
drivers/iommu/iommu-sva.c
300
struct iommu_fault_param *fault_param = group->fault_param;
drivers/iommu/iommu-sva.c
302
INIT_WORK(&group->work, iommu_sva_handle_iopf);
drivers/iommu/iommu-sva.c
303
if (!queue_work(fault_param->queue->wq, &group->work))
drivers/iommu/iommu-sva.c
74
struct iommu_group *group = dev->iommu_group;
drivers/iommu/iommu-sva.c
81
if (!group)
drivers/iommu/iommu-sva.c
94
attach_handle = iommu_attach_handle_get(group, iommu_mm->pasid, IOMMU_DOMAIN_SVA);
drivers/iommu/iommu.c
1003
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
1018
struct iommu_group *group = to_iommu_group(kobj);
drivers/iommu/iommu.c
1020
pr_debug("Releasing group %d\n", group->id);
drivers/iommu/iommu.c
1022
if (group->iommu_data_release)
drivers/iommu/iommu.c
1023
group->iommu_data_release(group->iommu_data);
drivers/iommu/iommu.c
1025
ida_free(&iommu_group_ida, group->id);
drivers/iommu/iommu.c
1028
WARN_ON(group->default_domain);
drivers/iommu/iommu.c
1029
WARN_ON(group->blocking_domain);
drivers/iommu/iommu.c
1031
kfree(group->name);
drivers/iommu/iommu.c
1032
kfree(group);
drivers/iommu/iommu.c
1053
struct iommu_group *group;
drivers/iommu/iommu.c
1056
group = kzalloc_obj(*group);
drivers/iommu/iommu.c
1057
if (!group)
drivers/iommu/iommu.c
1060
group->kobj.kset = iommu_group_kset;
drivers/iommu/iommu.c
1061
mutex_init(&group->mutex);
drivers/iommu/iommu.c
1062
INIT_LIST_HEAD(&group->devices);
drivers/iommu/iommu.c
1063
INIT_LIST_HEAD(&group->entry);
drivers/iommu/iommu.c
1064
xa_init(&group->pasid_array);
drivers/iommu/iommu.c
1068
kfree(group);
drivers/iommu/iommu.c
1071
group->id = ret;
drivers/iommu/iommu.c
1073
ret = kobject_init_and_add(&group->kobj, &iommu_group_ktype,
drivers/iommu/iommu.c
1074
NULL, "%d", group->id);
drivers/iommu/iommu.c
1076
kobject_put(&group->kobj);
drivers/iommu/iommu.c
1080
group->devices_kobj = kobject_create_and_add("devices", &group->kobj);
drivers/iommu/iommu.c
1081
if (!group->devices_kobj) {
drivers/iommu/iommu.c
1082
kobject_put(&group->kobj); /* triggers .release & free */
drivers/iommu/iommu.c
1091
kobject_put(&group->kobj);
drivers/iommu/iommu.c
1093
ret = iommu_group_create_file(group,
drivers/iommu/iommu.c
1096
kobject_put(group->devices_kobj);
drivers/iommu/iommu.c
110
struct iommu_group *group);
drivers/iommu/iommu.c
1100
ret = iommu_group_create_file(group, &iommu_group_attr_type);
drivers/iommu/iommu.c
1102
kobject_put(group->devices_kobj);
drivers/iommu/iommu.c
1106
pr_debug("Allocated group %d\n", group->id);
drivers/iommu/iommu.c
1108
return group;
drivers/iommu/iommu.c
1120
void *iommu_group_get_iommudata(struct iommu_group *group)
drivers/iommu/iommu.c
1122
return group->iommu_data;
drivers/iommu/iommu.c
1136
void iommu_group_set_iommudata(struct iommu_group *group, void *iommu_data,
drivers/iommu/iommu.c
1139
group->iommu_data = iommu_data;
drivers/iommu/iommu.c
1140
group->iommu_data_release = release;
drivers/iommu/iommu.c
1152
int iommu_group_set_name(struct iommu_group *group, const char *name)
drivers/iommu/iommu.c
1156
if (group->name) {
drivers/iommu/iommu.c
1157
iommu_group_remove_file(group, &iommu_group_attr_name);
drivers/iommu/iommu.c
1158
kfree(group->name);
drivers/iommu/iommu.c
1159
group->name = NULL;
drivers/iommu/iommu.c
1164
group->name = kstrdup(name, GFP_KERNEL);
drivers/iommu/iommu.c
1165
if (!group->name)
drivers/iommu/iommu.c
1168
ret = iommu_group_create_file(group, &iommu_group_attr_name);
drivers/iommu/iommu.c
1170
kfree(group->name);
drivers/iommu/iommu.c
1171
group->name = NULL;
drivers/iommu/iommu.c
119
static int __iommu_device_set_domain(struct iommu_group *group,
drivers/iommu/iommu.c
124
static int __iommu_group_set_domain_internal(struct iommu_group *group,
drivers/iommu/iommu.c
1245
static struct group_device *iommu_group_alloc_device(struct iommu_group *group,
drivers/iommu/iommu.c
1257
ret = sysfs_create_link(&dev->kobj, &group->kobj, "iommu_group");
drivers/iommu/iommu.c
1268
ret = sysfs_create_link_nowarn(group->devices_kobj,
drivers/iommu/iommu.c
127
static int __iommu_group_set_domain(struct iommu_group *group,
drivers/iommu/iommu.c
1284
trace_add_device_to_group(group->id, dev);
drivers/iommu/iommu.c
1286
dev_info(dev, "Adding to iommu group %d\n", group->id);
drivers/iommu/iommu.c
1296
dev_err(dev, "Failed to add to iommu group %d: %d\n", group->id, ret);
drivers/iommu/iommu.c
130
return __iommu_group_set_domain_internal(group, new_domain, 0);
drivers/iommu/iommu.c
1308
int iommu_group_add_device(struct iommu_group *group, struct device *dev)
drivers/iommu/iommu.c
1312
gdev = iommu_group_alloc_device(group, dev);
drivers/iommu/iommu.c
1316
iommu_group_ref_get(group);
drivers/iommu/iommu.c
1317
dev->iommu_group = group;
drivers/iommu/iommu.c
1319
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
132
static void __iommu_group_set_domain_nofail(struct iommu_group *group,
drivers/iommu/iommu.c
1320
list_add_tail(&gdev->list, &group->devices);
drivers/iommu/iommu.c
1321
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
1335
struct iommu_group *group = dev->iommu_group;
drivers/iommu/iommu.c
1337
if (!group)
drivers/iommu/iommu.c
1340
dev_info(dev, "Removing from iommu group %d\n", group->id);
drivers/iommu/iommu.c
1358
struct iommu_group *group = dev->iommu_group;
drivers/iommu/iommu.c
136
group, new_domain, IOMMU_SET_DOMAIN_MUST_SUCCEED));
drivers/iommu/iommu.c
1360
lockdep_assert_held(&group->mutex);
drivers/iommu/iommu.c
1365
static struct device *iommu_group_first_dev(struct iommu_group *group)
drivers/iommu/iommu.c
1367
lockdep_assert_held(&group->mutex);
drivers/iommu/iommu.c
1368
return list_first_entry(&group->devices, struct group_device, list)->dev;
drivers/iommu/iommu.c
1382
int iommu_group_for_each_dev(struct iommu_group *group, void *data,
drivers/iommu/iommu.c
1388
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
1389
for_each_group_device(group, device) {
drivers/iommu/iommu.c
139
static int iommu_setup_default_domain(struct iommu_group *group,
drivers/iommu/iommu.c
1394
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
1410
struct iommu_group *group = dev->iommu_group;
drivers/iommu/iommu.c
1412
if (group)
drivers/iommu/iommu.c
1413
kobject_get(group->devices_kobj);
drivers/iommu/iommu.c
1415
return group;
drivers/iommu/iommu.c
1426
struct iommu_group *iommu_group_ref_get(struct iommu_group *group)
drivers/iommu/iommu.c
1428
kobject_get(group->devices_kobj);
drivers/iommu/iommu.c
1429
return group;
drivers/iommu/iommu.c
143
static ssize_t iommu_group_store_type(struct iommu_group *group,
drivers/iommu/iommu.c
1440
void iommu_group_put(struct iommu_group *group)
drivers/iommu/iommu.c
1442
if (group)
drivers/iommu/iommu.c
1443
kobject_put(group->devices_kobj);
drivers/iommu/iommu.c
145
static struct group_device *iommu_group_alloc_device(struct iommu_group *group,
drivers/iommu/iommu.c
1453
int iommu_group_id(struct iommu_group *group)
drivers/iommu/iommu.c
1455
return group->id;
drivers/iommu/iommu.c
147
static void __iommu_group_free_device(struct iommu_group *group,
drivers/iommu/iommu.c
1482
struct iommu_group *group;
drivers/iommu/iommu.c
1493
group = get_pci_alias_group(tmp, devfns);
drivers/iommu/iommu.c
1494
if (group) {
drivers/iommu/iommu.c
1496
return group;
drivers/iommu/iommu.c
1516
struct iommu_group *group;
drivers/iommu/iommu.c
1521
group = iommu_group_get(&pdev->dev);
drivers/iommu/iommu.c
1522
if (group)
drivers/iommu/iommu.c
1523
return group;
drivers/iommu/iommu.c
1531
group = get_pci_alias_group(tmp, devfns);
drivers/iommu/iommu.c
1532
if (group) {
drivers/iommu/iommu.c
1534
return group;
drivers/iommu/iommu.c
1537
group = get_pci_function_alias_group(tmp, devfns);
drivers/iommu/iommu.c
1538
if (group) {
drivers/iommu/iommu.c
1540
return group;
drivers/iommu/iommu.c
1550
struct iommu_group *group;
drivers/iommu/iommu.c
1562
data->group = iommu_group_get(&pdev->dev);
drivers/iommu/iommu.c
1564
return data->group != NULL;
drivers/iommu/iommu.c
1587
struct iommu_group *group;
drivers/iommu/iommu.c
1589
group = iommu_group_alloc();
drivers/iommu/iommu.c
1590
if (IS_ERR(group))
drivers/iommu/iommu.c
1591
return group;
drivers/iommu/iommu.c
1592
iommu->singleton_group = group;
drivers/iommu/iommu.c
1607
struct iommu_group *group = NULL;
drivers/iommu/iommu.c
1620
return data.group;
drivers/iommu/iommu.c
1639
group = iommu_group_get(&pdev->dev);
drivers/iommu/iommu.c
1640
if (group)
drivers/iommu/iommu.c
1641
return group;
drivers/iommu/iommu.c
1648
group = get_pci_alias_group(pdev, (unsigned long *)devfns);
drivers/iommu/iommu.c
1649
if (group)
drivers/iommu/iommu.c
1650
return group;
drivers/iommu/iommu.c
1657
group = get_pci_function_alias_group(pdev, (unsigned long *)devfns);
drivers/iommu/iommu.c
1658
if (group)
drivers/iommu/iommu.c
1659
return group;
drivers/iommu/iommu.c
1670
struct iommu_group *group;
drivers/iommu/iommu.c
1672
group = iommu_group_get(cont_dev);
drivers/iommu/iommu.c
1673
if (!group)
drivers/iommu/iommu.c
1674
group = iommu_group_alloc();
drivers/iommu/iommu.c
1675
return group;
drivers/iommu/iommu.c
1700
__iommu_group_alloc_default_domain(struct iommu_group *group, int req_type)
drivers/iommu/iommu.c
1702
struct device *dev = iommu_group_first_dev(group);
drivers/iommu/iommu.c
1705
if (group->default_domain && group->default_domain->type == req_type)
drivers/iommu/iommu.c
1706
return group->default_domain;
drivers/iommu/iommu.c
1737
iommu_group_alloc_default_domain(struct iommu_group *group, int req_type)
drivers/iommu/iommu.c
1739
const struct iommu_ops *ops = dev_iommu_ops(iommu_group_first_dev(group));
drivers/iommu/iommu.c
1742
lockdep_assert_held(&group->mutex);
drivers/iommu/iommu.c
1756
return __iommu_group_alloc_default_domain(group, req_type);
drivers/iommu/iommu.c
1759
dom = __iommu_group_alloc_default_domain(group, iommu_def_domain_type);
drivers/iommu/iommu.c
1766
dom = __iommu_group_alloc_default_domain(group, IOMMU_DOMAIN_DMA);
drivers/iommu/iommu.c
1771
iommu_def_domain_type, group->name);
drivers/iommu/iommu.c
1775
struct iommu_domain *iommu_group_default_domain(struct iommu_group *group)
drivers/iommu/iommu.c
1777
return group->default_domain;
drivers/iommu/iommu.c
1816
static int iommu_get_def_domain_type(struct iommu_group *group,
drivers/iommu/iommu.c
1843
group->id);
drivers/iommu/iommu.c
1858
static int iommu_get_default_domain_type(struct iommu_group *group,
drivers/iommu/iommu.c
1865
lockdep_assert_held(&group->mutex);
drivers/iommu/iommu.c
1879
for_each_group_device(group, gdev) {
drivers/iommu/iommu.c
1880
driver_type = iommu_get_def_domain_type(group, gdev->dev,
drivers/iommu/iommu.c
1911
group->id, iommu_domain_type_str(driver_type));
drivers/iommu/iommu.c
1935
struct iommu_group *group, *next;
drivers/iommu/iommu.c
1943
list_for_each_entry_safe(group, next, &group_list, entry) {
drivers/iommu/iommu.c
1946
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
1949
list_del_init(&group->entry);
drivers/iommu/iommu.c
1956
ret = iommu_setup_default_domain(group, 0);
drivers/iommu/iommu.c
1958
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
1961
for_each_group_device(group, gdev)
drivers/iommu/iommu.c
1962
iommu_setup_dma_ops(gdev->dev, group->default_domain);
drivers/iommu/iommu.c
1963
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
1971
for_each_group_device(group, gdev)
drivers/iommu/iommu.c
2011
bool iommu_group_has_isolated_msi(struct iommu_group *group)
drivers/iommu/iommu.c
2016
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
2017
for_each_group_device(group, group_dev)
drivers/iommu/iommu.c
2019
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
2130
static void __iommu_group_set_core_domain(struct iommu_group *group)
drivers/iommu/iommu.c
2134
if (group->owner)
drivers/iommu/iommu.c
2135
new_domain = group->blocking_domain;
drivers/iommu/iommu.c
2137
new_domain = group->default_domain;
drivers/iommu/iommu.c
2139
__iommu_group_set_domain_nofail(group, new_domain);
drivers/iommu/iommu.c
2173
struct iommu_group *group = dev->iommu_group;
drivers/iommu/iommu.c
2176
if (!group)
drivers/iommu/iommu.c
2183
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
2185
if (list_count_nodes(&group->devices) != 1)
drivers/iommu/iommu.c
2188
ret = __iommu_attach_group(domain, group);
drivers/iommu/iommu.c
2191
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
2223
struct iommu_group *group = dev->iommu_group;
drivers/iommu/iommu.c
2225
if (!group)
drivers/iommu/iommu.c
2228
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
2229
if (WARN_ON(domain != group->domain) ||
drivers/iommu/iommu.c
2230
WARN_ON(list_count_nodes(&group->devices) != 1))
drivers/iommu/iommu.c
2232
__iommu_group_set_core_domain(group);
drivers/iommu/iommu.c
2235
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
2251
struct iommu_group *group = dev->iommu_group;
drivers/iommu/iommu.c
2253
if (!group)
drivers/iommu/iommu.c
2256
lockdep_assert_not_held(&group->mutex);
drivers/iommu/iommu.c
2258
return group->domain;
drivers/iommu/iommu.c
2271
struct iommu_group *group = dev->iommu_group;
drivers/iommu/iommu.c
2273
lockdep_assert_held(&group->mutex);
drivers/iommu/iommu.c
2283
if (group->resetting_domain)
drivers/iommu/iommu.c
2284
return group->resetting_domain;
drivers/iommu/iommu.c
2286
return group->domain;
drivers/iommu/iommu.c
2324
struct iommu_group *group)
drivers/iommu/iommu.c
2328
if (group->domain && group->domain != group->default_domain &&
drivers/iommu/iommu.c
2329
group->domain != group->blocking_domain)
drivers/iommu/iommu.c
2332
dev = iommu_group_first_dev(group);
drivers/iommu/iommu.c
2337
return __iommu_group_set_domain(group, domain);
drivers/iommu/iommu.c
2352
int iommu_attach_group(struct iommu_domain *domain, struct iommu_group *group)
drivers/iommu/iommu.c
2356
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
2357
ret = __iommu_attach_group(domain, group);
drivers/iommu/iommu.c
2358
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
2364
static int __iommu_device_set_domain(struct iommu_group *group,
drivers/iommu/iommu.c
2381
new_domain == group->blocking_domain)) {
drivers/iommu/iommu.c
2388
if (new_domain == group->default_domain)
drivers/iommu/iommu.c
2401
group->blocking_domain &&
drivers/iommu/iommu.c
2402
group->blocking_domain != new_domain)
drivers/iommu/iommu.c
2403
__iommu_attach_device(group->blocking_domain, dev,
drivers/iommu/iommu.c
2425
static int __iommu_group_set_domain_internal(struct iommu_group *group,
drivers/iommu/iommu.c
2434
lockdep_assert_held(&group->mutex);
drivers/iommu/iommu.c
2436
if (group->domain == new_domain)
drivers/iommu/iommu.c
2446
if (group->resetting_domain)
drivers/iommu/iommu.c
2456
for_each_group_device(group, gdev) {
drivers/iommu/iommu.c
2457
ret = __iommu_device_set_domain(group, gdev->dev, new_domain,
drivers/iommu/iommu.c
2458
group->domain, flags);
drivers/iommu/iommu.c
2473
group->domain = new_domain;
drivers/iommu/iommu.c
2482
for_each_group_device(group, gdev) {
drivers/iommu/iommu.c
2491
if (group->domain)
drivers/iommu/iommu.c
2493
group, gdev->dev, group->domain, new_domain,
drivers/iommu/iommu.c
2499
void iommu_detach_group(struct iommu_domain *domain, struct iommu_group *group)
drivers/iommu/iommu.c
2501
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
2502
__iommu_group_set_core_domain(group);
drivers/iommu/iommu.c
2503
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
3075
static int iommu_setup_default_domain(struct iommu_group *group,
drivers/iommu/iommu.c
3078
struct iommu_domain *old_dom = group->default_domain;
drivers/iommu/iommu.c
3085
lockdep_assert_held(&group->mutex);
drivers/iommu/iommu.c
3087
req_type = iommu_get_default_domain_type(group, target_type);
drivers/iommu/iommu.c
3091
dom = iommu_group_alloc_default_domain(group, req_type);
drivers/iommu/iommu.c
3095
if (group->default_domain == dom)
drivers/iommu/iommu.c
3112
for_each_group_device(group, gdev) {
drivers/iommu/iommu.c
3122
group->default_domain = dom;
drivers/iommu/iommu.c
3123
if (!group->domain) {
drivers/iommu/iommu.c
3131
group, dom, IOMMU_SET_DOMAIN_MUST_SUCCEED);
drivers/iommu/iommu.c
3135
ret = __iommu_group_set_domain(group, dom);
drivers/iommu/iommu.c
3147
for_each_group_device(group, gdev) {
drivers/iommu/iommu.c
3162
group, old_dom, IOMMU_SET_DOMAIN_MUST_SUCCEED);
drivers/iommu/iommu.c
3166
group->default_domain = old_dom;
drivers/iommu/iommu.c
3180
static ssize_t iommu_group_store_type(struct iommu_group *group,
drivers/iommu/iommu.c
3189
if (WARN_ON(!group) || !group->default_domain)
drivers/iommu/iommu.c
3203
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
3206
group->default_domain->type == IOMMU_DOMAIN_DMA) {
drivers/iommu/iommu.c
3207
ret = iommu_dma_init_fq(group->default_domain);
drivers/iommu/iommu.c
3211
group->default_domain->type = IOMMU_DOMAIN_DMA_FQ;
drivers/iommu/iommu.c
3217
if (list_empty(&group->devices) || group->owner_cnt) {
drivers/iommu/iommu.c
3222
ret = iommu_setup_default_domain(group, req_type);
drivers/iommu/iommu.c
3227
for_each_group_device(group, gdev)
drivers/iommu/iommu.c
3228
iommu_setup_dma_ops(gdev->dev, group->default_domain);
drivers/iommu/iommu.c
3231
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
3246
struct iommu_group *group = dev->iommu_group;
drivers/iommu/iommu.c
3249
if (!group)
drivers/iommu/iommu.c
3252
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
3254
if (!group->default_domain) {
drivers/iommu/iommu.c
3258
if (group->owner_cnt) {
drivers/iommu/iommu.c
3259
if (group->domain != group->default_domain || group->owner ||
drivers/iommu/iommu.c
3260
!xa_empty(&group->pasid_array)) {
drivers/iommu/iommu.c
3266
group->owner_cnt++;
drivers/iommu/iommu.c
3269
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
3284
struct iommu_group *group = dev->iommu_group;
drivers/iommu/iommu.c
3286
if (!group)
drivers/iommu/iommu.c
3289
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
3290
if (!WARN_ON(!group->owner_cnt || !xa_empty(&group->pasid_array)))
drivers/iommu/iommu.c
3291
group->owner_cnt--;
drivers/iommu/iommu.c
3293
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
3296
static int __iommu_group_alloc_blocking_domain(struct iommu_group *group)
drivers/iommu/iommu.c
3298
struct device *dev = iommu_group_first_dev(group);
drivers/iommu/iommu.c
3302
if (group->blocking_domain)
drivers/iommu/iommu.c
3306
group->blocking_domain = ops->blocked_domain;
drivers/iommu/iommu.c
3317
group->blocking_domain = domain;
drivers/iommu/iommu.c
3321
static int __iommu_take_dma_ownership(struct iommu_group *group, void *owner)
drivers/iommu/iommu.c
3325
if ((group->domain && group->domain != group->default_domain) ||
drivers/iommu/iommu.c
3326
!xa_empty(&group->pasid_array))
drivers/iommu/iommu.c
3329
ret = __iommu_group_alloc_blocking_domain(group);
drivers/iommu/iommu.c
3332
ret = __iommu_group_set_domain(group, group->blocking_domain);
drivers/iommu/iommu.c
3336
group->owner = owner;
drivers/iommu/iommu.c
3337
group->owner_cnt++;
drivers/iommu/iommu.c
3350
int iommu_group_claim_dma_owner(struct iommu_group *group, void *owner)
drivers/iommu/iommu.c
3357
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
3358
if (group->owner_cnt) {
drivers/iommu/iommu.c
3363
ret = __iommu_take_dma_ownership(group, owner);
drivers/iommu/iommu.c
3365
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
3383
struct iommu_group *group = dev->iommu_group;
drivers/iommu/iommu.c
3389
if (!group)
drivers/iommu/iommu.c
3392
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
3393
if (group->owner_cnt) {
drivers/iommu/iommu.c
3394
if (group->owner != owner) {
drivers/iommu/iommu.c
3398
group->owner_cnt++;
drivers/iommu/iommu.c
3402
ret = __iommu_take_dma_ownership(group, owner);
drivers/iommu/iommu.c
3404
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
3409
static void __iommu_release_dma_ownership(struct iommu_group *group)
drivers/iommu/iommu.c
3411
if (WARN_ON(!group->owner_cnt || !group->owner ||
drivers/iommu/iommu.c
3412
!xa_empty(&group->pasid_array)))
drivers/iommu/iommu.c
3415
group->owner_cnt = 0;
drivers/iommu/iommu.c
3416
group->owner = NULL;
drivers/iommu/iommu.c
3417
__iommu_group_set_domain_nofail(group, group->default_domain);
drivers/iommu/iommu.c
3426
void iommu_group_release_dma_owner(struct iommu_group *group)
drivers/iommu/iommu.c
3428
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
3429
__iommu_release_dma_ownership(group);
drivers/iommu/iommu.c
3430
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
3443
struct iommu_group *group = dev->iommu_group;
drivers/iommu/iommu.c
3445
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
3446
if (group->owner_cnt > 1)
drivers/iommu/iommu.c
3447
group->owner_cnt--;
drivers/iommu/iommu.c
3449
__iommu_release_dma_ownership(group);
drivers/iommu/iommu.c
3450
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
3461
bool iommu_group_dma_owner_claimed(struct iommu_group *group)
drivers/iommu/iommu.c
3465
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
3466
user = group->owner_cnt;
drivers/iommu/iommu.c
3467
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
3484
struct iommu_group *group, ioasid_t pasid,
drivers/iommu/iommu.c
3490
for_each_group_device(group, device) {
drivers/iommu/iommu.c
3503
for_each_group_device(group, device) {
drivers/iommu/iommu.c
3522
static void __iommu_remove_group_pasid(struct iommu_group *group,
drivers/iommu/iommu.c
3528
for_each_group_device(group, device) {
drivers/iommu/iommu.c
3551
struct iommu_group *group = dev->iommu_group;
drivers/iommu/iommu.c
3557
if (!group)
drivers/iommu/iommu.c
3571
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
3577
if (group->resetting_domain) {
drivers/iommu/iommu.c
3582
for_each_group_device(group, device) {
drivers/iommu/iommu.c
3601
ret = xa_insert(&group->pasid_array, pasid, XA_ZERO_ENTRY, GFP_KERNEL);
drivers/iommu/iommu.c
3605
ret = __iommu_set_group_pasid(domain, group, pasid, NULL);
drivers/iommu/iommu.c
3607
xa_release(&group->pasid_array, pasid);
drivers/iommu/iommu.c
3617
WARN_ON(xa_is_err(xa_store(&group->pasid_array,
drivers/iommu/iommu.c
3621
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
3648
struct iommu_group *group = dev->iommu_group;
drivers/iommu/iommu.c
3654
if (!group)
drivers/iommu/iommu.c
3664
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
3670
if (group->resetting_domain) {
drivers/iommu/iommu.c
3676
curr = xa_cmpxchg(&group->pasid_array, pasid, NULL,
drivers/iommu/iommu.c
3688
xa_release(&group->pasid_array, pasid);
drivers/iommu/iommu.c
3708
ret = __iommu_set_group_pasid(domain, group,
drivers/iommu/iommu.c
3718
WARN_ON(xa_is_err(xa_store(&group->pasid_array,
drivers/iommu/iommu.c
3722
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
3740
struct iommu_group *group = dev->iommu_group;
drivers/iommu/iommu.c
3742
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
3743
__iommu_remove_group_pasid(group, pasid, domain);
drivers/iommu/iommu.c
3744
xa_erase(&group->pasid_array, pasid);
drivers/iommu/iommu.c
3745
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
3791
iommu_attach_handle_get(struct iommu_group *group, ioasid_t pasid, unsigned int type)
drivers/iommu/iommu.c
3796
xa_lock(&group->pasid_array);
drivers/iommu/iommu.c
3797
entry = xa_load(&group->pasid_array, pasid);
drivers/iommu/iommu.c
3805
xa_unlock(&group->pasid_array);
drivers/iommu/iommu.c
3827
struct iommu_group *group,
drivers/iommu/iommu.c
3836
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
3838
ret = xa_insert(&group->pasid_array,
drivers/iommu/iommu.c
3843
ret = __iommu_attach_group(domain, group);
drivers/iommu/iommu.c
3845
xa_release(&group->pasid_array, IOMMU_NO_PASID);
drivers/iommu/iommu.c
3855
WARN_ON(xa_is_err(xa_store(&group->pasid_array,
drivers/iommu/iommu.c
3859
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
3873
struct iommu_group *group)
drivers/iommu/iommu.c
3875
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
3876
__iommu_group_set_core_domain(group);
drivers/iommu/iommu.c
3877
xa_erase(&group->pasid_array, IOMMU_NO_PASID);
drivers/iommu/iommu.c
3878
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
3898
int iommu_replace_group_handle(struct iommu_group *group,
drivers/iommu/iommu.c
3908
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
3910
ret = xa_reserve(&group->pasid_array, IOMMU_NO_PASID, GFP_KERNEL);
drivers/iommu/iommu.c
3914
ret = __iommu_group_set_domain(group, new_domain);
drivers/iommu/iommu.c
3918
curr = xa_store(&group->pasid_array, IOMMU_NO_PASID, entry, GFP_KERNEL);
drivers/iommu/iommu.c
3921
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
3925
xa_release(&group->pasid_array, IOMMU_NO_PASID);
drivers/iommu/iommu.c
3927
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
3958
struct iommu_group *group = pdev->dev.iommu_group;
drivers/iommu/iommu.c
3966
guard(mutex)(&group->mutex);
drivers/iommu/iommu.c
3969
if (WARN_ON(group->resetting_domain))
drivers/iommu/iommu.c
3972
ret = __iommu_group_alloc_blocking_domain(group);
drivers/iommu/iommu.c
3977
if (group->domain != group->blocking_domain) {
drivers/iommu/iommu.c
3978
ret = __iommu_attach_device(group->blocking_domain, &pdev->dev,
drivers/iommu/iommu.c
3979
group->domain);
drivers/iommu/iommu.c
3990
xa_for_each_start(&group->pasid_array, pasid, entry, 1)
drivers/iommu/iommu.c
3994
group->resetting_domain = group->blocking_domain;
drivers/iommu/iommu.c
4015
struct iommu_group *group = pdev->dev.iommu_group;
drivers/iommu/iommu.c
4022
guard(mutex)(&group->mutex);
drivers/iommu/iommu.c
4025
if (!group->resetting_domain)
drivers/iommu/iommu.c
4029
if (WARN_ON(!group->blocking_domain))
drivers/iommu/iommu.c
4033
if (group->domain != group->blocking_domain) {
drivers/iommu/iommu.c
4034
WARN_ON(__iommu_attach_device(group->domain, &pdev->dev,
drivers/iommu/iommu.c
4035
group->blocking_domain));
drivers/iommu/iommu.c
4044
xa_for_each_start(&group->pasid_array, pasid, entry, 1)
drivers/iommu/iommu.c
4046
pasid_array_entry_to_domain(entry), group, pasid,
drivers/iommu/iommu.c
4047
group->blocking_domain));
drivers/iommu/iommu.c
4049
group->resetting_domain = NULL;
drivers/iommu/iommu.c
4068
struct iommu_group *group = dev->iommu_group;
drivers/iommu/iommu.c
4071
if (!group)
drivers/iommu/iommu.c
4074
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
4076
if (group->domain && group->domain->type != IOMMU_DOMAIN_IDENTITY) {
drivers/iommu/iommu.c
4077
switch (group->domain->cookie_type) {
drivers/iommu/iommu.c
4080
ret = iommu_dma_sw_msi(group->domain, desc, msi_addr);
drivers/iommu/iommu.c
4083
ret = iommufd_sw_msi(group->domain, desc, msi_addr);
drivers/iommu/iommu.c
4090
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
451
struct iommu_group *group;
drivers/iommu/iommu.c
500
group = ops->device_group(dev);
drivers/iommu/iommu.c
501
if (WARN_ON_ONCE(group == NULL))
drivers/iommu/iommu.c
502
group = ERR_PTR(-EINVAL);
drivers/iommu/iommu.c
503
if (IS_ERR(group)) {
drivers/iommu/iommu.c
504
ret = PTR_ERR(group);
drivers/iommu/iommu.c
507
dev->iommu_group = group;
drivers/iommu/iommu.c
529
struct iommu_group *group = dev->iommu_group;
drivers/iommu/iommu.c
532
lockdep_assert_held(&group->mutex);
drivers/iommu/iommu.c
566
group->domain);
drivers/iommu/iommu.c
576
if (list_empty(&group->devices)) {
drivers/iommu/iommu.c
577
if (group->default_domain) {
drivers/iommu/iommu.c
578
iommu_domain_free(group->default_domain);
drivers/iommu/iommu.c
579
group->default_domain = NULL;
drivers/iommu/iommu.c
581
if (group->blocking_domain) {
drivers/iommu/iommu.c
582
iommu_domain_free(group->blocking_domain);
drivers/iommu/iommu.c
583
group->blocking_domain = NULL;
drivers/iommu/iommu.c
585
group->domain = NULL;
drivers/iommu/iommu.c
608
struct iommu_group *group;
drivers/iommu/iommu.c
635
group = dev->iommu_group;
drivers/iommu/iommu.c
636
gdev = iommu_group_alloc_device(group, dev);
drivers/iommu/iommu.c
637
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
647
list_add_tail(&gdev->list, &group->devices);
drivers/iommu/iommu.c
648
WARN_ON(group->default_domain && !group->domain);
drivers/iommu/iommu.c
649
if (group->default_domain)
drivers/iommu/iommu.c
650
iommu_create_device_direct_mappings(group->default_domain, dev);
drivers/iommu/iommu.c
651
if (group->domain) {
drivers/iommu/iommu.c
652
ret = __iommu_device_set_domain(group, dev, group->domain, NULL,
drivers/iommu/iommu.c
656
} else if (!group->default_domain && !group_list) {
drivers/iommu/iommu.c
657
ret = iommu_setup_default_domain(group, 0);
drivers/iommu/iommu.c
660
} else if (!group->default_domain) {
drivers/iommu/iommu.c
666
if (list_empty(&group->entry))
drivers/iommu/iommu.c
667
list_add_tail(&group->entry, group_list);
drivers/iommu/iommu.c
670
if (group->default_domain)
drivers/iommu/iommu.c
671
iommu_setup_dma_ops(dev, group->default_domain);
drivers/iommu/iommu.c
673
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
679
__iommu_group_free_device(group, gdev);
drivers/iommu/iommu.c
682
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
683
iommu_group_put(group);
drivers/iommu/iommu.c
706
static void __iommu_group_free_device(struct iommu_group *group,
drivers/iommu/iommu.c
711
sysfs_remove_link(group->devices_kobj, grp_dev->name);
drivers/iommu/iommu.c
714
trace_remove_device_from_group(group->id, dev);
drivers/iommu/iommu.c
721
if (list_empty(&group->devices))
drivers/iommu/iommu.c
722
WARN_ON(group->owner_cnt ||
drivers/iommu/iommu.c
723
group->domain != group->default_domain);
drivers/iommu/iommu.c
732
struct iommu_group *group = dev->iommu_group;
drivers/iommu/iommu.c
735
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
736
for_each_group_device(group, device) {
drivers/iommu/iommu.c
741
__iommu_group_free_device(group, device);
drivers/iommu/iommu.c
748
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
754
iommu_group_put(group);
drivers/iommu/iommu.c
759
struct iommu_group *group = dev->iommu_group;
drivers/iommu/iommu.c
761
if (group)
drivers/iommu/iommu.c
808
struct iommu_group *group = to_iommu_group(kobj);
drivers/iommu/iommu.c
812
ret = attr->show(group, buf);
drivers/iommu/iommu.c
82
#define for_each_group_device(group, pos) \
drivers/iommu/iommu.c
821
struct iommu_group *group = to_iommu_group(kobj);
drivers/iommu/iommu.c
825
ret = attr->store(group, buf, count);
drivers/iommu/iommu.c
83
list_for_each_entry(pos, &(group)->devices, list)
drivers/iommu/iommu.c
834
static int iommu_group_create_file(struct iommu_group *group,
drivers/iommu/iommu.c
837
return sysfs_create_file(&group->kobj, &attr->attr);
drivers/iommu/iommu.c
840
static void iommu_group_remove_file(struct iommu_group *group,
drivers/iommu/iommu.c
843
sysfs_remove_file(&group->kobj, &attr->attr);
drivers/iommu/iommu.c
846
static ssize_t iommu_group_show_name(struct iommu_group *group, char *buf)
drivers/iommu/iommu.c
848
return sysfs_emit(buf, "%s\n", group->name);
drivers/iommu/iommu.c
87
ssize_t (*show)(struct iommu_group *group, char *buf);
drivers/iommu/iommu.c
88
ssize_t (*store)(struct iommu_group *group,
drivers/iommu/iommu.c
927
int iommu_get_group_resv_regions(struct iommu_group *group,
drivers/iommu/iommu.c
933
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
934
for_each_group_device(group, device) {
drivers/iommu/iommu.c
951
mutex_unlock(&group->mutex);
drivers/iommu/iommu.c
956
static ssize_t iommu_group_show_resv_regions(struct iommu_group *group,
drivers/iommu/iommu.c
964
iommu_get_group_resv_regions(group, &group_resv_regions);
drivers/iommu/iommu.c
978
static ssize_t iommu_group_show_type(struct iommu_group *group,
drivers/iommu/iommu.c
983
mutex_lock(&group->mutex);
drivers/iommu/iommu.c
984
if (group->default_domain) {
drivers/iommu/iommu.c
985
switch (group->default_domain->type) {
drivers/iommu/iommufd/device.c
101
new_igroup->group = group;
drivers/iommu/iommufd/device.c
131
if (iommufd_group_try_get(igroup, group)) {
drivers/iommu/iommufd/device.c
243
!iommu_group_has_isolated_msi(igroup->group)) {
drivers/iommu/iommufd/device.c
303
bool iommufd_ctx_has_group(struct iommufd_ctx *ictx, struct iommu_group *group)
drivers/iommu/iommufd/device.c
308
if (!ictx || !group)
drivers/iommu/iommufd/device.c
315
->igroup->group == group) {
drivers/iommu/iommufd/device.c
33
xa_cmpxchg(&igroup->ictx->groups, iommu_group_id(igroup->group), igroup,
drivers/iommu/iommufd/device.c
35
iommu_group_put(igroup->group);
drivers/iommu/iommufd/device.c
40
static void iommufd_put_group(struct iommufd_group *group)
drivers/iommu/iommufd/device.c
42
kref_put(&group->ref, iommufd_group_release);
drivers/iommu/iommufd/device.c
46
struct iommu_group *group)
drivers/iommu/iommufd/device.c
517
rc = iommu_attach_group_handle(hwpt->domain, idev->igroup->group,
drivers/iommu/iommufd/device.c
539
handle = iommu_attach_handle_get(idev->igroup->group, pasid, 0);
drivers/iommu/iommufd/device.c
54
if (WARN_ON(igroup->group != group))
drivers/iommu/iommufd/device.c
553
iommu_detach_group_handle(hwpt->domain, idev->igroup->group);
drivers/iommu/iommufd/device.c
584
rc = iommu_replace_group_handle(idev->igroup->group,
drivers/iommu/iommufd/device.c
72
struct iommu_group *group;
drivers/iommu/iommufd/device.c
75
group = iommu_group_get(dev);
drivers/iommu/iommufd/device.c
76
if (!group)
drivers/iommu/iommufd/device.c
79
id = iommu_group_id(group);
drivers/iommu/iommufd/device.c
83
if (iommufd_group_try_get(igroup, group)) {
drivers/iommu/iommufd/device.c
85
iommu_group_put(group);
drivers/iommu/iommufd/device.c
92
iommu_group_put(group);
drivers/iommu/iommufd/eventq.c
104
struct iopf_group *group = NULL;
drivers/iommu/iommufd/eventq.c
108
group = list_first_entry(list, struct iopf_group, node);
drivers/iommu/iommufd/eventq.c
109
list_del(&group->node);
drivers/iommu/iommufd/eventq.c
112
return group;
drivers/iommu/iommufd/eventq.c
117
struct iopf_group *group)
drivers/iommu/iommufd/eventq.c
120
list_add(&group->node, &fault->common.deliver);
drivers/iommu/iommufd/eventq.c
132
struct iopf_group *group;
drivers/iommu/iommufd/eventq.c
141
while ((group = iommufd_fault_deliver_fetch(fault))) {
drivers/iommu/iommufd/eventq.c
143
group->fault_count * fault_size > count - done) {
drivers/iommu/iommufd/eventq.c
144
iommufd_fault_deliver_restore(fault, group);
drivers/iommu/iommufd/eventq.c
148
rc = xa_alloc(&fault->response, &group->cookie, group,
drivers/iommu/iommufd/eventq.c
151
iommufd_fault_deliver_restore(fault, group);
drivers/iommu/iommufd/eventq.c
155
idev = to_iommufd_handle(group->attach_handle)->idev;
drivers/iommu/iommufd/eventq.c
156
list_for_each_entry(iopf, &group->faults, list) {
drivers/iommu/iommufd/eventq.c
159
group->cookie);
drivers/iommu/iommufd/eventq.c
161
xa_erase(&fault->response, group->cookie);
drivers/iommu/iommufd/eventq.c
162
iommufd_fault_deliver_restore(fault, group);
drivers/iommu/iommufd/eventq.c
181
struct iopf_group *group;
drivers/iommu/iommufd/eventq.c
204
group = xa_erase(&fault->response, response.cookie);
drivers/iommu/iommufd/eventq.c
205
if (!group) {
drivers/iommu/iommufd/eventq.c
210
iopf_group_response(group, response.code);
drivers/iommu/iommufd/eventq.c
211
iopf_free_group(group);
drivers/iommu/iommufd/eventq.c
23
struct iopf_group *group, *next;
drivers/iommu/iommufd/eventq.c
33
list_for_each_entry_safe(group, next, &fault->common.deliver, node) {
drivers/iommu/iommufd/eventq.c
34
if (group->attach_handle != &handle->handle)
drivers/iommu/iommufd/eventq.c
36
list_move(&group->node, &free_list);
drivers/iommu/iommufd/eventq.c
40
list_for_each_entry_safe(group, next, &free_list, node) {
drivers/iommu/iommufd/eventq.c
41
list_del(&group->node);
drivers/iommu/iommufd/eventq.c
42
iopf_group_response(group, IOMMU_PAGE_RESP_INVALID);
drivers/iommu/iommufd/eventq.c
43
iopf_free_group(group);
drivers/iommu/iommufd/eventq.c
455
int iommufd_fault_iopf_handler(struct iopf_group *group)
drivers/iommu/iommufd/eventq.c
46
xa_for_each(&fault->response, index, group) {
drivers/iommu/iommufd/eventq.c
460
hwpt = group->attach_handle->domain->iommufd_hwpt;
drivers/iommu/iommufd/eventq.c
464
list_add_tail(&group->node, &fault->common.deliver);
drivers/iommu/iommufd/eventq.c
47
if (group->attach_handle != &handle->handle)
drivers/iommu/iommufd/eventq.c
50
iopf_group_response(group, IOMMU_PAGE_RESP_INVALID);
drivers/iommu/iommufd/eventq.c
51
iopf_free_group(group);
drivers/iommu/iommufd/eventq.c
61
struct iopf_group *group, *next;
drivers/iommu/iommufd/eventq.c
70
list_for_each_entry_safe(group, next, &fault->common.deliver, node) {
drivers/iommu/iommufd/eventq.c
71
list_del(&group->node);
drivers/iommu/iommufd/eventq.c
72
iopf_group_response(group, IOMMU_PAGE_RESP_INVALID);
drivers/iommu/iommufd/eventq.c
73
iopf_free_group(group);
drivers/iommu/iommufd/eventq.c
75
xa_for_each(&fault->response, index, group) {
drivers/iommu/iommufd/eventq.c
77
iopf_group_response(group, IOMMU_PAGE_RESP_INVALID);
drivers/iommu/iommufd/eventq.c
78
iopf_free_group(group);
drivers/iommu/iommufd/iommufd_private.h
473
struct iommu_group *group;
drivers/iommu/iommufd/iommufd_private.h
596
int iommufd_fault_iopf_handler(struct iopf_group *group);
drivers/iommu/mtk_iommu.c
978
struct iommu_group *group;
drivers/iommu/mtk_iommu.c
990
group = data->m4u_group[groupid];
drivers/iommu/mtk_iommu.c
991
if (!group) {
drivers/iommu/mtk_iommu.c
992
group = iommu_group_alloc();
drivers/iommu/mtk_iommu.c
993
if (!IS_ERR(group))
drivers/iommu/mtk_iommu.c
994
data->m4u_group[groupid] = group;
drivers/iommu/mtk_iommu.c
996
iommu_group_ref_get(group);
drivers/iommu/mtk_iommu.c
999
return group;
drivers/iommu/tegra-smmu.c
1040
const struct tegra_smmu_swgroup *group = &smmu->soc->swgroups[i];
drivers/iommu/tegra-smmu.c
1044
value = smmu_readl(smmu, group->reg);
drivers/iommu/tegra-smmu.c
1053
seq_printf(s, "%-9s %-7s %#04x\n", group->name, status,
drivers/iommu/tegra-smmu.c
28
struct iommu_group *group;
drivers/iommu/tegra-smmu.c
346
const struct tegra_smmu_swgroup *group = NULL;
drivers/iommu/tegra-smmu.c
351
group = &smmu->soc->swgroups[i];
drivers/iommu/tegra-smmu.c
356
return group;
drivers/iommu/tegra-smmu.c
362
const struct tegra_smmu_swgroup *group;
drivers/iommu/tegra-smmu.c
366
group = tegra_smmu_find_swgroup(smmu, swgroup);
drivers/iommu/tegra-smmu.c
367
if (group) {
drivers/iommu/tegra-smmu.c
368
value = smmu_readl(smmu, group->reg);
drivers/iommu/tegra-smmu.c
372
smmu_writel(smmu, value, group->reg);
drivers/iommu/tegra-smmu.c
395
const struct tegra_smmu_swgroup *group;
drivers/iommu/tegra-smmu.c
399
group = tegra_smmu_find_swgroup(smmu, swgroup);
drivers/iommu/tegra-smmu.c
400
if (group) {
drivers/iommu/tegra-smmu.c
401
value = smmu_readl(smmu, group->reg);
drivers/iommu/tegra-smmu.c
405
smmu_writel(smmu, value, group->reg);
drivers/iommu/tegra-smmu.c
907
struct tegra_smmu_group *group = iommu_data;
drivers/iommu/tegra-smmu.c
908
struct tegra_smmu *smmu = group->smmu;
drivers/iommu/tegra-smmu.c
911
list_del(&group->list);
drivers/iommu/tegra-smmu.c
921
struct tegra_smmu_group *group;
drivers/iommu/tegra-smmu.c
930
list_for_each_entry(group, &smmu->groups, list)
drivers/iommu/tegra-smmu.c
931
if ((group->swgroup == swgroup) || (soc && group->soc == soc)) {
drivers/iommu/tegra-smmu.c
932
grp = iommu_group_ref_get(group->group);
drivers/iommu/tegra-smmu.c
937
group = devm_kzalloc(smmu->dev, sizeof(*group), GFP_KERNEL);
drivers/iommu/tegra-smmu.c
938
if (!group) {
drivers/iommu/tegra-smmu.c
943
INIT_LIST_HEAD(&group->list);
drivers/iommu/tegra-smmu.c
944
group->swgroup = swgroup;
drivers/iommu/tegra-smmu.c
945
group->smmu = smmu;
drivers/iommu/tegra-smmu.c
946
group->soc = soc;
drivers/iommu/tegra-smmu.c
949
group->group = pci_device_group(dev);
drivers/iommu/tegra-smmu.c
951
group->group = generic_device_group(dev);
drivers/iommu/tegra-smmu.c
953
if (IS_ERR(group->group)) {
drivers/iommu/tegra-smmu.c
954
devm_kfree(smmu->dev, group);
drivers/iommu/tegra-smmu.c
959
iommu_group_set_iommudata(group->group, group, tegra_smmu_group_release);
drivers/iommu/tegra-smmu.c
961
iommu_group_set_name(group->group, soc->name);
drivers/iommu/tegra-smmu.c
962
list_add_tail(&group->list, &smmu->groups);
drivers/iommu/tegra-smmu.c
965
return group->group;
drivers/irqchip/irq-gic-v3-its.c
1085
its_encode_sgi_group(cmd, desc->its_vsgi_cmd.group);
drivers/irqchip/irq-gic-v3-its.c
4345
desc.its_vsgi_cmd.group = vpe->sgi_config[d->hwirq].group;
drivers/irqchip/irq-gic-v3-its.c
4468
vpe->sgi_config[d->hwirq].group = info->group;
drivers/irqchip/irq-gic-v3-its.c
4500
vpe->sgi_config[i].group = false;
drivers/irqchip/irq-gic-v3-its.c
523
bool group;
drivers/irqchip/irq-gic-v4.c
363
int its_prop_update_vsgi(int irq, u8 priority, bool group)
drivers/irqchip/irq-gic-v4.c
369
.group = group,
drivers/irqchip/irq-sifive-plic.c
112
int group = hwirq / 32;
drivers/irqchip/irq-sifive-plic.c
115
value = readl(base + group);
drivers/irqchip/irq-sifive-plic.c
122
handler->enable_save[group] = value;
drivers/irqchip/irq-sifive-plic.c
123
writel(value, base + group);
drivers/irqchip/irq-sp7021-intc.c
151
u32 group;
drivers/irqchip/irq-sp7021-intc.c
159
group = fls(pending_group) - 1;
drivers/irqchip/irq-sp7021-intc.c
160
pending_irq = readl_relaxed(base + group * 4);
drivers/irqchip/irq-sp7021-intc.c
164
return (group * 32) + fls(pending_irq) - 1;
drivers/leds/blink/leds-lgm-sso.c
176
static u32 sso_led_pin_blink_off(u32 pin, unsigned int group)
drivers/leds/blink/leds-lgm-sso.c
178
if (group == LED_GRP2_29_31)
drivers/leds/blink/leds-lgm-sso.c
180
else if (group == LED_GRP1_24_28)
drivers/leds/blink/leds-lgm-sso.c
196
unsigned int group;
drivers/leds/blink/leds-lgm-sso.c
201
group = sso_led_pin_to_group(pin);
drivers/leds/blink/leds-lgm-sso.c
203
off = sso_led_pin_blink_off(pin, group);
drivers/leds/blink/leds-lgm-sso.c
205
if (group == LED_GRP0_0_23)
drivers/leds/blink/leds-lgm-sso.c
207
else if (group == LED_GRP1_24_28)
drivers/mcb/mcb-core.c
238
bus->bus_nr, device_id, dev->inst, dev->group, dev->var);
drivers/mcb/mcb-internal.h
109
unsigned int group:6;
drivers/mcb/mcb-parse.c
58
mdev->group = GDD_GRP(reg2);
drivers/md/dm-log-userspace-base.c
491
uint64_t group[MAX_FLUSH_GROUP_COUNT];
drivers/md/dm-log-userspace-base.c
500
group[count] = fe->region;
drivers/md/dm-log-userspace-base.c
512
(char *)(group),
drivers/md/dm-log-userspace-base.c
522
(char *)(group),
drivers/md/md-bitmap.c
3016
.group = &md_bitmap_group,
drivers/md/md-bitmap.h
128
struct attribute_group *group;
drivers/md/md-llbitmap.c
1600
.group = &md_llbitmap_group,
drivers/md/md.c
706
if (!mddev_is_dm(mddev) && mddev->bitmap_ops->group) {
drivers/md/md.c
707
if (sysfs_create_group(&mddev->kobj, mddev->bitmap_ops->group))
drivers/md/md.c
727
mddev->bitmap_ops->group)
drivers/md/md.c
728
sysfs_remove_group(&mddev->kobj, mddev->bitmap_ops->group);
drivers/md/raid5.c
182
struct r5worker_group *group;
drivers/md/raid5.c
192
struct r5worker_group *group;
drivers/md/raid5.c
193
group = conf->worker_groups + cpu_to_group(cpu);
drivers/md/raid5.c
195
list_add_tail(&sh->lru, &group->loprio_list);
drivers/md/raid5.c
197
list_add_tail(&sh->lru, &group->handle_list);
drivers/md/raid5.c
198
group->stripes_cnt++;
drivers/md/raid5.c
199
sh->group = group;
drivers/md/raid5.c
207
group = conf->worker_groups + cpu_to_group(sh->cpu);
drivers/md/raid5.c
209
group->workers[0].working = true;
drivers/md/raid5.c
211
queue_work_on(sh->cpu, raid5_wq, &group->workers[0].work);
drivers/md/raid5.c
213
thread_cnt = group->stripes_cnt / MAX_STRIPE_BATCH - 1;
drivers/md/raid5.c
216
if (group->workers[i].working == false) {
drivers/md/raid5.c
217
group->workers[i].working = true;
drivers/md/raid5.c
219
&group->workers[i].work);
drivers/md/raid5.c
5523
static struct stripe_head *__get_priority_stripe(struct r5conf *conf, int group)
drivers/md/raid5.c
5540
} else if (group != ANY_GROUP) {
drivers/md/raid5.c
5541
handle_list = try_loprio ? &conf->worker_groups[group].loprio_list :
drivers/md/raid5.c
5542
&conf->worker_groups[group].handle_list;
drivers/md/raid5.c
5543
wg = &conf->worker_groups[group];
drivers/md/raid5.c
5583
group == ANY_GROUP ||
drivers/md/raid5.c
5585
cpu_to_group(tmp->cpu) == group) {
drivers/md/raid5.c
5609
sh->group = NULL;
drivers/md/raid5.c
6663
static int handle_active_stripes(struct r5conf *conf, int group,
drivers/md/raid5.c
667
if (sh->group) {
drivers/md/raid5.c
6673
(sh = __get_priority_stripe(conf, group)) != NULL)
drivers/md/raid5.c
668
sh->group->stripes_cnt--;
drivers/md/raid5.c
669
sh->group = NULL;
drivers/md/raid5.c
6716
struct r5worker_group *group = worker->group;
drivers/md/raid5.c
6717
struct r5conf *conf = group->conf;
drivers/md/raid5.c
6719
int group_id = group - conf->worker_groups;
drivers/md/raid5.c
7279
struct r5worker_group *group;
drivers/md/raid5.c
7281
group = &(*worker_groups)[i];
drivers/md/raid5.c
7282
INIT_LIST_HEAD(&group->handle_list);
drivers/md/raid5.c
7283
INIT_LIST_HEAD(&group->loprio_list);
drivers/md/raid5.c
7284
group->conf = conf;
drivers/md/raid5.c
7285
group->workers = workers + i * cnt;
drivers/md/raid5.c
7288
struct r5worker *worker = group->workers + j;
drivers/md/raid5.c
7289
worker->group = group;
drivers/md/raid5.h
224
struct r5worker_group *group;
drivers/md/raid5.h
505
struct r5worker_group *group;
drivers/media/pci/intel/ipu-bridge.c
478
sensor->group[SWNODE_SENSOR_HID] = &nodes[SWNODE_SENSOR_HID];
drivers/media/pci/intel/ipu-bridge.c
479
sensor->group[SWNODE_SENSOR_PORT] = &nodes[SWNODE_SENSOR_PORT];
drivers/media/pci/intel/ipu-bridge.c
480
sensor->group[SWNODE_SENSOR_ENDPOINT] = &nodes[SWNODE_SENSOR_ENDPOINT];
drivers/media/pci/intel/ipu-bridge.c
481
sensor->group[SWNODE_IPU_PORT] = &nodes[SWNODE_IPU_PORT];
drivers/media/pci/intel/ipu-bridge.c
482
sensor->group[SWNODE_IPU_ENDPOINT] = &nodes[SWNODE_IPU_ENDPOINT];
drivers/media/pci/intel/ipu-bridge.c
484
sensor->group[SWNODE_VCM] = &nodes[SWNODE_VCM];
drivers/media/pci/intel/ipu-bridge.c
487
sensor->group[SWNODE_IVSC_HID] =
drivers/media/pci/intel/ipu-bridge.c
489
sensor->group[SWNODE_IVSC_SENSOR_PORT] =
drivers/media/pci/intel/ipu-bridge.c
491
sensor->group[SWNODE_IVSC_SENSOR_ENDPOINT] =
drivers/media/pci/intel/ipu-bridge.c
493
sensor->group[SWNODE_IVSC_IPU_PORT] =
drivers/media/pci/intel/ipu-bridge.c
495
sensor->group[SWNODE_IVSC_IPU_ENDPOINT] =
drivers/media/pci/intel/ipu-bridge.c
499
sensor->group[SWNODE_VCM] = &nodes[SWNODE_VCM];
drivers/media/pci/intel/ipu-bridge.c
502
sensor->group[SWNODE_IVSC_HID] = &nodes[SWNODE_VCM];
drivers/media/pci/intel/ipu-bridge.c
688
software_node_unregister_node_group(sensor->group);
drivers/media/pci/intel/ipu-bridge.c
729
ret = software_node_register_node_group(sensor->group);
drivers/media/pci/intel/ipu-bridge.c
758
software_node_unregister_node_group(sensor->group);
drivers/media/platform/nvidia/tegra-vde/iommu.c
133
iommu_detach_group(vde->domain, vde->group);
drivers/media/platform/nvidia/tegra-vde/iommu.c
140
iommu_group_put(vde->group);
drivers/media/platform/nvidia/tegra-vde/iommu.c
150
iommu_detach_group(vde->domain, vde->group);
drivers/media/platform/nvidia/tegra-vde/iommu.c
154
iommu_group_put(vde->group);
drivers/media/platform/nvidia/tegra-vde/iommu.c
69
vde->group = iommu_group_get(dev);
drivers/media/platform/nvidia/tegra-vde/iommu.c
70
if (!vde->group)
drivers/media/platform/nvidia/tegra-vde/iommu.c
95
err = iommu_attach_group(vde->domain, vde->group);
drivers/media/platform/nvidia/tegra-vde/vde.h
114
struct iommu_group *group;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
104
struct rvin_group *group =
drivers/media/platform/renesas/rcar-vin/rcar-core.c
111
rvin_group_cleanup(group);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
113
kfree(group);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
122
struct rvin_group *group;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
1265
if (&vin->v4l2_dev == vin->group->notifier.v4l2_dev) {
drivers/media/platform/renesas/rcar-vin/rcar-core.c
1266
v4l2_async_nf_unregister(&vin->group->notifier);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
1267
v4l2_async_nf_cleanup(&vin->group->notifier);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
128
group = rvin_group_data;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
129
kref_get(&group->refcount);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
131
group = kzalloc_obj(*group);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
132
if (!group) {
drivers/media/platform/renesas/rcar-vin/rcar-core.c
137
ret = rvin_group_init(group, vin, link_setup, ops);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
139
kfree(group);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
144
kref_init(&group->refcount);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
145
group->info = vin->info;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
147
rvin_group_data = group;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
152
mutex_lock(&group->lock);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
154
if (group->vin[vin->id]) {
drivers/media/platform/renesas/rcar-vin/rcar-core.c
156
mutex_unlock(&group->lock);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
157
kref_put(&group->refcount, rvin_group_release);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
161
group->vin[vin->id] = vin;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
163
vin->group = group;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
164
vin->v4l2_dev.mdev = &group->mdev;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
166
mutex_unlock(&group->lock);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
176
struct rvin_group *group = vin->group;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
178
mutex_lock(&group->lock);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
180
vin->group = NULL;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
183
if (WARN_ON(group->vin[vin->id] != vin))
drivers/media/platform/renesas/rcar-vin/rcar-core.c
186
group->vin[vin->id] = NULL;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
188
mutex_unlock(&group->lock);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
190
kref_put(&group->refcount, rvin_group_release);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
194
static int rvin_group_entity_to_remote_id(struct rvin_group *group,
drivers/media/platform/renesas/rcar-vin/rcar-core.c
202
for (i = 0; i < ARRAY_SIZE(group->remotes); i++)
drivers/media/platform/renesas/rcar-vin/rcar-core.c
203
if (group->remotes[i].subdev == sd)
drivers/media/platform/renesas/rcar-vin/rcar-core.c
215
ret = media_device_register(&vin->group->mdev);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
227
if (vin->group->vin[i] &&
drivers/media/platform/renesas/rcar-vin/rcar-core.c
228
!video_is_registered(&vin->group->vin[i]->vdev)) {
drivers/media/platform/renesas/rcar-vin/rcar-core.c
229
ret = rvin_v4l2_register(vin->group->vin[i]);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
235
return vin->group->link_setup(vin->group);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
243
struct rvin_group *group = vin->group;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
246
if (group->vin[i])
drivers/media/platform/renesas/rcar-vin/rcar-core.c
247
rvin_v4l2_unregister(group->vin[i]);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
250
mutex_lock(&vin->group->lock);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
253
if (!group->vin[i] || group->vin[i]->parallel.asc != asc)
drivers/media/platform/renesas/rcar-vin/rcar-core.c
256
group->vin[i]->parallel.subdev = NULL;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
258
vin_dbg(group->vin[i], "Unbind parallel subdev %s\n",
drivers/media/platform/renesas/rcar-vin/rcar-core.c
262
for (unsigned int i = 0; i < ARRAY_SIZE(group->remotes); i++) {
drivers/media/platform/renesas/rcar-vin/rcar-core.c
263
if (group->remotes[i].asc != asc)
drivers/media/platform/renesas/rcar-vin/rcar-core.c
266
group->remotes[i].subdev = NULL;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
271
mutex_unlock(&vin->group->lock);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
273
media_device_unregister(&vin->group->mdev);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
281
struct rvin_group *group = vin->group;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
283
guard(mutex)(&group->lock);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
286
struct rvin_dev *pvin = group->vin[i];
drivers/media/platform/renesas/rcar-vin/rcar-core.c
302
for (unsigned int i = 0; i < ARRAY_SIZE(group->remotes); i++) {
drivers/media/platform/renesas/rcar-vin/rcar-core.c
303
if (vin->group->remotes[i].asc != asc)
drivers/media/platform/renesas/rcar-vin/rcar-core.c
306
vin->group->remotes[i].subdev = subdev;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
344
asc = v4l2_async_nf_add_fwnode(&vin->group->notifier, fwnode,
drivers/media/platform/renesas/rcar-vin/rcar-core.c
351
vin->group->remotes[vep.base.id].asc = asc;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
394
asc = v4l2_async_nf_add_fwnode(&vin->group->notifier, fwnode,
drivers/media/platform/renesas/rcar-vin/rcar-core.c
413
mutex_lock(&vin->group->lock);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
417
if (vin->group->vin[i]) {
drivers/media/platform/renesas/rcar-vin/rcar-core.c
423
if (vin->group->count != count) {
drivers/media/platform/renesas/rcar-vin/rcar-core.c
424
mutex_unlock(&vin->group->lock);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
428
mutex_unlock(&vin->group->lock);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
430
v4l2_async_nf_init(&vin->group->notifier, &vin->v4l2_dev);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
441
ret = rvin_parallel_parse_of(vin->group->vin[i]);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
447
if (vin->group->remotes[id].asc)
drivers/media/platform/renesas/rcar-vin/rcar-core.c
450
ret = rvin_group_parse_of(vin->group->vin[i], port, id);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
456
if (list_empty(&vin->group->notifier.waiting_list))
drivers/media/platform/renesas/rcar-vin/rcar-core.c
459
vin->group->notifier.ops = &rvin_group_notify_ops;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
460
ret = v4l2_async_nf_register(&vin->group->notifier);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
463
v4l2_async_nf_cleanup(&vin->group->notifier);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
552
struct rvin_group *group = container_of(link->graph_obj.mdev,
drivers/media/platform/renesas/rcar-vin/rcar-core.c
573
media_device_for_each_entity(entity, &group->mdev)
drivers/media/platform/renesas/rcar-vin/rcar-core.c
581
mutex_lock(&group->lock);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
583
csi_id = rvin_group_entity_to_remote_id(group, link->source->entity);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
597
if (group->vin[i] &&
drivers/media/platform/renesas/rcar-vin/rcar-core.c
598
group->vin[i]->parallel.subdev == sd) {
drivers/media/platform/renesas/rcar-vin/rcar-core.c
599
group->vin[i]->is_csi = false;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
615
if (WARN_ON(!group->vin[master_id])) {
drivers/media/platform/renesas/rcar-vin/rcar-core.c
62
static void rvin_group_cleanup(struct rvin_group *group)
drivers/media/platform/renesas/rcar-vin/rcar-core.c
624
if (!group->vin[i])
drivers/media/platform/renesas/rcar-vin/rcar-core.c
629
&group->vin[i]->vdev.entity.pads[0]);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
64
media_device_cleanup(&group->mdev);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
65
mutex_destroy(&group->lock);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
654
ret = rvin_set_channel_routing(group->vin[master_id], chsel);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
661
mutex_unlock(&group->lock);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
670
static int rvin_csi2_create_link(struct rvin_group *group, unsigned int id,
drivers/media/platform/renesas/rcar-vin/rcar-core.c
673
struct media_entity *source = &group->remotes[route->csi].subdev->entity;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
674
struct media_entity *sink = &group->vin[id]->vdev.entity;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
68
static int rvin_group_init(struct rvin_group *group, struct rvin_dev *vin,
drivers/media/platform/renesas/rcar-vin/rcar-core.c
695
static int rvin_parallel_setup_links(struct rvin_group *group)
drivers/media/platform/renesas/rcar-vin/rcar-core.c
699
guard(mutex)(&group->lock);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
702
for (unsigned int i = 0; i < ARRAY_SIZE(group->remotes); i++) {
drivers/media/platform/renesas/rcar-vin/rcar-core.c
703
if (group->remotes[i].subdev) {
drivers/media/platform/renesas/rcar-vin/rcar-core.c
711
struct rvin_dev *vin = group->vin[i];
drivers/media/platform/renesas/rcar-vin/rcar-core.c
72
struct media_device *mdev = &group->mdev;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
732
static int rvin_csi2_setup_links(struct rvin_group *group)
drivers/media/platform/renesas/rcar-vin/rcar-core.c
738
ret = rvin_parallel_setup_links(group);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
743
mutex_lock(&group->lock);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
744
for (route = group->info->routes; route->chsel; route++) {
drivers/media/platform/renesas/rcar-vin/rcar-core.c
746
if (!group->vin[route->master])
drivers/media/platform/renesas/rcar-vin/rcar-core.c
750
if (!group->remotes[route->csi].subdev)
drivers/media/platform/renesas/rcar-vin/rcar-core.c
755
if (!group->vin[id])
drivers/media/platform/renesas/rcar-vin/rcar-core.c
758
ret = rvin_csi2_create_link(group, id, route);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
76
mutex_init(&group->lock);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
764
mutex_unlock(&group->lock);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
788
static int rvin_isp_setup_links(struct rvin_group *group)
drivers/media/platform/renesas/rcar-vin/rcar-core.c
79
group->count = 0;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
794
mutex_lock(&group->lock);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
800
struct rvin_dev *vin = group->vin[i];
drivers/media/platform/renesas/rcar-vin/rcar-core.c
806
if (!group->remotes[source_slot].subdev)
drivers/media/platform/renesas/rcar-vin/rcar-core.c
809
source = &group->remotes[source_slot].subdev->entity;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
82
group->count++;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
828
mutex_unlock(&group->lock);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
84
vin_dbg(vin, "found %u enabled VIN's in DT", group->count);
drivers/media/platform/renesas/rcar-vin/rcar-core.c
86
group->link_setup = link_setup;
drivers/media/platform/renesas/rcar-vin/rcar-core.c
880
struct rvin_dev *master = vin->group->vin[master_id];
drivers/media/platform/renesas/rcar-vin/rcar-v4l2.c
663
struct rvin_group *group;
drivers/media/platform/renesas/rcar-vin/rcar-v4l2.c
669
group = vin->group;
drivers/media/platform/renesas/rcar-vin/rcar-v4l2.c
672
vin = group->vin[i];
drivers/media/platform/renesas/rcar-vin/rcar-vin.h
187
struct rvin_group *group;
drivers/media/platform/renesas/rcar-vin/rcar-vin.h
252
int (*link_setup)(struct rvin_group *group);
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2097
unsigned int group;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2102
.group = RKISP1_EXT_PARAMS_BLOCK_GROUP_OTHERS,
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2107
.group = RKISP1_EXT_PARAMS_BLOCK_GROUP_OTHERS,
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2111
.group = RKISP1_EXT_PARAMS_BLOCK_GROUP_OTHERS,
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2115
.group = RKISP1_EXT_PARAMS_BLOCK_GROUP_OTHERS,
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2119
.group = RKISP1_EXT_PARAMS_BLOCK_GROUP_OTHERS,
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2123
.group = RKISP1_EXT_PARAMS_BLOCK_GROUP_OTHERS,
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2127
.group = RKISP1_EXT_PARAMS_BLOCK_GROUP_OTHERS,
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2131
.group = RKISP1_EXT_PARAMS_BLOCK_GROUP_OTHERS,
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2135
.group = RKISP1_EXT_PARAMS_BLOCK_GROUP_OTHERS,
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2139
.group = RKISP1_EXT_PARAMS_BLOCK_GROUP_OTHERS,
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2143
.group = RKISP1_EXT_PARAMS_BLOCK_GROUP_OTHERS,
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2147
.group = RKISP1_EXT_PARAMS_BLOCK_GROUP_OTHERS,
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2151
.group = RKISP1_EXT_PARAMS_BLOCK_GROUP_LSC,
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2155
.group = RKISP1_EXT_PARAMS_BLOCK_GROUP_OTHERS,
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2159
.group = RKISP1_EXT_PARAMS_BLOCK_GROUP_OTHERS,
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2163
.group = RKISP1_EXT_PARAMS_BLOCK_GROUP_OTHERS,
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2167
.group = RKISP1_EXT_PARAMS_BLOCK_GROUP_OTHERS,
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2171
.group = RKISP1_EXT_PARAMS_BLOCK_GROUP_OTHERS,
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2176
.group = RKISP1_EXT_PARAMS_BLOCK_GROUP_OTHERS,
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2181
.group = RKISP1_EXT_PARAMS_BLOCK_GROUP_OTHERS,
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2186
.group = RKISP1_EXT_PARAMS_BLOCK_GROUP_OTHERS,
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2246
if (!(block_handler->group & block_group_mask))
drivers/mfd/max8997-irq.c
140
max8997->irq_masks_cur[irq_data->group] |= irq_data->mask;
drivers/mfd/max8997-irq.c
149
max8997->irq_masks_cur[irq_data->group] &= ~irq_data->mask;
drivers/mfd/max8997-irq.c
258
if (irq_reg[max8997_irqs[i].group] & max8997_irqs[i].mask) {
drivers/mfd/max8997-irq.c
51
enum max8997_irq_source group;
drivers/mfd/max8997-irq.c
55
[(idx)] = { .group = (_group), .mask = (_mask) }
drivers/mfd/twl4030-power.c
168
#define TWL_RESOURCE_GROUP_RESET(group, type1, type2) \
drivers/mfd/twl4030-power.c
169
{ MSG_BROADCAST(DEV_GRP_NULL, (group), (type1), (type2), \
drivers/mfd/twl4030-power.c
171
#define TWL_RESOURCE_GROUP_SLEEP(group, type, type2) \
drivers/mfd/twl4030-power.c
172
{ MSG_BROADCAST(DEV_GRP_NULL, (group), (type), (type2), \
drivers/mfd/twl4030-power.c
174
#define TWL_RESOURCE_GROUP_ACTIVE(group, type, type2) \
drivers/mfd/twl4030-power.c
175
{ MSG_BROADCAST(DEV_GRP_NULL, (group), (type), (type2), \
drivers/mmc/core/sd_ops.c
363
int mmc_sd_switch(struct mmc_card *card, bool mode, int group,
drivers/mmc/core/sd_ops.c
372
cmd_args &= ~(0xF << (group * 4));
drivers/mmc/core/sd_ops.c
373
cmd_args |= value << (group * 4);
drivers/most/configfs.c
413
struct config_group group;
drivers/most/configfs.c
423
static struct config_item *most_common_make_item(struct config_group *group,
drivers/most/configfs.c
427
struct most_common *mc = to_most_common(group->cg_subsys);
drivers/most/configfs.c
440
if (!strcmp(group->cg_item.ci_namebuf, "most_cdev"))
drivers/most/configfs.c
442
else if (!strcmp(group->cg_item.ci_namebuf, "most_net"))
drivers/most/configfs.c
444
else if (!strcmp(group->cg_item.ci_namebuf, "most_video"))
drivers/most/configfs.c
452
struct config_group *group = to_config_group(item);
drivers/most/configfs.c
454
kfree(to_most_common(group->cg_subsys));
drivers/most/configfs.c
461
static void most_common_disconnect(struct config_group *group,
drivers/most/configfs.c
464
struct most_common *mc = to_most_common(group->cg_subsys);
drivers/most/configfs.c
514
struct config_group group;
drivers/most/configfs.c
521
return container_of(to_config_group(item), struct most_snd_grp, group);
drivers/most/configfs.c
524
static struct config_item *most_snd_grp_make_item(struct config_group *group,
drivers/most/configfs.c
568
struct most_snd_grp *group = to_most_snd_grp(item);
drivers/most/configfs.c
570
list_del(&group->list);
drivers/most/configfs.c
571
kfree(group);
drivers/most/configfs.c
595
static struct config_group *most_sound_make_group(struct config_group *group,
drivers/most/configfs.c
599
struct most_sound *ms = container_of(group->cg_subsys,
drivers/most/configfs.c
615
config_group_init_type_name(&most->group, name, &most_snd_grp_type);
drivers/most/configfs.c
617
return &most->group;
drivers/most/configfs.c
620
static void most_sound_disconnect(struct config_group *group,
drivers/most/configfs.c
623
struct most_sound *ms = container_of(group->cg_subsys,
drivers/mtd/mtdcore.c
1706
info.group = 0;
drivers/mtd/mtdcore.c
513
info->group = 0;
drivers/mtd/mtdcore.c
552
info->group < 0 || info->group >= ngroups)
drivers/mtd/nand/raw/fsmc_nand.c
717
int off, len, ret, group = 0;
drivers/mtd/nand/raw/fsmc_nand.c
737
ret = mtd_ooblayout_ecc(mtd, group++, &oobregion);
drivers/mtd/nand/raw/nand_base.c
59
info->group = 0;
drivers/mtd/nand/raw/nand_base.c
62
info->group = 1;
drivers/mtd/nand/raw/nand_base.c
76
if (!info->group && !info->pair)
drivers/mtd/nand/raw/nand_base.c
79
if (info->pair == lastpair && info->group)
drivers/mtd/nand/raw/nand_base.c
82
if (!info->group)
drivers/net/amt.c
1179
union amt_addr group = {0,};
drivers/net/amt.c
1214
group.ip4 = iph->daddr;
drivers/net/amt.c
1238
group.ip6 = ip6h->daddr;
drivers/net/amt.c
1277
hash = amt_group_hash(tunnel, &group);
drivers/net/amt.c
1948
union amt_addr group, host;
drivers/net/amt.c
1950
memset(&group, 0, sizeof(union amt_addr));
drivers/net/amt.c
1951
group.ip4 = ih->group;
drivers/net/amt.c
1955
gnode = amt_lookup_group(tunnel, &group, &host, false);
drivers/net/amt.c
1957
gnode = amt_add_group(amt, tunnel, &group, &host, false);
drivers/net/amt.c
1985
union amt_addr group, host;
drivers/net/amt.c
1987
memset(&group, 0, sizeof(union amt_addr));
drivers/net/amt.c
1988
group.ip4 = ih->group;
drivers/net/amt.c
1992
gnode = amt_lookup_group(tunnel, &group, &host, false);
drivers/net/amt.c
200
static u32 amt_group_hash(struct amt_tunnel_list *tunnel, union amt_addr *group)
drivers/net/amt.c
2005
union amt_addr group, host;
drivers/net/amt.c
202
u32 hash = jhash(group, sizeof(*group), tunnel->amt->hash_seed);
drivers/net/amt.c
2022
memset(&group, 0, sizeof(union amt_addr));
drivers/net/amt.c
2023
group.ip4 = grec->grec_mca;
drivers/net/amt.c
2026
gnode = amt_lookup_group(tunnel, &group, &host, false);
drivers/net/amt.c
2028
gnode = amt_add_group(amt, tunnel, &group, &host,
drivers/net/amt.c
208
union amt_addr *group,
drivers/net/amt.c
2109
union amt_addr group, host;
drivers/net/amt.c
2111
memcpy(&group.ip6, &mld->mld_mca, sizeof(struct in6_addr));
drivers/net/amt.c
2114
gnode = amt_lookup_group(tunnel, &group, &host, true);
drivers/net/amt.c
2116
gnode = amt_add_group(amt, tunnel, &group, &host, true);
drivers/net/amt.c
212
u32 hash = amt_group_hash(tunnel, group);
drivers/net/amt.c
2146
union amt_addr group, host;
drivers/net/amt.c
2148
memcpy(&group.ip6, &mld->mld_mca, sizeof(struct in6_addr));
drivers/net/amt.c
2152
gnode = amt_lookup_group(tunnel, &group, &host, true);
drivers/net/amt.c
216
if (amt_addr_equal(&gnode->group_addr, group) &&
drivers/net/amt.c
2167
union amt_addr group, host;
drivers/net/amt.c
2184
memset(&group, 0, sizeof(union amt_addr));
drivers/net/amt.c
2185
group.ip6 = grec->grec_mca;
drivers/net/amt.c
2188
gnode = amt_lookup_group(tunnel, &group, &host, true);
drivers/net/amt.c
2190
gnode = amt_add_group(amt, tunnel, &group, &host,
drivers/net/amt.c
467
union amt_addr *group,
drivers/net/amt.c
485
gnode->group_addr = *group;
drivers/net/amt.c
495
hash = amt_group_hash(tunnel, group);
drivers/net/amt.c
563
ihv3->group = 0;
drivers/net/ethernet/airoha/airoha_eth.c
2263
u32 idx = QDMA_METER_IDX(queue_id), group = QDMA_METER_GROUP(queue_id);
drivers/net/ethernet/airoha/airoha_eth.c
2265
FIELD_PREP(RATE_LIMIT_METER_GROUP_MASK, group) |
drivers/net/ethernet/airoha/airoha_eth.c
2286
u32 idx = QDMA_METER_IDX(queue_id), group = QDMA_METER_GROUP(queue_id);
drivers/net/ethernet/airoha/airoha_eth.c
2289
FIELD_PREP(RATE_LIMIT_METER_GROUP_MASK, group) |
drivers/net/ethernet/airoha/airoha_eth.c
2383
u32 idx = QDMA_METER_IDX(channel), group = QDMA_METER_GROUP(channel);
drivers/net/ethernet/airoha/airoha_eth.c
2385
FIELD_PREP(TRTCM_METER_GROUP_MASK, group) |
drivers/net/ethernet/airoha/airoha_eth.c
2407
u32 idx = QDMA_METER_IDX(channel), group = QDMA_METER_GROUP(channel);
drivers/net/ethernet/airoha/airoha_eth.c
2410
FIELD_PREP(TRTCM_METER_GROUP_MASK, group) |
drivers/net/ethernet/amazon/ena/ena_admin_defs.h
1153
u16 group;
drivers/net/ethernet/amazon/ena/ena_com.c
2279
u16 group)
drivers/net/ethernet/amazon/ena/ena_com.c
2283
if ((group < ENA_MAX_HANDLERS) && aenq_handlers->handlers[group])
drivers/net/ethernet/amazon/ena/ena_com.c
2284
return aenq_handlers->handlers[group];
drivers/net/ethernet/amazon/ena/ena_com.c
2319
aenq_common->group, aenq_common->syndrome, timestamp);
drivers/net/ethernet/amazon/ena/ena_com.c
2323
aenq_common->group);
drivers/net/ethernet/amazon/ena/ena_netdev.c
4343
WARN(aenq_e->aenq_common_desc.group != ENA_ADMIN_NOTIFICATION,
drivers/net/ethernet/amazon/ena/ena_netdev.c
4345
aenq_e->aenq_common_desc.group,
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth-devlink.c
138
const struct devlink_trap_group *group,
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth-devlink.c
149
if (group->id != DEVLINK_TRAP_GROUP_GENERIC_ID_PARSER_ERROR_DROPS)
drivers/net/ethernet/freescale/enetc/ntmp.c
421
u8 *group;
drivers/net/ethernet/freescale/enetc/ntmp.c
444
group = (u8 *)req;
drivers/net/ethernet/freescale/enetc/ntmp.c
445
group += NTMP_ENTRY_ID_SIZE + RSST_STSE_DATA_SIZE(count);
drivers/net/ethernet/freescale/enetc/ntmp.c
447
table[i] = group[i];
drivers/net/ethernet/hisilicon/hip04_eth.c
217
unsigned int group;
drivers/net/ethernet/hisilicon/hip04_eth.c
320
val = BIT(priv->group);
drivers/net/ethernet/hisilicon/hip04_eth.c
323
val = priv->group << PPE_CFG_QOS_VMID_GRP_SHIFT;
drivers/net/ethernet/hisilicon/hip04_eth.c
935
priv->group = arg.args[2];
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4648
static void hns3_add_ring_to_group(struct hns3_enet_ring_group *group,
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4651
ring->next = group->ring;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4652
group->ring = ring;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4654
group->count++;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4677
struct hns3_enet_ring_group *group = container_of(dim,
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4679
struct hns3_enet_tqp_vector *tqp_vector = group->ring->tqp_vector;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4683
hns3_set_vector_coalesce_rx_gl(group->ring->tqp_vector, cur_moder.usec);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4697
struct hns3_enet_ring_group *group = container_of(dim,
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4699
struct hns3_enet_tqp_vector *tqp_vector = group->ring->tqp_vector;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4864
static void hns3_clear_ring_group(struct hns3_enet_ring_group *group)
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4866
group->ring = NULL;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4867
group->count = 0;
drivers/net/ethernet/intel/igb/igb.h
475
struct attribute_group group;
drivers/net/ethernet/intel/igb/igb_hwmon.c
210
igb_hwmon->groups[0] = &igb_hwmon->group;
drivers/net/ethernet/intel/igb/igb_hwmon.c
211
igb_hwmon->group.attrs = igb_hwmon->attrs;
drivers/net/ethernet/intel/ixgbe/ixgbe.h
539
struct attribute_group group;
drivers/net/ethernet/intel/ixgbe/ixgbe_sysfs.c
193
ixgbe_hwmon->groups[0] = &ixgbe_hwmon->group;
drivers/net/ethernet/intel/ixgbe/ixgbe_sysfs.c
194
ixgbe_hwmon->group.attrs = ixgbe_hwmon->attrs;
drivers/net/ethernet/marvell/octeontx2/af/mbox.h
1240
u8 group; /* RSS context or group */
drivers/net/ethernet/marvell/octeontx2/af/rvu.h
1074
int group, int alg_idx, int mcam_index);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
4365
rvu_npc_update_flowkey_alg_idx(rvu, pcifunc, nixlf, req->group,
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
1003
int group, int alg_idx, int mcam_index)
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
1016
if (group != DEFAULT_RSS_CONTEXT_GROUP)
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
1039
action.index = group;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
304
req->group = DEFAULT_RSS_CONTEXT_GROUP;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
2234
struct mlx5_flow_group *group;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
2249
group = mlx5_create_flow_group(from, flow_group_in);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
2250
if (IS_ERR(group)) {
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
2251
err = PTR_ERR(group);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
2262
*miss_group = group;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
2268
mlx5_destroy_flow_group(group);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec.h
221
struct mlx5_flow_group *group;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1108
mlx5_destroy_flow_group(rx->sa.group);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1263
mlx5_destroy_flow_group(tx->pol.group);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1269
mlx5_destroy_flow_group(tx->sa.group);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1373
mlx5_destroy_flow_group(tx->pol.group);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1379
mlx5_destroy_flow_group(tx->sa.group);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
566
miss->group = mlx5_create_flow_group(ft, flow_group_in);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
567
if (IS_ERR(miss->group)) {
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
568
err = PTR_ERR(miss->group);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
577
mlx5_destroy_flow_group(miss->group);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
703
mlx5_destroy_flow_group(rx->pol.group);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
719
mlx5_destroy_flow_group(rx->sa_sel.group);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
731
mlx5_destroy_flow_group(rx->sa.group);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
998
rx->sa_sel.group = fg;
drivers/net/ethernet/mellanox/mlx5/core/fs_cmd.c
101
struct mlx5_flow_group *group,
drivers/net/ethernet/mellanox/mlx5/core/fs_cmd.c
777
struct mlx5_flow_group *group,
drivers/net/ethernet/mellanox/mlx5/core/fs_cmd.c
781
unsigned int group_id = group->id;
drivers/net/ethernet/mellanox/mlx5/core/fs_cmd.c
93
struct mlx5_flow_group *group,
drivers/net/ethernet/mellanox/mlx5/core/fw.c
129
enum mlx5_mcam_reg_groups group)
drivers/net/ethernet/mellanox/mlx5/core/fw.c
131
return mlx5_query_mcam_reg(dev, dev->caps.mcam[group],
drivers/net/ethernet/mellanox/mlx5/core/fw.c
132
MLX5_MCAM_FEATURE_ENHANCED_FEATURES, group);
drivers/net/ethernet/mellanox/mlx5/core/lib/ipsec_fs_roce.c
12
struct mlx5_flow_group *group;
drivers/net/ethernet/mellanox/mlx5/core/lib/ipsec_fs_roce.c
702
mlx5_destroy_flow_group(rx_roce->roce_miss.group);
drivers/net/ethernet/mellanox/mlx5/core/lib/ipsec_fs_roce.c
776
roce->roce_miss.group = g;
drivers/net/ethernet/mellanox/mlx5/core/lib/ipsec_fs_roce.c
815
mlx5_destroy_flow_group(roce->roce_miss.group);
drivers/net/ethernet/mellanox/mlx5/core/port.c
716
MLX5_SET(qetc_reg, in, tc_configuration[i].group, tc_group[i]);
drivers/net/ethernet/mellanox/mlx5/core/port.c
737
group);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/fs_hws.c
1069
struct mlx5_flow_group *group,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/fs_hws.c
1078
return mlx5_fs_cmd_get_fw_cmds()->create_fte(ns, ft, group, fte);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/fs_hws.c
1080
err = mlx5_fs_fte_get_hws_actions(ns, ft, group, fte, &ractions);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/fs_hws.c
1087
rule = mlx5hws_bwc_rule_create(group->fs_hws_matcher.matcher, ¶ms,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/fs_hws.c
1128
struct mlx5_flow_group *group,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/fs_hws.c
1141
return mlx5_fs_cmd_get_fw_cmds()->update_fte(ns, ft, group,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/fs_hws.c
1150
ret = mlx5_fs_fte_get_hws_actions(ns, ft, group, fte, &ractions);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/fs_hws.c
741
struct mlx5_flow_group *group,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/fs_dr.c
254
struct mlx5_flow_group *group,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/fs_dr.c
276
return mlx5_fs_cmd_get_fw_cmds()->create_fte(ns, ft, group, fte);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/fs_dr.c
597
rule = mlx5dr_rule_create(group->fs_dr_matcher.dr_matcher,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/fs_dr.c
758
struct mlx5_flow_group *group,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/fs_dr.c
766
return mlx5_fs_cmd_get_fw_cmds()->update_fte(ns, ft, group, modify_mask, fte);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/fs_dr.c
773
ret = mlx5_cmd_dr_create_fte(ns, ft, group, fte);
drivers/net/ethernet/mellanox/mlxsw/core.c
1657
const struct devlink_trap_group *group)
drivers/net/ethernet/mellanox/mlxsw/core.c
1664
return mlxsw_driver->trap_group_init(mlxsw_core, group);
drivers/net/ethernet/mellanox/mlxsw/core.c
1669
const struct devlink_trap_group *group,
drivers/net/ethernet/mellanox/mlxsw/core.c
1678
return mlxsw_driver->trap_group_set(mlxsw_core, group, policer, extack);
drivers/net/ethernet/mellanox/mlxsw/core.h
419
const struct devlink_trap_group *group);
drivers/net/ethernet/mellanox/mlxsw/core.h
421
const struct devlink_trap_group *group,
drivers/net/ethernet/mellanox/mlxsw/core_hwmon.c
48
struct attribute_group group;
drivers/net/ethernet/mellanox/mlxsw/core_hwmon.c
820
linecard->groups[0] = &linecard->group;
drivers/net/ethernet/mellanox/mlxsw/core_hwmon.c
821
linecard->group.attrs = linecard->attrs;
drivers/net/ethernet/mellanox/mlxsw/core_hwmon.c
899
mlxsw_hwmon->line_cards[0].groups[0] = &mlxsw_hwmon->line_cards[0].group;
drivers/net/ethernet/mellanox/mlxsw/core_hwmon.c
900
mlxsw_hwmon->line_cards[0].group.attrs = mlxsw_hwmon->line_cards[0].attrs;
drivers/net/ethernet/mellanox/mlxsw/reg.h
6577
static inline void mlxsw_reg_htgt_pack(char *payload, u8 group, u8 policer_id,
drivers/net/ethernet/mellanox/mlxsw/reg.h
6592
mlxsw_reg_htgt_trap_group_set(payload, group);
drivers/net/ethernet/mellanox/mlxsw/reg.h
6597
mlxsw_reg_htgt_local_path_rdq_set(payload, group);
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
1409
const struct devlink_trap_group *group);
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
1411
const struct devlink_trap_group *group,
drivers/net/ethernet/mellanox/mlxsw/spectrum1_mr_tcam.c
40
ntohl(key->group.addr4),
drivers/net/ethernet/mellanox/mlxsw/spectrum1_mr_tcam.c
50
key->group.addr6,
drivers/net/ethernet/mellanox/mlxsw/spectrum2_mr_tcam.c
151
(char *) &key->group.addr4,
drivers/net/ethernet/mellanox/mlxsw/spectrum2_mr_tcam.c
179
&key->group.addr6.s6_addr[0x0],
drivers/net/ethernet/mellanox/mlxsw/spectrum2_mr_tcam.c
182
&key->group.addr6.s6_addr[0x4],
drivers/net/ethernet/mellanox/mlxsw/spectrum2_mr_tcam.c
185
&key->group.addr6.s6_addr[0x8],
drivers/net/ethernet/mellanox/mlxsw/spectrum2_mr_tcam.c
188
&key->group.addr6.s6_addr[0xc],
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
115
struct mlxsw_sp_acl_tcam_group group;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
1421
vregion->region2->group,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
1722
return mlxsw_sp_acl_tcam_group_bind(mlxsw_sp, &ruleset->vgroup.group,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
1734
mlxsw_sp_acl_tcam_group_unbind(mlxsw_sp, &ruleset->vgroup.group,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
1743
return mlxsw_sp_acl_tcam_group_id(&ruleset->vgroup.group);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
1878
return mlxsw_sp_acl_tcam_group_id(&ruleset->vgroup.group);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
208
struct mlxsw_sp_acl_tcam_group *group)
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
214
mlxsw_reg_pagt_pack(pagt_pl, group->id);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
215
list_for_each_entry(region, &group->region_list, list) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
219
if (region->list.next != &group->region_list &&
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
231
struct mlxsw_sp_acl_tcam_group *group)
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
235
group->tcam = tcam;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
236
INIT_LIST_HEAD(&group->region_list);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
238
err = mlxsw_sp_acl_tcam_group_id_get(tcam, &group->id);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
242
mutex_init(&group->lock);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
247
static void mlxsw_sp_acl_tcam_group_del(struct mlxsw_sp_acl_tcam_group *group)
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
249
struct mlxsw_sp_acl_tcam *tcam = group->tcam;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
251
mutex_destroy(&group->lock);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
252
mlxsw_sp_acl_tcam_group_id_put(tcam, group->id);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
253
WARN_ON(!list_empty(&group->region_list));
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
282
err = mlxsw_sp_acl_tcam_group_add(tcam, &vgroup->group);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
294
mlxsw_sp_acl_tcam_group_del(&vgroup->group);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
302
mlxsw_sp_acl_tcam_group_del(&vgroup->group);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
308
struct mlxsw_sp_acl_tcam_group *group,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
317
group->id);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
323
struct mlxsw_sp_acl_tcam_group *group,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
332
group->id);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
337
mlxsw_sp_acl_tcam_group_id(struct mlxsw_sp_acl_tcam_group *group)
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
339
return group->id;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
384
struct mlxsw_sp_acl_tcam_group *group,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
393
mutex_lock(&group->lock);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
394
if (group->region_count == group->tcam->max_group_size) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
406
list_for_each(pos, &group->region_list) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
414
region->group = group;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
416
err = mlxsw_sp_acl_tcam_group_update(mlxsw_sp, group);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
420
group->region_count++;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
421
mutex_unlock(&group->lock);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
427
mutex_unlock(&group->lock);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
435
struct mlxsw_sp_acl_tcam_group *group = region->group;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
437
mutex_lock(&group->lock);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
439
group->region_count--;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
440
mlxsw_sp_acl_tcam_group_update(mlxsw_sp, group);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
441
mutex_unlock(&group->lock);
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
462
err = mlxsw_sp_acl_tcam_group_region_attach(mlxsw_sp, &vgroup->group,
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.c
777
struct mlxsw_sp_acl_tcam *tcam = vgroup->group.tcam;
drivers/net/ethernet/mellanox/mlxsw/spectrum_acl_tcam.h
79
struct mlxsw_sp_acl_tcam_group *group;
drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c
806
key->group.addr4 = mfc->mfc_mcastgrp;
drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.c
850
key->group.addr6 = mfc->mf6c_mcastgrp;
drivers/net/ethernet/mellanox/mlxsw/spectrum_mr.h
21
union mlxsw_sp_l3addr group;
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
1211
if (trap->group_items_arr[i].group.id == id)
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
1409
err = devl_trap_groups_register(devlink, &group_item->group, 1);
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
1419
devl_trap_groups_unregister(devlink, &group_item->group, 1);
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
1435
devl_trap_groups_unregister(devlink, &group_item->group, 1);
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
1662
const struct devlink_trap_group *group,
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
1670
group_item = mlxsw_sp_trap_group_item_lookup(mlxsw_sp, group->id);
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
1674
if (group_item->fixed_policer && policer_id != group->init_policer_id) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
1695
const struct devlink_trap_group *group)
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
1697
return __mlxsw_sp_trap_group_init(mlxsw_core, group,
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
1698
group->init_policer_id, NULL);
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
1702
const struct devlink_trap_group *group,
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
1708
return __mlxsw_sp_trap_group_init(mlxsw_core, group, policer_id,
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
1866
pol_id = gr_item->group.init_policer_id;
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
1884
.group = DEVLINK_TRAP_GROUP_GENERIC(ACL_SAMPLE, 0),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
1931
.group = DEVLINK_TRAP_GROUP_GENERIC(BUFFER_DROPS, 20),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
1937
.group = DEVLINK_TRAP_GROUP_GENERIC(ACL_SAMPLE, 0),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
21
struct devlink_trap_group group;
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
520
.group = DEVLINK_TRAP_GROUP_GENERIC(L2_DROPS, 1),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
525
.group = DEVLINK_TRAP_GROUP_GENERIC(L3_DROPS, 1),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
530
.group = DEVLINK_TRAP_GROUP_GENERIC(L3_EXCEPTIONS, 1),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
535
.group = DEVLINK_TRAP_GROUP_GENERIC(TUNNEL_DROPS, 1),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
540
.group = DEVLINK_TRAP_GROUP_GENERIC(ACL_DROPS, 1),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
545
.group = DEVLINK_TRAP_GROUP_GENERIC(STP, 2),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
550
.group = DEVLINK_TRAP_GROUP_GENERIC(LACP, 3),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
555
.group = DEVLINK_TRAP_GROUP_GENERIC(LLDP, 4),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
560
.group = DEVLINK_TRAP_GROUP_GENERIC(MC_SNOOPING, 5),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
565
.group = DEVLINK_TRAP_GROUP_GENERIC(DHCP, 6),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
570
.group = DEVLINK_TRAP_GROUP_GENERIC(NEIGH_DISCOVERY, 7),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
575
.group = DEVLINK_TRAP_GROUP_GENERIC(BFD, 8),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
580
.group = DEVLINK_TRAP_GROUP_GENERIC(OSPF, 9),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
585
.group = DEVLINK_TRAP_GROUP_GENERIC(BGP, 10),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
590
.group = DEVLINK_TRAP_GROUP_GENERIC(VRRP, 11),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
595
.group = DEVLINK_TRAP_GROUP_GENERIC(PIM, 12),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
600
.group = DEVLINK_TRAP_GROUP_GENERIC(UC_LB, 13),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
605
.group = DEVLINK_TRAP_GROUP_GENERIC(LOCAL_DELIVERY, 14),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
610
.group = DEVLINK_TRAP_GROUP_GENERIC(EXTERNAL_DELIVERY, 19),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
615
.group = DEVLINK_TRAP_GROUP_GENERIC(IPV6, 15),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
620
.group = DEVLINK_TRAP_GROUP_GENERIC(PTP_EVENT, 16),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
625
.group = DEVLINK_TRAP_GROUP_GENERIC(PTP_GENERAL, 17),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
630
.group = DEVLINK_TRAP_GROUP_GENERIC(ACL_TRAP, 18),
drivers/net/ethernet/mellanox/mlxsw/spectrum_trap.c
635
.group = DEVLINK_TRAP_GROUP_GENERIC(EAPOL, 21),
drivers/net/ethernet/microchip/sparx5/sparx5_main.h
621
int sparx5_sdlb_group_get_by_index(struct sparx5 *sparx5, u32 idx, u32 *group);
drivers/net/ethernet/microchip/sparx5/sparx5_main.h
623
int sparx5_sdlb_group_add(struct sparx5 *sparx5, u32 group, u32 idx);
drivers/net/ethernet/microchip/sparx5/sparx5_main.h
624
int sparx5_sdlb_group_del(struct sparx5 *sparx5, u32 group, u32 idx);
drivers/net/ethernet/microchip/sparx5/sparx5_main.h
640
u32 group;
drivers/net/ethernet/microchip/sparx5/sparx5_police.c
18
g = ops->get_sdlb_group(pol->group);
drivers/net/ethernet/microchip/sparx5/sparx5_psfp.c
194
int (*sparx5_sdlb_group_action)(struct sparx5 *sparx5, u32 group,
drivers/net/ethernet/microchip/sparx5/sparx5_psfp.c
204
return sparx5_sdlb_group_action(sparx5, fm->pol.group, fm->pol.idx);
drivers/net/ethernet/microchip/sparx5/sparx5_psfp.c
288
fm->pol.group = ret;
drivers/net/ethernet/microchip/sparx5/sparx5_psfp.c
306
ret = sparx5_sdlb_group_get_by_index(sparx5, id, &fm.pol.group);
drivers/net/ethernet/microchip/sparx5/sparx5_psfp.c
323
const struct sparx5_sdlb_group *group;
drivers/net/ethernet/microchip/sparx5/sparx5_psfp.c
327
group = ops->get_sdlb_group(i);
drivers/net/ethernet/microchip/sparx5/sparx5_psfp.c
328
sparx5_sdlb_group_init(sparx5, group->max_rate,
drivers/net/ethernet/microchip/sparx5/sparx5_psfp.c
329
group->min_burst, group->frame_size, i);
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
103
value = spx5_rd(sparx5, HSCH_HSCH_LEAK_CFG(layer, group));
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
107
static u32 sparx5_lg_get_next(struct sparx5 *sparx5, u32 layer, u32 group,
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
117
static u32 sparx5_lg_get_last(struct sparx5 *sparx5, u32 layer, u32 group)
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
121
itr = sparx5_lg_get_first(sparx5, layer, group);
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
124
next = sparx5_lg_get_next(sparx5, layer, group, itr);
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
132
static bool sparx5_lg_is_last(struct sparx5 *sparx5, u32 layer, u32 group,
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
135
return idx == sparx5_lg_get_next(sparx5, layer, group, idx);
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
138
static bool sparx5_lg_is_first(struct sparx5 *sparx5, u32 layer, u32 group,
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
141
return idx == sparx5_lg_get_first(sparx5, layer, group);
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
144
static bool sparx5_lg_is_empty(struct sparx5 *sparx5, u32 layer, u32 group)
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
146
return sparx5_lg_get_leak_time(sparx5, layer, group) == 0;
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
149
static bool sparx5_lg_is_singular(struct sparx5 *sparx5, u32 layer, u32 group)
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
151
if (sparx5_lg_is_empty(sparx5, layer, group))
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
154
return sparx5_lg_get_first(sparx5, layer, group) ==
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
155
sparx5_lg_get_last(sparx5, layer, group);
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
158
static void sparx5_lg_enable(struct sparx5 *sparx5, u32 layer, u32 group,
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
161
sparx5_lg_set_leak_time(sparx5, layer, group, leak_time);
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
164
static void sparx5_lg_disable(struct sparx5 *sparx5, u32 layer, u32 group)
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
166
sparx5_lg_set_leak_time(sparx5, layer, group, 0);
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
170
u32 idx, u32 *group)
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
185
*group = i;
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
198
static int sparx5_lg_get_group_by_rate(u32 layer, u32 rate, u32 *group)
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
207
*group = i;
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
215
static int sparx5_lg_get_adjacent(struct sparx5 *sparx5, u32 layer, u32 group,
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
220
*first = sparx5_lg_get_first(sparx5, layer, group);
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
226
*next = sparx5_lg_get_next(sparx5, layer, group, itr);
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
241
static int sparx5_lg_conf_set(struct sparx5 *sparx5, u32 layer, u32 group,
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
244
u32 leak_time = layers[layer].leak_groups[group].leak_time;
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
247
sparx5_lg_disable(sparx5, layer, group);
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
263
HSCH_HSCH_LEAK_CFG(layer, group));
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
266
sparx5_lg_enable(sparx5, layer, group, leak_time);
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
271
static int sparx5_lg_del(struct sparx5 *sparx5, u32 layer, u32 group, u32 idx)
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
277
WARN_ON(sparx5_lg_get_adjacent(sparx5, layer, group, idx, &prev, &next,
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
280
if (sparx5_lg_is_singular(sparx5, layer, group)) {
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
282
} else if (sparx5_lg_is_last(sparx5, layer, group, idx)) {
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
286
} else if (sparx5_lg_is_first(sparx5, layer, group, idx)) {
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
295
return sparx5_lg_conf_set(sparx5, layer, group, first, idx, next,
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
332
u32 idx, u32 group)
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
356
sparx5_lg_action(sparx5, layer, group, idx);
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
506
u32 group;
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
509
if (sparx5_lg_get_group_by_rate(layer, sh.rate, &group) < 0) {
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
515
lg = &layers[layer].leak_groups[group];
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
529
return sparx5_shaper_conf_set(port, &sh, layer, idx, group);
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
535
u32 group;
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
537
sparx5_lg_get_group_by_index(port->sparx5, layer, idx, &group);
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
539
return sparx5_shaper_conf_set(port, &sh, layer, idx, group);
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
84
static u32 sparx5_lg_get_leak_time(struct sparx5 *sparx5, u32 layer, u32 group)
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
88
value = spx5_rd(sparx5, HSCH_HSCH_TIMER_CFG(layer, group));
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
92
static void sparx5_lg_set_leak_time(struct sparx5 *sparx5, u32 layer, u32 group,
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
96
HSCH_HSCH_TIMER_CFG(layer, group));
drivers/net/ethernet/microchip/sparx5/sparx5_qos.c
99
static u32 sparx5_lg_get_first(struct sparx5 *sparx5, u32 layer, u32 group)
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
102
return lb == sparx5_sdlb_group_get_next(sparx5, group, lb);
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
105
static bool sparx5_sdlb_group_is_empty(struct sparx5 *sparx5, u32 group)
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
109
val = spx5_rd(sparx5, ANA_AC_SDLB_PUP_CTRL(group));
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
114
static u32 sparx5_sdlb_group_get_last(struct sparx5 *sparx5, u32 group)
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
118
itr = sparx5_sdlb_group_get_first(sparx5, group);
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
121
next = sparx5_sdlb_group_get_next(sparx5, group, itr);
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
129
static bool sparx5_sdlb_group_is_singular(struct sparx5 *sparx5, u32 group)
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
131
if (sparx5_sdlb_group_is_empty(sparx5, group))
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
134
return sparx5_sdlb_group_get_first(sparx5, group) ==
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
135
sparx5_sdlb_group_get_last(sparx5, group);
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
138
static int sparx5_sdlb_group_get_adjacent(struct sparx5 *sparx5, u32 group,
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
144
*first = sparx5_sdlb_group_get_first(sparx5, group);
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
150
*next = sparx5_sdlb_group_get_next(sparx5, group, itr);
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
163
static int sparx5_sdlb_group_get_count(struct sparx5 *sparx5, u32 group)
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
168
itr = sparx5_sdlb_group_get_first(sparx5, group);
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
171
next = sparx5_sdlb_group_get_next(sparx5, group, itr);
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
183
const struct sparx5_sdlb_group *group;
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
190
group = ops->get_sdlb_group(i);
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
198
if (count > ((group->pup_interval / 4) - 1))
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
201
if (rate_bps < group->max_rate)
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
208
int sparx5_sdlb_group_get_by_index(struct sparx5 *sparx5, u32 idx, u32 *group)
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
223
*group = i;
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
236
static int sparx5_sdlb_group_link(struct sparx5 *sparx5, u32 group, u32 idx,
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
240
sparx5_sdlb_group_disable(sparx5, group);
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
247
ANA_AC_SDLB_XLB_NEXT_LBGRP_SET(group),
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
252
ANA_AC_SDLB_XLB_START(group));
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
255
sparx5_sdlb_group_enable(sparx5, group);
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
260
int sparx5_sdlb_group_add(struct sparx5 *sparx5, u32 group, u32 idx)
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
267
if (sparx5_sdlb_group_is_empty(sparx5, group))
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
270
next = sparx5_sdlb_group_get_first(sparx5, group);
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
272
return sparx5_sdlb_group_link(sparx5, group, idx, first, next, false);
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
275
int sparx5_sdlb_group_del(struct sparx5 *sparx5, u32 group, u32 idx)
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
280
if (sparx5_sdlb_group_get_adjacent(sparx5, group, idx, &prev, &next,
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
283
__LINE__, idx, group);
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
287
if (sparx5_sdlb_group_is_singular(sparx5, group)) {
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
289
} else if (sparx5_sdlb_group_is_last(sparx5, group, idx)) {
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
293
} else if (sparx5_sdlb_group_is_first(sparx5, group, idx)) {
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
302
return sparx5_sdlb_group_link(sparx5, group, idx, first, next, empty);
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
310
struct sparx5_sdlb_group *group;
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
313
group = ops->get_sdlb_group(idx);
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
325
group->pup_interval =
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
328
group->frame_size = frame_size;
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
330
spx5_wr(ANA_AC_SDLB_PUP_INTERVAL_PUP_INTERVAL_SET(group->pup_interval),
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
60
static void sparx5_sdlb_group_disable(struct sparx5 *sparx5, u32 group)
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
64
ANA_AC_SDLB_PUP_CTRL(group));
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
67
static void sparx5_sdlb_group_enable(struct sparx5 *sparx5, u32 group)
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
71
ANA_AC_SDLB_PUP_CTRL(group));
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
74
static u32 sparx5_sdlb_group_get_first(struct sparx5 *sparx5, u32 group)
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
78
val = spx5_rd(sparx5, ANA_AC_SDLB_XLB_START(group));
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
83
static u32 sparx5_sdlb_group_get_next(struct sparx5 *sparx5, u32 group,
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
93
static bool sparx5_sdlb_group_is_first(struct sparx5 *sparx5, u32 group,
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
96
return lb == sparx5_sdlb_group_get_first(sparx5, group);
drivers/net/ethernet/microchip/sparx5/sparx5_sdlb.c
99
static bool sparx5_sdlb_group_is_last(struct sparx5 *sparx5, u32 group,
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
108
struct nfp_fl_lag_group *group;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
122
group = kmalloc_obj(*group);
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
123
if (!group) {
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
128
group->group_id = id;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
129
group->master_ndev = master;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
130
group->dirty = true;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
131
group->offloaded = false;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
132
group->to_remove = false;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
133
group->to_destroy = false;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
134
group->slave_cnt = 0;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
135
group->group_inst = ++lag->global_inst;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
136
list_add_tail(&group->list, &lag->group_list);
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
138
return group;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
164
struct nfp_fl_lag_group *group = NULL;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
168
group = nfp_fl_lag_find_group_for_master_with_lag(&priv->nfp_lag,
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
170
if (!group) {
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
176
*group_id = cpu_to_be16(group->group_id);
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
185
*group_inst = group->group_inst;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
218
struct nfp_fl_lag_group *group = NULL;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
222
group = nfp_fl_lag_find_group_for_master_with_lag(&priv->nfp_lag,
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
224
if (group)
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
225
group_id = group->group_id;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
232
nfp_fl_lag_config_group(struct nfp_fl_lag *lag, struct nfp_fl_lag_group *group,
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
275
cmsg_payload->group_id = cpu_to_be32(group->group_id);
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
276
cmsg_payload->group_inst = cpu_to_be32(group->group_inst);
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
503
struct nfp_fl_lag_group *group)
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
505
group->to_remove = true;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
514
struct nfp_fl_lag_group *group;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
523
group = nfp_fl_lag_find_group_for_master_with_lag(lag, master);
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
524
if (!group) {
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
531
group->to_remove = true;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
532
group->to_destroy = true;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
544
struct nfp_fl_lag_group *group;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
590
group = nfp_fl_lag_find_group_for_master_with_lag(lag, upper);
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
594
if (group && group->offloaded)
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
595
nfp_fl_lag_schedule_group_remove(lag, group);
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
601
if (!group) {
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
602
group = nfp_fl_lag_group_create(lag, upper);
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
603
if (IS_ERR(group)) {
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
605
return PTR_ERR(group);
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
609
group->dirty = true;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
610
group->slave_cnt = slave_count;
drivers/net/ethernet/netronome/nfp/flower/lag_conf.c
613
group->to_remove = false;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
1052
int group;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
1067
for (i = 0; i < ARRAY_SIZE(nfp->expl.group); i++) {
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
1068
if (!nfp->expl.group[i].bitsize)
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
1071
for (j = 0; j < ARRAY_SIZE(nfp->expl.group[i].free); j++) {
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
1074
if (!nfp->expl.group[i].free[j])
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
1078
priv->bar.group = i;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
1080
priv->bitsize = nfp->expl.group[i].bitsize - 2;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
1082
data_offset = (priv->bar.group << 9) +
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
1085
priv->addr = nfp->expl.group[i].addr +
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
1087
nfp->expl.group[i].free[j] = false;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
1104
nfp->expl.group[priv->bar.group].free[priv->bar.area] = true;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
1144
((priv->bar.group * 4 + priv->bar.area) << 1);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
1153
(priv->bar.group << 9) + (priv->bar.area << 7);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
1174
NFP_PCIE_BAR_EXPLICIT_BAR0(priv->bar.group,
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
1177
NFP_PCIE_BAR_EXPLICIT_BAR1(priv->bar.group,
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
1180
NFP_PCIE_BAR_EXPLICIT_BAR2(priv->bar.group,
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
1184
NFP_PCIE_BAR_EXPLICIT_BAR0(priv->bar.group,
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
1187
NFP_PCIE_BAR_EXPLICIT_BAR1(priv->bar.group,
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
1190
NFP_PCIE_BAR_EXPLICIT_BAR2(priv->bar.group,
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
1195
priv->bar.group, priv->bar.area),
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
1200
priv->bar.group, priv->bar.area),
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
1205
priv->bar.group, priv->bar.area),
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
170
} group[4];
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
673
nfp->expl.group[i].bitsize = 0;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
686
nfp->expl.group[i].bitsize = bar->bitsize;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
687
nfp->expl.group[i].addr = bar->iomem;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp6000_pcie.c
691
nfp->expl.group[i].free[j] = true;
drivers/net/ethernet/pensando/ionic/ionic_if.h
2457
u8 group;
drivers/net/ethernet/pensando/ionic/ionic_if.h
2473
u8 group;
drivers/net/ethernet/sfc/mcdi.h
98
struct attribute_group group;
drivers/net/ethernet/sfc/mcdi_mon.c
300
hwmon->group.attrs[hwmon->n_attrs++] = &attr->dev_attr.attr;
drivers/net/ethernet/sfc/mcdi_mon.c
358
hwmon->group.attrs = kzalloc_objs(struct attribute *, n_attrs + 1);
drivers/net/ethernet/sfc/mcdi_mon.c
359
if (!hwmon->group.attrs) {
drivers/net/ethernet/sfc/mcdi_mon.c
503
hwmon->groups[0] = &hwmon->group;
drivers/net/ethernet/sfc/mcdi_mon.c
526
kfree(hwmon->group.attrs);
drivers/net/ethernet/sfc/siena/mcdi.h
98
struct attribute_group group;
drivers/net/ethernet/sfc/siena/mcdi_mon.c
300
hwmon->group.attrs[hwmon->n_attrs++] = &attr->dev_attr.attr;
drivers/net/ethernet/sfc/siena/mcdi_mon.c
358
hwmon->group.attrs = kzalloc_objs(struct attribute *, n_attrs + 1);
drivers/net/ethernet/sfc/siena/mcdi_mon.c
359
if (!hwmon->group.attrs) {
drivers/net/ethernet/sfc/siena/mcdi_mon.c
503
hwmon->groups[0] = &hwmon->group;
drivers/net/ethernet/sfc/siena/mcdi_mon.c
526
kfree(hwmon->group.attrs);
drivers/net/ethernet/tehuti/tn40.h
132
const struct software_node *group[SWNODE_MAX + 1];
drivers/net/ethernet/tehuti/tn40_mdio.c
151
nodes->group[SWNODE_PHY] = &swnodes[SWNODE_PHY];
drivers/net/ethernet/tehuti/tn40_mdio.c
152
nodes->group[SWNODE_MDIO] = &swnodes[SWNODE_MDIO];
drivers/net/ethernet/tehuti/tn40_mdio.c
153
return software_node_register_node_group(nodes->group);
drivers/net/ethernet/tehuti/tn40_mdio.c
162
software_node_unregister_node_group(priv->nodes.group);
drivers/net/ethernet/tehuti/tn40_mdio.c
195
priv->nodes.group[SWNODE_MDIO]);
drivers/net/ethernet/tehuti/tn40_mdio.c
213
software_node_unregister_node_group(priv->nodes.group);
drivers/net/ethernet/wangxun/txgbe/txgbe_phy.c
287
fwnode = software_node_fwnode(txgbe->nodes.group[SWNODE_PHYLINK]);
drivers/net/ethernet/wangxun/txgbe/txgbe_phy.c
406
gc->fwnode = software_node_fwnode(txgbe->nodes.group[SWNODE_GPIO]);
drivers/net/ethernet/wangxun/txgbe/txgbe_phy.c
490
info.fwnode = software_node_fwnode(txgbe->nodes.group[SWNODE_I2C]);
drivers/net/ethernet/wangxun/txgbe/txgbe_phy.c
512
info.fwnode = software_node_fwnode(txgbe->nodes.group[SWNODE_SFP]);
drivers/net/ethernet/wangxun/txgbe/txgbe_phy.c
646
software_node_unregister_node_group(txgbe->nodes.group);
drivers/net/ethernet/wangxun/txgbe/txgbe_phy.c
675
software_node_unregister_node_group(txgbe->nodes.group);
drivers/net/ethernet/wangxun/txgbe/txgbe_phy.c
80
nodes->group[SWNODE_GPIO] = &swnodes[SWNODE_GPIO];
drivers/net/ethernet/wangxun/txgbe/txgbe_phy.c
81
nodes->group[SWNODE_I2C] = &swnodes[SWNODE_I2C];
drivers/net/ethernet/wangxun/txgbe/txgbe_phy.c
82
nodes->group[SWNODE_SFP] = &swnodes[SWNODE_SFP];
drivers/net/ethernet/wangxun/txgbe/txgbe_phy.c
83
nodes->group[SWNODE_PHYLINK] = &swnodes[SWNODE_PHYLINK];
drivers/net/ethernet/wangxun/txgbe/txgbe_phy.c
85
return software_node_register_node_group(nodes->group);
drivers/net/ethernet/wangxun/txgbe/txgbe_type.h
440
const struct software_node *group[SWNODE_MAX + 1];
drivers/net/hamradio/scc.c
1098
grp1 = scc->kiss.group;
drivers/net/hamradio/scc.c
1103
grp2 = scc2->kiss.group;
drivers/net/hamradio/scc.c
1147
if (scc->dcd || (scc->kiss.persist) < Rand || (scc->kiss.group && is_grouped(scc)) )
drivers/net/hamradio/scc.c
1314
case PARAM_GROUP: scc->kiss.group=arg; break;
drivers/net/hamradio/scc.c
1386
case PARAM_GROUP: return CAST(scc->kiss.group);
drivers/net/hamradio/scc.c
2059
K(idletime), K(maxdefer), K(tx_inhibit), K(group));
drivers/net/netconsole.c
1258
static struct config_item *userdatum_make_item(struct config_group *group,
drivers/net/netconsole.c
1268
ud = to_userdata(&group->cg_item);
drivers/net/netconsole.c
1281
static void userdatum_drop(struct config_group *group, struct config_item *item)
drivers/net/netconsole.c
1286
ud = to_userdata(&group->cg_item);
drivers/net/netconsole.c
1367
config_group_init_type_name(&nt->group, name, &netconsole_target_type);
drivers/net/netconsole.c
1370
configfs_add_default_group(&nt->userdata_group, &nt->group);
drivers/net/netconsole.c
1380
if (!strcmp(nt->group.cg_item.ci_name, name)) {
drivers/net/netconsole.c
1394
static struct config_group *make_netconsole_target(struct config_group *group,
drivers/net/netconsole.c
1409
return &nt->group;
drivers/net/netconsole.c
1425
return &nt->group;
drivers/net/netconsole.c
1428
static void drop_netconsole_target(struct config_group *group,
drivers/net/netconsole.c
1461
config_item_put(&nt->group.cg_item);
drivers/net/netconsole.c
171
struct config_group group;
drivers/net/netconsole.c
216
if (config_item_name(&nt->group.cg_item))
drivers/net/netconsole.c
217
config_group_get(&nt->group);
drivers/net/netconsole.c
222
if (config_item_name(&nt->group.cg_item))
drivers/net/netconsole.c
223
config_group_put(&nt->group);
drivers/net/netconsole.c
478
struct netconsole_target, group);
drivers/net/netconsole.c
771
config_item_name(&nt->group.cg_item));
drivers/net/netconsole.c
798
config_item_name(&nt->group.cg_item));
drivers/net/netconsole.c
822
config_item_name(&nt->group.cg_item));
drivers/net/netconsole.c
843
config_item_name(&nt->group.cg_item));
drivers/net/netconsole.c
865
config_item_name(&nt->group.cg_item));
drivers/net/netconsole.c
888
config_item_name(&nt->group.cg_item));
drivers/net/netconsole.c
913
config_item_name(&nt->group.cg_item));
drivers/net/netconsole.c
948
config_item_name(&nt->group.cg_item));
drivers/net/netconsole.c
975
struct config_group group;
drivers/net/netconsole.c
980
return container_of(to_config_group(item), struct userdata, group);
drivers/net/netconsole.c
987
netconsole_group = to_config_group(ud->group.cg_item.ci_parent);
drivers/net/netdevsim/dev.c
1171
const struct devlink_trap_group *group,
drivers/net/netdevsim/netdev.c
641
.group = nsim_shaper_group,
drivers/net/netdevsim/psample.c
125
psample_sample_packet(psample->group, skb, psample->rate, &md);
drivers/net/netdevsim/psample.c
143
psample->group = psample_group_get(devlink_net(devlink),
drivers/net/netdevsim/psample.c
145
if (!psample->group)
drivers/net/netdevsim/psample.c
166
psample_group_put(psample->group);
drivers/net/netdevsim/psample.c
26
struct psample_group *group;
drivers/net/netdevsim/psample.c
262
psample_group_put(nsim_dev->psample->group);
drivers/net/tun.c
184
kgid_t group;
drivers/net/tun.c
2266
tun->group = INVALID_GID;
drivers/net/tun.c
2314
if (gid_valid(tun->group) &&
drivers/net/tun.c
2316
from_kgid_munged(current_user_ns(), tun->group)))
drivers/net/tun.c
2671
return gid_valid(tun->group) ?
drivers/net/tun.c
2673
from_kgid_munged(current_user_ns(), tun->group)) :
drivers/net/tun.c
2679
static DEVICE_ATTR_RO(group);
drivers/net/tun.c
3050
kgid_t group;
drivers/net/tun.c
3176
group = make_kgid(current_user_ns(), arg);
drivers/net/tun.c
3177
if (!gid_valid(group)) {
drivers/net/tun.c
3181
tun->group = group;
drivers/net/tun.c
3184
from_kgid(&init_user_ns, tun->group));
drivers/net/tun.c
519
(gid_valid(tun->group) && !in_egroup_p(tun->group))) &&
drivers/net/vxlan/vxlan_mdb.c
1181
const struct vxlan_mdb_entry_key *group)
drivers/net/vxlan/vxlan_mdb.c
1186
mdb_entry = vxlan_mdb_entry_lookup(vxlan, group);
drivers/net/vxlan/vxlan_mdb.c
1195
memcpy(&mdb_entry->key, group, sizeof(mdb_entry->key));
drivers/net/vxlan/vxlan_mdb.c
1237
mdb_entry = vxlan_mdb_entry_get(vxlan, &cfg->group);
drivers/net/vxlan/vxlan_mdb.c
1261
mdb_entry = vxlan_mdb_entry_lookup(vxlan, &cfg->group);
drivers/net/vxlan/vxlan_mdb.c
1464
struct vxlan_mdb_entry_key *group,
drivers/net/vxlan/vxlan_mdb.c
1472
memset(group, 0, sizeof(*group));
drivers/net/vxlan/vxlan_mdb.c
1473
group->vni = vxlan->default_dst.remote_vni;
drivers/net/vxlan/vxlan_mdb.c
1476
vxlan_mdb_group_set(group, entry, NULL);
drivers/net/vxlan/vxlan_mdb.c
1491
vxlan_mdb_group_set(group, entry, mdbe_attrs[MDBE_ATTR_SOURCE]);
drivers/net/vxlan/vxlan_mdb.c
1494
group->vni =
drivers/net/vxlan/vxlan_mdb.c
1573
struct vxlan_mdb_entry_key group;
drivers/net/vxlan/vxlan_mdb.c
1579
err = vxlan_mdb_get_parse(dev, tb, &group, extack);
drivers/net/vxlan/vxlan_mdb.c
1583
mdb_entry = vxlan_mdb_entry_lookup(vxlan, &group);
drivers/net/vxlan/vxlan_mdb.c
1611
struct vxlan_mdb_entry_key group;
drivers/net/vxlan/vxlan_mdb.c
1623
memset(&group, 0, sizeof(group));
drivers/net/vxlan/vxlan_mdb.c
1624
group.vni = src_vni;
drivers/net/vxlan/vxlan_mdb.c
1630
group.dst.sa.sa_family = AF_INET;
drivers/net/vxlan/vxlan_mdb.c
1631
group.dst.sin.sin_addr.s_addr = ip_hdr(skb)->daddr;
drivers/net/vxlan/vxlan_mdb.c
1632
group.src.sa.sa_family = AF_INET;
drivers/net/vxlan/vxlan_mdb.c
1633
group.src.sin.sin_addr.s_addr = ip_hdr(skb)->saddr;
drivers/net/vxlan/vxlan_mdb.c
1639
group.dst.sa.sa_family = AF_INET6;
drivers/net/vxlan/vxlan_mdb.c
1640
group.dst.sin6.sin6_addr = ipv6_hdr(skb)->daddr;
drivers/net/vxlan/vxlan_mdb.c
1641
group.src.sa.sa_family = AF_INET6;
drivers/net/vxlan/vxlan_mdb.c
1642
group.src.sin6.sin6_addr = ipv6_hdr(skb)->saddr;
drivers/net/vxlan/vxlan_mdb.c
1649
mdb_entry = vxlan_mdb_entry_lookup(vxlan, &group);
drivers/net/vxlan/vxlan_mdb.c
1653
memset(&group.src, 0, sizeof(group.src));
drivers/net/vxlan/vxlan_mdb.c
1654
mdb_entry = vxlan_mdb_entry_lookup(vxlan, &group);
drivers/net/vxlan/vxlan_mdb.c
1665
if (ipv4_is_local_multicast(group.dst.sin.sin_addr.s_addr))
drivers/net/vxlan/vxlan_mdb.c
1667
group.dst.sin.sin_addr.s_addr = 0;
drivers/net/vxlan/vxlan_mdb.c
1671
if (ipv6_addr_type(&group.dst.sin6.sin6_addr) &
drivers/net/vxlan/vxlan_mdb.c
1674
memset(&group.dst.sin6.sin6_addr, 0,
drivers/net/vxlan/vxlan_mdb.c
1675
sizeof(group.dst.sin6.sin6_addr));
drivers/net/vxlan/vxlan_mdb.c
1682
return vxlan_mdb_entry_lookup(vxlan, &group);
drivers/net/vxlan/vxlan_mdb.c
381
static void vxlan_mdb_group_set(struct vxlan_mdb_entry_key *group,
drivers/net/vxlan/vxlan_mdb.c
387
group->dst.sa.sa_family = AF_INET;
drivers/net/vxlan/vxlan_mdb.c
388
group->dst.sin.sin_addr.s_addr = entry->addr.u.ip4;
drivers/net/vxlan/vxlan_mdb.c
392
group->dst.sa.sa_family = AF_INET6;
drivers/net/vxlan/vxlan_mdb.c
393
group->dst.sin6.sin6_addr = entry->addr.u.ip6;
drivers/net/vxlan/vxlan_mdb.c
399
vxlan_nla_get_addr(&group->src, source_attr);
drivers/net/vxlan/vxlan_mdb.c
402
static bool vxlan_mdb_is_star_g(const struct vxlan_mdb_entry_key *group)
drivers/net/vxlan/vxlan_mdb.c
404
return !vxlan_addr_any(&group->dst) && vxlan_addr_any(&group->src);
drivers/net/vxlan/vxlan_mdb.c
407
static bool vxlan_mdb_is_sg(const struct vxlan_mdb_entry_key *group)
drivers/net/vxlan/vxlan_mdb.c
409
return !vxlan_addr_any(&group->dst) && !vxlan_addr_any(&group->src);
drivers/net/vxlan/vxlan_mdb.c
512
vxlan_mdb_group_set(&cfg->group, entry, mdbe_attrs[MDBE_ATTR_SOURCE]);
drivers/net/vxlan/vxlan_mdb.c
517
if (!vxlan_addr_is_multicast(&cfg->group.dst) &&
drivers/net/vxlan/vxlan_mdb.c
518
!vxlan_addr_any(&cfg->group.dst)) {
drivers/net/vxlan/vxlan_mdb.c
523
if (vxlan_addr_any(&cfg->group.dst) &&
drivers/net/vxlan/vxlan_mdb.c
529
if (vxlan_mdb_is_sg(&cfg->group))
drivers/net/vxlan/vxlan_mdb.c
533
if (!vxlan_mdb_is_star_g(&cfg->group)) {
drivers/net/vxlan/vxlan_mdb.c
541
if (!vxlan_mdb_is_star_g(&cfg->group)) {
drivers/net/vxlan/vxlan_mdb.c
556
if (vxlan_mdb_is_star_g(&cfg->group) && list_empty(&cfg->src_list) &&
drivers/net/vxlan/vxlan_mdb.c
590
cfg->group.vni =
drivers/net/vxlan/vxlan_mdb.c
610
cfg->group.vni = vxlan->default_dst.remote_vni;
drivers/net/vxlan/vxlan_mdb.c
65
struct vxlan_mdb_entry_key group;
drivers/net/vxlan/vxlan_mdb.c
664
const struct vxlan_mdb_entry_key *group)
drivers/net/vxlan/vxlan_mdb.c
666
return rhashtable_lookup_fast(&vxlan->mdb_tbl, group,
drivers/net/vxlan/vxlan_mdb.c
796
sg_cfg.group.src = *addr;
drivers/net/vxlan/vxlan_mdb.c
797
sg_cfg.group.dst = cfg->group.dst;
drivers/net/vxlan/vxlan_mdb.c
798
sg_cfg.group.vni = cfg->group.vni;
drivers/net/vxlan/vxlan_mdb.c
815
const struct vxlan_mdb_entry_key *group,
drivers/net/vxlan/vxlan_mdb.c
824
sg_cfg.group.src = *addr;
drivers/net/vxlan/vxlan_mdb.c
825
sg_cfg.group.dst = group->dst;
drivers/net/vxlan/vxlan_mdb.c
826
sg_cfg.group.vni = group->vni;
drivers/net/vxlan/vxlan_mdb.c
869
const struct vxlan_mdb_entry_key *group,
drivers/net/vxlan/vxlan_mdb.c
873
vxlan_mdb_remote_src_fwd_del(vxlan, group, remote, &ent->addr);
drivers/net/vxlan/vxlan_mdb.c
896
vxlan_mdb_remote_src_del(cfg->vxlan, &cfg->group, remote, ent);
drivers/net/vxlan/vxlan_mdb.c
901
const struct vxlan_mdb_entry_key *group,
drivers/net/vxlan/vxlan_mdb.c
908
vxlan_mdb_remote_src_del(vxlan, group, remote, ent);
drivers/net/vxlan/vxlan_mdb.c
912
vxlan_mdb_nlmsg_src_list_size(const struct vxlan_mdb_entry_key *group,
drivers/net/vxlan/vxlan_mdb.c
928
nla_total_size(vxlan_addr_size(&group->dst)) +
drivers/net/vxlan/vxlan_mdb.c
941
const struct vxlan_mdb_entry_key *group = &mdb_entry->key;
drivers/net/vxlan/vxlan_mdb.c
951
if (vxlan_mdb_is_sg(group))
drivers/net/vxlan/vxlan_mdb.c
952
nlmsg_size += nla_total_size(vxlan_addr_size(&group->dst));
drivers/net/vxlan/vxlan_mdb.c
956
nlmsg_size += vxlan_mdb_nlmsg_src_list_size(group, remote);
drivers/net/vxlan/vxlan_mdb.c
971
if ((vxlan->cfg.flags & VXLAN_F_COLLECT_METADATA) && group->vni)
drivers/net/vxlan/vxlan_vnifilter.c
515
union vxlan_addr *group,
drivers/net/vxlan/vxlan_vnifilter.c
530
if (group && !vxlan_addr_any(group)) {
drivers/net/vxlan/vxlan_vnifilter.c
531
newrip = group;
drivers/net/vxlan/vxlan_vnifilter.c
555
if (group)
drivers/net/vxlan/vxlan_vnifilter.c
556
memcpy(&vninode->remote_ip, group, sizeof(vninode->remote_ip));
drivers/net/vxlan/vxlan_vnifilter.c
647
__be32 vni, union vxlan_addr *group,
drivers/net/vxlan/vxlan_vnifilter.c
659
ret = vxlan_vni_update_group(vxlan, vninode, group, false, changed,
drivers/net/vxlan/vxlan_vnifilter.c
725
u32 vni, union vxlan_addr *group,
drivers/net/vxlan/vxlan_vnifilter.c
734
return vxlan_vni_update(vxlan, vg, v, group, &changed, extack);
drivers/net/vxlan/vxlan_vnifilter.c
759
err = vxlan_vni_update_group(vxlan, vninode, group, true, &changed,
drivers/net/vxlan/vxlan_vnifilter.c
816
__u32 end_vni, union vxlan_addr *group,
drivers/net/vxlan/vxlan_vnifilter.c
827
err = vxlan_vni_add(vxlan, vg, v, group, extack);
drivers/net/vxlan/vxlan_vnifilter.c
851
union vxlan_addr group;
drivers/net/vxlan/vxlan_vnifilter.c
876
group.sin.sin_addr.s_addr =
drivers/net/vxlan/vxlan_vnifilter.c
878
group.sa.sa_family = AF_INET;
drivers/net/vxlan/vxlan_vnifilter.c
880
group.sin6.sin6_addr =
drivers/net/vxlan/vxlan_vnifilter.c
882
group.sa.sa_family = AF_INET6;
drivers/net/vxlan/vxlan_vnifilter.c
884
memset(&group, 0, sizeof(group));
drivers/net/vxlan/vxlan_vnifilter.c
887
if (vxlan_addr_multicast(&group) && !vxlan->default_dst.remote_ifindex) {
drivers/net/vxlan/vxlan_vnifilter.c
894
err = vxlan_vni_add_del(vxlan, vni_start, vni_end, &group, cmd,
drivers/net/wireless/ath/ath10k/htc.h
270
#define SVC(group, idx) \
drivers/net/wireless/ath/ath10k/htc.h
271
(int)(((int)(group) << 8) | (int)(idx))
drivers/net/wireless/ath/ath11k/htc.h
170
#define SVC(group, idx) \
drivers/net/wireless/ath/ath11k/htc.h
171
(int)(((int)(group) << 8) | (int)(idx))
drivers/net/wireless/ath/ath12k/htc.h
188
#define SVC(group, idx) \
drivers/net/wireless/ath/ath12k/htc.h
189
(int)(((int)(group) << 8) | (int)(idx))
drivers/net/wireless/ath/ath6kl/htc.h
85
#define MAKE_SERVICE_ID(group, index) \
drivers/net/wireless/ath/ath6kl/htc.h
86
(int)(((int)group << 8) | (int)(index))
drivers/net/wireless/ath/ath9k/htc_hst.h
157
#define MAKE_SERVICE_ID(group, index) \
drivers/net/wireless/ath/ath9k/htc_hst.h
158
(int)(((int)group << 8) | (int)(index))
drivers/net/wireless/intel/iwlegacy/3945.c
1892
u8 group;
drivers/net/wireless/intel/iwlegacy/3945.c
1898
for (group = 1; group < 5; group++) {
drivers/net/wireless/intel/iwlegacy/3945.c
1899
grp_channel = ch_grp[group].group_channel;
drivers/net/wireless/intel/iwlegacy/3945.c
1901
group_idx = group;
drivers/net/wireless/intel/iwlegacy/3945.c
1906
if (group == 5)
drivers/net/wireless/intel/iwlegacy/3945.c
1977
const struct il3945_eeprom_txpower_group *group;
drivers/net/wireless/intel/iwlegacy/3945.c
1984
group = &eeprom->groups[i];
drivers/net/wireless/intel/iwlegacy/3945.c
1987
if (group->saturation_power < 40) {
drivers/net/wireless/intel/iwlegacy/3945.c
1990
group->saturation_power);
drivers/net/wireless/intel/iwlegacy/3945.c
2006
satur_pwr = (s8) (group->saturation_power >> 1);
drivers/net/wireless/intel/iwlwifi/fw/api/dbg-tlv.h
28
u8 group;
drivers/net/wireless/intel/iwlwifi/fw/debugfs.c
404
cmd_id = WIDE_ID(ver->group, ver->cmd);
drivers/net/wireless/intel/iwlwifi/fw/file.h
1038
u8 group;
drivers/net/wireless/intel/iwlwifi/fw/img.c
23
if (entry->group == grp && entry->cmd == cmd) {
drivers/net/wireless/intel/iwlwifi/fw/img.c
45
if (entry->group == grp && entry->cmd == cmd) {
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
842
.id = WIDE_ID(hcmd_data->group, hcmd_data->id),
drivers/net/wireless/mediatek/mt76/mt7615/eeprom.c
177
int index, group;
drivers/net/wireless/mediatek/mt76/mt7615/eeprom.c
185
group = mt7615_get_channel_group(chan->hw_value);
drivers/net/wireless/mediatek/mt76/mt7615/eeprom.c
191
return index + group * 3;
drivers/net/wireless/mediatek/mt76/mt7615/eeprom.c
219
int group = mt7615_get_channel_group(chan->hw_value);
drivers/net/wireless/mediatek/mt76/mt7615/eeprom.c
236
index += 5 * group;
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
229
enum mt76x2_cal_channel_group group;
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
231
group = mt76x2_get_cal_channel_group(channel);
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
232
switch (group) {
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
380
enum mt76x2_cal_channel_group group;
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
385
group = mt76x2_get_cal_channel_group(channel);
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
386
offset += group * MT_TX_POWER_GROUP_SIZE_5G;
drivers/net/wireless/mediatek/mt76/mt7915/eeprom.c
316
int group = mt7915_get_channel_group_5g(chan->hw_value, is_7976);
drivers/net/wireless/mediatek/mt76/mt7915/eeprom.c
320
target_power = eeprom[index + group];
drivers/net/wireless/mediatek/mt76/mt7915/eeprom.c
323
target_power = eeprom[index + group];
drivers/net/wireless/mediatek/mt76/mt7915/eeprom.c
329
int group = mt7915_get_channel_group_6g(chan->hw_value);
drivers/net/wireless/mediatek/mt76/mt7915/eeprom.c
332
target_power = is_7976 ? eeprom[index + group] : 0;
drivers/net/wireless/realtek/rtl8xxxu/8188f.c
338
void rtl8188f_channel_to_group(int channel, int *group, int *cck_group)
drivers/net/wireless/realtek/rtl8xxxu/8188f.c
341
*group = 0;
drivers/net/wireless/realtek/rtl8xxxu/8188f.c
343
*group = 1;
drivers/net/wireless/realtek/rtl8xxxu/8188f.c
345
*group = 2;
drivers/net/wireless/realtek/rtl8xxxu/8188f.c
347
*group = 3;
drivers/net/wireless/realtek/rtl8xxxu/8188f.c
349
*group = 4;
drivers/net/wireless/realtek/rtl8xxxu/8188f.c
354
*cck_group = *group;
drivers/net/wireless/realtek/rtl8xxxu/8188f.c
362
int group, cck_group;
drivers/net/wireless/realtek/rtl8xxxu/8188f.c
364
rtl8188f_channel_to_group(channel, &group, &cck_group);
drivers/net/wireless/realtek/rtl8xxxu/8188f.c
378
ofdmbase = priv->ht40_1s_tx_power_index_A[group];
drivers/net/wireless/realtek/rtl8xxxu/8188f.c
385
mcsbase = priv->ht40_1s_tx_power_index_A[group];
drivers/net/wireless/realtek/rtl8xxxu/8192e.c
517
int group, tx_idx;
drivers/net/wireless/realtek/rtl8xxxu/8192e.c
520
group = rtl8xxxu_gen2_channel_to_group(channel);
drivers/net/wireless/realtek/rtl8xxxu/8192e.c
522
cck = priv->cck_tx_power_index_A[group];
drivers/net/wireless/realtek/rtl8xxxu/8192e.c
534
ofdmbase = priv->ht40_1s_tx_power_index_A[group];
drivers/net/wireless/realtek/rtl8xxxu/8192e.c
541
mcsbase = priv->ht40_1s_tx_power_index_A[group];
drivers/net/wireless/realtek/rtl8xxxu/8192e.c
554
cck = priv->cck_tx_power_index_B[group];
drivers/net/wireless/realtek/rtl8xxxu/8192e.c
566
ofdmbase = priv->ht40_1s_tx_power_index_B[group];
drivers/net/wireless/realtek/rtl8xxxu/8192e.c
574
mcsbase = priv->ht40_1s_tx_power_index_B[group];
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
430
int group, cck_group;
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
432
rtl8188f_channel_to_group(channel, &group, &cck_group);
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
442
ofdmbase = priv->ht40_1s_tx_power_index_A[group];
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
449
mcsbase = priv->ht40_1s_tx_power_index_A[group];
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
473
ofdmbase = priv->ht40_1s_tx_power_index_B[group];
drivers/net/wireless/realtek/rtl8xxxu/8192f.c
480
mcsbase = priv->ht40_1s_tx_power_index_B[group];
drivers/net/wireless/realtek/rtl8xxxu/8723b.c
393
int group, tx_idx;
drivers/net/wireless/realtek/rtl8xxxu/8723b.c
396
group = rtl8xxxu_gen2_channel_to_group(channel);
drivers/net/wireless/realtek/rtl8xxxu/8723b.c
398
cck = priv->cck_tx_power_index_B[group];
drivers/net/wireless/realtek/rtl8xxxu/8723b.c
409
ofdmbase = priv->ht40_1s_tx_power_index_B[group];
drivers/net/wireless/realtek/rtl8xxxu/8723b.c
416
mcsbase = priv->ht40_1s_tx_power_index_B[group];
drivers/net/wireless/realtek/rtl8xxxu/core.c
1173
int group;
drivers/net/wireless/realtek/rtl8xxxu/core.c
1176
group = 0;
drivers/net/wireless/realtek/rtl8xxxu/core.c
1178
group = 1;
drivers/net/wireless/realtek/rtl8xxxu/core.c
1180
group = 2;
drivers/net/wireless/realtek/rtl8xxxu/core.c
1182
return group;
drivers/net/wireless/realtek/rtl8xxxu/core.c
1190
int group;
drivers/net/wireless/realtek/rtl8xxxu/core.c
1193
group = 0;
drivers/net/wireless/realtek/rtl8xxxu/core.c
1195
group = 1;
drivers/net/wireless/realtek/rtl8xxxu/core.c
1197
group = 2;
drivers/net/wireless/realtek/rtl8xxxu/core.c
1199
group = 3;
drivers/net/wireless/realtek/rtl8xxxu/core.c
1201
group = 4;
drivers/net/wireless/realtek/rtl8xxxu/core.c
1203
return group;
drivers/net/wireless/realtek/rtl8xxxu/core.c
1468
int group, i;
drivers/net/wireless/realtek/rtl8xxxu/core.c
1470
group = rtl8xxxu_gen1_channel_to_group(channel);
drivers/net/wireless/realtek/rtl8xxxu/core.c
1472
cck[0] = priv->cck_tx_power_index_A[group];
drivers/net/wireless/realtek/rtl8xxxu/core.c
1473
cck[1] = priv->cck_tx_power_index_B[group];
drivers/net/wireless/realtek/rtl8xxxu/core.c
1482
ofdm[0] = priv->ht40_1s_tx_power_index_A[group];
drivers/net/wireless/realtek/rtl8xxxu/core.c
1483
ofdm[1] = priv->ht40_1s_tx_power_index_B[group];
drivers/net/wireless/realtek/rtl8xxxu/core.c
1485
ofdmbase[0] = ofdm[0] + priv->ofdm_tx_power_index_diff[group].a;
drivers/net/wireless/realtek/rtl8xxxu/core.c
1486
ofdmbase[1] = ofdm[1] + priv->ofdm_tx_power_index_diff[group].b;
drivers/net/wireless/realtek/rtl8xxxu/core.c
1491
mcsbase[0] += priv->ht20_tx_power_index_diff[group].a;
drivers/net/wireless/realtek/rtl8xxxu/core.c
1492
mcsbase[1] += priv->ht20_tx_power_index_diff[group].b;
drivers/net/wireless/realtek/rtl8xxxu/core.c
1496
if (ofdm[0] > priv->ht40_2s_tx_power_index_diff[group].a)
drivers/net/wireless/realtek/rtl8xxxu/core.c
1497
ofdm[0] -= priv->ht40_2s_tx_power_index_diff[group].a;
drivers/net/wireless/realtek/rtl8xxxu/core.c
1498
if (ofdm[1] > priv->ht40_2s_tx_power_index_diff[group].b)
drivers/net/wireless/realtek/rtl8xxxu/core.c
1499
ofdm[1] -= priv->ht40_2s_tx_power_index_diff[group].b;
drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu.h
2102
void rtl8188f_channel_to_group(int channel, int *group, int *cck_group);
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1512
u8 group = 0;
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1515
group = 0;
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1517
group = 1;
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1519
group = 2;
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1521
group = 3;
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1523
group = 4;
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1525
group = 5;
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1527
return group;
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1532
int group, txcnt;
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1534
for (group = 0 ; group < MAX_CHNL_GROUP_24G; group++) {
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1535
pwrinfo24g->index_cck_base[rfpath][group] = 0x2D;
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1536
pwrinfo24g->index_bw40_base[rfpath][group] = 0x2D;
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1557
u32 rfpath, eeaddr = EEPROM_TX_PWR_INX, group, txcnt = 0;
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1577
for (group = 0 ; group < MAX_CHNL_GROUP_24G; group++) {
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1578
pwrinfo24g->index_cck_base[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1580
if (pwrinfo24g->index_cck_base[rfpath][group] == 0xFF)
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1581
pwrinfo24g->index_cck_base[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1584
for (group = 0 ; group < MAX_CHNL_GROUP_24G-1; group++) {
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1585
pwrinfo24g->index_bw40_base[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1587
if (pwrinfo24g->index_bw40_base[rfpath][group] == 0xFF)
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1588
pwrinfo24g->index_bw40_base[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1663
for (group = 0 ; group < MAX_CHNL_GROUP_5G; group++) {
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1664
pwrinfo5g->index_bw40_base[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1666
if (pwrinfo5g->index_bw40_base[rfpath][group] == 0xFF)
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
1667
pwrinfo5g->index_bw40_base[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8192ce/hw.h
12
group = 0;
drivers/net/wireless/realtek/rtlwifi/rtl8192ce/hw.h
14
group = 1;
drivers/net/wireless/realtek/rtlwifi/rtl8192ce/hw.h
16
group = 2;
drivers/net/wireless/realtek/rtlwifi/rtl8192ce/hw.h
17
return group;
drivers/net/wireless/realtek/rtlwifi/rtl8192ce/hw.h
9
u8 group;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
368
u32 rfpath, eeaddr, group, offset, offset1, offset2;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
373
for (group = 0; group < CHANNEL_GROUP_MAX; group++) {
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
375
if (group < CHANNEL_GROUP_MAX_2G) {
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
376
pwrinfo->cck_index[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
378
pwrinfo->ht40_1sindex[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
381
pwrinfo->ht40_1sindex[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
384
pwrinfo->ht40_2sindexdiff[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
386
pwrinfo->ht20indexdiff[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
388
pwrinfo->ofdmindexdiff[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
390
pwrinfo->ht40maxoffset[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
392
pwrinfo->ht20maxoffset[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
407
for (group = 0; group < CHANNEL_GROUP_MAX_2G; group++) {
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
408
eeaddr = EEPROM_CCK_TX_PWR_INX_2G + (rfpath * 3) + group;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
410
pwrinfo->cck_index[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
419
for (group = 0; group < CHANNEL_GROUP_MAX; group++) {
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
420
offset1 = group / 3;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
421
offset2 = group % 3;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
425
pwrinfo->ht40_1sindex[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
435
for (group = 0; group < CHANNEL_GROUP_MAX; group++) {
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
437
offset1 = group / 3;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
438
offset2 = group % 3;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
443
pwrinfo->ht40_2sindexdiff[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
446
pwrinfo->ht40_2sindexdiff[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
451
pwrinfo->ht20indexdiff[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
454
pwrinfo->ht20indexdiff[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
459
pwrinfo->ofdmindexdiff[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
462
pwrinfo->ofdmindexdiff[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
467
pwrinfo->ht40maxoffset[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
470
pwrinfo->ht40maxoffset[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
475
pwrinfo->ht20maxoffset[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
478
pwrinfo->ht20maxoffset[rfpath][group] =
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
510
u32 ch, rfpath, group;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
594
group = rtl92d_get_chnlgroup_fromarray((u8)ch);
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
597
pwrinfo.cck_index[rfpath][group];
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
599
pwrinfo.ht40_1sindex[rfpath][group];
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
601
pwrinfo.ht20indexdiff[rfpath][group];
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
603
pwrinfo.ofdmindexdiff[rfpath][group];
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
605
pwrinfo.ht20maxoffset[rfpath][group];
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
607
pwrinfo.ht40maxoffset[rfpath][group];
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
608
pwr = pwrinfo.ht40_1sindex[rfpath][group];
drivers/net/wireless/realtek/rtlwifi/rtl8192d/hw_common.c
609
diff = pwrinfo.ht40_2sindexdiff[rfpath][group];
drivers/net/wireless/realtek/rtlwifi/rtl8192d/phy_common.c
783
u8 group;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/phy_common.c
786
group = 0;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/phy_common.c
788
group = 1;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/phy_common.c
790
group = 2;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/phy_common.c
792
group = 3;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/phy_common.c
794
group = 4;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/phy_common.c
796
group = 5;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/phy_common.c
798
group = 6;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/phy_common.c
800
group = 7;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/phy_common.c
802
group = 8;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/phy_common.c
804
group = 9;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/phy_common.c
806
group = 10;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/phy_common.c
808
group = 11;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/phy_common.c
809
return group;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/phy_common.c
815
u8 group;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/phy_common.c
818
group = 0;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/phy_common.c
820
group = 1;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/phy_common.c
822
group = 2;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/phy_common.c
824
group = 6;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/phy_common.c
826
group = 7;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/phy_common.c
828
group = 8;
drivers/net/wireless/realtek/rtlwifi/rtl8192d/phy_common.c
829
return group;
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
661
u8 group, i;
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
677
group = channel <= 64 ? 1 : 2;
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
682
rf_imr_param_normal[0][group][i]);
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
660
u8 group, i;
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
681
group = channel <= 64 ? 1 : 2;
drivers/net/wireless/realtek/rtlwifi/rtl8192du/phy.c
686
rf_imr_param_normal[group][i]);
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1736
u8 group = 0;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1740
group = 0;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1742
group = 1;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1744
group = 2;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1746
group = 3;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1748
group = 4;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1751
group = 0;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1753
group = 1;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1755
group = 2;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1757
group = 3;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1759
group = 4;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1761
group = 5;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1763
group = 6;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1765
group = 7;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1767
group = 8;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1769
group = 9;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1771
group = 10;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1773
group = 11;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1775
group = 12;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1777
group = 13;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1779
return group;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1788
u32 rf, addr = EEPROM_TX_PWR_INX, group, i = 0;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1801
for (group = 0 ; group < MAX_CHNL_GROUP_24G; group++) {
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1802
pwr2g->index_cck_base[rf][group] = 0x2D;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1803
pwr2g->index_bw40_base[rf][group] = 0x2D;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1818
for (group = 0 ; group < MAX_CHNL_GROUP_5G; group++)
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1819
pwr5g->index_bw40_base[rf][group] = 0x2A;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1843
for (group = 0 ; group < MAX_CHNL_GROUP_24G; group++) {
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1844
pwr2g->index_cck_base[rf][group] = hwinfo[addr++];
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1845
if (pwr2g->index_cck_base[rf][group] == 0xFF)
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1846
pwr2g->index_cck_base[rf][group] = 0x2D;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1848
for (group = 0 ; group < MAX_CHNL_GROUP_24G - 1; group++) {
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1849
pwr2g->index_bw40_base[rf][group] = hwinfo[addr++];
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1850
if (pwr2g->index_bw40_base[rf][group] == 0xFF)
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1851
pwr2g->index_bw40_base[rf][group] = 0x2D;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1917
for (group = 0 ; group < MAX_CHNL_GROUP_5G; group++) {
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1918
pwr5g->index_bw40_base[rf][group] = hwinfo[addr++];
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1919
if (pwr5g->index_bw40_base[rf][group] == 0xFF)
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
1920
pwr5g->index_bw40_base[rf][group] = 0xFE;
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/hw.c
1373
u8 group;
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/hw.c
1376
group = 0;
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/hw.c
1378
group = 1;
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/hw.c
1380
group = 2;
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/hw.c
1381
return group;
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
1728
u8 group;
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
1731
group = 0;
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
1733
group = 1;
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
1735
group = 2;
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
1736
return group;
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
1745
u32 path, addr = EEPROM_TX_PWR_INX, group, cnt = 0;
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
1758
for (group = 0 ; group < MAX_CHNL_GROUP_24G; group++) {
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
1759
pw2g->index_cck_base[path][group] = 0x2D;
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
1760
pw2g->index_bw40_base[path][group] = 0x2D;
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
1779
for (group = 0; group < MAX_CHNL_GROUP_24G; group++) {
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
1780
pw2g->index_cck_base[path][group] = hwinfo[addr++];
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
1781
if (pw2g->index_cck_base[path][group] == 0xFF)
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
1782
pw2g->index_cck_base[path][group] = 0x2D;
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
1785
for (group = 0; group < MAX_CHNL_GROUP_24G - 1; group++) {
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
1786
pw2g->index_bw40_base[path][group] = hwinfo[addr++];
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
1787
if (pw2g->index_bw40_base[path][group] == 0xFF)
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
1788
pw2g->index_bw40_base[path][group] = 0x2D;
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
1862
for (group = 0; group < MAX_CHNL_GROUP_5G; group++) {
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
1863
pw5g->index_bw40_base[path][group] = hwinfo[addr++];
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
1864
if (pw5g->index_bw40_base[path][group] == 0xFF)
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
1865
pw5g->index_bw40_base[path][group] = 0xFE;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2488
u8 group = 0;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2492
group = 0;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2494
group = 1;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2496
group = 2;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2498
group = 3;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2500
group = 4;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2503
group = 0;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2505
group = 1;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2507
group = 2;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2509
group = 3;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2511
group = 4;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2513
group = 5;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2515
group = 6;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2517
group = 7;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2519
group = 8;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2521
group = 9;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2523
group = 10;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2525
group = 11;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2527
group = 12;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2529
group = 13;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2535
return group;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2545
u32 rfpath, eeaddr = EEPROM_TX_PWR_INX, group, txcount = 0;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2558
for (group = 0 ; group < MAX_CHNL_GROUP_24G; group++) {
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2559
pwrinfo24g->index_cck_base[rfpath][group] = 0x2D;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2560
pwrinfo24g->index_bw40_base[rfpath][group] = 0x2D;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2574
for (group = 0 ; group < MAX_CHNL_GROUP_5G; group++)
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2575
pwrinfo5g->index_bw40_base[rfpath][group] = 0x2A;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2599
for (group = 0 ; group < MAX_CHNL_GROUP_24G; group++) {
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2600
pwrinfo24g->index_cck_base[rfpath][group] = hwinfo[eeaddr++];
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2601
if (pwrinfo24g->index_cck_base[rfpath][group] == 0xFF)
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2602
pwrinfo24g->index_cck_base[rfpath][group] = 0x2D;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2604
for (group = 0 ; group < MAX_CHNL_GROUP_24G - 1; group++) {
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2605
pwrinfo24g->index_bw40_base[rfpath][group] = hwinfo[eeaddr++];
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2606
if (pwrinfo24g->index_bw40_base[rfpath][group] == 0xFF)
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2607
pwrinfo24g->index_bw40_base[rfpath][group] = 0x2D;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2647
for (group = 0 ; group < MAX_CHNL_GROUP_5G; group++) {
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2648
pwrinfo5g->index_bw40_base[rfpath][group] = hwinfo[eeaddr++];
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2649
if (pwrinfo5g->index_bw40_base[rfpath][group] == 0xFF)
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
2650
pwrinfo5g->index_bw40_base[rfpath][group] = 0xFE;
drivers/net/wireless/realtek/rtw88/main.h
726
bool group;
drivers/net/wireless/realtek/rtw88/phy.c
2016
u8 rate, u8 group)
drivers/net/wireless/realtek/rtw88/phy.c
2025
tx_power = pwr_idx_2g->cck_base[group];
drivers/net/wireless/realtek/rtw88/phy.c
2027
tx_power = pwr_idx_2g->bw40_base[group];
drivers/net/wireless/realtek/rtw88/phy.c
2075
u8 rate, u8 group)
drivers/net/wireless/realtek/rtw88/phy.c
2084
tx_power = pwr_idx_5g->bw40_base[group];
drivers/net/wireless/realtek/rtw88/phy.c
2125
lower = pwr_idx_5g->bw40_base[group];
drivers/net/wireless/realtek/rtw88/phy.c
2126
upper = pwr_idx_5g->bw40_base[group + 1];
drivers/net/wireless/realtek/rtw88/phy.c
2245
u8 group, band;
drivers/net/wireless/realtek/rtw88/phy.c
2253
group = rtw_get_channel_group(ch, rate);
drivers/net/wireless/realtek/rtw88/phy.c
2260
bw, rate, group);
drivers/net/wireless/realtek/rtw88/phy.c
2266
bw, rate, group);
drivers/net/wireless/realtek/rtw88/sec.c
37
cam->group = !(key->flags & IEEE80211_KEY_FLAG_PAIRWISE);
drivers/net/wireless/realtek/rtw88/sec.c
52
(cam->group << 6) |
drivers/net/wireless/realtek/rtw89/chan.c
1856
req.group = mcc->group;
drivers/net/wireless/realtek/rtw89/chan.c
1879
ret = rtw89_fw_h2c_mcc_macid_bitmap(rtwdev, mcc->group,
drivers/net/wireless/realtek/rtw89/chan.c
1929
req.group = mcc->group;
drivers/net/wireless/realtek/rtw89/chan.c
1969
req.old_group = mcc->group;
drivers/net/wireless/realtek/rtw89/chan.c
1971
mcc->group = RTW89_MCC_NEXT_GROUP(mcc->group);
drivers/net/wireless/realtek/rtw89/chan.c
1974
req.group = mcc->group;
drivers/net/wireless/realtek/rtw89/chan.c
2019
ret = rtw89_fw_h2c_mcc_sync(rtwdev, req.group, sync->macid_src,
drivers/net/wireless/realtek/rtw89/chan.c
2083
start_arg.old_sch_idx = mcc->group;
drivers/net/wireless/realtek/rtw89/chan.c
2085
mcc->group = RTW89_MCC_NEXT_GROUP(mcc->group);
drivers/net/wireless/realtek/rtw89/chan.c
2088
add_arg.sch_idx = mcc->group;
drivers/net/wireless/realtek/rtw89/chan.c
2150
start_arg.sch_idx = mcc->group;
drivers/net/wireless/realtek/rtw89/chan.c
2171
.group = mcc->group,
drivers/net/wireless/realtek/rtw89/chan.c
2193
ret = rtw89_fw_h2c_mcc_sync(rtwdev, mcc->group, sync->macid_src,
drivers/net/wireless/realtek/rtw89/chan.c
2212
.sch_idx = mcc->group,
drivers/net/wireless/realtek/rtw89/chan.c
2377
mcc->group = RTW89_MCC_DFLT_GROUP;
drivers/net/wireless/realtek/rtw89/chan.c
2482
ret = rtw89_fw_h2c_mrc_del(rtwdev, mcc->group, sel.slot_idx);
drivers/net/wireless/realtek/rtw89/chan.c
2487
ret = rtw89_fw_h2c_stop_mcc(rtwdev, mcc->group,
drivers/net/wireless/realtek/rtw89/chan.c
2493
ret = rtw89_fw_h2c_del_mcc_group(rtwdev, mcc->group, true);
drivers/net/wireless/realtek/rtw89/chan.c
2748
ret = rtw89_fw_h2c_mcc_macid_bitmap(rtwdev, mcc->group,
drivers/net/wireless/realtek/rtw89/chan.c
2773
arg.sch_idx = mcc->group;
drivers/net/wireless/realtek/rtw89/chan.c
730
req.group = mcc->group;
drivers/net/wireless/realtek/rtw89/core.h
6095
u8 group;
drivers/net/wireless/realtek/rtw89/fw.c
10010
cond = RTW89_MCC_WAIT_COND(p->group, H2C_FUNC_START_MCC);
drivers/net/wireless/realtek/rtw89/fw.c
10015
int rtw89_fw_h2c_stop_mcc(struct rtw89_dev *rtwdev, u8 group, u8 macid,
drivers/net/wireless/realtek/rtw89/fw.c
10031
RTW89_SET_FWCMD_STOP_MCC_GROUP(skb->data, group);
drivers/net/wireless/realtek/rtw89/fw.c
10040
cond = RTW89_MCC_WAIT_COND(group, H2C_FUNC_STOP_MCC);
drivers/net/wireless/realtek/rtw89/fw.c
10045
int rtw89_fw_h2c_del_mcc_group(struct rtw89_dev *rtwdev, u8 group,
drivers/net/wireless/realtek/rtw89/fw.c
10060
RTW89_SET_FWCMD_DEL_MCC_GROUP_GROUP(skb->data, group);
drivers/net/wireless/realtek/rtw89/fw.c
10069
cond = RTW89_MCC_WAIT_COND(group, H2C_FUNC_DEL_MCC_GROUP);
drivers/net/wireless/realtek/rtw89/fw.c
10074
int rtw89_fw_h2c_reset_mcc_group(struct rtw89_dev *rtwdev, u8 group)
drivers/net/wireless/realtek/rtw89/fw.c
10088
RTW89_SET_FWCMD_RESET_MCC_GROUP_GROUP(skb->data, group);
drivers/net/wireless/realtek/rtw89/fw.c
10096
cond = RTW89_MCC_WAIT_COND(group, H2C_FUNC_RESET_MCC_GROUP);
drivers/net/wireless/realtek/rtw89/fw.c
10119
RTW89_SET_FWCMD_MCC_REQ_TSF_GROUP(skb->data, req->group);
drivers/net/wireless/realtek/rtw89/fw.c
10129
cond = RTW89_MCC_WAIT_COND(req->group, H2C_FUNC_MCC_REQ_TSF);
drivers/net/wireless/realtek/rtw89/fw.c
10141
int rtw89_fw_h2c_mcc_macid_bitmap(struct rtw89_dev *rtwdev, u8 group, u8 macid,
drivers/net/wireless/realtek/rtw89/fw.c
10161
RTW89_SET_FWCMD_MCC_MACID_BITMAP_GROUP(skb->data, group);
drivers/net/wireless/realtek/rtw89/fw.c
10172
cond = RTW89_MCC_WAIT_COND(group, H2C_FUNC_MCC_MACID_BITMAP);
drivers/net/wireless/realtek/rtw89/fw.c
10177
int rtw89_fw_h2c_mcc_sync(struct rtw89_dev *rtwdev, u8 group, u8 source,
drivers/net/wireless/realtek/rtw89/fw.c
10192
RTW89_SET_FWCMD_MCC_SYNC_GROUP(skb->data, group);
drivers/net/wireless/realtek/rtw89/fw.c
10203
cond = RTW89_MCC_WAIT_COND(group, H2C_FUNC_MCC_SYNC);
drivers/net/wireless/realtek/rtw89/fw.c
10223
RTW89_SET_FWCMD_MCC_SET_DURATION_GROUP(skb->data, p->group);
drivers/net/wireless/realtek/rtw89/fw.c
10241
cond = RTW89_MCC_WAIT_COND(p->group, H2C_FUNC_MCC_SET_DURATION);
drivers/net/wireless/realtek/rtw89/fw.c
9952
RTW89_SET_FWCMD_ADD_MCC_GROUP(skb->data, p->group);
drivers/net/wireless/realtek/rtw89/fw.c
9974
cond = RTW89_MCC_WAIT_COND(p->group, H2C_FUNC_ADD_MCC);
drivers/net/wireless/realtek/rtw89/fw.c
9994
RTW89_SET_FWCMD_START_MCC_GROUP(skb->data, p->group);
drivers/net/wireless/realtek/rtw89/fw.h
3215
u32 group: 2;
drivers/net/wireless/realtek/rtw89/fw.h
3340
u32 group: 2;
drivers/net/wireless/realtek/rtw89/fw.h
3430
u8 group: 2;
drivers/net/wireless/realtek/rtw89/fw.h
3494
u32 group: 2;
drivers/net/wireless/realtek/rtw89/fw.h
4646
#define RTW89_MCC_WAIT_COND(group, func) \
drivers/net/wireless/realtek/rtw89/fw.h
4647
((group) * NUM_OF_RTW89_MCC_H2C_FUNC + (func))
drivers/net/wireless/realtek/rtw89/fw.h
5431
int rtw89_fw_h2c_stop_mcc(struct rtw89_dev *rtwdev, u8 group, u8 macid,
drivers/net/wireless/realtek/rtw89/fw.h
5433
int rtw89_fw_h2c_del_mcc_group(struct rtw89_dev *rtwdev, u8 group,
drivers/net/wireless/realtek/rtw89/fw.h
5435
int rtw89_fw_h2c_reset_mcc_group(struct rtw89_dev *rtwdev, u8 group);
drivers/net/wireless/realtek/rtw89/fw.h
5439
int rtw89_fw_h2c_mcc_macid_bitmap(struct rtw89_dev *rtwdev, u8 group, u8 macid,
drivers/net/wireless/realtek/rtw89/fw.h
5441
int rtw89_fw_h2c_mcc_sync(struct rtw89_dev *rtwdev, u8 group, u8 source,
drivers/net/wireless/realtek/rtw89/mac.c
5499
u8 group = RTW89_GET_MAC_C2H_MCC_RCV_ACK_GROUP(c2h->data);
drivers/net/wireless/realtek/rtw89/mac.c
5520
"MCC C2H RCV ACK: group %d, func %d\n", group, func);
drivers/net/wireless/realtek/rtw89/mac.c
5526
u8 group = RTW89_GET_MAC_C2H_MCC_REQ_ACK_GROUP(c2h->data);
drivers/net/wireless/realtek/rtw89/mac.c
5554
group, func, retcode);
drivers/net/wireless/realtek/rtw89/mac.c
5560
cond = RTW89_MCC_WAIT_COND(group, func);
drivers/net/wireless/realtek/rtw89/mac.c
5567
u8 group = RTW89_GET_MAC_C2H_MCC_TSF_RPT_GROUP(c2h->data);
drivers/net/wireless/realtek/rtw89/mac.c
5585
cond = RTW89_MCC_WAIT_COND(group, H2C_FUNC_MCC_REQ_TSF);
drivers/net/wireless/realtek/rtw89/mac.c
5592
u8 group = RTW89_GET_MAC_C2H_MCC_STATUS_RPT_GROUP(c2h->data);
drivers/net/wireless/realtek/rtw89/mac.c
5649
group, macid, status, (u64)tsf_high << 32 | tsf_low);
drivers/net/wireless/realtek/rtw89/mac.c
5655
cond = RTW89_MCC_WAIT_COND(group, func);
drivers/net/wireless/realtek/rtw89/phy.c
4224
#define PHY_IS_TSSI_EXTRA_GROUP(group) ((group) & PHY_TSSI_EXTRA_GROUP_BIT)
drivers/net/wireless/realtek/rtw89/phy.c
4225
#define PHY_TSSI_EXTRA_GET_GROUP_IDX1(group) \
drivers/net/wireless/realtek/rtw89/phy.c
4226
((group) & ~PHY_TSSI_EXTRA_GROUP_BIT)
drivers/net/wireless/realtek/rtw89/phy.c
4227
#define PHY_TSSI_EXTRA_GET_GROUP_IDX2(group) \
drivers/net/wireless/realtek/rtw89/phy.c
4228
(PHY_TSSI_EXTRA_GET_GROUP_IDX1(group) + 1)
drivers/net/wireless/realtek/rtw89/rtw8851b.c
2154
void rtw8851b_set_trx_mask(struct rtw89_dev *rtwdev, u8 path, u8 group, u32 val)
drivers/net/wireless/realtek/rtw89/rtw8851b.c
2156
if (group > BTC_BT_SS_GROUP)
drivers/net/wireless/realtek/rtw89/rtw8851b.c
2157
group--; /* Tx-group=1, Rx-group=2 */
drivers/net/wireless/realtek/rtw89/rtw8851b.c
2160
group += 3;
drivers/net/wireless/realtek/rtw89/rtw8851b.c
2162
rtw89_write_rf(rtwdev, path, RR_LUTWA, RFREG_MASK, group);
drivers/net/wireless/realtek/rtw89/rtw8851b_rfk.c
3008
#define IS_TSSI_EXTRA_GROUP(group) ((group) & TSSI_EXTRA_GROUP_BIT)
drivers/net/wireless/realtek/rtw89/rtw8851b_rfk.c
3009
#define TSSI_EXTRA_GET_GROUP_IDX1(group) ((group) & ~TSSI_EXTRA_GROUP_BIT)
drivers/net/wireless/realtek/rtw89/rtw8851b_rfk.c
3010
#define TSSI_EXTRA_GET_GROUP_IDX2(group) (TSSI_EXTRA_GET_GROUP_IDX1(group) + 1)
drivers/net/wireless/realtek/rtw89/rtw8852a.c
1835
void rtw8852a_set_trx_mask(struct rtw89_dev *rtwdev, u8 path, u8 group, u32 val)
drivers/net/wireless/realtek/rtw89/rtw8852a.c
1838
rtw89_write_rf(rtwdev, path, RR_LUTWA, 0xfffff, group);
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
1110
u8 group = 0x2;
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
1137
rtw89_phy_write32_mask(rtwdev, R_CFIR_LUT + (path << 8), B_CFIR_LUT_GP, group);
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
3184
#define IS_TSSI_EXTRA_GROUP(group) ((group) & TSSI_EXTRA_GROUP_BIT)
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
3185
#define TSSI_EXTRA_GET_GROUP_IDX1(group) ((group) & ~TSSI_EXTRA_GROUP_BIT)
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
3186
#define TSSI_EXTRA_GET_GROUP_IDX2(group) (TSSI_EXTRA_GET_GROUP_IDX1(group) + 1)
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
599
u8 group)
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
613
if (group >= RTW8852A_IQK_CFIR_GROUP_NR) {
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
614
rtw89_warn(rtwdev, "cfir group %d out of range\n", group);
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
621
base_addr = base_addrs[path][group];
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
660
u8 group)
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
674
if (group >= RTW8852A_IQK_CFIR_GROUP_NR) {
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
675
rtw89_warn(rtwdev, "cfir group %d out of range\n", group);
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
682
base_addr = base_addrs[path][group];
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
962
u8 group = 0x0;
drivers/net/wireless/realtek/rtw89/rtw8852a_rfk.c
995
B_CFIR_LUT_GP, group);
drivers/net/wireless/realtek/rtw89/rtw8852b_common.c
1789
void rtw8852bx_set_trx_mask(struct rtw89_dev *rtwdev, u8 path, u8 group, u32 val)
drivers/net/wireless/realtek/rtw89/rtw8852b_common.c
1792
rtw89_write_rf(rtwdev, path, RR_LUTWA, RFREG_MASK, group);
drivers/net/wireless/realtek/rtw89/rtw8852b_rfk.c
3128
#define IS_TSSI_EXTRA_GROUP(group) ((group) & TSSI_EXTRA_GROUP_BIT)
drivers/net/wireless/realtek/rtw89/rtw8852b_rfk.c
3129
#define TSSI_EXTRA_GET_GROUP_IDX1(group) ((group) & ~TSSI_EXTRA_GROUP_BIT)
drivers/net/wireless/realtek/rtw89/rtw8852b_rfk.c
3130
#define TSSI_EXTRA_GET_GROUP_IDX2(group) (TSSI_EXTRA_GET_GROUP_IDX1(group) + 1)
drivers/net/wireless/realtek/rtw89/rtw8852bt_rfk.c
3197
#define IS_TSSI_EXTRA_GROUP(group) ((group) & TSSI_EXTRA_GROUP_BIT)
drivers/net/wireless/realtek/rtw89/rtw8852bt_rfk.c
3198
#define TSSI_EXTRA_GET_GROUP_IDX1(group) ((group) & ~TSSI_EXTRA_GROUP_BIT)
drivers/net/wireless/realtek/rtw89/rtw8852bt_rfk.c
3199
#define TSSI_EXTRA_GET_GROUP_IDX2(group) (TSSI_EXTRA_GET_GROUP_IDX1(group) + 1)
drivers/net/wireless/realtek/rtw89/rtw8852c.c
2655
void rtw8852c_set_trx_mask(struct rtw89_dev *rtwdev, u8 path, u8 group, u32 val)
drivers/net/wireless/realtek/rtw89/rtw8852c.c
2658
rtw89_write_rf(rtwdev, path, RR_LUTWA, RFREG_MASK, group);
drivers/net/wireless/realtek/rtw89/rtw8852c_rfk.c
3350
#define IS_TSSI_EXTRA_GROUP(group) ((group) & TSSI_EXTRA_GROUP_BIT)
drivers/net/wireless/realtek/rtw89/rtw8852c_rfk.c
3351
#define TSSI_EXTRA_GET_GROUP_IDX1(group) ((group) & ~TSSI_EXTRA_GROUP_BIT)
drivers/net/wireless/realtek/rtw89/rtw8852c_rfk.c
3352
#define TSSI_EXTRA_GET_GROUP_IDX2(group) (TSSI_EXTRA_GET_GROUP_IDX1(group) + 1)
drivers/net/wireless/realtek/rtw89/rtw8922a.c
2539
void rtw8922a_set_trx_mask(struct rtw89_dev *rtwdev, u8 path, u8 group, u32 val)
drivers/net/wireless/realtek/rtw89/rtw8922a.c
2541
rtw89_write_rf(rtwdev, path, RR_LUTWA, RFREG_MASK, group);
drivers/net/wireless/virtual/mac80211_hwsim.c
1170
*val = data->group;
drivers/net/wireless/virtual/mac80211_hwsim.c
1177
data->group = val;
drivers/net/wireless/virtual/mac80211_hwsim.c
1857
if (!(data->group & data2->group))
drivers/net/wireless/virtual/mac80211_hwsim.c
3373
data[i++] = ar->group;
drivers/net/wireless/virtual/mac80211_hwsim.c
5768
data->group = 1;
drivers/net/wireless/virtual/mac80211_hwsim.c
726
u64 group;
drivers/nvme/target/configfs.c
1019
&subsys->group);
drivers/nvme/target/configfs.c
1710
static struct config_group *nvmet_subsys_make(struct config_group *group,
drivers/nvme/target/configfs.c
1729
config_group_init_type_name(&subsys->group, name, &nvmet_subsys_type);
drivers/nvme/target/configfs.c
1733
configfs_add_default_group(&subsys->namespaces_group, &subsys->group);
drivers/nvme/target/configfs.c
1738
&subsys->group);
drivers/nvme/target/configfs.c
1742
return &subsys->group;
drivers/nvme/target/configfs.c
1797
static void nvmet_referral_notify(struct config_group *group,
drivers/nvme/target/configfs.c
1824
struct config_group *group, const char *name)
drivers/nvme/target/configfs.c
1834
config_group_init_type_name(&port->group, name, &nvmet_referral_type);
drivers/nvme/target/configfs.c
1836
return &port->group;
drivers/nvme/target/configfs.c
1930
struct config_group *group, const char *name)
drivers/nvme/target/configfs.c
1932
struct nvmet_port *port = ana_groups_to_port(&group->cg_item);
drivers/nvme/target/configfs.c
1959
config_group_init_type_name(&grp->group, name, &nvmet_ana_group_type);
drivers/nvme/target/configfs.c
1960
return &grp->group;
drivers/nvme/target/configfs.c
2015
static struct config_group *nvmet_ports_make(struct config_group *group,
drivers/nvme/target/configfs.c
2062
config_group_init_type_name(&port->group, name, &nvmet_port_type);
drivers/nvme/target/configfs.c
2066
configfs_add_default_group(&port->subsys_group, &port->group);
drivers/nvme/target/configfs.c
2070
configfs_add_default_group(&port->referrals_group, &port->group);
drivers/nvme/target/configfs.c
2074
configfs_add_default_group(&port->ana_groups_group, &port->group);
drivers/nvme/target/configfs.c
2078
config_group_init_type_name(&port->ana_default_group.group,
drivers/nvme/target/configfs.c
2081
configfs_add_default_group(&port->ana_default_group.group,
drivers/nvme/target/configfs.c
2084
return &port->group;
drivers/nvme/target/configfs.c
2254
static struct config_group *nvmet_hosts_make_group(struct config_group *group,
drivers/nvme/target/configfs.c
2268
config_group_init_type_name(&host->group, name, &nvmet_host_type);
drivers/nvme/target/configfs.c
2270
return &host->group;
drivers/nvme/target/configfs.c
833
static struct config_group *nvmet_ns_make(struct config_group *group,
drivers/nvme/target/configfs.c
836
struct nvmet_subsys *subsys = namespaces_to_subsys(&group->cg_item);
drivers/nvme/target/configfs.c
855
config_group_init_type_name(&ns->group, name, &nvmet_ns_type);
drivers/nvme/target/configfs.c
859
return &ns->group;
drivers/nvme/target/nvmet.h
119
struct config_group group;
drivers/nvme/target/nvmet.h
135
return container_of(to_config_group(item), struct nvmet_ns, group);
drivers/nvme/target/nvmet.h
180
struct config_group group;
drivers/nvme/target/nvmet.h
188
group);
drivers/nvme/target/nvmet.h
203
struct config_group group;
drivers/nvme/target/nvmet.h
224
group);
drivers/nvme/target/nvmet.h
342
struct config_group group;
drivers/nvme/target/nvmet.h
369
return container_of(to_config_group(item), struct nvmet_subsys, group);
drivers/nvme/target/nvmet.h
380
struct config_group group;
drivers/nvme/target/nvmet.h
391
return container_of(to_config_group(item), struct nvmet_host, group);
drivers/nvme/target/nvmet.h
396
return config_item_name(&host->group.cg_item);
drivers/nvme/target/pci-epf.c
220
struct config_group group;
drivers/nvme/target/pci-epf.c
2478
container_of(epf_group, struct nvmet_pci_epf, group)
drivers/nvme/target/pci-epf.c
2482
struct config_group *group = to_config_group(item);
drivers/nvme/target/pci-epf.c
2483
struct nvmet_pci_epf *nvme_epf = to_nvme_epf(group);
drivers/nvme/target/pci-epf.c
2491
struct config_group *group = to_config_group(item);
drivers/nvme/target/pci-epf.c
2492
struct nvmet_pci_epf *nvme_epf = to_nvme_epf(group);
drivers/nvme/target/pci-epf.c
2515
struct config_group *group = to_config_group(item);
drivers/nvme/target/pci-epf.c
2516
struct nvmet_pci_epf *nvme_epf = to_nvme_epf(group);
drivers/nvme/target/pci-epf.c
2524
struct config_group *group = to_config_group(item);
drivers/nvme/target/pci-epf.c
2525
struct nvmet_pci_epf *nvme_epf = to_nvme_epf(group);
drivers/nvme/target/pci-epf.c
2543
struct config_group *group = to_config_group(item);
drivers/nvme/target/pci-epf.c
2544
struct nvmet_pci_epf *nvme_epf = to_nvme_epf(group);
drivers/nvme/target/pci-epf.c
2552
struct config_group *group = to_config_group(item);
drivers/nvme/target/pci-epf.c
2553
struct nvmet_pci_epf *nvme_epf = to_nvme_epf(group);
drivers/nvme/target/pci-epf.c
2591
struct config_group *group)
drivers/nvme/target/pci-epf.c
2595
config_group_init_type_name(&nvme_epf->group, "nvme",
drivers/nvme/target/pci-epf.c
2598
return &nvme_epf->group;
drivers/nvmem/core.c
457
struct attribute_group group = {
drivers/nvmem/core.c
506
group.bin_attrs = pattrs;
drivers/nvmem/core.c
508
ret = device_add_group(&nvmem->dev, &group);
drivers/pci/endpoint/functions/pci-epf-ntb.c
1937
struct config_group *group = to_config_group(item); \
drivers/pci/endpoint/functions/pci-epf-ntb.c
1938
struct epf_ntb *ntb = to_epf_ntb(group); \
drivers/pci/endpoint/functions/pci-epf-ntb.c
1947
struct config_group *group = to_config_group(item); \
drivers/pci/endpoint/functions/pci-epf-ntb.c
1948
struct epf_ntb *ntb = to_epf_ntb(group); \
drivers/pci/endpoint/functions/pci-epf-ntb.c
1963
struct config_group *group = to_config_group(item); \
drivers/pci/endpoint/functions/pci-epf-ntb.c
1964
struct epf_ntb *ntb = to_epf_ntb(group); \
drivers/pci/endpoint/functions/pci-epf-ntb.c
1976
struct config_group *group = to_config_group(item); \
drivers/pci/endpoint/functions/pci-epf-ntb.c
1977
struct epf_ntb *ntb = to_epf_ntb(group); \
drivers/pci/endpoint/functions/pci-epf-ntb.c
2001
struct config_group *group = to_config_group(item);
drivers/pci/endpoint/functions/pci-epf-ntb.c
2002
struct epf_ntb *ntb = to_epf_ntb(group);
drivers/pci/endpoint/functions/pci-epf-ntb.c
2064
struct config_group *group)
drivers/pci/endpoint/functions/pci-epf-ntb.c
2067
struct config_group *ntb_group = &ntb->group;
drivers/pci/endpoint/functions/pci-epf-ntb.c
82
struct config_group group;
drivers/pci/endpoint/functions/pci-epf-ntb.c
86
#define to_epf_ntb(epf_group) container_of((epf_group), struct epf_ntb, group)
drivers/pci/endpoint/functions/pci-epf-test.c
1327
struct config_group *group = to_config_group(item); \
drivers/pci/endpoint/functions/pci-epf-test.c
1329
container_of(group, struct pci_epf_test, group); \
drivers/pci/endpoint/functions/pci-epf-test.c
1339
struct config_group *group = to_config_group(item); \
drivers/pci/endpoint/functions/pci-epf-test.c
1341
container_of(group, struct pci_epf_test, group); \
drivers/pci/endpoint/functions/pci-epf-test.c
1399
struct config_group *group)
drivers/pci/endpoint/functions/pci-epf-test.c
1402
struct config_group *epf_group = &epf_test->group;
drivers/pci/endpoint/functions/pci-epf-test.c
75
struct config_group group;
drivers/pci/endpoint/functions/pci-epf-vntb.c
1015
struct config_group *group = to_config_group(item); \
drivers/pci/endpoint/functions/pci-epf-vntb.c
1016
struct epf_ntb *ntb = to_epf_ntb(group); \
drivers/pci/endpoint/functions/pci-epf-vntb.c
1043
struct config_group *group = to_config_group(item); \
drivers/pci/endpoint/functions/pci-epf-vntb.c
1044
struct epf_ntb *ntb = to_epf_ntb(group); \
drivers/pci/endpoint/functions/pci-epf-vntb.c
1053
struct config_group *group = to_config_group(item); \
drivers/pci/endpoint/functions/pci-epf-vntb.c
1054
struct epf_ntb *ntb = to_epf_ntb(group); \
drivers/pci/endpoint/functions/pci-epf-vntb.c
1073
struct config_group *group = to_config_group(item);
drivers/pci/endpoint/functions/pci-epf-vntb.c
1074
struct epf_ntb *ntb = to_epf_ntb(group);
drivers/pci/endpoint/functions/pci-epf-vntb.c
1176
struct config_group *group)
drivers/pci/endpoint/functions/pci-epf-vntb.c
1179
struct config_group *ntb_group = &ntb->group;
drivers/pci/endpoint/functions/pci-epf-vntb.c
125
struct config_group group;
drivers/pci/endpoint/functions/pci-epf-vntb.c
152
#define to_epf_ntb(epf_group) container_of((epf_group), struct epf_ntb, group)
drivers/pci/endpoint/functions/pci-epf-vntb.c
967
struct config_group *group = to_config_group(item); \
drivers/pci/endpoint/functions/pci-epf-vntb.c
968
struct epf_ntb *ntb = to_epf_ntb(group); \
drivers/pci/endpoint/functions/pci-epf-vntb.c
977
struct config_group *group = to_config_group(item); \
drivers/pci/endpoint/functions/pci-epf-vntb.c
978
struct epf_ntb *ntb = to_epf_ntb(group); \
drivers/pci/endpoint/functions/pci-epf-vntb.c
995
struct config_group *group = to_config_group(item); \
drivers/pci/endpoint/functions/pci-epf-vntb.c
996
struct epf_ntb *ntb = to_epf_ntb(group); \
drivers/pci/endpoint/pci-ep-cfs.c
105
configfs_add_default_group(secondary_epc_group, &epf_group->group);
drivers/pci/endpoint/pci-ep-cfs.c
168
configfs_add_default_group(primary_epc_group, &epf_group->group);
drivers/pci/endpoint/pci-ep-cfs.c
23
struct config_group group;
drivers/pci/endpoint/pci-ep-cfs.c
274
struct config_group *group;
drivers/pci/endpoint/pci-ep-cfs.c
283
group = &epc_group->group;
drivers/pci/endpoint/pci-ep-cfs.c
285
config_group_init_type_name(group, name, &pci_epc_type);
drivers/pci/endpoint/pci-ep-cfs.c
286
ret = configfs_register_group(controllers_group, group);
drivers/pci/endpoint/pci-ep-cfs.c
300
return group;
drivers/pci/endpoint/pci-ep-cfs.c
303
configfs_unregister_group(group);
drivers/pci/endpoint/pci-ep-cfs.c
31
struct config_group group;
drivers/pci/endpoint/pci-ep-cfs.c
313
void pci_ep_cfs_remove_epc_group(struct config_group *group)
drivers/pci/endpoint/pci-ep-cfs.c
317
if (!group)
drivers/pci/endpoint/pci-ep-cfs.c
320
epc_group = container_of(group, struct pci_epc_group, group);
drivers/pci/endpoint/pci-ep-cfs.c
322
configfs_unregister_group(&epc_group->group);
drivers/pci/endpoint/pci-ep-cfs.c
38
return container_of(to_config_group(item), struct pci_epf_group, group);
drivers/pci/endpoint/pci-ep-cfs.c
43
return container_of(to_config_group(item), struct pci_epc_group, group);
drivers/pci/endpoint/pci-ep-cfs.c
539
struct config_group *group)
drivers/pci/endpoint/pci-ep-cfs.c
552
epf_type_group = epf->driver->ops->add_cfs(epf, group);
drivers/pci/endpoint/pci-ep-cfs.c
560
struct config_group *group;
drivers/pci/endpoint/pci-ep-cfs.c
562
group = pci_epf_type_add_cfs(epf_group->epf, &epf_group->group);
drivers/pci/endpoint/pci-ep-cfs.c
563
if (!group)
drivers/pci/endpoint/pci-ep-cfs.c
566
if (IS_ERR(group)) {
drivers/pci/endpoint/pci-ep-cfs.c
572
configfs_add_default_group(group, &epf_group->group);
drivers/pci/endpoint/pci-ep-cfs.c
577
struct config_group *group;
drivers/pci/endpoint/pci-ep-cfs.c
579
group = pci_ep_cfs_add_primary_group(epf_group);
drivers/pci/endpoint/pci-ep-cfs.c
580
if (IS_ERR(group)) {
drivers/pci/endpoint/pci-ep-cfs.c
585
group = pci_ep_cfs_add_secondary_group(epf_group);
drivers/pci/endpoint/pci-ep-cfs.c
586
if (IS_ERR(group)) {
drivers/pci/endpoint/pci-ep-cfs.c
594
static struct config_group *pci_epf_make(struct config_group *group,
drivers/pci/endpoint/pci-ep-cfs.c
616
config_group_init_type_name(&epf_group->group, name, &pci_epf_type);
drivers/pci/endpoint/pci-ep-cfs.c
619
group->cg_item.ci_name, epf_group->index);
drivers/pci/endpoint/pci-ep-cfs.c
632
epf->group = &epf_group->group;
drivers/pci/endpoint/pci-ep-cfs.c
639
return &epf_group->group;
drivers/pci/endpoint/pci-ep-cfs.c
655
static void pci_epf_drop(struct config_group *group, struct config_item *item)
drivers/pci/endpoint/pci-ep-cfs.c
672
struct config_group *group;
drivers/pci/endpoint/pci-ep-cfs.c
674
group = configfs_register_default_group(functions_group, name,
drivers/pci/endpoint/pci-ep-cfs.c
676
if (IS_ERR(group))
drivers/pci/endpoint/pci-ep-cfs.c
680
return group;
drivers/pci/endpoint/pci-ep-cfs.c
684
void pci_ep_cfs_remove_epf_group(struct config_group *group)
drivers/pci/endpoint/pci-ep-cfs.c
686
if (IS_ERR_OR_NULL(group))
drivers/pci/endpoint/pci-ep-cfs.c
689
list_del(&group->group_entry);
drivers/pci/endpoint/pci-ep-cfs.c
690
configfs_unregister_default_group(group);
drivers/pci/endpoint/pci-epc-core.c
1019
epc->group = pci_ep_cfs_add_epc_group(dev_name(dev));
drivers/pci/endpoint/pci-epc-core.c
952
pci_ep_cfs_remove_epc_group(epc->group);
drivers/pci/endpoint/pci-epf-core.c
428
struct config_group *group, *tmp;
drivers/pci/endpoint/pci-epf-core.c
434
list_for_each_entry_safe(group, tmp, &driver->epf_group, group_entry)
drivers/pci/endpoint/pci-epf-core.c
435
pci_ep_cfs_remove_epf_group(group);
drivers/pci/endpoint/pci-epf-core.c
455
struct config_group *group;
drivers/pci/endpoint/pci-epf-core.c
465
group = pci_ep_cfs_add_epf_group(id->name);
drivers/pci/endpoint/pci-epf-core.c
466
if (IS_ERR(group)) {
drivers/pci/endpoint/pci-epf-core.c
468
return PTR_ERR(group);
drivers/pci/endpoint/pci-epf-core.c
472
list_add_tail(&group->group_entry, &driver->epf_group);
drivers/perf/arm_v7_pmu.c
1353
unsigned int group = EVENT_GROUP(config_base);
drivers/perf/arm_v7_pmu.c
1358
group_shift = group * 8;
drivers/perf/arm_v7_pmu.c
1366
val += group;
drivers/perf/arm_v7_pmu.c
1388
static u32 clear_pmresrn_group(u32 val, int group)
drivers/perf/arm_v7_pmu.c
1393
group_shift = group * 8;
drivers/perf/arm_v7_pmu.c
1409
unsigned int group = EVENT_GROUP(config_base);
drivers/perf/arm_v7_pmu.c
1415
val = clear_pmresrn_group(val, group);
drivers/perf/arm_v7_pmu.c
1420
val = clear_pmresrn_group(val, group);
drivers/perf/arm_v7_pmu.c
1490
unsigned int group)
drivers/perf/arm_v7_pmu.c
1501
bit += group;
drivers/perf/arm_v7_pmu.c
1523
unsigned int group = EVENT_GROUP(hwc->config_base);
drivers/perf/arm_v7_pmu.c
1529
if (group > 3 || region > 2)
drivers/perf/arm_v7_pmu.c
1534
bit = krait_event_to_bit(event, region, group);
drivers/perf/arm_v7_pmu.c
1552
unsigned int group = EVENT_GROUP(hwc->config_base);
drivers/perf/arm_v7_pmu.c
1558
bit = krait_event_to_bit(event, region, group);
drivers/perf/arm_v7_pmu.c
1677
unsigned int group = EVENT_GROUP(config_base);
drivers/perf/arm_v7_pmu.c
1682
group_shift = group * 8;
drivers/perf/arm_v7_pmu.c
1690
val += group;
drivers/perf/arm_v7_pmu.c
1719
unsigned int group = EVENT_GROUP(config_base);
drivers/perf/arm_v7_pmu.c
1725
val = clear_pmresrn_group(val, group);
drivers/perf/arm_v7_pmu.c
1730
val = clear_pmresrn_group(val, group);
drivers/perf/arm_v7_pmu.c
1800
unsigned int group)
drivers/perf/arm_v7_pmu.c
1811
bit += group;
drivers/perf/arm_v7_pmu.c
1832
unsigned int group = EVENT_GROUP(hwc->config_base);
drivers/perf/arm_v7_pmu.c
1838
if (group > 3 || region > 3)
drivers/perf/arm_v7_pmu.c
1841
bit = scorpion_event_to_bit(event, region, group);
drivers/perf/arm_v7_pmu.c
1859
unsigned int group = EVENT_GROUP(hwc->config_base);
drivers/perf/arm_v7_pmu.c
1865
bit = scorpion_event_to_bit(event, region, group);
drivers/perf/dwc_pcie_pmu.c
383
int event_id, event_nr, group;
drivers/perf/dwc_pcie_pmu.c
387
group = FIELD_GET(DWC_PCIE_CNT_EVENT_SEL_GROUP, event_id);
drivers/perf/dwc_pcie_pmu.c
389
if (group != DWC_PCIE_LANE_GROUP_6 && group != DWC_PCIE_LANE_GROUP_7)
drivers/perf/dwc_pcie_pmu.c
392
group -= DWC_PCIE_LANE_GROUP_6;
drivers/perf/dwc_pcie_pmu.c
394
if (test_and_set_bit(group * DWC_PCIE_LANE_MAX_EVENTS_PER_GROUP + event_nr,
drivers/perf/dwc_pcie_pmu.c
521
int group = FIELD_GET(DWC_PCIE_CNT_EVENT_SEL_GROUP, event_id) -
drivers/perf/dwc_pcie_pmu.c
524
if (test_and_set_bit(group * DWC_PCIE_LANE_MAX_EVENTS_PER_GROUP + event_nr,
drivers/perf/dwc_pcie_pmu.c
575
int group = FIELD_GET(DWC_PCIE_CNT_EVENT_SEL_GROUP, event_id) -
drivers/perf/dwc_pcie_pmu.c
578
clear_bit(group * DWC_PCIE_LANE_MAX_EVENTS_PER_GROUP + event_nr,
drivers/perf/qcom_l2_pmu.c
345
unsigned int group;
drivers/perf/qcom_l2_pmu.c
364
group = L2_EVT_GROUP(hwc->config_base);
drivers/perf/qcom_l2_pmu.c
365
if (test_bit(group, cluster->used_groups))
drivers/perf/qcom_l2_pmu.c
369
set_bit(group, cluster->used_groups);
drivers/pinctrl/actions/pinctrl-owl.c
103
return pctrl->soc->groups[group].name;
drivers/pinctrl/actions/pinctrl-owl.c
107
unsigned int group,
drivers/pinctrl/actions/pinctrl-owl.c
113
*pins = pctrl->soc->groups[group].pads;
drivers/pinctrl/actions/pinctrl-owl.c
114
*num_pins = pctrl->soc->groups[group].npads;
drivers/pinctrl/actions/pinctrl-owl.c
194
unsigned int group)
drivers/pinctrl/actions/pinctrl-owl.c
201
g = &pctrl->soc->groups[group];
drivers/pinctrl/actions/pinctrl-owl.c
423
unsigned int group,
drivers/pinctrl/actions/pinctrl-owl.c
432
g = &pctrl->soc->groups[group];
drivers/pinctrl/actions/pinctrl-owl.c
450
unsigned int group,
drivers/pinctrl/actions/pinctrl-owl.c
461
g = &pctrl->soc->groups[group];
drivers/pinctrl/actions/pinctrl-owl.c
99
unsigned int group)
drivers/pinctrl/aspeed/pinctrl-aspeed.c
22
unsigned int group)
drivers/pinctrl/aspeed/pinctrl-aspeed.c
224
unsigned int group)
drivers/pinctrl/aspeed/pinctrl-aspeed.c
229
const struct aspeed_pin_group *pgroup = &pdata->pinmux.groups[group];
drivers/pinctrl/aspeed/pinctrl-aspeed.c
26
return pdata->pinmux.groups[group].name;
drivers/pinctrl/aspeed/pinctrl-aspeed.c
30
unsigned int group, const unsigned int **pins,
drivers/pinctrl/aspeed/pinctrl-aspeed.c
35
*pins = &pdata->pinmux.groups[group].pins[0];
drivers/pinctrl/aspeed/pinctrl-aspeed.c
36
*npins = pdata->pinmux.groups[group].npins;
drivers/pinctrl/aspeed/pinctrl-aspeed.h
100
unsigned int group);
drivers/pinctrl/aspeed/pinctrl-aspeed.h
87
unsigned int group);
drivers/pinctrl/aspeed/pinctrl-aspeed.h
89
unsigned int group, const unsigned int **pins,
drivers/pinctrl/aspeed/pinmux-aspeed.h
513
#define SIG_DESC_LIST_SYM(sig, group) sig_descs_ ## sig ## _ ## group
drivers/pinctrl/aspeed/pinmux-aspeed.h
514
#define SIG_DESC_LIST_DECL(sig, group, ...) \
drivers/pinctrl/aspeed/pinmux-aspeed.h
515
static const struct aspeed_sig_desc SIG_DESC_LIST_SYM(sig, group)[] = \
drivers/pinctrl/aspeed/pinmux-aspeed.h
518
#define SIG_EXPR_SYM(sig, group) sig_expr_ ## sig ## _ ## group
drivers/pinctrl/aspeed/pinmux-aspeed.h
519
#define SIG_EXPR_DECL_(sig, group, func) \
drivers/pinctrl/aspeed/pinmux-aspeed.h
520
static const struct aspeed_sig_expr SIG_EXPR_SYM(sig, group) = \
drivers/pinctrl/aspeed/pinmux-aspeed.h
524
.ndescs = ARRAY_SIZE(SIG_DESC_LIST_SYM(sig, group)), \
drivers/pinctrl/aspeed/pinmux-aspeed.h
525
.descs = &(SIG_DESC_LIST_SYM(sig, group))[0], \
drivers/pinctrl/aspeed/pinmux-aspeed.h
545
#define SIG_EXPR_DECL(sig, group, func, ...) \
drivers/pinctrl/aspeed/pinmux-aspeed.h
546
SIG_DESC_LIST_DECL(sig, group, __VA_ARGS__); \
drivers/pinctrl/aspeed/pinmux-aspeed.h
547
SIG_EXPR_DECL_(sig, group, func)
drivers/pinctrl/aspeed/pinmux-aspeed.h
555
#define SIG_EXPR_PTR(sig, group) (&SIG_EXPR_SYM(sig, group))
drivers/pinctrl/aspeed/pinmux-aspeed.h
557
#define SIG_EXPR_LIST_SYM(sig, group) sig_exprs_ ## sig ## _ ## group
drivers/pinctrl/aspeed/pinmux-aspeed.h
574
#define SIG_EXPR_LIST_DECL(sig, group, ...) \
drivers/pinctrl/aspeed/pinmux-aspeed.h
575
static const struct aspeed_sig_expr *SIG_EXPR_LIST_SYM(sig, group)[] =\
drivers/pinctrl/aspeed/pinmux-aspeed.h
594
#define SIG_EXPR_LIST_ALIAS(pin, sig, group) \
drivers/pinctrl/aspeed/pinmux-aspeed.h
596
SIG_EXPR_LIST_SYM(pin, sig)[ARRAY_SIZE(SIG_EXPR_LIST_SYM(sig, group))] \
drivers/pinctrl/aspeed/pinmux-aspeed.h
597
__attribute__((alias(istringify(SIG_EXPR_LIST_SYM(sig, group)))))
drivers/pinctrl/aspeed/pinmux-aspeed.h
628
#define SIG_EXPR_LIST_DECL_SEMG(pin, sig, group, func, ...) \
drivers/pinctrl/aspeed/pinmux-aspeed.h
629
SIG_DESC_LIST_DECL(sig, group, __VA_ARGS__); \
drivers/pinctrl/aspeed/pinmux-aspeed.h
630
SIG_EXPR_DECL_(sig, group, func); \
drivers/pinctrl/aspeed/pinmux-aspeed.h
631
SIG_EXPR_LIST_DECL(sig, group, SIG_EXPR_PTR(sig, group)); \
drivers/pinctrl/aspeed/pinmux-aspeed.h
632
SIG_EXPR_LIST_ALIAS(pin, sig, group)
drivers/pinctrl/aspeed/pinmux-aspeed.h
650
#define SIG_EXPR_LIST_PTR(sig, group) SIG_EXPR_LIST_SYM(sig, group)
drivers/pinctrl/aspeed/pinmux-aspeed.h
741
#define GROUP_SYM(group) group_pins_ ## group
drivers/pinctrl/aspeed/pinmux-aspeed.h
742
#define GROUP_DECL(group, ...) \
drivers/pinctrl/aspeed/pinmux-aspeed.h
743
static const int GROUP_SYM(group)[] = { __VA_ARGS__ }
drivers/pinctrl/aspeed/pinmux-aspeed.h
749
#define FUNC_DECL_1(func, group) FUNC_DECL_(func, #group)
drivers/pinctrl/bcm/pinctrl-bcm281xx.c
1584
unsigned int group)
drivers/pinctrl/bcm/pinctrl-bcm281xx.c
1588
return pdata->info->pins[group].name;
drivers/pinctrl/bcm/pinctrl-bcm281xx.c
1592
unsigned int group,
drivers/pinctrl/bcm/pinctrl-bcm281xx.c
1598
*pins = &pdata->info->pins[group].number;
drivers/pinctrl/bcm/pinctrl-bcm281xx.c
1650
unsigned int group)
drivers/pinctrl/bcm/pinctrl-bcm281xx.c
1655
unsigned int pin = pdata->info->pins[group].number;
drivers/pinctrl/bcm/pinctrl-bcm281xx.c
1661
__func__, f->name, function, pdata->info->pins[group].name,
drivers/pinctrl/bcm/pinctrl-bcm281xx.c
1678
pdata->info->pins[group].name, pin);
drivers/pinctrl/bcm/pinctrl-bcm2835.c
448
int group = 0;
drivers/pinctrl/bcm/pinctrl-bcm2835.c
453
group = i;
drivers/pinctrl/bcm/pinctrl-bcm2835.c
462
switch (group) {
drivers/pinctrl/bcm/pinctrl-bcm2835.c
787
map->data.mux.group = bcm2835_gpio_groups[pin];
drivers/pinctrl/bcm/pinctrl-bcm4908.c
422
const struct bcm4908_pinctrl_grp *group;
drivers/pinctrl/bcm/pinctrl-bcm4908.c
429
group = group_desc->data;
drivers/pinctrl/bcm/pinctrl-bcm4908.c
432
for (i = 0; i < group->num_pins; i++) {
drivers/pinctrl/bcm/pinctrl-bcm4908.c
435
lsb |= group->pins[i].number;
drivers/pinctrl/bcm/pinctrl-bcm4908.c
436
lsb |= group->pins[i].function << BCM4908_TEST_PORT_LSB_PINMUX_DATA_SHIFT;
drivers/pinctrl/bcm/pinctrl-bcm4908.c
523
const struct bcm4908_pinctrl_grp *group = &bcm4908_pinctrl_grps[i];
drivers/pinctrl/bcm/pinctrl-bcm4908.c
527
pins = devm_kcalloc(dev, group->num_pins, sizeof(*pins), GFP_KERNEL);
drivers/pinctrl/bcm/pinctrl-bcm4908.c
530
for (j = 0; j < group->num_pins; j++)
drivers/pinctrl/bcm/pinctrl-bcm4908.c
531
pins[j] = group->pins[j].number;
drivers/pinctrl/bcm/pinctrl-bcm4908.c
533
pinctrl_generic_add_group(bcm4908_pinctrl->pctldev, group->name,
drivers/pinctrl/bcm/pinctrl-bcm4908.c
534
pins, group->num_pins, (void *)group);
drivers/pinctrl/bcm/pinctrl-bcm6318.c
351
unsigned group)
drivers/pinctrl/bcm/pinctrl-bcm6318.c
353
return bcm6318_groups[group].name;
drivers/pinctrl/bcm/pinctrl-bcm6318.c
357
unsigned group, const unsigned **pins,
drivers/pinctrl/bcm/pinctrl-bcm6318.c
360
*pins = bcm6318_groups[group].pins;
drivers/pinctrl/bcm/pinctrl-bcm6318.c
361
*npins = bcm6318_groups[group].npins;
drivers/pinctrl/bcm/pinctrl-bcm6318.c
411
unsigned selector, unsigned group)
drivers/pinctrl/bcm/pinctrl-bcm6318.c
414
const struct pingroup *pg = &bcm6318_groups[group];
drivers/pinctrl/bcm/pinctrl-bcm63268.c
469
unsigned group)
drivers/pinctrl/bcm/pinctrl-bcm63268.c
471
return bcm63268_groups[group].name;
drivers/pinctrl/bcm/pinctrl-bcm63268.c
475
unsigned group,
drivers/pinctrl/bcm/pinctrl-bcm63268.c
479
*pins = bcm63268_groups[group].pins;
drivers/pinctrl/bcm/pinctrl-bcm63268.c
480
*npins = bcm63268_groups[group].npins;
drivers/pinctrl/bcm/pinctrl-bcm63268.c
532
unsigned selector, unsigned group)
drivers/pinctrl/bcm/pinctrl-bcm63268.c
535
const struct pingroup *pg = &bcm63268_groups[group];
drivers/pinctrl/bcm/pinctrl-bcm6328.c
275
unsigned group)
drivers/pinctrl/bcm/pinctrl-bcm6328.c
277
return bcm6328_groups[group].name;
drivers/pinctrl/bcm/pinctrl-bcm6328.c
281
unsigned group, const unsigned **pins,
drivers/pinctrl/bcm/pinctrl-bcm6328.c
284
*pins = bcm6328_groups[group].pins;
drivers/pinctrl/bcm/pinctrl-bcm6328.c
285
*npins = bcm6328_groups[group].npins;
drivers/pinctrl/bcm/pinctrl-bcm6328.c
325
unsigned selector, unsigned group)
drivers/pinctrl/bcm/pinctrl-bcm6328.c
328
const struct pingroup *pg = &bcm6328_groups[group];
drivers/pinctrl/bcm/pinctrl-bcm6358.c
216
unsigned group)
drivers/pinctrl/bcm/pinctrl-bcm6358.c
218
return bcm6358_groups[group].grp.name;
drivers/pinctrl/bcm/pinctrl-bcm6358.c
222
unsigned group, const unsigned **pins,
drivers/pinctrl/bcm/pinctrl-bcm6358.c
225
*pins = bcm6358_groups[group].grp.pins;
drivers/pinctrl/bcm/pinctrl-bcm6358.c
226
*npins = bcm6358_groups[group].grp.npins;
drivers/pinctrl/bcm/pinctrl-bcm6358.c
254
unsigned selector, unsigned group)
drivers/pinctrl/bcm/pinctrl-bcm6358.c
258
const struct bcm6358_pingroup *pg = &bcm6358_groups[group];
drivers/pinctrl/bcm/pinctrl-bcm6362.c
446
unsigned group)
drivers/pinctrl/bcm/pinctrl-bcm6362.c
448
return bcm6362_groups[group].name;
drivers/pinctrl/bcm/pinctrl-bcm6362.c
452
unsigned group, const unsigned **pins,
drivers/pinctrl/bcm/pinctrl-bcm6362.c
455
*pins = bcm6362_groups[group].pins;
drivers/pinctrl/bcm/pinctrl-bcm6362.c
456
*npins = bcm6362_groups[group].npins;
drivers/pinctrl/bcm/pinctrl-bcm6362.c
506
unsigned selector, unsigned group)
drivers/pinctrl/bcm/pinctrl-bcm6362.c
509
const struct pingroup *pg = &bcm6362_groups[group];
drivers/pinctrl/bcm/pinctrl-bcm6368.c
341
unsigned group)
drivers/pinctrl/bcm/pinctrl-bcm6368.c
343
return bcm6368_groups[group].name;
drivers/pinctrl/bcm/pinctrl-bcm6368.c
347
unsigned group, const unsigned **pins,
drivers/pinctrl/bcm/pinctrl-bcm6368.c
350
*pins = bcm6368_groups[group].pins;
drivers/pinctrl/bcm/pinctrl-bcm6368.c
351
*npins = bcm6368_groups[group].npins;
drivers/pinctrl/bcm/pinctrl-bcm6368.c
379
unsigned selector, unsigned group)
drivers/pinctrl/bcm/pinctrl-bcm6368.c
383
const struct pingroup *pg = &bcm6368_groups[group];
drivers/pinctrl/bcm/pinctrl-ns.c
165
struct group_desc *group;
drivers/pinctrl/bcm/pinctrl-ns.c
170
group = pinctrl_generic_get_group(pctrl_dev, group_selector);
drivers/pinctrl/bcm/pinctrl-ns.c
171
if (!group)
drivers/pinctrl/bcm/pinctrl-ns.c
174
for (i = 0; i < group->grp.npins; i++)
drivers/pinctrl/bcm/pinctrl-ns.c
175
unset |= BIT(group->grp.pins[i]);
drivers/pinctrl/bcm/pinctrl-ns.c
264
const struct ns_pinctrl_group *group = &ns_pinctrl_groups[i];
drivers/pinctrl/bcm/pinctrl-ns.c
266
if (!(group->chipsets & ns_pinctrl->chipset_flag))
drivers/pinctrl/bcm/pinctrl-ns.c
269
pinctrl_generic_add_group(ns_pinctrl->pctldev, group->name,
drivers/pinctrl/bcm/pinctrl-ns.c
270
group->pins, group->num_pins, NULL);
drivers/pinctrl/berlin/berlin.c
135
const struct berlin_desc_group *group,
drivers/pinctrl/berlin/berlin.c
138
struct berlin_desc_function *function = group->functions;
drivers/pinctrl/berlin/berlin.c
152
unsigned group)
drivers/pinctrl/berlin/berlin.c
155
const struct berlin_desc_group *group_desc = pctrl->desc->groups + group;
drivers/pinctrl/berlin/berlin.c
43
unsigned group)
drivers/pinctrl/berlin/berlin.c
47
return pctrl->desc->groups[group].name;
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
1004
group->name, master ? "master" : "slave");
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
1009
group->name, ret);
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
866
const struct lochnagar_group *group,
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
870
const struct lochnagar_aif *aif = group->priv;
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
877
group->name, ret);
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
885
group->name, ret);
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
889
for (i = 0; i < group->npins; i++) {
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
890
pin = priv->pins[group->pins[i]].drv_data;
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
915
const struct lochnagar_group *group = &priv->groups[group_idx];
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
921
group->name, func->name);
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
923
return lochnagar_aif_set_mux(priv, group, func->op);
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
925
pin = priv->pins[*group->pins].drv_data;
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
992
const struct lochnagar_group *group = &priv->groups[group_idx];
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
993
const struct lochnagar_aif *aif = group->priv;
drivers/pinctrl/cirrus/pinctrl-lochnagar.c
997
if (group->type != LN_FTYPE_AIF)
drivers/pinctrl/cirrus/pinctrl-madera-core.c
612
unsigned int group)
drivers/pinctrl/cirrus/pinctrl-madera-core.c
623
__func__, selector, func_name, group,
drivers/pinctrl/cirrus/pinctrl-madera-core.c
624
madera_get_group_name(pctldev, group));
drivers/pinctrl/cirrus/pinctrl-madera-core.c
656
group -= n_chip_groups;
drivers/pinctrl/cirrus/pinctrl-madera-core.c
657
reg = MADERA_GPIO1_CTRL_1 + (2 * group);
drivers/pinctrl/cix/pinctrl-sky1-base.c
248
unsigned int group)
drivers/pinctrl/cix/pinctrl-sky1-base.c
252
return spctl->groups[group].name;
drivers/pinctrl/cix/pinctrl-sky1-base.c
256
unsigned int group,
drivers/pinctrl/cix/pinctrl-sky1-base.c
262
*pins = (unsigned int *)&spctl->groups[group].pin;
drivers/pinctrl/cix/pinctrl-sky1-base.c
302
unsigned int group)
drivers/pinctrl/cix/pinctrl-sky1-base.c
307
(struct sky1_pinctrl_group *)spctl->groups + group;
drivers/pinctrl/cix/pinctrl-sky1-base.c
312
function, group);
drivers/pinctrl/cix/pinctrl-sky1-base.c
439
unsigned int group,
drivers/pinctrl/cix/pinctrl-sky1-base.c
443
struct sky1_pinctrl_group *g = &spctl->groups[group];
drivers/pinctrl/cix/pinctrl-sky1-base.c
450
static int sky1_pconf_group_set(struct pinctrl_dev *pctldev, unsigned int group,
drivers/pinctrl/cix/pinctrl-sky1-base.c
454
struct sky1_pinctrl_group *g = &spctl->groups[group];
drivers/pinctrl/cix/pinctrl-sky1-base.c
495
struct sky1_pinctrl_group *group =
drivers/pinctrl/cix/pinctrl-sky1-base.c
498
group->name = pin->pin.name;
drivers/pinctrl/cix/pinctrl-sky1-base.c
499
group->pin = pin->pin.number;
drivers/pinctrl/cix/pinctrl-sky1-base.c
80
(*map)[*num_maps].data.mux.group = grp->name;
drivers/pinctrl/core.c
561
struct group_desc *group;
drivers/pinctrl/core.c
563
group = radix_tree_lookup(&pctldev->pin_group_tree,
drivers/pinctrl/core.c
565
if (!group)
drivers/pinctrl/core.c
568
return group->grp.name;
drivers/pinctrl/core.c
584
struct group_desc *group;
drivers/pinctrl/core.c
586
group = radix_tree_lookup(&pctldev->pin_group_tree,
drivers/pinctrl/core.c
588
if (!group) {
drivers/pinctrl/core.c
594
*pins = group->grp.pins;
drivers/pinctrl/core.c
595
*num_pins = group->grp.npins;
drivers/pinctrl/core.c
609
struct group_desc *group;
drivers/pinctrl/core.c
611
group = radix_tree_lookup(&pctldev->pin_group_tree,
drivers/pinctrl/core.c
613
if (!group)
drivers/pinctrl/core.c
616
return group;
drivers/pinctrl/core.c
653
struct group_desc *group;
drivers/pinctrl/core.c
665
group = devm_kzalloc(pctldev->dev, sizeof(*group), GFP_KERNEL);
drivers/pinctrl/core.c
666
if (!group)
drivers/pinctrl/core.c
669
*group = PINCTRL_GROUP_DESC(name, pins, num_pins, data);
drivers/pinctrl/core.c
671
error = radix_tree_insert(&pctldev->pin_group_tree, selector, group);
drivers/pinctrl/core.c
691
struct group_desc *group;
drivers/pinctrl/core.c
693
group = radix_tree_lookup(&pctldev->pin_group_tree,
drivers/pinctrl/core.c
695
if (!group)
drivers/pinctrl/core.c
699
devm_kfree(pctldev->dev, group);
drivers/pinctrl/core.h
114
unsigned int group;
drivers/pinctrl/freescale/pinctrl-imx-scmi.c
172
unsigned int selector, unsigned int group)
drivers/pinctrl/freescale/pinctrl-imx.c
106
new_map[0].data.mux.group = np->name;
drivers/pinctrl/freescale/pinctrl-imx.c
144
(*map)->data.mux.function, (*map)->data.mux.group, map_num);
drivers/pinctrl/freescale/pinctrl-imx.c
243
unsigned group)
drivers/pinctrl/freescale/pinctrl-imx.c
257
grp = pinctrl_generic_get_group(pctldev, group);
drivers/pinctrl/freescale/pinctrl-imx.c
410
struct seq_file *s, unsigned group)
drivers/pinctrl/freescale/pinctrl-imx.c
417
if (group >= pctldev->num_groups)
drivers/pinctrl/freescale/pinctrl-imx.c
421
grp = pinctrl_generic_get_group(pctldev, group);
drivers/pinctrl/freescale/pinctrl-imx1-core.c
262
new_map[0].data.mux.group = np->name;
drivers/pinctrl/freescale/pinctrl-imx1-core.c
277
(*map)->data.mux.function, (*map)->data.mux.group, map_num);
drivers/pinctrl/freescale/pinctrl-imx1-core.c
298
unsigned group)
drivers/pinctrl/freescale/pinctrl-imx1-core.c
310
pins = info->groups[group].pins;
drivers/pinctrl/freescale/pinctrl-imx1-core.c
311
npins = info->groups[group].npins;
drivers/pinctrl/freescale/pinctrl-imx1-core.c
316
info->functions[selector].name, info->groups[group].name);
drivers/pinctrl/freescale/pinctrl-imx1-core.c
424
struct seq_file *s, unsigned group)
drivers/pinctrl/freescale/pinctrl-imx1-core.c
433
if (group >= info->ngroups)
drivers/pinctrl/freescale/pinctrl-imx1-core.c
437
grp = &info->groups[group];
drivers/pinctrl/freescale/pinctrl-mxs.c
104
group = kzalloc(length, GFP_KERNEL);
drivers/pinctrl/freescale/pinctrl-mxs.c
105
if (!group) {
drivers/pinctrl/freescale/pinctrl-mxs.c
109
snprintf(group, length, "%s.%d", np->name, reg);
drivers/pinctrl/freescale/pinctrl-mxs.c
110
new_map[i].data.mux.group = group;
drivers/pinctrl/freescale/pinctrl-mxs.c
123
group;
drivers/pinctrl/freescale/pinctrl-mxs.c
135
kfree(group);
drivers/pinctrl/freescale/pinctrl-mxs.c
148
kfree(map[i].data.mux.group);
drivers/pinctrl/freescale/pinctrl-mxs.c
181
unsigned group,
drivers/pinctrl/freescale/pinctrl-mxs.c
187
*groups = d->soc->functions[group].groups;
drivers/pinctrl/freescale/pinctrl-mxs.c
188
*num_groups = d->soc->functions[group].ngroups;
drivers/pinctrl/freescale/pinctrl-mxs.c
204
unsigned group)
drivers/pinctrl/freescale/pinctrl-mxs.c
207
struct mxs_group *g = &d->soc->groups[group];
drivers/pinctrl/freescale/pinctrl-mxs.c
247
unsigned group, unsigned long *config)
drivers/pinctrl/freescale/pinctrl-mxs.c
251
*config = d->soc->groups[group].config;
drivers/pinctrl/freescale/pinctrl-mxs.c
257
unsigned group, unsigned long *configs,
drivers/pinctrl/freescale/pinctrl-mxs.c
261
struct mxs_group *g = &d->soc->groups[group];
drivers/pinctrl/freescale/pinctrl-mxs.c
326
struct seq_file *s, unsigned group)
drivers/pinctrl/freescale/pinctrl-mxs.c
330
if (!mxs_pinconf_group_get(pctldev, group, &config))
drivers/pinctrl/freescale/pinctrl-mxs.c
358
char *group;
drivers/pinctrl/freescale/pinctrl-mxs.c
362
group = devm_kzalloc(&pdev->dev, length, GFP_KERNEL);
drivers/pinctrl/freescale/pinctrl-mxs.c
363
if (!group)
drivers/pinctrl/freescale/pinctrl-mxs.c
366
snprintf(group, length, "%s", np->name);
drivers/pinctrl/freescale/pinctrl-mxs.c
368
snprintf(group, length, "%s.%d", np->name, val);
drivers/pinctrl/freescale/pinctrl-mxs.c
369
g->name = group;
drivers/pinctrl/freescale/pinctrl-mxs.c
39
unsigned group)
drivers/pinctrl/freescale/pinctrl-mxs.c
43
return d->soc->groups[group].name;
drivers/pinctrl/freescale/pinctrl-mxs.c
46
static int mxs_get_group_pins(struct pinctrl_dev *pctldev, unsigned group,
drivers/pinctrl/freescale/pinctrl-mxs.c
51
*pins = d->soc->groups[group].pins;
drivers/pinctrl/freescale/pinctrl-mxs.c
52
*num_pins = d->soc->groups[group].npins;
drivers/pinctrl/freescale/pinctrl-mxs.c
68
char *group = NULL;
drivers/pinctrl/intel/pinctrl-baytrail.c
611
const struct intel_pingroup group,
drivers/pinctrl/intel/pinctrl-baytrail.c
618
for (i = 0; i < group.grp.npins; i++) {
drivers/pinctrl/intel/pinctrl-baytrail.c
622
padcfg0 = byt_gpio_reg(vg, group.grp.pins[i], BYT_CONF0_REG);
drivers/pinctrl/intel/pinctrl-baytrail.c
625
group.grp.name, i);
drivers/pinctrl/intel/pinctrl-baytrail.c
637
const struct intel_pingroup group,
drivers/pinctrl/intel/pinctrl-baytrail.c
644
for (i = 0; i < group.grp.npins; i++) {
drivers/pinctrl/intel/pinctrl-baytrail.c
648
padcfg0 = byt_gpio_reg(vg, group.grp.pins[i], BYT_CONF0_REG);
drivers/pinctrl/intel/pinctrl-baytrail.c
651
group.grp.name, i);
drivers/pinctrl/intel/pinctrl-baytrail.c
667
const struct intel_pingroup group = vg->soc->groups[group_selector];
drivers/pinctrl/intel/pinctrl-baytrail.c
669
if (group.modes)
drivers/pinctrl/intel/pinctrl-baytrail.c
670
byt_set_group_mixed_mux(vg, group, group.modes);
drivers/pinctrl/intel/pinctrl-baytrail.c
672
byt_set_group_simple_mux(vg, group, BYT_DEFAULT_GPIO_MUX);
drivers/pinctrl/intel/pinctrl-baytrail.c
674
byt_set_group_simple_mux(vg, group, group.mode);
drivers/pinctrl/intel/pinctrl-cherryview.c
1039
unsigned int group,
drivers/pinctrl/intel/pinctrl-cherryview.c
1046
ret = intel_get_group_pins(pctldev, group, &pins, &npins);
drivers/pinctrl/intel/pinctrl-cherryview.c
1058
unsigned int group, unsigned long *configs,
drivers/pinctrl/intel/pinctrl-cherryview.c
1065
ret = intel_get_group_pins(pctldev, group, &pins, &npins);
drivers/pinctrl/intel/pinctrl-cherryview.c
656
unsigned int function, unsigned int group)
drivers/pinctrl/intel/pinctrl-cherryview.c
663
grp = &pctrl->soc->groups[group];
drivers/pinctrl/intel/pinctrl-intel-platform.c
132
group = 0;
drivers/pinctrl/intel/pinctrl-intel-platform.c
134
struct intel_padgroup *gpp = &gpps[group];
drivers/pinctrl/intel/pinctrl-intel-platform.c
136
gpp->reg_num = group;
drivers/pinctrl/intel/pinctrl-intel-platform.c
142
group++;
drivers/pinctrl/intel/pinctrl-intel-platform.c
94
unsigned int group;
drivers/pinctrl/intel/pinctrl-intel.c
321
const char *intel_get_group_name(struct pinctrl_dev *pctldev, unsigned int group)
drivers/pinctrl/intel/pinctrl-intel.c
325
return pctrl->soc->groups[group].grp.name;
drivers/pinctrl/intel/pinctrl-intel.c
329
int intel_get_group_pins(struct pinctrl_dev *pctldev, unsigned int group,
drivers/pinctrl/intel/pinctrl-intel.c
334
*pins = pctrl->soc->groups[group].grp.pins;
drivers/pinctrl/intel/pinctrl-intel.c
335
*npins = pctrl->soc->groups[group].grp.npins;
drivers/pinctrl/intel/pinctrl-intel.c
426
unsigned int function, unsigned int group)
drivers/pinctrl/intel/pinctrl-intel.c
429
const struct intel_pingroup *grp = &pctrl->soc->groups[group];
drivers/pinctrl/intel/pinctrl-intel.h
282
const char *intel_get_group_name(struct pinctrl_dev *pctldev, unsigned int group);
drivers/pinctrl/intel/pinctrl-intel.h
283
int intel_get_group_pins(struct pinctrl_dev *pctldev, unsigned int group,
drivers/pinctrl/intel/pinctrl-lynxpoint.c
299
unsigned int function, unsigned int group)
drivers/pinctrl/intel/pinctrl-lynxpoint.c
302
const struct intel_pingroup *grp = &lg->soc->groups[group];
drivers/pinctrl/intel/pinctrl-tangier.c
143
unsigned int group)
drivers/pinctrl/intel/pinctrl-tangier.c
147
return tp->groups[group].grp.name;
drivers/pinctrl/intel/pinctrl-tangier.c
150
static int tng_get_group_pins(struct pinctrl_dev *pctldev, unsigned int group,
drivers/pinctrl/intel/pinctrl-tangier.c
155
*pins = tp->groups[group].grp.pins;
drivers/pinctrl/intel/pinctrl-tangier.c
156
*npins = tp->groups[group].grp.npins;
drivers/pinctrl/intel/pinctrl-tangier.c
218
unsigned int group)
drivers/pinctrl/intel/pinctrl-tangier.c
221
const struct intel_pingroup *grp = &tp->groups[group];
drivers/pinctrl/intel/pinctrl-tangier.c
472
unsigned int group, unsigned long *config)
drivers/pinctrl/intel/pinctrl-tangier.c
478
ret = tng_get_group_pins(pctldev, group, &pins, &npins);
drivers/pinctrl/intel/pinctrl-tangier.c
486
unsigned int group, unsigned long *configs,
drivers/pinctrl/intel/pinctrl-tangier.c
493
ret = tng_get_group_pins(pctldev, group, &pins, &npins);
drivers/pinctrl/mediatek/pinctrl-airoha.c
2459
unsigned int group)
drivers/pinctrl/mediatek/pinctrl-airoha.c
2471
grp = pinctrl_generic_get_group(pctrl_dev, group);
drivers/pinctrl/mediatek/pinctrl-airoha.c
2480
const struct airoha_pinctrl_func_group *group;
drivers/pinctrl/mediatek/pinctrl-airoha.c
2483
group = &func->groups[i];
drivers/pinctrl/mediatek/pinctrl-airoha.c
2484
if (strcmp(group->name, grp->grp.name))
drivers/pinctrl/mediatek/pinctrl-airoha.c
2487
for (j = 0; j < group->regmap_size; j++) {
drivers/pinctrl/mediatek/pinctrl-airoha.c
2488
switch (group->regmap[j].mux) {
drivers/pinctrl/mediatek/pinctrl-airoha.c
2492
group->regmap[j].offset,
drivers/pinctrl/mediatek/pinctrl-airoha.c
2493
group->regmap[j].mask,
drivers/pinctrl/mediatek/pinctrl-airoha.c
2494
group->regmap[j].val);
drivers/pinctrl/mediatek/pinctrl-airoha.c
2498
group->regmap[j].offset,
drivers/pinctrl/mediatek/pinctrl-airoha.c
2499
group->regmap[j].mask,
drivers/pinctrl/mediatek/pinctrl-airoha.c
2500
group->regmap[j].val);
drivers/pinctrl/mediatek/pinctrl-airoha.c
2811
unsigned int group, unsigned long *config)
drivers/pinctrl/mediatek/pinctrl-airoha.c
2817
for (i = 0; i < pinctrl->grps[group].npins; i++) {
drivers/pinctrl/mediatek/pinctrl-airoha.c
2819
pinctrl->grps[group].pins[i],
drivers/pinctrl/mediatek/pinctrl-airoha.c
2833
unsigned int group, unsigned long *configs,
drivers/pinctrl/mediatek/pinctrl-airoha.c
2839
for (i = 0; i < pinctrl->grps[group].npins; i++) {
drivers/pinctrl/mediatek/pinctrl-airoha.c
2843
pinctrl->grps[group].pins[i],
drivers/pinctrl/mediatek/pinctrl-moore.c
402
unsigned int group, unsigned long *config)
drivers/pinctrl/mediatek/pinctrl-moore.c
408
ret = pinctrl_generic_get_group_pins(pctldev, group, &pins, &npins);
drivers/pinctrl/mediatek/pinctrl-moore.c
427
unsigned int group, unsigned long *configs,
drivers/pinctrl/mediatek/pinctrl-moore.c
43
unsigned int selector, unsigned int group)
drivers/pinctrl/mediatek/pinctrl-moore.c
434
ret = pinctrl_generic_get_group_pins(pctldev, group, &pins, &npins);
drivers/pinctrl/mediatek/pinctrl-moore.c
54
grp = pinctrl_generic_get_group(pctldev, group);
drivers/pinctrl/mediatek/pinctrl-moore.c
606
const struct group_desc *group = hw->soc->grps + i;
drivers/pinctrl/mediatek/pinctrl-moore.c
607
const struct pingroup *grp = &group->grp;
drivers/pinctrl/mediatek/pinctrl-moore.c
610
group->data);
drivers/pinctrl/mediatek/pinctrl-mtk-common.c
406
unsigned group,
drivers/pinctrl/mediatek/pinctrl-mtk-common.c
411
*config = pctl->groups[group].config;
drivers/pinctrl/mediatek/pinctrl-mtk-common.c
416
static int mtk_pconf_group_set(struct pinctrl_dev *pctldev, unsigned group,
drivers/pinctrl/mediatek/pinctrl-mtk-common.c
420
struct mtk_pinctrl_group *g = &pctl->groups[group];
drivers/pinctrl/mediatek/pinctrl-mtk-common.c
507
(*map)[*num_maps].data.mux.group = grp->name;
drivers/pinctrl/mediatek/pinctrl-mtk-common.c
651
unsigned group)
drivers/pinctrl/mediatek/pinctrl-mtk-common.c
655
return pctl->groups[group].name;
drivers/pinctrl/mediatek/pinctrl-mtk-common.c
659
unsigned group,
drivers/pinctrl/mediatek/pinctrl-mtk-common.c
665
*pins = (unsigned *)&pctl->groups[group].pin;
drivers/pinctrl/mediatek/pinctrl-mtk-common.c
744
unsigned group)
drivers/pinctrl/mediatek/pinctrl-mtk-common.c
749
struct mtk_pinctrl_group *g = pctl->groups + group;
drivers/pinctrl/mediatek/pinctrl-mtk-common.c
754
function, group);
drivers/pinctrl/mediatek/pinctrl-mtk-common.c
945
struct mtk_pinctrl_group *group = pctl->groups + i;
drivers/pinctrl/mediatek/pinctrl-mtk-common.c
947
group->name = pin->pin.name;
drivers/pinctrl/mediatek/pinctrl-mtk-common.c
948
group->pin = pin->pin.number;
drivers/pinctrl/mediatek/pinctrl-mtmips.c
118
unsigned int func, unsigned int group)
drivers/pinctrl/mediatek/pinctrl-mtmips.c
127
if (p->groups[group].enabled) {
drivers/pinctrl/mediatek/pinctrl-mtmips.c
129
p->groups[group].name);
drivers/pinctrl/mediatek/pinctrl-mtmips.c
133
p->groups[group].enabled = 1;
drivers/pinctrl/mediatek/pinctrl-mtmips.c
136
shift = p->groups[group].shift;
drivers/pinctrl/mediatek/pinctrl-mtmips.c
142
mode &= ~(p->groups[group].mask << shift);
drivers/pinctrl/mediatek/pinctrl-mtmips.c
145
for (i = 0; i < p->groups[group].func[0].pin_count; i++)
drivers/pinctrl/mediatek/pinctrl-mtmips.c
146
p->gpio[p->groups[group].func[0].pins[i]] = 1;
drivers/pinctrl/mediatek/pinctrl-mtmips.c
150
mode |= p->groups[group].gpio << shift;
drivers/pinctrl/mediatek/pinctrl-mtmips.c
54
unsigned int group)
drivers/pinctrl/mediatek/pinctrl-mtmips.c
58
return (group >= p->group_count) ? NULL : p->group_names[group];
drivers/pinctrl/mediatek/pinctrl-mtmips.c
62
unsigned int group,
drivers/pinctrl/mediatek/pinctrl-mtmips.c
68
if (group >= p->group_count)
drivers/pinctrl/mediatek/pinctrl-mtmips.c
71
*pins = p->groups[group].func[0].pins;
drivers/pinctrl/mediatek/pinctrl-mtmips.c
72
*num_pins = p->groups[group].func[0].pin_count;
drivers/pinctrl/mediatek/pinctrl-paris.c
420
(*map)[*num_maps].data.mux.group = grp->name;
drivers/pinctrl/mediatek/pinctrl-paris.c
567
unsigned group)
drivers/pinctrl/mediatek/pinctrl-paris.c
571
return hw->groups[group].name;
drivers/pinctrl/mediatek/pinctrl-paris.c
575
unsigned group, const unsigned **pins,
drivers/pinctrl/mediatek/pinctrl-paris.c
580
*pins = (unsigned *)&hw->groups[group].pin;
drivers/pinctrl/mediatek/pinctrl-paris.c
746
unsigned group)
drivers/pinctrl/mediatek/pinctrl-paris.c
749
struct mtk_pinctrl_group *grp = hw->groups + group;
drivers/pinctrl/mediatek/pinctrl-paris.c
757
function, group);
drivers/pinctrl/mediatek/pinctrl-paris.c
778
static int mtk_pconf_group_get(struct pinctrl_dev *pctldev, unsigned group,
drivers/pinctrl/mediatek/pinctrl-paris.c
782
struct mtk_pinctrl_group *grp = &hw->groups[group];
drivers/pinctrl/mediatek/pinctrl-paris.c
788
static int mtk_pconf_group_set(struct pinctrl_dev *pctldev, unsigned group,
drivers/pinctrl/mediatek/pinctrl-paris.c
792
struct mtk_pinctrl_group *grp = &hw->groups[group];
drivers/pinctrl/mediatek/pinctrl-paris.c
984
struct mtk_pinctrl_group *group = hw->groups + i;
drivers/pinctrl/mediatek/pinctrl-paris.c
986
group->name = pin->name;
drivers/pinctrl/mediatek/pinctrl-paris.c
987
group->pin = pin->number;
drivers/pinctrl/meson/pinctrl-amlogic-a4.c
248
struct aml_pctl_group *group = &info->groups[group_id];
drivers/pinctrl/meson/pinctrl-amlogic-a4.c
252
for (i = 0; i < group->npins; i++) {
drivers/pinctrl/meson/pinctrl-amlogic-a4.c
253
range = pinctrl_find_gpio_range_from_pin(pctldev, group->pins[i]);
drivers/pinctrl/meson/pinctrl-amlogic-a4.c
254
aml_pctl_set_function(info, range, group->pins[i], group->func[i]);
drivers/pinctrl/meson/pinctrl-amlogic-a4.c
627
unsigned int group, unsigned long *config)
drivers/pinctrl/meson/pinctrl-meson-axg-pmx.c
86
const struct meson_pmx_group *group = &pc->data->groups[group_num];
drivers/pinctrl/meson/pinctrl-meson-axg-pmx.c
88
(struct meson_pmx_axg_data *)group->data;
drivers/pinctrl/meson/pinctrl-meson-axg-pmx.c
91
group->name);
drivers/pinctrl/meson/pinctrl-meson-axg-pmx.c
93
for (i = 0; i < group->num_pins; i++) {
drivers/pinctrl/meson/pinctrl-meson-axg-pmx.c
94
ret = meson_axg_pmx_update_function(pc, group->pins[i],
drivers/pinctrl/meson/pinctrl-meson.c
532
const struct meson_pmx_group *group = &pc->data->groups[num_group];
drivers/pinctrl/meson/pinctrl-meson.c
535
dev_dbg(pc->dev, "set pinconf for group %s\n", group->name);
drivers/pinctrl/meson/pinctrl-meson.c
537
for (i = 0; i < group->num_pins; i++) {
drivers/pinctrl/meson/pinctrl-meson.c
538
meson_pinconf_set(pcdev, group->pins[i], configs,
drivers/pinctrl/meson/pinctrl-meson.c
546
unsigned int group, unsigned long *config)
drivers/pinctrl/meson/pinctrl-meson8-pmx.c
35
const struct meson_pmx_group *group;
drivers/pinctrl/meson/pinctrl-meson8-pmx.c
40
group = &pc->data->groups[i];
drivers/pinctrl/meson/pinctrl-meson8-pmx.c
41
pmx_data = (struct meson8_pmx_data *)group->data;
drivers/pinctrl/meson/pinctrl-meson8-pmx.c
45
for (j = 0; j < group->num_pins; j++) {
drivers/pinctrl/meson/pinctrl-meson8-pmx.c
46
if (group->pins[j] == pin) {
drivers/pinctrl/meson/pinctrl-meson8-pmx.c
61
const struct meson_pmx_group *group = &pc->data->groups[group_num];
drivers/pinctrl/meson/pinctrl-meson8-pmx.c
63
(struct meson8_pmx_data *)group->data;
drivers/pinctrl/meson/pinctrl-meson8-pmx.c
67
group->name);
drivers/pinctrl/meson/pinctrl-meson8-pmx.c
73
for (i = 0; i < group->num_pins; i++)
drivers/pinctrl/meson/pinctrl-meson8-pmx.c
74
meson8_pmx_disable_other_groups(pc, group->pins[i], group_num);
drivers/pinctrl/microchip/pinctrl-mpfs-iomux0.c
189
const struct mpfs_iomux0_pin_group *group;
drivers/pinctrl/microchip/pinctrl-mpfs-iomux0.c
192
group = &mpfs_iomux0_pin_groups[gsel];
drivers/pinctrl/microchip/pinctrl-mpfs-iomux0.c
196
function->name, group->mask, group->setting);
drivers/pinctrl/microchip/pinctrl-mpfs-iomux0.c
197
regmap_assign_bits(pctrl->regmap, MPFS_IOMUX0_REG, group->mask, group->setting);
drivers/pinctrl/microchip/pinctrl-mpfs-mssio.c
330
const struct group_desc *group;
drivers/pinctrl/microchip/pinctrl-mpfs-mssio.c
333
group = pinctrl_generic_get_group(pctrl_dev, gsel);
drivers/pinctrl/microchip/pinctrl-mpfs-mssio.c
334
if (!group)
drivers/pinctrl/microchip/pinctrl-mpfs-mssio.c
337
functions = group->data;
drivers/pinctrl/microchip/pinctrl-mpfs-mssio.c
339
for (int i = 0; i < group->grp.npins; i++) {
drivers/pinctrl/microchip/pinctrl-mpfs-mssio.c
348
mpfs_pinctrl_set_pin_func(pctrl, group->grp.pins[i], function);
drivers/pinctrl/microchip/pinctrl-mpfs-mssio.c
577
const struct group_desc *group;
drivers/pinctrl/microchip/pinctrl-mpfs-mssio.c
582
group = pinctrl_generic_get_group(pctrl_dev, gsel);
drivers/pinctrl/microchip/pinctrl-mpfs-mssio.c
583
if (!group)
drivers/pinctrl/microchip/pinctrl-mpfs-mssio.c
590
pin = group->grp.pins[0];
drivers/pinctrl/microchip/pinctrl-mpfs-mssio.c
597
for (int i = 0; i < group->grp.npins; i++)
drivers/pinctrl/microchip/pinctrl-mpfs-mssio.c
598
mpfs_pinctrl_pin_set_config(pctrl, group->grp.pins[i], val);
drivers/pinctrl/microchip/pinctrl-mpfs-mssio.c
601
mpfs_pinctrl_set_bank_voltage(pctrl, group->grp.pins[0], bank_voltage);
drivers/pinctrl/microchip/pinctrl-pic64gx-gpio2.c
279
const struct pic64gx_gpio2_pin_group *group;
drivers/pinctrl/microchip/pinctrl-pic64gx-gpio2.c
282
group = &pic64gx_gpio2_pin_groups[gsel];
drivers/pinctrl/microchip/pinctrl-pic64gx-gpio2.c
286
function->name, group->mask, group->setting);
drivers/pinctrl/microchip/pinctrl-pic64gx-gpio2.c
287
regmap_assign_bits(pctrl->regmap, PIC64GX_PINMUX_REG, group->mask, group->setting);
drivers/pinctrl/mvebu/pinctrl-armada-37xx.c
241
struct armada_37xx_pin_group *group = &info->groups[*grp];
drivers/pinctrl/mvebu/pinctrl-armada-37xx.c
245
for (j = 0; j < (group->npins + group->extra_npins); j++)
drivers/pinctrl/mvebu/pinctrl-armada-37xx.c
246
if (group->pins[j] == pin)
drivers/pinctrl/mvebu/pinctrl-armada-37xx.c
247
return group;
drivers/pinctrl/mvebu/pinctrl-armada-37xx.c
279
unsigned int group)
drivers/pinctrl/mvebu/pinctrl-armada-37xx.c
283
return info->groups[group].name;
drivers/pinctrl/mvebu/pinctrl-armada-37xx.c
366
unsigned int group)
drivers/pinctrl/mvebu/pinctrl-armada-37xx.c
370
struct armada_37xx_pin_group *grp = &info->groups[group];
drivers/pinctrl/mvebu/pinctrl-armada-37xx.c
495
struct armada_37xx_pin_group *group;
drivers/pinctrl/mvebu/pinctrl-armada-37xx.c
501
while ((group = armada_37xx_find_next_grp_by_pin(info, offset, &grp))) {
drivers/pinctrl/mvebu/pinctrl-armada-37xx.c
502
ret = armada_37xx_pmx_set_by_name(pctldev, "gpio", group);
drivers/pinctrl/mvebu/pinctrl-mvebu.c
413
const char *group;
drivers/pinctrl/mvebu/pinctrl-mvebu.c
438
of_property_for_each_string(np, "marvell,pins", prop, group) {
drivers/pinctrl/mvebu/pinctrl-mvebu.c
440
mvebu_pinctrl_find_group_by_name(pctl, group);
drivers/pinctrl/mvebu/pinctrl-mvebu.c
443
dev_err(pctl->dev, "unknown pin %s", group);
drivers/pinctrl/mvebu/pinctrl-mvebu.c
449
function, group);
drivers/pinctrl/mvebu/pinctrl-mvebu.c
454
(*map)[n].data.mux.group = group;
drivers/pinctrl/nomadik/pinctrl-abx500.c
573
unsigned group)
drivers/pinctrl/nomadik/pinctrl-abx500.c
581
g = &pct->soc->groups[group];
drivers/pinctrl/nomadik/pinctrl-abx500.c
692
unsigned *num_maps, const char *group,
drivers/pinctrl/nomadik/pinctrl-abx500.c
699
(*map)[*num_maps].data.mux.group = group;
drivers/pinctrl/nomadik/pinctrl-abx500.c
708
unsigned *num_maps, const char *group,
drivers/pinctrl/nomadik/pinctrl-abx500.c
722
(*map)[*num_maps].data.configs.group_or_pin = group;
drivers/pinctrl/nomadik/pinctrl-abx500.c
757
const char *group;
drivers/pinctrl/nomadik/pinctrl-abx500.c
768
of_property_for_each_string(np, "groups", prop, group) {
drivers/pinctrl/nomadik/pinctrl-abx500.c
770
num_maps, group, function);
drivers/pinctrl/nomadik/pinctrl-nomadik.c
591
unsigned int *num_maps, const char *group,
drivers/pinctrl/nomadik/pinctrl-nomadik.c
598
(*map)[*num_maps].data.mux.group = group;
drivers/pinctrl/nomadik/pinctrl-nomadik.c
607
unsigned int *num_maps, const char *group,
drivers/pinctrl/nomadik/pinctrl-nomadik.c
621
(*map)[*num_maps].data.configs.group_or_pin = group;
drivers/pinctrl/nomadik/pinctrl-nomadik.c
757
const char *group;
drivers/pinctrl/nomadik/pinctrl-nomadik.c
769
of_property_for_each_string(np, "groups", prop, group) {
drivers/pinctrl/nomadik/pinctrl-nomadik.c
771
group, function);
drivers/pinctrl/nomadik/pinctrl-nomadik.c
873
unsigned int group)
drivers/pinctrl/nomadik/pinctrl-nomadik.c
883
g = &npct->soc->groups[group];
drivers/pinctrl/nuvoton/pinctrl-ma35.c
220
new_map[0].data.mux.group = np->name;
drivers/pinctrl/nuvoton/pinctrl-ma35.c
231
(*map)->data.mux.function, (*map)->data.mux.group, map_num);
drivers/pinctrl/nuvoton/pinctrl-ma35.c
273
unsigned int group)
drivers/pinctrl/nuvoton/pinctrl-ma35.c
276
struct group_desc *grp = &npctl->groups[group];
drivers/pinctrl/nuvoton/pinctrl-ma35.c
607
static void ma35_gpio_cla_port(unsigned int gpio_num, unsigned int *group,
drivers/pinctrl/nuvoton/pinctrl-ma35.c
610
*group = gpio_num / MA35_GPIO_PORT_MAX;
drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c
1601
unsigned int group)
drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c
1605
dev_dbg(npcm->dev, "set_mux: %d, %d[%s]\n", function, group,
drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c
1606
npcm7xx_groups[group].name);
drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c
1608
npcm7xx_setfunc(npcm->gcr_regmap, npcm7xx_groups[group].pins,
drivers/pinctrl/nuvoton/pinctrl-npcm7xx.c
1609
npcm7xx_groups[group].npins, group);
drivers/pinctrl/nuvoton/pinctrl-npcm8xx.c
2009
unsigned int group)
drivers/pinctrl/nuvoton/pinctrl-npcm8xx.c
2013
npcm8xx_setfunc(npcm->gcr_regmap, npcm8xx_pingroups[group].pins,
drivers/pinctrl/nuvoton/pinctrl-npcm8xx.c
2014
npcm8xx_pingroups[group].npins, group);
drivers/pinctrl/nuvoton/pinctrl-wpcm450.c
895
unsigned int group)
drivers/pinctrl/nuvoton/pinctrl-wpcm450.c
899
wpcm450_setfunc(pctrl->gcr_regmap, wpcm450_groups[group].pins,
drivers/pinctrl/nuvoton/pinctrl-wpcm450.c
900
wpcm450_groups[group].npins, function);
drivers/pinctrl/nxp/pinctrl-s32cc.c
305
unsigned int group)
drivers/pinctrl/nxp/pinctrl-s32cc.c
316
grp = &info->groups[group];
drivers/pinctrl/nxp/pinctrl-s32cc.c
325
grp->data.pins[i], group);
drivers/pinctrl/pinconf-generic.c
400
const char *group;
drivers/pinctrl/pinconf-generic.c
447
of_property_for_each_string(np, subnode_target_type, prop, group) {
drivers/pinctrl/pinconf-generic.c
450
reserved_maps, num_maps, group,
drivers/pinctrl/pinconf-generic.c
458
reserved_maps, num_maps, group, configs,
drivers/pinctrl/pinctrl-amd.c
1077
static int amd_set_mux(struct pinctrl_dev *pctrldev, unsigned int function, unsigned int group)
drivers/pinctrl/pinctrl-amd.c
1088
if (strcmp(gpio_dev->groups[group].name, pmx_functions[function].groups[index]))
drivers/pinctrl/pinctrl-amd.c
1107
for (ind = 0; ind < gpio_dev->groups[group].npins; ind++) {
drivers/pinctrl/pinctrl-amd.c
1108
if (strncmp(gpio_dev->groups[group].name, "IMX_F", strlen("IMX_F")))
drivers/pinctrl/pinctrl-amd.c
1111
pd = pin_desc_get(gpio_dev->pctrl, gpio_dev->groups[group].pins[ind]);
drivers/pinctrl/pinctrl-amd.c
1112
pd->mux_owner = gpio_dev->groups[group].name;
drivers/pinctrl/pinctrl-amd.c
706
unsigned group)
drivers/pinctrl/pinctrl-amd.c
710
return gpio_dev->groups[group].name;
drivers/pinctrl/pinctrl-amd.c
714
unsigned group,
drivers/pinctrl/pinctrl-amd.c
720
*pins = gpio_dev->groups[group].pins;
drivers/pinctrl/pinctrl-amd.c
721
*num_pins = gpio_dev->groups[group].npins;
drivers/pinctrl/pinctrl-amd.c
830
unsigned int group,
drivers/pinctrl/pinctrl-amd.c
837
ret = amd_get_group_pins(pctldev, group, &pins, &npins);
drivers/pinctrl/pinctrl-amd.c
848
unsigned group, unsigned long *configs,
drivers/pinctrl/pinctrl-amd.c
855
ret = amd_get_group_pins(pctldev, group, &pins, &npins);
drivers/pinctrl/pinctrl-amdisp.c
64
unsigned int group)
drivers/pinctrl/pinctrl-amdisp.c
68
return pctrl->data->groups[group].name;
drivers/pinctrl/pinctrl-amdisp.c
72
unsigned int group,
drivers/pinctrl/pinctrl-amdisp.c
78
*pins = pctrl->data->groups[group].pins;
drivers/pinctrl/pinctrl-amdisp.c
79
*num_pins = pctrl->data->groups[group].npins;
drivers/pinctrl/pinctrl-apple-gpio.c
181
unsigned int group)
drivers/pinctrl/pinctrl-apple-gpio.c
186
pctl, group, REG_GPIOx_PERIPH | REG_GPIOx_INPUT_ENABLE,
drivers/pinctrl/pinctrl-artpec6.c
430
unsigned int group)
drivers/pinctrl/pinctrl-artpec6.c
432
return artpec6_pin_groups[group].name;
drivers/pinctrl/pinctrl-artpec6.c
436
unsigned int group,
drivers/pinctrl/pinctrl-artpec6.c
440
*pins = (unsigned int *)artpec6_pin_groups[group].pins;
drivers/pinctrl/pinctrl-artpec6.c
441
*num_pins = artpec6_pin_groups[group].num_pins;
drivers/pinctrl/pinctrl-artpec6.c
654
unsigned int function, unsigned int group,
drivers/pinctrl/pinctrl-artpec6.c
662
for (i = 0; i < artpec6_pin_groups[group].num_pins; i++) {
drivers/pinctrl/pinctrl-artpec6.c
667
if (artpec6_pin_groups[group].pins[i] > ARTPEC6_MAX_MUXABLE)
drivers/pinctrl/pinctrl-artpec6.c
675
val = artpec6_pin_groups[group].config
drivers/pinctrl/pinctrl-artpec6.c
682
reg = artpec6_pmx_reg_offset(artpec6_pin_groups[group].pins[i]);
drivers/pinctrl/pinctrl-artpec6.c
693
unsigned int group)
drivers/pinctrl/pinctrl-artpec6.c
699
artpec6_get_group_name(pctldev, group));
drivers/pinctrl/pinctrl-artpec6.c
701
artpec6_pmx_select_func(pctldev, function, group, true);
drivers/pinctrl/pinctrl-artpec6.c
880
unsigned int group, unsigned long *configs,
drivers/pinctrl/pinctrl-artpec6.c
887
artpec6_get_group_name(pctldev, group));
drivers/pinctrl/pinctrl-artpec6.c
889
num_pins = artpec6_pin_groups[group].num_pins;
drivers/pinctrl/pinctrl-artpec6.c
893
artpec6_pin_groups[group].pins[current_pin],
drivers/pinctrl/pinctrl-as3722.c
170
unsigned group)
drivers/pinctrl/pinctrl-as3722.c
174
return as_pci->pin_groups[group].name;
drivers/pinctrl/pinctrl-as3722.c
178
unsigned group, const unsigned **pins, unsigned *num_pins)
drivers/pinctrl/pinctrl-as3722.c
182
*pins = as_pci->pin_groups[group].pins;
drivers/pinctrl/pinctrl-as3722.c
183
*num_pins = as_pci->pin_groups[group].npins;
drivers/pinctrl/pinctrl-as3722.c
222
unsigned group)
drivers/pinctrl/pinctrl-as3722.c
225
int gpio_cntr_reg = AS3722_GPIOn_CONTROL_REG(group);
drivers/pinctrl/pinctrl-as3722.c
230
__func__, group, function, val);
drivers/pinctrl/pinctrl-as3722.c
236
group, ret);
drivers/pinctrl/pinctrl-as3722.c
239
as_pci->gpio_control[group].io_function = function;
drivers/pinctrl/pinctrl-as3722.c
251
group, ret);
drivers/pinctrl/pinctrl-as3722.c
254
as_pci->gpio_control[group].mode_prop =
drivers/pinctrl/pinctrl-at91-pio4.c
1157
struct atmel_group *group = atmel_pioctrl->groups + i;
drivers/pinctrl/pinctrl-at91-pio4.c
1177
group->name = group_names[i] = pin_desc[i].name;
drivers/pinctrl/pinctrl-at91-pio4.c
1178
group->pin = pin_desc[i].number;
drivers/pinctrl/pinctrl-at91-pio4.c
605
const char *group, *func;
drivers/pinctrl/pinctrl-at91-pio4.c
611
ret = atmel_pctl_xlate_pinfunc(pctldev, np, pinfunc, &group,
drivers/pinctrl/pinctrl-at91-pio4.c
617
group, func);
drivers/pinctrl/pinctrl-at91-pio4.c
623
reserved_maps, num_maps, group,
drivers/pinctrl/pinctrl-at91-pio4.c
707
unsigned int group)
drivers/pinctrl/pinctrl-at91-pio4.c
714
atmel_functions[function], atmel_pioctrl->groups[group].name);
drivers/pinctrl/pinctrl-at91-pio4.c
716
pin = atmel_pioctrl->groups[group].pin;
drivers/pinctrl/pinctrl-at91-pio4.c
734
unsigned int group,
drivers/pinctrl/pinctrl-at91-pio4.c
739
struct atmel_group *grp = atmel_pioctrl->groups + group;
drivers/pinctrl/pinctrl-at91-pio4.c
801
unsigned int group,
drivers/pinctrl/pinctrl-at91-pio4.c
806
struct atmel_group *grp = atmel_pioctrl->groups + group;
drivers/pinctrl/pinctrl-at91.c
1126
struct seq_file *s, unsigned group)
drivers/pinctrl/pinctrl-at91.c
327
new_map[0].data.mux.group = np->name;
drivers/pinctrl/pinctrl-at91.c
341
(*map)->data.mux.function, (*map)->data.mux.group, map_num);
drivers/pinctrl/pinctrl-at91.c
843
unsigned group)
drivers/pinctrl/pinctrl-at91.c
846
const struct at91_pmx_pin *pins_conf = info->groups[group].pins_conf;
drivers/pinctrl/pinctrl-at91.c
848
uint32_t npins = info->groups[group].npins;
drivers/pinctrl/pinctrl-at91.c
854
info->functions[selector].name, info->groups[group].name);
drivers/pinctrl/pinctrl-at91.c
860
ret = pin_check_config(info, info->groups[group].name, i, pin);
drivers/pinctrl/pinctrl-axp209.c
271
unsigned int function, unsigned int group)
drivers/pinctrl/pinctrl-axp209.c
278
return axp20x_pmx_set(pctldev, group,
drivers/pinctrl/pinctrl-axp209.c
286
if (!(BIT(group) & mask))
drivers/pinctrl/pinctrl-axp209.c
298
return axp20x_pmx_set(pctldev, group, pctl->funcs[function].muxval);
drivers/pinctrl/pinctrl-axp209.c
364
const char **group;
drivers/pinctrl/pinctrl-axp209.c
375
group = func->groups;
drivers/pinctrl/pinctrl-axp209.c
377
*group = pins[bit].name;
drivers/pinctrl/pinctrl-axp209.c
378
group++;
drivers/pinctrl/pinctrl-bm1880.c
985
unsigned int group)
drivers/pinctrl/pinctrl-bm1880.c
988
const struct bm1880_pctrl_group *pgrp = &pctrl->groups[group];
drivers/pinctrl/pinctrl-cy8c95x0.c
1140
unsigned int group)
drivers/pinctrl/pinctrl-cy8c95x0.c
1142
return cy8c95x0_groups[group];
drivers/pinctrl/pinctrl-cy8c95x0.c
1146
unsigned int group,
drivers/pinctrl/pinctrl-cy8c95x0.c
1150
*pins = &cy8c9560_pins[group].number;
drivers/pinctrl/pinctrl-cy8c95x0.c
1222
unsigned int selector, unsigned int group)
drivers/pinctrl/pinctrl-cy8c95x0.c
1224
u8 port = cypress_get_port(chip, group);
drivers/pinctrl/pinctrl-cy8c95x0.c
1225
u8 bit = cypress_get_pin_mask(chip, group);
drivers/pinctrl/pinctrl-cy8c95x0.c
1228
ret = cy8c95x0_set_mode(chip, group, selector);
drivers/pinctrl/pinctrl-cy8c95x0.c
1244
unsigned int group)
drivers/pinctrl/pinctrl-cy8c95x0.c
1248
return cy8c95x0_pinmux_mode(chip, selector, group);
drivers/pinctrl/pinctrl-digicolor.c
127
unsigned group)
drivers/pinctrl/pinctrl-digicolor.c
133
dc_client_sel(group, ®_off, &bit_off);
drivers/pinctrl/pinctrl-ep93xx.c
1270
unsigned int group)
drivers/pinctrl/pinctrl-ep93xx.c
1283
grp = &ep9301_pin_groups[group];
drivers/pinctrl/pinctrl-ep93xx.c
1286
grp = &ep9307_pin_groups[group];
drivers/pinctrl/pinctrl-ep93xx.c
1289
grp = &ep9312_pin_groups[group];
drivers/pinctrl/pinctrl-equilibrium.c
332
unsigned int selector, unsigned int group)
drivers/pinctrl/pinctrl-equilibrium.c
344
grp = pinctrl_generic_get_group(pctldev, group);
drivers/pinctrl/pinctrl-equilibrium.c
533
unsigned int group, unsigned long *config)
drivers/pinctrl/pinctrl-equilibrium.c
539
ret = pinctrl_generic_get_group_pins(pctldev, group, &pins, &npins);
drivers/pinctrl/pinctrl-equilibrium.c
556
unsigned int group, unsigned long *configs,
drivers/pinctrl/pinctrl-equilibrium.c
563
ret = pinctrl_generic_get_group_pins(pctldev, group, &pins, &npins);
drivers/pinctrl/pinctrl-equilibrium.c
720
struct pingroup group, *grp = &group;
drivers/pinctrl/pinctrl-equilibrium.c
773
memset(&group, 0, sizeof(group));
drivers/pinctrl/pinctrl-falcon.c
229
unsigned group, unsigned long *config)
drivers/pinctrl/pinctrl-falcon.c
235
unsigned group, unsigned long *configs,
drivers/pinctrl/pinctrl-gemini.c
2192
unsigned int group)
drivers/pinctrl/pinctrl-gemini.c
2205
grp = &gemini_3512_pin_groups[group];
drivers/pinctrl/pinctrl-gemini.c
2207
grp = &gemini_3516_pin_groups[group];
drivers/pinctrl/pinctrl-ingenic.c
4000
unsigned int selector, unsigned int group)
drivers/pinctrl/pinctrl-ingenic.c
4013
grp = pinctrl_generic_get_group(pctldev, group);
drivers/pinctrl/pinctrl-ingenic.c
4334
unsigned int group, unsigned long *config)
drivers/pinctrl/pinctrl-ingenic.c
4340
ret = pinctrl_generic_get_group_pins(pctldev, group, &pins, &npins);
drivers/pinctrl/pinctrl-ingenic.c
4359
unsigned int group, unsigned long *configs,
drivers/pinctrl/pinctrl-ingenic.c
4366
ret = pinctrl_generic_get_group_pins(pctldev, group, &pins, &npins);
drivers/pinctrl/pinctrl-ingenic.c
4559
const struct group_desc *group = &chip_info->groups[i];
drivers/pinctrl/pinctrl-ingenic.c
4560
const struct pingroup *grp = &group->grp;
drivers/pinctrl/pinctrl-ingenic.c
4563
group->data);
drivers/pinctrl/pinctrl-k210.c
705
unsigned int group)
drivers/pinctrl/pinctrl-k210.c
708
if (group >= K210_NPINS)
drivers/pinctrl/pinctrl-k210.c
711
k210_pinmux_set_pin_function(pctldev, group, function);
drivers/pinctrl/pinctrl-k210.c
730
unsigned int group)
drivers/pinctrl/pinctrl-k210.c
732
return k210_group_names[group];
drivers/pinctrl/pinctrl-k210.c
736
unsigned int group,
drivers/pinctrl/pinctrl-k210.c
740
if (group >= K210_NPINS) {
drivers/pinctrl/pinctrl-k210.c
746
*pins = &k210_pins[group].number;
drivers/pinctrl/pinctrl-k230.c
225
new_map[idx].data.mux.group = grp->name;
drivers/pinctrl/pinctrl-k230.c
438
unsigned int group)
drivers/pinctrl/pinctrl-k230.c
441
const struct k230_pin_conf *data = info->groups[group].data;
drivers/pinctrl/pinctrl-k230.c
442
struct k230_pin_group *grp = &info->groups[group];
drivers/pinctrl/pinctrl-lantiq.c
126
of_property_for_each_string(np, "lantiq,groups", prop, group) {
drivers/pinctrl/pinctrl-lantiq.c
131
(*map)->name = group;
drivers/pinctrl/pinctrl-lantiq.c
132
(*map)->data.configs.group_or_pin = group;
drivers/pinctrl/pinctrl-lantiq.c
261
unsigned group)
drivers/pinctrl/pinctrl-lantiq.c
264
const struct ltq_pin_group *pin_grp = &info->grps[group];
drivers/pinctrl/pinctrl-lantiq.c
270
info->grps[group].name);
drivers/pinctrl/pinctrl-lantiq.c
76
const char *group, *pin;
drivers/pinctrl/pinctrl-lantiq.c
94
of_property_for_each_string(np, "lantiq,groups", prop, group) {
drivers/pinctrl/pinctrl-lantiq.c
97
(*map)->data.mux.group = group;
drivers/pinctrl/pinctrl-loongson2.c
30
#define SPECIFIC_GROUP(group) \
drivers/pinctrl/pinctrl-loongson2.c
31
static const char * const group##_groups[] = { \
drivers/pinctrl/pinctrl-loongson2.c
32
#group \
drivers/pinctrl/pinctrl-lpc18xx.c
1155
unsigned group)
drivers/pinctrl/pinctrl-lpc18xx.c
1158
struct lpc18xx_pin_caps *pin = lpc18xx_pins[group].drv_data;
drivers/pinctrl/pinctrl-lpc18xx.c
1218
dev_err(pctldev->dev, "Pin %s can't be %s\n", lpc18xx_pins[group].name,
drivers/pinctrl/pinctrl-lpc18xx.c
1236
unsigned group)
drivers/pinctrl/pinctrl-lpc18xx.c
1238
return lpc18xx_pins[group].name;
drivers/pinctrl/pinctrl-lpc18xx.c
1242
unsigned group,
drivers/pinctrl/pinctrl-lpc18xx.c
1246
*pins = &lpc18xx_pins[group].number;
drivers/pinctrl/pinctrl-max7360.c
136
unsigned int group)
drivers/pinctrl/pinctrl-max7360.c
145
if (max7360_groups[group].pins[0] >= 6) {
drivers/pinctrl/pinctrl-max7360.c
71
unsigned int group)
drivers/pinctrl/pinctrl-max7360.c
73
return max7360_groups[group].name;
drivers/pinctrl/pinctrl-max7360.c
77
unsigned int group,
drivers/pinctrl/pinctrl-max7360.c
81
*pins = max7360_groups[group].pins;
drivers/pinctrl/pinctrl-max7360.c
82
*num_pins = max7360_groups[group].npins;
drivers/pinctrl/pinctrl-max77620.c
183
struct pinctrl_dev *pctldev, unsigned int group)
drivers/pinctrl/pinctrl-max77620.c
187
return mpci->pin_groups[group].name;
drivers/pinctrl/pinctrl-max77620.c
191
struct pinctrl_dev *pctldev, unsigned int group,
drivers/pinctrl/pinctrl-max77620.c
196
*pins = mpci->pin_groups[group].pins;
drivers/pinctrl/pinctrl-max77620.c
197
*num_pins = mpci->pin_groups[group].npins;
drivers/pinctrl/pinctrl-max77620.c
239
unsigned int function, unsigned int group)
drivers/pinctrl/pinctrl-max77620.c
247
} else if (function == mpci->pin_groups[group].alt_option) {
drivers/pinctrl/pinctrl-max77620.c
248
val = 1 << group;
drivers/pinctrl/pinctrl-max77620.c
251
group, function);
drivers/pinctrl/pinctrl-max77620.c
255
BIT(group), val);
drivers/pinctrl/pinctrl-mcp23s08.c
176
unsigned int group)
drivers/pinctrl/pinctrl-mcp23s08.c
182
unsigned int group,
drivers/pinctrl/pinctrl-microchip-sgpio.c
447
unsigned int selector, unsigned int group)
drivers/pinctrl/pinctrl-microchip-sgpio.c
497
unsigned int group)
drivers/pinctrl/pinctrl-microchip-sgpio.c
501
return bank->pctl_desc.pins[group].name;
drivers/pinctrl/pinctrl-microchip-sgpio.c
505
unsigned int group,
drivers/pinctrl/pinctrl-microchip-sgpio.c
511
*pins = &bank->pctl_desc.pins[group].number;
drivers/pinctrl/pinctrl-mlxbf3.c
191
unsigned int group)
drivers/pinctrl/pinctrl-mlxbf3.c
196
if (group < MLXBF3_NGPIOS_GPIO0)
drivers/pinctrl/pinctrl-mlxbf3.c
197
writel(BIT(group), priv->fw_ctrl_clr0);
drivers/pinctrl/pinctrl-mlxbf3.c
199
writel(BIT(group % MLXBF3_NGPIOS_GPIO0), priv->fw_ctrl_clr1);
drivers/pinctrl/pinctrl-mlxbf3.c
203
if (group < MLXBF3_NGPIOS_GPIO0)
drivers/pinctrl/pinctrl-mlxbf3.c
204
writel(BIT(group), priv->fw_ctrl_set0);
drivers/pinctrl/pinctrl-mlxbf3.c
206
writel(BIT(group % MLXBF3_NGPIOS_GPIO0), priv->fw_ctrl_set1);
drivers/pinctrl/pinctrl-ocelot.c
1512
unsigned int selector, unsigned int group)
drivers/pinctrl/pinctrl-ocelot.c
1515
struct ocelot_pin_caps *pin = info->desc->pins[group].drv_data;
drivers/pinctrl/pinctrl-ocelot.c
1519
f = ocelot_pin_function_idx(info, group, selector);
drivers/pinctrl/pinctrl-ocelot.c
1540
unsigned int selector, unsigned int group)
drivers/pinctrl/pinctrl-ocelot.c
1543
struct ocelot_pin_caps *pin = info->desc->pins[group].drv_data;
drivers/pinctrl/pinctrl-ocelot.c
1547
f = ocelot_pin_function_idx(info, group, selector);
drivers/pinctrl/pinctrl-ocelot.c
1658
unsigned int group)
drivers/pinctrl/pinctrl-ocelot.c
1662
return info->desc->pins[group].name;
drivers/pinctrl/pinctrl-ocelot.c
1666
unsigned int group,
drivers/pinctrl/pinctrl-ocelot.c
1672
*pins = &info->desc->pins[group].number;
drivers/pinctrl/pinctrl-palmas.c
623
unsigned group)
drivers/pinctrl/pinctrl-palmas.c
627
return pci->pin_groups[group].name;
drivers/pinctrl/pinctrl-palmas.c
631
unsigned group, const unsigned **pins, unsigned *num_pins)
drivers/pinctrl/pinctrl-palmas.c
635
*pins = pci->pin_groups[group].pins;
drivers/pinctrl/pinctrl-palmas.c
636
*num_pins = pci->pin_groups[group].npins;
drivers/pinctrl/pinctrl-palmas.c
676
unsigned group)
drivers/pinctrl/pinctrl-palmas.c
683
g = &pci->pin_groups[group];
drivers/pinctrl/pinctrl-palmas.c
724
pci->pins_current_opt[group] = i;
drivers/pinctrl/pinctrl-pic32.c
1712
unsigned group)
drivers/pinctrl/pinctrl-pic32.c
1716
return pctl->groups[group].name;
drivers/pinctrl/pinctrl-pic32.c
1720
unsigned group,
drivers/pinctrl/pinctrl-pic32.c
1726
*pins = &pctl->groups[group].pin;
drivers/pinctrl/pinctrl-pic32.c
1769
unsigned func, unsigned group)
drivers/pinctrl/pinctrl-pic32.c
1772
const struct pic32_pin_group *pg = &pctl->groups[group];
drivers/pinctrl/pinctrl-pic32.c
1791
dev_err(pctl->dev, "cannot mux pin %u to function %u\n", group, func);
drivers/pinctrl/pinctrl-pistachio.c
888
unsigned group)
drivers/pinctrl/pinctrl-pistachio.c
892
return pctl->groups[group].name;
drivers/pinctrl/pinctrl-pistachio.c
896
unsigned group,
drivers/pinctrl/pinctrl-pistachio.c
902
*pins = &pctl->groups[group].pin;
drivers/pinctrl/pinctrl-pistachio.c
945
unsigned func, unsigned group)
drivers/pinctrl/pinctrl-pistachio.c
948
const struct pistachio_pin_group *pg = &pctl->groups[group];
drivers/pinctrl/pinctrl-pistachio.c
961
group, func);
drivers/pinctrl/pinctrl-pistachio.c
972
if (pf->scenarios[i] == group)
drivers/pinctrl/pinctrl-rk805.c
398
unsigned int group)
drivers/pinctrl/pinctrl-rk805.c
402
return pci->groups[group].name;
drivers/pinctrl/pinctrl-rk805.c
406
unsigned int group,
drivers/pinctrl/pinctrl-rk805.c
412
*pins = pci->groups[group].pins;
drivers/pinctrl/pinctrl-rk805.c
413
*num_pins = pci->groups[group].npins;
drivers/pinctrl/pinctrl-rk805.c
477
unsigned int group)
drivers/pinctrl/pinctrl-rk805.c
481
int offset = group;
drivers/pinctrl/pinctrl-rockchip.c
3511
unsigned group)
drivers/pinctrl/pinctrl-rockchip.c
3514
const unsigned int *pins = info->groups[group].pins;
drivers/pinctrl/pinctrl-rockchip.c
3515
const struct rockchip_pin_config *data = info->groups[group].data;
drivers/pinctrl/pinctrl-rockchip.c
3521
info->functions[selector].name, info->groups[group].name);
drivers/pinctrl/pinctrl-rockchip.c
3527
for (cnt = 0; cnt < info->groups[group].npins; cnt++) {
drivers/pinctrl/pinctrl-rockchip.c
433
new_map[0].data.mux.group = np->name;
drivers/pinctrl/pinctrl-rockchip.c
447
(*map)->data.mux.function, (*map)->data.mux.group, map_num);
drivers/pinctrl/pinctrl-rp1.c
1128
map->data.mux.group = rp1_pctl_get_group_name(pc->pctl_dev,
drivers/pinctrl/pinctrl-scmi.c
169
unsigned int selector, unsigned int group)
drivers/pinctrl/pinctrl-scmi.c
173
return pinctrl_ops->mux_set(pmx->ph, selector, group);
drivers/pinctrl/pinctrl-scmi.c
386
unsigned int group,
drivers/pinctrl/pinctrl-scmi.c
417
ret = pinctrl_ops->settings_conf(pmx->ph, group, GROUP_TYPE,
drivers/pinctrl/pinctrl-scmi.c
430
unsigned int group,
drivers/pinctrl/pinctrl-scmi.c
449
ret = pinctrl_ops->settings_get_one(pmx->ph, group, GROUP_TYPE, type,
drivers/pinctrl/pinctrl-single.c
1085
(*map)->data.mux.group = np->name;
drivers/pinctrl/pinctrl-single.c
1233
(*map)->data.mux.group = np->name;
drivers/pinctrl/pinctrl-single.c
360
unsigned group)
drivers/pinctrl/pinctrl-single.c
620
unsigned group, unsigned long *config)
drivers/pinctrl/pinctrl-single.c
626
ret = pinctrl_generic_get_group_pins(pctldev, group, &pins, &npins);
drivers/pinctrl/pinctrl-single.c
641
unsigned group, unsigned long *configs,
drivers/pinctrl/pinctrl-single.c
648
ret = pinctrl_generic_get_group_pins(pctldev, group, &pins, &npins);
drivers/pinctrl/pinctrl-st.c
836
new_map[0].data.mux.group = np->name;
drivers/pinctrl/pinctrl-st.c
894
unsigned group)
drivers/pinctrl/pinctrl-st.c
897
struct st_pinconf *conf = info->groups[group].pin_conf;
drivers/pinctrl/pinctrl-st.c
901
for (i = 0; i < info->groups[group].npins; i++) {
drivers/pinctrl/pinctrl-sx150x.c
349
unsigned int group)
drivers/pinctrl/pinctrl-sx150x.c
355
unsigned int group,
drivers/pinctrl/pinctrl-tb10x.c
452
const char *group;
drivers/pinctrl/pinctrl-tb10x.c
595
*groups = &state->pinfuncs[n].group;
drivers/pinctrl/pinctrl-tb10x.c
779
state->pinfuncs[state->pinfuncnt].group = name;
drivers/pinctrl/pinctrl-th1520.c
525
map[nmaps].data.mux.group = thp->desc.pins[i].name;
drivers/pinctrl/pinctrl-th1520.c
852
const struct th1520_pad_group *group;
drivers/pinctrl/pinctrl-th1520.c
876
group = &th1520_group1;
drivers/pinctrl/pinctrl-th1520.c
878
group = &th1520_group2;
drivers/pinctrl/pinctrl-th1520.c
880
group = &th1520_group3;
drivers/pinctrl/pinctrl-th1520.c
884
thp->desc.name = group->name;
drivers/pinctrl/pinctrl-th1520.c
885
thp->desc.pins = group->pins;
drivers/pinctrl/pinctrl-th1520.c
886
thp->desc.npins = group->npins;
drivers/pinctrl/pinctrl-tps6594.c
239
unsigned int group;
drivers/pinctrl/pinctrl-tps6594.c
416
unsigned int function, unsigned int group)
drivers/pinctrl/pinctrl-tps6594.c
424
if (group == remap[i].group && muxval == remap[i].muxval) {
drivers/pinctrl/pinctrl-tps6594.c
430
return tps6594_pmx_set(pinctrl, group, muxval);
drivers/pinctrl/pinctrl-utils.c
47
unsigned int *num_maps, const char *group,
drivers/pinctrl/pinctrl-utils.c
54
(*map)[*num_maps].data.mux.group = group;
drivers/pinctrl/pinctrl-utils.c
64
unsigned int *num_maps, const char *group,
drivers/pinctrl/pinctrl-utils.c
79
(*map)[*num_maps].data.configs.group_or_pin = group;
drivers/pinctrl/pinctrl-utils.h
22
unsigned int *num_maps, const char *group,
drivers/pinctrl/pinctrl-utils.h
26
unsigned int *num_maps, const char *group,
drivers/pinctrl/pinctrl-zynq.c
891
unsigned int group)
drivers/pinctrl/pinctrl-zynq.c
895
const struct zynq_pctrl_group *pgrp = &pctrl->groups[group];
drivers/pinctrl/pinctrl-zynqmp.c
214
unsigned int group)
drivers/pinctrl/pinctrl-zynqmp.c
220
zynqmp_pctrl_get_group_pins(pctldev, group, &pins, &npins);
drivers/pinctrl/pinctrl-zynqmp.c
691
static void zynqmp_pinctrl_group_add_pin(struct zynqmp_pctrl_group *group,
drivers/pinctrl/pinctrl-zynqmp.c
694
group->pins[group->npins++] = pin;
drivers/pinctrl/pinmux.c
376
const char *group;
drivers/pinctrl/pinmux.c
404
if (map->data.mux.group) {
drivers/pinctrl/pinmux.c
405
group = map->data.mux.group;
drivers/pinctrl/pinmux.c
406
ret = match_string(groups, num_groups, group);
drivers/pinctrl/pinmux.c
410
group, map->data.mux.function);
drivers/pinctrl/pinmux.c
414
group = groups[0];
drivers/pinctrl/pinmux.c
417
ret = pinctrl_get_group_selector(pctldev, group);
drivers/pinctrl/pinmux.c
420
map->data.mux.group);
drivers/pinctrl/pinmux.c
423
setting->data.mux.group = ret;
drivers/pinctrl/pinmux.c
445
ret = pctlops->get_group_pins(pctldev, setting->data.mux.group,
drivers/pinctrl/pinmux.c
453
setting->data.mux.group);
drivers/pinctrl/pinmux.c
470
setting->data.mux.group);
drivers/pinctrl/pinmux.c
493
setting->data.mux.group);
drivers/pinctrl/pinmux.c
528
ret = pctlops->get_group_pins(pctldev, setting->data.mux.group,
drivers/pinctrl/pinmux.c
535
setting->data.mux.group);
drivers/pinctrl/pinmux.c
560
setting->data.mux.group);
drivers/pinctrl/pinmux.c
677
desc->mux_setting->group));
drivers/pinctrl/pinmux.c
692
map->data.mux.group ? map->data.mux.group : "(default)",
drivers/pinctrl/pinmux.c
704
pctlops->get_group_name(pctldev, setting->data.mux.group),
drivers/pinctrl/pinmux.c
705
setting->data.mux.group,
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
183
unsigned group,
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
187
struct pingroup *g = pctl->groups + group;
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
204
unsigned group,
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
209
struct pingroup *g = pctl->groups + group;
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
330
struct pingroup *group;
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
35
struct pingroup *group = pctl->groups + tgroup;
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
352
group = pctl->groups + i;
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
353
group->name = ppins[i].pin.name;
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
354
group->pins = &ppins[i].pin.number;
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
355
group->npins = 1;
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
37
return group->name;
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
46
struct pingroup *group = pctl->groups + tgroup;
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
48
*pins = group->pins;
drivers/pinctrl/pxa/pinctrl-pxa2xx.c
49
*num_pins = group->npins;
drivers/pinctrl/qcom/pinctrl-lpass-lpi.c
106
unsigned int group)
drivers/pinctrl/qcom/pinctrl-lpass-lpi.c
109
const struct lpi_pingroup *g = &pctrl->data->groups[group];
drivers/pinctrl/qcom/pinctrl-lpass-lpi.c
131
!test_and_set_bit(group, pctrl->ever_gpio)) {
drivers/pinctrl/qcom/pinctrl-lpass-lpi.c
132
u32 io_val = lpi_gpio_read(pctrl, group, LPI_GPIO_VALUE_REG);
drivers/pinctrl/qcom/pinctrl-lpass-lpi.c
136
lpi_gpio_write(pctrl, group, LPI_GPIO_VALUE_REG,
drivers/pinctrl/qcom/pinctrl-lpass-lpi.c
140
lpi_gpio_write(pctrl, group, LPI_GPIO_VALUE_REG,
drivers/pinctrl/qcom/pinctrl-lpass-lpi.c
205
unsigned int group, unsigned int slew)
drivers/pinctrl/qcom/pinctrl-lpass-lpi.c
213
slew, group);
drivers/pinctrl/qcom/pinctrl-lpass-lpi.c
222
reg = pctrl->tlmm_base + LPI_TLMM_REG_OFFSET * group + LPI_GPIO_CFG_REG;
drivers/pinctrl/qcom/pinctrl-lpass-lpi.c
238
static int lpi_config_set(struct pinctrl_dev *pctldev, unsigned int group,
drivers/pinctrl/qcom/pinctrl-lpass-lpi.c
248
g = &pctrl->data->groups[group];
drivers/pinctrl/qcom/pinctrl-lpass-lpi.c
277
ret = lpi_config_set_slew_rate(pctrl, g, group, arg);
drivers/pinctrl/qcom/pinctrl-lpass-lpi.c
292
lpi_gpio_write(pctrl, group, LPI_GPIO_VALUE_REG, val);
drivers/pinctrl/qcom/pinctrl-lpass-lpi.c
296
val = lpi_gpio_read(pctrl, group, LPI_GPIO_CFG_REG);
drivers/pinctrl/qcom/pinctrl-lpass-lpi.c
303
lpi_gpio_write(pctrl, group, LPI_GPIO_CFG_REG, val);
drivers/pinctrl/qcom/pinctrl-msm.c
119
unsigned group)
drivers/pinctrl/qcom/pinctrl-msm.c
123
return pctrl->soc->groups[group].grp.name;
drivers/pinctrl/qcom/pinctrl-msm.c
127
unsigned group,
drivers/pinctrl/qcom/pinctrl-msm.c
133
*pins = pctrl->soc->groups[group].grp.pins;
drivers/pinctrl/qcom/pinctrl-msm.c
134
*num_pins = pctrl->soc->groups[group].grp.npins;
drivers/pinctrl/qcom/pinctrl-msm.c
156
unsigned group)
drivers/pinctrl/qcom/pinctrl-msm.c
160
unsigned int irq = irq_find_mapping(gc->irq.domain, group);
drivers/pinctrl/qcom/pinctrl-msm.c
169
g = &pctrl->soc->groups[group];
drivers/pinctrl/qcom/pinctrl-msm.c
205
!test_and_set_bit(group, pctrl->ever_gpio)) {
drivers/pinctrl/qcom/pinctrl-msm.c
324
unsigned int group,
drivers/pinctrl/qcom/pinctrl-msm.c
337
if (!gpiochip_line_is_valid(&pctrl->chip, group))
drivers/pinctrl/qcom/pinctrl-msm.c
340
g = &pctrl->soc->groups[group];
drivers/pinctrl/qcom/pinctrl-msm.c
410
unsigned group,
drivers/pinctrl/qcom/pinctrl-msm.c
425
g = &pctrl->soc->groups[group];
drivers/pinctrl/qcom/pinctrl-ssbi-gpio.c
174
unsigned group)
drivers/pinctrl/qcom/pinctrl-ssbi-gpio.c
176
return pm8xxx_groups[group];
drivers/pinctrl/qcom/pinctrl-ssbi-gpio.c
181
unsigned group,
drivers/pinctrl/qcom/pinctrl-ssbi-gpio.c
187
*pins = &pctrl->desc.pins[group].number;
drivers/pinctrl/qcom/pinctrl-ssbi-gpio.c
226
unsigned group)
drivers/pinctrl/qcom/pinctrl-ssbi-gpio.c
229
struct pm8xxx_pin_data *pin = pctrl->desc.pins[group].drv_data;
drivers/pinctrl/qcom/pinctrl-ssbi-mpp.c
251
unsigned group)
drivers/pinctrl/qcom/pinctrl-ssbi-mpp.c
253
return pm8xxx_groups[group];
drivers/pinctrl/qcom/pinctrl-ssbi-mpp.c
258
unsigned group,
drivers/pinctrl/qcom/pinctrl-ssbi-mpp.c
264
*pins = &pctrl->desc.pins[group].number;
drivers/pinctrl/qcom/pinctrl-ssbi-mpp.c
303
unsigned group)
drivers/pinctrl/qcom/pinctrl-ssbi-mpp.c
306
struct pm8xxx_pin_data *pin = pctrl->desc.pins[group].drv_data;
drivers/pinctrl/realtek/pinctrl-rtd.c
194
unsigned int function, unsigned int group)
drivers/pinctrl/realtek/pinctrl-rtd.c
204
group_name = data->info->groups[group].name;
drivers/pinctrl/realtek/pinctrl-rtd.c
206
ret = rtd_pinctrl_get_group_pins(pcdev, group, &pins, &num_pins);
drivers/pinctrl/realtek/pinctrl-rtd.c
503
static int rtd_pin_config_group_set(struct pinctrl_dev *pcdev, unsigned int group,
drivers/pinctrl/realtek/pinctrl-rtd.c
512
group_name = data->info->groups[group].name;
drivers/pinctrl/realtek/pinctrl-rtd.c
514
ret = rtd_pinctrl_get_group_pins(pcdev, group, &pins, &num_pins);
drivers/pinctrl/renesas/core.c
1170
const struct sh_pfc_pin_group *group = &info->groups[i];
drivers/pinctrl/renesas/core.c
1172
if (!group->name) {
drivers/pinctrl/renesas/core.c
1177
sh_pfc_compare_groups(drvname, group, &info->groups[j]);
drivers/pinctrl/renesas/core.c
1180
sh_pfc_err("orphan group %s\n", group->name);
drivers/pinctrl/renesas/core.c
1183
group->name, refcnts[i]);
drivers/pinctrl/renesas/pfc-r8a7740.c
3507
const struct r8a7740_portcr_group *group =
drivers/pinctrl/renesas/pfc-r8a7740.c
3510
if (pin <= group->end_pin)
drivers/pinctrl/renesas/pfc-r8a7740.c
3511
return group->offset + pin;
drivers/pinctrl/renesas/pinctrl-rza1.c
1072
(*map)->data.mux.group = np->name;
drivers/pinctrl/renesas/pinctrl-rza1.c
1118
unsigned int group)
drivers/pinctrl/renesas/pinctrl-rza1.c
1126
grp = pinctrl_generic_get_group(pctldev, group);
drivers/pinctrl/renesas/pinctrl-rza2.c
405
(*map)->data.mux.group = np->name;
drivers/pinctrl/renesas/pinctrl-rza2.c
442
unsigned int group)
drivers/pinctrl/renesas/pinctrl-rza2.c
449
grp = pinctrl_generic_get_group(pctldev, group);
drivers/pinctrl/renesas/pinctrl-rzg2l.c
1621
unsigned int group,
drivers/pinctrl/renesas/pinctrl-rzg2l.c
1629
ret = pinctrl_generic_get_group_pins(pctldev, group, &pins, &npins);
drivers/pinctrl/renesas/pinctrl-rzg2l.c
1644
unsigned int group,
drivers/pinctrl/renesas/pinctrl-rzg2l.c
1651
ret = pinctrl_generic_get_group_pins(pctldev, group, &pins, &npins);
drivers/pinctrl/renesas/pinctrl-rzg2l.c
587
struct group_desc *group;
drivers/pinctrl/renesas/pinctrl-rzg2l.c
594
group = pinctrl_generic_get_group(pctldev, group_selector);
drivers/pinctrl/renesas/pinctrl-rzg2l.c
595
if (!group)
drivers/pinctrl/renesas/pinctrl-rzg2l.c
599
pins = group->grp.pins;
drivers/pinctrl/renesas/pinctrl-rzg2l.c
601
for (i = 0; i < group->grp.npins; i++) {
drivers/pinctrl/renesas/pinctrl-rzg2l.c
70
#define RZG2L_MPXED_COMMON_PIN_FUNCS(group) \
drivers/pinctrl/renesas/pinctrl-rzg2l.c
71
(PIN_CFG_IOLH_##group | \
drivers/pinctrl/renesas/pinctrl-rzg2l.c
78
#define RZG3S_MPXED_PIN_FUNCS(group) (RZG2L_MPXED_COMMON_PIN_FUNCS(group) | \
drivers/pinctrl/renesas/pinctrl-rzg2l.c
790
maps[idx].data.mux.group = name;
drivers/pinctrl/renesas/pinctrl-rzn1.c
462
unsigned int group)
drivers/pinctrl/renesas/pinctrl-rzn1.c
465
struct rzn1_pin_group *grp = &ipctl->groups[group];
drivers/pinctrl/renesas/pinctrl-rzn1.c
469
ipctl->functions[selector].name, selector, grp->name, group);
drivers/pinctrl/renesas/pinctrl-rzt2h.c
173
struct group_desc *group;
drivers/pinctrl/renesas/pinctrl-rzt2h.c
183
group = pinctrl_generic_get_group(pctldev, group_selector);
drivers/pinctrl/renesas/pinctrl-rzt2h.c
184
if (!group)
drivers/pinctrl/renesas/pinctrl-rzt2h.c
188
pins = group->grp.pins;
drivers/pinctrl/renesas/pinctrl-rzt2h.c
190
for (i = 0; i < group->grp.npins; i++) {
drivers/pinctrl/renesas/pinctrl-rzt2h.c
375
maps[idx].data.mux.group = name;
drivers/pinctrl/renesas/pinctrl-rzv2m.c
167
struct group_desc *group;
drivers/pinctrl/renesas/pinctrl-rzv2m.c
173
group = pinctrl_generic_get_group(pctldev, group_selector);
drivers/pinctrl/renesas/pinctrl-rzv2m.c
174
if (!group)
drivers/pinctrl/renesas/pinctrl-rzv2m.c
178
pins = group->grp.pins;
drivers/pinctrl/renesas/pinctrl-rzv2m.c
180
for (i = 0; i < group->grp.npins; i++) {
drivers/pinctrl/renesas/pinctrl-rzv2m.c
349
maps[idx].data.mux.group = name;
drivers/pinctrl/renesas/pinctrl-rzv2m.c
672
unsigned int group,
drivers/pinctrl/renesas/pinctrl-rzv2m.c
680
ret = pinctrl_generic_get_group_pins(pctldev, group, &pins, &npins);
drivers/pinctrl/renesas/pinctrl-rzv2m.c
695
unsigned int group,
drivers/pinctrl/renesas/pinctrl-rzv2m.c
702
ret = pinctrl_generic_get_group_pins(pctldev, group, &pins, &npins);
drivers/pinctrl/renesas/pinctrl.c
114
const char *group;
drivers/pinctrl/renesas/pinctrl.c
180
of_property_for_each_string(np, "groups", prop, group) {
drivers/pinctrl/renesas/pinctrl.c
183
maps[idx].data.mux.group = group;
drivers/pinctrl/renesas/pinctrl.c
189
ret = sh_pfc_map_add_config(&maps[idx], group,
drivers/pinctrl/renesas/pinctrl.c
319
unsigned group)
drivers/pinctrl/renesas/pinctrl.c
323
const struct sh_pfc_pin_group *grp = &pfc->info->groups[group];
drivers/pinctrl/renesas/pinctrl.c
722
static int sh_pfc_pinconf_group_set(struct pinctrl_dev *pctldev, unsigned group,
drivers/pinctrl/renesas/pinctrl.c
732
pins = pmx->pfc->info->groups[group].pins;
drivers/pinctrl/renesas/pinctrl.c
733
num_pins = pmx->pfc->info->groups[group].nr_pins;
drivers/pinctrl/samsung/pinctrl-exynos.c
342
unsigned int svc, group, pin;
drivers/pinctrl/samsung/pinctrl-exynos.c
358
group = EXYNOS_SVC_GROUP(svc);
drivers/pinctrl/samsung/pinctrl-exynos.c
361
if (!group)
drivers/pinctrl/samsung/pinctrl-exynos.c
363
bank += (group - 1);
drivers/pinctrl/samsung/pinctrl-s3c64xx.c
428
unsigned int group;
drivers/pinctrl/samsung/pinctrl-s3c64xx.c
433
group = SVC_GROUP(svc);
drivers/pinctrl/samsung/pinctrl-s3c64xx.c
436
if (!group)
drivers/pinctrl/samsung/pinctrl-s3c64xx.c
440
if (group == 1) {
drivers/pinctrl/samsung/pinctrl-s3c64xx.c
442
group = 0;
drivers/pinctrl/samsung/pinctrl-s3c64xx.c
447
ret = generic_handle_domain_irq(data->domains[group], pin);
drivers/pinctrl/samsung/pinctrl-samsung.c
101
unsigned *num_maps, const char *group,
drivers/pinctrl/samsung/pinctrl-samsung.c
108
(*map)[*num_maps].data.mux.group = group;
drivers/pinctrl/samsung/pinctrl-samsung.c
117
const char *group, unsigned long *configs,
drivers/pinctrl/samsung/pinctrl-samsung.c
131
(*map)[*num_maps].data.configs.group_or_pin = group;
drivers/pinctrl/samsung/pinctrl-samsung.c
186
const char *group;
drivers/pinctrl/samsung/pinctrl-samsung.c
223
of_property_for_each_string(np, "samsung,pins", prop, group) {
drivers/pinctrl/samsung/pinctrl-samsung.c
226
num_maps, group, np->full_name);
drivers/pinctrl/samsung/pinctrl-samsung.c
233
num_maps, group, configs,
drivers/pinctrl/samsung/pinctrl-samsung.c
374
unsigned group)
drivers/pinctrl/samsung/pinctrl-samsung.c
388
grp = &drvdata->pin_groups[group];
drivers/pinctrl/samsung/pinctrl-samsung.c
423
unsigned group)
drivers/pinctrl/samsung/pinctrl-samsung.c
425
return samsung_pinmux_setup(pctldev, selector, group);
drivers/pinctrl/samsung/pinctrl-samsung.c
514
unsigned group, unsigned long *configs,
drivers/pinctrl/samsung/pinctrl-samsung.c
522
pins = drvdata->pin_groups[group].pins;
drivers/pinctrl/samsung/pinctrl-samsung.c
524
for (cnt = 0; cnt < drvdata->pin_groups[group].num_pins; cnt++)
drivers/pinctrl/samsung/pinctrl-samsung.c
532
unsigned int group, unsigned long *config)
drivers/pinctrl/samsung/pinctrl-samsung.c
538
pins = drvdata->pin_groups[group].pins;
drivers/pinctrl/samsung/pinctrl-samsung.c
57
unsigned group)
drivers/pinctrl/samsung/pinctrl-samsung.c
61
return pmx->pin_groups[group].name;
drivers/pinctrl/samsung/pinctrl-samsung.c
65
unsigned group,
drivers/pinctrl/samsung/pinctrl-samsung.c
71
*pins = pmx->pin_groups[group].pins;
drivers/pinctrl/samsung/pinctrl-samsung.c
72
*num_pins = pmx->pin_groups[group].num_pins;
drivers/pinctrl/sophgo/pinctrl-sophgo-common.c
174
map[nmaps].data.mux.group = grpname;
drivers/pinctrl/sophgo/pinctrl-sophgo-common.c
225
const struct group_desc *group;
drivers/pinctrl/sophgo/pinctrl-sophgo-common.c
229
group = pinctrl_generic_get_group(pctldev, gsel);
drivers/pinctrl/sophgo/pinctrl-sophgo-common.c
230
if (!group)
drivers/pinctrl/sophgo/pinctrl-sophgo-common.c
233
configs = group->data;
drivers/pinctrl/sophgo/pinctrl-sophgo-common.c
235
for (i = 0; i < group->grp.npins; i++) {
drivers/pinctrl/sophgo/pinctrl-sophgo-common.c
283
const struct group_desc *group;
drivers/pinctrl/sophgo/pinctrl-sophgo-common.c
288
group = pinctrl_generic_get_group(pctldev, gsel);
drivers/pinctrl/sophgo/pinctrl-sophgo-common.c
289
if (!group)
drivers/pinctrl/sophgo/pinctrl-sophgo-common.c
292
pinmuxs = group->data;
drivers/pinctrl/sophgo/pinctrl-sophgo-common.c
299
for (i = 0; i < group->grp.npins; i++)
drivers/pinctrl/sophgo/pinctrl-sophgo-common.c
300
sophgo_pin_set_config(pctrl, group->grp.pins[i], value, mask);
drivers/pinctrl/spacemit/pinctrl-k1.c
558
map[nmaps].data.mux.group = grpname;
drivers/pinctrl/spacemit/pinctrl-k1.c
607
const struct group_desc *group;
drivers/pinctrl/spacemit/pinctrl-k1.c
612
group = pinctrl_generic_get_group(pctldev, gsel);
drivers/pinctrl/spacemit/pinctrl-k1.c
613
if (!group)
drivers/pinctrl/spacemit/pinctrl-k1.c
616
configs = group->data;
drivers/pinctrl/spacemit/pinctrl-k1.c
618
for (i = 0; i < group->grp.npins; i++) {
drivers/pinctrl/spacemit/pinctrl-k1.c
832
const struct group_desc *group;
drivers/pinctrl/spacemit/pinctrl-k1.c
836
group = pinctrl_generic_get_group(pctldev, gsel);
drivers/pinctrl/spacemit/pinctrl-k1.c
837
if (!group)
drivers/pinctrl/spacemit/pinctrl-k1.c
840
spin = spacemit_get_pin(pctrl, group->grp.pins[0]);
drivers/pinctrl/spacemit/pinctrl-k1.c
845
for (i = 0; i < group->grp.npins; i++)
drivers/pinctrl/spacemit/pinctrl-k1.c
846
spacemit_pin_set_config(pctrl, group->grp.pins[i], value);
drivers/pinctrl/spear/pinctrl-spear.c
101
int i, j, group;
drivers/pinctrl/spear/pinctrl-spear.c
103
for (group = 0; group < machdata->ngroups; group++) {
drivers/pinctrl/spear/pinctrl-spear.c
104
pgroup = machdata->groups[group];
drivers/pinctrl/spear/pinctrl-spear.c
124
unsigned group)
drivers/pinctrl/spear/pinctrl-spear.c
128
return pmx->machdata->groups[group]->name;
drivers/pinctrl/spear/pinctrl-spear.c
132
unsigned group, const unsigned **pins, unsigned *num_pins)
drivers/pinctrl/spear/pinctrl-spear.c
136
*pins = pmx->machdata->groups[group]->pins;
drivers/pinctrl/spear/pinctrl-spear.c
137
*num_pins = pmx->machdata->groups[group]->npins;
drivers/pinctrl/spear/pinctrl-spear.c
155
const char *function, *group;
drivers/pinctrl/spear/pinctrl-spear.c
182
of_property_for_each_string(np, "st,pins", prop, group) {
drivers/pinctrl/spear/pinctrl-spear.c
184
(*map)[index].data.mux.group = group;
drivers/pinctrl/spear/pinctrl-spear.c
239
unsigned function, unsigned group, bool enable)
drivers/pinctrl/spear/pinctrl-spear.c
247
pgroup = pmx->machdata->groups[group];
drivers/pinctrl/spear/pinctrl-spear.c
273
unsigned group)
drivers/pinctrl/spear/pinctrl-spear.c
275
return spear_pinctrl_endisable(pctldev, function, group, true);
drivers/pinctrl/starfive/pinctrl-starfive-jh7100.c
564
map[nmaps].data.mux.group = grpname;
drivers/pinctrl/starfive/pinctrl-starfive-jh7100.c
643
const struct group_desc *group;
drivers/pinctrl/starfive/pinctrl-starfive-jh7100.c
647
group = pinctrl_generic_get_group(pctldev, gsel);
drivers/pinctrl/starfive/pinctrl-starfive-jh7100.c
648
if (!group)
drivers/pinctrl/starfive/pinctrl-starfive-jh7100.c
651
pinmux = group->data;
drivers/pinctrl/starfive/pinctrl-starfive-jh7100.c
652
for (i = 0; i < group->grp.npins; i++) {
drivers/pinctrl/starfive/pinctrl-starfive-jh7100.c
789
const struct group_desc *group;
drivers/pinctrl/starfive/pinctrl-starfive-jh7100.c
791
group = pinctrl_generic_get_group(pctldev, gsel);
drivers/pinctrl/starfive/pinctrl-starfive-jh7100.c
792
if (!group)
drivers/pinctrl/starfive/pinctrl-starfive-jh7100.c
795
return starfive_pinconf_get(pctldev, group->grp.pins[0], config);
drivers/pinctrl/starfive/pinctrl-starfive-jh7100.c
804
const struct group_desc *group;
drivers/pinctrl/starfive/pinctrl-starfive-jh7100.c
808
group = pinctrl_generic_get_group(pctldev, gsel);
drivers/pinctrl/starfive/pinctrl-starfive-jh7100.c
809
if (!group)
drivers/pinctrl/starfive/pinctrl-starfive-jh7100.c
874
for (i = 0; i < group->grp.npins; i++)
drivers/pinctrl/starfive/pinctrl-starfive-jh7100.c
875
starfive_padctl_rmw(sfp, group->grp.pins[i], mask, value);
drivers/pinctrl/starfive/pinctrl-starfive-jh7110-sys.c
232
u8 group;
drivers/pinctrl/starfive/pinctrl-starfive-jh7110-sys.c
323
grp = gs->group << gs->shift;
drivers/pinctrl/starfive/pinctrl-starfive-jh7110.c
196
map[nmaps].data.mux.group = grpname;
drivers/pinctrl/starfive/pinctrl-starfive-jh7110.c
299
const struct group_desc *group;
drivers/pinctrl/starfive/pinctrl-starfive-jh7110.c
303
group = pinctrl_generic_get_group(pctldev, gsel);
drivers/pinctrl/starfive/pinctrl-starfive-jh7110.c
304
if (!group)
drivers/pinctrl/starfive/pinctrl-starfive-jh7110.c
307
pinmux = group->data;
drivers/pinctrl/starfive/pinctrl-starfive-jh7110.c
308
for (i = 0; i < group->grp.npins; i++) {
drivers/pinctrl/starfive/pinctrl-starfive-jh7110.c
432
const struct group_desc *group;
drivers/pinctrl/starfive/pinctrl-starfive-jh7110.c
434
group = pinctrl_generic_get_group(pctldev, gsel);
drivers/pinctrl/starfive/pinctrl-starfive-jh7110.c
435
if (!group)
drivers/pinctrl/starfive/pinctrl-starfive-jh7110.c
438
return jh7110_pinconf_get(pctldev, group->grp.pins[0], config);
drivers/pinctrl/starfive/pinctrl-starfive-jh7110.c
447
const struct group_desc *group;
drivers/pinctrl/starfive/pinctrl-starfive-jh7110.c
451
group = pinctrl_generic_get_group(pctldev, gsel);
drivers/pinctrl/starfive/pinctrl-starfive-jh7110.c
452
if (!group)
drivers/pinctrl/starfive/pinctrl-starfive-jh7110.c
509
for (i = 0; i < group->grp.npins; i++)
drivers/pinctrl/starfive/pinctrl-starfive-jh7110.c
510
jh7110_padcfg_rmw(sfp, group->grp.pins[i], mask, value);
drivers/pinctrl/stm32/pinctrl-stm32.c
1453
unsigned group,
drivers/pinctrl/stm32/pinctrl-stm32.c
1458
*config = pctl->groups[group].config;
drivers/pinctrl/stm32/pinctrl-stm32.c
1463
static int stm32_pconf_group_set(struct pinctrl_dev *pctldev, unsigned group,
drivers/pinctrl/stm32/pinctrl-stm32.c
1467
struct stm32_pinctrl_group *g = &pctl->groups[group];
drivers/pinctrl/stm32/pinctrl-stm32.c
1824
struct stm32_pinctrl_group *group = pctl->groups + i;
drivers/pinctrl/stm32/pinctrl-stm32.c
1826
group->name = pin->pin.name;
drivers/pinctrl/stm32/pinctrl-stm32.c
1827
group->pin = pin->pin.number;
drivers/pinctrl/stm32/pinctrl-stm32.c
743
(*map)[*num_maps].data.mux.group = grp->name;
drivers/pinctrl/stm32/pinctrl-stm32.c
880
unsigned group)
drivers/pinctrl/stm32/pinctrl-stm32.c
884
return pctl->groups[group].name;
drivers/pinctrl/stm32/pinctrl-stm32.c
888
unsigned group,
drivers/pinctrl/stm32/pinctrl-stm32.c
894
*pins = (unsigned *)&pctl->groups[group].pin;
drivers/pinctrl/stm32/pinctrl-stm32.c
994
unsigned group)
drivers/pinctrl/stm32/pinctrl-stm32.c
998
struct stm32_pinctrl_group *g = pctl->groups + group;
drivers/pinctrl/sunplus/sppctl.c
912
(*map)[i].data.mux.group = pin_get_name(pctldev, pin_num);
drivers/pinctrl/sunplus/sppctl.c
914
dev_dbg(pctldev->dev, "%s: %s\n", (*map)[i].data.mux.group,
drivers/pinctrl/sunplus/sppctl.c
928
(*map)[*num_maps].data.mux.group = s_g;
drivers/pinctrl/sunxi/pinctrl-sunxi.c
123
sunxi_pinctrl_find_group_by_name(struct sunxi_pinctrl *pctl, const char *group)
drivers/pinctrl/sunxi/pinctrl-sunxi.c
130
if (!strcmp(grp->name, group))
drivers/pinctrl/sunxi/pinctrl-sunxi.c
1383
struct sunxi_pinctrl_group *group = pctl->groups + pctl->ngroups;
drivers/pinctrl/sunxi/pinctrl-sunxi.c
1388
group->name = pin->pin.name;
drivers/pinctrl/sunxi/pinctrl-sunxi.c
1389
group->pin = pin->pin.number;
drivers/pinctrl/sunxi/pinctrl-sunxi.c
244
unsigned group)
drivers/pinctrl/sunxi/pinctrl-sunxi.c
248
return pctl->groups[group].name;
drivers/pinctrl/sunxi/pinctrl-sunxi.c
252
unsigned group,
drivers/pinctrl/sunxi/pinctrl-sunxi.c
258
*pins = (unsigned *)&pctl->groups[group].pin;
drivers/pinctrl/sunxi/pinctrl-sunxi.c
438
const char *group;
drivers/pinctrl/sunxi/pinctrl-sunxi.c
477
of_property_for_each_string(node, pin_prop, prop, group) {
drivers/pinctrl/sunxi/pinctrl-sunxi.c
479
sunxi_pinctrl_find_group_by_name(pctl, group);
drivers/pinctrl/sunxi/pinctrl-sunxi.c
482
dev_err(pctl->dev, "unknown pin %s", group);
drivers/pinctrl/sunxi/pinctrl-sunxi.c
490
function, group);
drivers/pinctrl/sunxi/pinctrl-sunxi.c
495
(*map)[i].data.mux.group = group;
drivers/pinctrl/sunxi/pinctrl-sunxi.c
502
(*map)[i].data.configs.group_or_pin = group;
drivers/pinctrl/sunxi/pinctrl-sunxi.c
635
unsigned group,
drivers/pinctrl/sunxi/pinctrl-sunxi.c
639
struct sunxi_pinctrl_group *g = &pctl->groups[group];
drivers/pinctrl/sunxi/pinctrl-sunxi.c
707
static int sunxi_pconf_group_set(struct pinctrl_dev *pctldev, unsigned group,
drivers/pinctrl/sunxi/pinctrl-sunxi.c
711
struct sunxi_pinctrl_group *g = &pctl->groups[group];
drivers/pinctrl/sunxi/pinctrl-sunxi.c
845
unsigned group)
drivers/pinctrl/sunxi/pinctrl-sunxi.c
848
struct sunxi_pinctrl_group *g = pctl->groups + group;
drivers/pinctrl/tegra/pinctrl-tegra-xusb.c
116
unsigned int group)
drivers/pinctrl/tegra/pinctrl-tegra-xusb.c
120
return padctl->soc->pins[group].name;
drivers/pinctrl/tegra/pinctrl-tegra-xusb.c
124
unsigned group,
drivers/pinctrl/tegra/pinctrl-tegra-xusb.c
132
*pins = &pinctrl->desc->pins[group].number;
drivers/pinctrl/tegra/pinctrl-tegra-xusb.c
161
const char *function, *group;
drivers/pinctrl/tegra/pinctrl-tegra-xusb.c
208
of_property_for_each_string(np, "nvidia,lanes", prop, group) {
drivers/pinctrl/tegra/pinctrl-tegra-xusb.c
211
reserved_maps, num_maps, group,
drivers/pinctrl/tegra/pinctrl-tegra-xusb.c
219
maps, reserved_maps, num_maps, group,
drivers/pinctrl/tegra/pinctrl-tegra-xusb.c
296
unsigned int group)
drivers/pinctrl/tegra/pinctrl-tegra-xusb.c
303
lane = &padctl->soc->lanes[group];
drivers/pinctrl/tegra/pinctrl-tegra-xusb.c
328
unsigned int group,
drivers/pinctrl/tegra/pinctrl-tegra-xusb.c
337
lane = &padctl->soc->lanes[group];
drivers/pinctrl/tegra/pinctrl-tegra-xusb.c
365
unsigned int group,
drivers/pinctrl/tegra/pinctrl-tegra-xusb.c
376
lane = &padctl->soc->lanes[group];
drivers/pinctrl/tegra/pinctrl-tegra-xusb.c
422
unsigned int group)
drivers/pinctrl/tegra/pinctrl-tegra-xusb.c
432
err = tegra_xusb_padctl_pinconf_group_get(pinctrl, group,
drivers/pinctrl/tegra/pinctrl-tegra.c
117
const char *group;
drivers/pinctrl/tegra/pinctrl-tegra.c
160
of_property_for_each_string(np, "nvidia,pins", prop, group) {
drivers/pinctrl/tegra/pinctrl-tegra.c
163
reserved_maps, num_maps, group,
drivers/pinctrl/tegra/pinctrl-tegra.c
171
reserved_maps, num_maps, group,
drivers/pinctrl/tegra/pinctrl-tegra.c
252
unsigned group)
drivers/pinctrl/tegra/pinctrl-tegra.c
259
g = &pmx->soc->groups[group];
drivers/pinctrl/tegra/pinctrl-tegra.c
286
unsigned int group, num_pins, j;
drivers/pinctrl/tegra/pinctrl-tegra.c
290
for (group = 0; group < pmx->soc->ngroups; ++group) {
drivers/pinctrl/tegra/pinctrl-tegra.c
291
ret = tegra_pinctrl_get_group_pins(pctldev, group, &pins, &num_pins);
drivers/pinctrl/tegra/pinctrl-tegra.c
296
return group;
drivers/pinctrl/tegra/pinctrl-tegra.c
332
const struct tegra_pingroup *group;
drivers/pinctrl/tegra/pinctrl-tegra.c
341
group = tegra_pinctrl_get_group(pctldev, offset, group_index);
drivers/pinctrl/tegra/pinctrl-tegra.c
343
if (!group)
drivers/pinctrl/tegra/pinctrl-tegra.c
346
if (group->mux_reg < 0 || group->sfsel_bit < 0)
drivers/pinctrl/tegra/pinctrl-tegra.c
352
value = pmx_readl(pmx, group->mux_bank, group->mux_reg);
drivers/pinctrl/tegra/pinctrl-tegra.c
353
config->is_sfsel = (value & BIT(group->sfsel_bit)) != 0;
drivers/pinctrl/tegra/pinctrl-tegra.c
354
value &= ~BIT(group->sfsel_bit);
drivers/pinctrl/tegra/pinctrl-tegra.c
355
pmx_writel(pmx, value, group->mux_bank, group->mux_reg);
drivers/pinctrl/tegra/pinctrl-tegra.c
365
const struct tegra_pingroup *group;
drivers/pinctrl/tegra/pinctrl-tegra.c
374
group = tegra_pinctrl_get_group(pctldev, offset, group_index);
drivers/pinctrl/tegra/pinctrl-tegra.c
376
if (!group)
drivers/pinctrl/tegra/pinctrl-tegra.c
379
if (group->mux_reg < 0 || group->sfsel_bit < 0)
drivers/pinctrl/tegra/pinctrl-tegra.c
385
value = pmx_readl(pmx, group->mux_bank, group->mux_reg);
drivers/pinctrl/tegra/pinctrl-tegra.c
387
value |= BIT(group->sfsel_bit);
drivers/pinctrl/tegra/pinctrl-tegra.c
388
pmx_writel(pmx, value, group->mux_bank, group->mux_reg);
drivers/pinctrl/tegra/pinctrl-tegra.c
50
unsigned group)
drivers/pinctrl/tegra/pinctrl-tegra.c
54
return pmx->soc->groups[group].name;
drivers/pinctrl/tegra/pinctrl-tegra.c
565
unsigned group, unsigned long *config)
drivers/pinctrl/tegra/pinctrl-tegra.c
576
g = &pmx->soc->groups[group];
drivers/pinctrl/tegra/pinctrl-tegra.c
58
unsigned group,
drivers/pinctrl/tegra/pinctrl-tegra.c
593
unsigned group, unsigned long *configs,
drivers/pinctrl/tegra/pinctrl-tegra.c
605
g = &pmx->soc->groups[group];
drivers/pinctrl/tegra/pinctrl-tegra.c
64
*pins = pmx->soc->groups[group].pins;
drivers/pinctrl/tegra/pinctrl-tegra.c
65
*num_pins = pmx->soc->groups[group].npins;
drivers/pinctrl/tegra/pinctrl-tegra.c
664
struct seq_file *s, unsigned group)
drivers/pinctrl/tegra/pinctrl-tegra.c
673
g = &pmx->soc->groups[group];
drivers/pinctrl/ti/pinctrl-ti-iodelay.c
589
struct ti_iodelay_pingroup *group;
drivers/pinctrl/ti/pinctrl-ti-iodelay.c
592
group = ti_iodelay_get_pingroup(iod, selector);
drivers/pinctrl/ti/pinctrl-ti-iodelay.c
594
if (!group)
drivers/pinctrl/ti/pinctrl-ti-iodelay.c
597
*config = group->config;
drivers/pinctrl/ti/pinctrl-ti-iodelay.c
617
struct ti_iodelay_pingroup *group;
drivers/pinctrl/ti/pinctrl-ti-iodelay.c
622
group = ti_iodelay_get_pingroup(iod, selector);
drivers/pinctrl/ti/pinctrl-ti-iodelay.c
635
for (i = 0; i < group->ncfg; i++) {
drivers/pinctrl/ti/pinctrl-ti-iodelay.c
636
if (ti_iodelay_pinconf_set(iod, &group->cfg[i]))
drivers/pinctrl/ti/pinctrl-ti-iodelay.c
705
struct ti_iodelay_pingroup *group;
drivers/pinctrl/ti/pinctrl-ti-iodelay.c
709
group = ti_iodelay_get_pingroup(iod, selector);
drivers/pinctrl/ti/pinctrl-ti-iodelay.c
710
if (!group)
drivers/pinctrl/ti/pinctrl-ti-iodelay.c
713
for (i = 0; i < group->ncfg; i++) {
drivers/pinctrl/ti/pinctrl-ti-iodelay.c
717
cfg = &group->cfg[i];
drivers/pinctrl/visconti/pinctrl-common.c
221
unsigned int function, unsigned int group)
drivers/pinctrl/visconti/pinctrl-common.c
225
const struct visconti_pin_group *grp = &priv->devdata->groups[group];
drivers/pinctrl/visconti/pinctrl-common.c
231
function, func->name, group, grp->name);
drivers/pinctrl/vt8500/pinctrl-wmt.c
212
int group;
drivers/pinctrl/vt8500/pinctrl-wmt.c
220
group = wmt_pctl_find_group_by_pin(data, pin);
drivers/pinctrl/vt8500/pinctrl-wmt.c
221
if (group < 0) {
drivers/pinctrl/vt8500/pinctrl-wmt.c
223
return group;
drivers/pinctrl/vt8500/pinctrl-wmt.c
227
map->data.mux.group = data->groups[group];
drivers/pinctrl/vt8500/pinctrl-wmt.c
239
int group;
drivers/pinctrl/vt8500/pinctrl-wmt.c
248
group = wmt_pctl_find_group_by_pin(data, pin);
drivers/pinctrl/vt8500/pinctrl-wmt.c
249
if (group < 0) {
drivers/pinctrl/vt8500/pinctrl-wmt.c
251
return group;
drivers/pinctrl/vt8500/pinctrl-wmt.c
274
map->data.configs.group_or_pin = data->groups[group];
drivers/platform/mellanox/mlxreg-hotplug.c
328
priv->group.attrs = devm_kcalloc(&priv->pdev->dev,
drivers/platform/mellanox/mlxreg-hotplug.c
332
if (!priv->group.attrs)
drivers/platform/mellanox/mlxreg-hotplug.c
335
priv->group.attrs = priv->mlxreg_hotplug_attr;
drivers/platform/mellanox/mlxreg-hotplug.c
336
priv->groups[0] = &priv->group;
drivers/platform/mellanox/mlxreg-hotplug.c
65
struct attribute_group group;
drivers/platform/mellanox/mlxreg-io.c
184
priv->group.attrs = devm_kcalloc(&priv->pdev->dev,
drivers/platform/mellanox/mlxreg-io.c
188
if (!priv->group.attrs)
drivers/platform/mellanox/mlxreg-io.c
216
priv->group.attrs = priv->mlxreg_io_attr;
drivers/platform/mellanox/mlxreg-io.c
217
priv->groups[0] = &priv->group;
drivers/platform/mellanox/mlxreg-io.c
41
struct attribute_group group;
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1402
poll_services_of_group(struct vchiq_state *state, int group)
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1404
u32 flags = atomic_xchg(&state->poll_services[group], 0);
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1414
service = find_service_by_port(state, (group << 5) + i);
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1452
int group;
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1454
for (group = 0; group < BITSET_SIZE(state->unused_service); group++)
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1455
poll_services_of_group(state, group);
drivers/platform/x86/dasharo-acpi.c
111
obj[0].integer.value = data->capabilities[feat][channel].group;
drivers/platform/x86/dasharo-acpi.c
304
for (int group = 0; group < MAX_GROUPS_PER_FEAT; ++group) {
drivers/platform/x86/dasharo-acpi.c
305
count = dasharo_get_feature_cap_count(data, feat, group);
drivers/platform/x86/dasharo-acpi.c
314
cap->group = group;
drivers/platform/x86/dasharo-acpi.c
317
dasharo_group_names[feat][group], i);
drivers/platform/x86/dasharo-acpi.c
60
unsigned int group;
drivers/ps3/ps3-lpm.c
660
static u64 pm_signal_group_to_ps3_lv1_signal_group(u64 group)
drivers/ps3/ps3-lpm.c
669
if (group < 1000) {
drivers/ps3/ps3-lpm.c
670
if (group < 100) {
drivers/ps3/ps3-lpm.c
671
if (20 <= group && group < 30) {
drivers/ps3/ps3-lpm.c
673
subgroup = group - 20;
drivers/ps3/ps3-lpm.c
674
} else if (30 <= group && group < 40) {
drivers/ps3/ps3-lpm.c
676
subgroup = group - 30;
drivers/ps3/ps3-lpm.c
677
} else if (40 <= group && group < 50) {
drivers/ps3/ps3-lpm.c
679
subgroup = group - 40;
drivers/ps3/ps3-lpm.c
680
} else if (50 <= group && group < 60) {
drivers/ps3/ps3-lpm.c
682
subgroup = group - 50;
drivers/ps3/ps3-lpm.c
683
} else if (60 <= group && group < 70) {
drivers/ps3/ps3-lpm.c
685
subgroup = group - 60;
drivers/ps3/ps3-lpm.c
686
} else if (70 <= group && group < 80) {
drivers/ps3/ps3-lpm.c
688
subgroup = group - 70;
drivers/ps3/ps3-lpm.c
689
} else if (80 <= group && group < 90) {
drivers/ps3/ps3-lpm.c
691
subgroup = group - 80;
drivers/ps3/ps3-lpm.c
693
} else if (200 <= group && group < 300) {
drivers/ps3/ps3-lpm.c
695
subgroup = group - 200;
drivers/ps3/ps3-lpm.c
696
} else if (600 <= group && group < 700) {
drivers/ps3/ps3-lpm.c
699
subsubgroup = group - 650;
drivers/ps3/ps3-lpm.c
701
} else if (6000 <= group && group < 7000) {
drivers/ps3/ps3-lpm.c
704
subsubgroup = group - 6500;
drivers/ps3/ps3-lpm.c
725
__LINE__, group);
drivers/ptp/ptp_ocp.c
2388
bp->attr_group[count++] = attr_tbl[i].group;
drivers/ptp/ptp_ocp.c
293
const struct attribute_group *group;
drivers/ptp/ptp_ocp.c
3908
{ .cap = OCP_CAP_BASIC, .group = &fb_timecard_group },
drivers/ptp/ptp_ocp.c
3909
{ .cap = OCP_CAP_BASIC, .group = &ptp_ocp_timecard_tty_group },
drivers/ptp/ptp_ocp.c
3910
{ .cap = OCP_CAP_SIGNAL, .group = &fb_timecard_signal0_group },
drivers/ptp/ptp_ocp.c
3911
{ .cap = OCP_CAP_SIGNAL, .group = &fb_timecard_signal1_group },
drivers/ptp/ptp_ocp.c
3912
{ .cap = OCP_CAP_SIGNAL, .group = &fb_timecard_signal2_group },
drivers/ptp/ptp_ocp.c
3913
{ .cap = OCP_CAP_SIGNAL, .group = &fb_timecard_signal3_group },
drivers/ptp/ptp_ocp.c
3914
{ .cap = OCP_CAP_FREQ, .group = &fb_timecard_freq0_group },
drivers/ptp/ptp_ocp.c
3915
{ .cap = OCP_CAP_FREQ, .group = &fb_timecard_freq1_group },
drivers/ptp/ptp_ocp.c
3916
{ .cap = OCP_CAP_FREQ, .group = &fb_timecard_freq2_group },
drivers/ptp/ptp_ocp.c
3917
{ .cap = OCP_CAP_FREQ, .group = &fb_timecard_freq3_group },
drivers/ptp/ptp_ocp.c
3948
{ .cap = OCP_CAP_BASIC, .group = &art_timecard_group },
drivers/ptp/ptp_ocp.c
3949
{ .cap = OCP_CAP_BASIC, .group = &ptp_ocp_timecard_tty_group },
drivers/ptp/ptp_ocp.c
3976
{ .cap = OCP_CAP_BASIC, .group = &adva_timecard_group },
drivers/ptp/ptp_ocp.c
3977
{ .cap = OCP_CAP_BASIC, .group = &ptp_ocp_timecard_tty_group },
drivers/ptp/ptp_ocp.c
3978
{ .cap = OCP_CAP_SIGNAL, .group = &fb_timecard_signal0_group },
drivers/ptp/ptp_ocp.c
3979
{ .cap = OCP_CAP_SIGNAL, .group = &fb_timecard_signal1_group },
drivers/ptp/ptp_ocp.c
3980
{ .cap = OCP_CAP_FREQ, .group = &fb_timecard_freq0_group },
drivers/ptp/ptp_ocp.c
3981
{ .cap = OCP_CAP_FREQ, .group = &fb_timecard_freq1_group },
drivers/ras/amd/atl/denormalize.c
656
u16 group, group_offset, log_coh_st_offset;
drivers/ras/amd/atl/denormalize.c
742
group = log_coh_st_offset / mod_value;
drivers/ras/amd/atl/denormalize.c
791
hashed_bit = !!(group & BIT(1));
drivers/ras/amd/atl/denormalize.c
793
hashed_bit = group & BIT(0);
drivers/ras/amd/atl/denormalize.c
805
hashed_bit = group & BIT(0);
drivers/remoteproc/imx_rproc.c
928
unsigned long event, void *group)
drivers/remoteproc/imx_rproc.c
933
if (!((event & BIT(priv->rproc_pt)) && (*(u8 *)group == IMX_SC_IRQ_GROUP_REBOOTED)))
drivers/rtc/rtc-imx-sc.c
125
unsigned long event, void *group)
drivers/rtc/rtc-imx-sc.c
128
if (!((event & SC_IRQ_RTC) && (*(u8 *)group == SC_IRQ_GROUP_RTC)))
drivers/rtc/rtc-omap.c
574
unsigned int group)
drivers/rtc/rtc-stm32.c
376
unsigned int group)
drivers/rtc/rtc-stm32.c
379
struct pinctrl_pin_desc pin = stm32_rtc_pinctrl_pins[group];
drivers/s390/block/dasd_alias.c
309
struct alias_pav_group *group;
drivers/s390/block/dasd_alias.c
323
group = _find_group(lcu, &uid);
drivers/s390/block/dasd_alias.c
324
if (!group) {
drivers/s390/block/dasd_alias.c
325
group = kzalloc_obj(*group, GFP_ATOMIC);
drivers/s390/block/dasd_alias.c
326
if (!group)
drivers/s390/block/dasd_alias.c
328
memcpy(group->uid.vendor, uid.vendor, sizeof(uid.vendor));
drivers/s390/block/dasd_alias.c
329
memcpy(group->uid.serial, uid.serial, sizeof(uid.serial));
drivers/s390/block/dasd_alias.c
330
group->uid.ssid = uid.ssid;
drivers/s390/block/dasd_alias.c
332
group->uid.base_unit_addr = uid.real_unit_addr;
drivers/s390/block/dasd_alias.c
334
group->uid.base_unit_addr = uid.base_unit_addr;
drivers/s390/block/dasd_alias.c
335
memcpy(group->uid.vduit, uid.vduit, sizeof(uid.vduit));
drivers/s390/block/dasd_alias.c
336
INIT_LIST_HEAD(&group->group);
drivers/s390/block/dasd_alias.c
337
INIT_LIST_HEAD(&group->baselist);
drivers/s390/block/dasd_alias.c
338
INIT_LIST_HEAD(&group->aliaslist);
drivers/s390/block/dasd_alias.c
339
list_add(&group->group, &lcu->grouplist);
drivers/s390/block/dasd_alias.c
342
list_move(&device->alias_list, &group->baselist);
drivers/s390/block/dasd_alias.c
344
list_move(&device->alias_list, &group->aliaslist);
drivers/s390/block/dasd_alias.c
345
private->pavgroup = group;
drivers/s390/block/dasd_alias.c
353
struct alias_pav_group *group;
drivers/s390/block/dasd_alias.c
356
group = private->pavgroup;
drivers/s390/block/dasd_alias.c
357
if (!group)
drivers/s390/block/dasd_alias.c
360
if (list_empty(&group->baselist) && list_empty(&group->aliaslist)) {
drivers/s390/block/dasd_alias.c
361
list_del(&group->group);
drivers/s390/block/dasd_alias.c
362
kfree(group);
drivers/s390/block/dasd_alias.c
365
if (group->next == device)
drivers/s390/block/dasd_alias.c
366
group->next = NULL;
drivers/s390/block/dasd_alias.c
482
list_for_each_entry_safe(pavgroup, tempgroup, &lcu->grouplist, group) {
drivers/s390/block/dasd_alias.c
495
list_del(&pavgroup->group);
drivers/s390/block/dasd_alias.c
571
struct alias_pav_group *group;
drivers/s390/block/dasd_alias.c
582
group = list_first_entry(&lcu->grouplist,
drivers/s390/block/dasd_alias.c
583
struct alias_pav_group, group);
drivers/s390/block/dasd_alias.c
584
if (!list_empty(&group->baselist))
drivers/s390/block/dasd_alias.c
585
usedev = list_first_entry(&group->baselist,
drivers/s390/block/dasd_alias.c
588
else if (!list_empty(&group->aliaslist))
drivers/s390/block/dasd_alias.c
589
usedev = list_first_entry(&group->aliaslist,
drivers/s390/block/dasd_alias.c
671
struct alias_pav_group *group;
drivers/s390/block/dasd_alias.c
691
group = private->pavgroup;
drivers/s390/block/dasd_alias.c
692
if (!group) {
drivers/s390/block/dasd_alias.c
696
alias_device = group->next;
drivers/s390/block/dasd_alias.c
698
if (list_empty(&group->aliaslist)) {
drivers/s390/block/dasd_alias.c
702
alias_device = list_first_entry(&group->aliaslist,
drivers/s390/block/dasd_alias.c
707
if (list_is_last(&alias_device->alias_list, &group->aliaslist))
drivers/s390/block/dasd_alias.c
708
group->next = list_first_entry(&group->aliaslist,
drivers/s390/block/dasd_alias.c
711
group->next = list_first_entry(&alias_device->alias_list,
drivers/s390/block/dasd_alias.c
778
list_for_each_entry(pavgroup, &lcu->grouplist, group) {
drivers/s390/block/dasd_alias.c
81
struct alias_pav_group, group);
drivers/s390/block/dasd_alias.c
814
list_for_each_entry(pavgroup, &lcu->grouplist, group) {
drivers/s390/block/dasd_alias.c
852
list_for_each_entry(pavgroup, &lcu->grouplist, group) {
drivers/s390/block/dasd_alias.c
881
list_for_each_entry(pavgroup, &lcu->grouplist, group) {
drivers/s390/block/dasd_alias.c
89
list_for_each_entry(pos, &lcu->grouplist, group) {
drivers/s390/block/dasd_eckd.c
6539
&private->lcu->grouplist, group) {
drivers/s390/block/dasd_eckd.c
6604
group) {
drivers/s390/block/dasd_eckd.c
6685
group) {
drivers/s390/block/dasd_eckd.h
654
struct list_head group;
drivers/s390/net/ctcm_main.c
1692
static DRIVER_ATTR_WO(group);
drivers/s390/net/qeth_core_main.c
6495
static DRIVER_ATTR_WO(group);
drivers/scsi/aic7xxx/aic79xx_core.c
7961
int group;
drivers/scsi/aic7xxx/aic79xx_core.c
7963
group = XPT_FC_GROUP(scb->io_ctx->ccb_h.func_code);
drivers/scsi/aic7xxx/aic79xx_core.c
7965
match = (group != XPT_FC_GROUP_TMODE)
drivers/scsi/aic7xxx/aic79xx_core.c
7969
match = (group == XPT_FC_GROUP_TMODE)
drivers/scsi/aic7xxx/aic7xxx_core.c
5706
int group;
drivers/scsi/aic7xxx/aic7xxx_core.c
5708
group = XPT_FC_GROUP(scb->io_ctx->ccb_h.func_code);
drivers/scsi/aic7xxx/aic7xxx_core.c
5710
match = (group != XPT_FC_GROUP_TMODE)
drivers/scsi/aic7xxx/aic7xxx_core.c
5714
match = (group == XPT_FC_GROUP_TMODE)
drivers/scsi/elx/efct/efct_lio.c
730
struct config_group *group, const char *name)
drivers/scsi/elx/efct/efct_lio.c
762
struct config_group *group, const char *name)
drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c
3831
struct config_group *group,
drivers/scsi/mpt3sas/mpt3sas_base.c
3279
unsigned int i, group = nr_cpus / nr_msix;
drivers/scsi/mpt3sas/mpt3sas_base.c
3289
group++;
drivers/scsi/mpt3sas/mpt3sas_base.c
3291
for (i = 0 ; i < group ; i++) {
drivers/scsi/nsp32_debug.c
65
const char **table = commands[ group(opcode) ];
drivers/scsi/pcmcia/nsp_debug.c
67
const char **table = commands[ group(opcode) ];
drivers/scsi/qla2xxx/tcm_qla2xxx.c
1602
struct config_group *group,
drivers/scsi/qla2xxx/tcm_qla2xxx.c
1714
struct config_group *group,
drivers/scsi/scsi_debug.c
4923
u8 group = 0;
drivers/scsi/scsi_debug.c
4944
group = cmd[14] & 0x3f;
drivers/scsi/scsi_debug.c
4950
group = cmd[6] & 0x3f;
drivers/scsi/scsi_debug.c
4965
group = cmd[6] & 0x3f;
drivers/scsi/scsi_debug.c
4971
group = cmd[6] & 0x1f;
drivers/scsi/scsi_debug.c
4976
group = cmd[6] & 0x3f;
drivers/scsi/scsi_debug.c
5038
ret = do_device_access(sip, scp, 0, lba, num, group, true, false);
drivers/scsi/scsi_debug.c
5093
u8 group;
drivers/scsi/scsi_debug.c
5100
group = cmd[6] & 0x3f;
drivers/scsi/scsi_debug.c
5111
group = cmd[14] & 0x3f;
drivers/scsi/scsi_debug.c
5203
ret = do_device_access(sip, scp, sg_off, lba, num, group, true, true);
drivers/scsi/scsi_transport_iscsi.c
2532
iscsi_multicast_skb(struct sk_buff *skb, uint32_t group, gfp_t gfp)
drivers/scsi/scsi_transport_iscsi.c
2534
return nlmsg_multicast(nls, skb, 0, group, gfp);
drivers/scsi/scsi_transport_iscsi.c
3883
iscsi_if_recv_msg(struct sk_buff *skb, struct nlmsghdr *nlh, uint32_t *group)
drivers/scsi/scsi_transport_iscsi.c
3898
*group = ISCSI_NL_GRP_UIP;
drivers/scsi/scsi_transport_iscsi.c
3900
*group = ISCSI_NL_GRP_ISCSID;
drivers/scsi/scsi_transport_iscsi.c
4063
uint32_t group;
drivers/scsi/scsi_transport_iscsi.c
4077
err = iscsi_if_recv_msg(skb, nlh, &group);
drivers/scsi/smartpqi/smartpqi_init.c
2976
u32 group;
drivers/scsi/smartpqi/smartpqi_init.c
2978
group = rmd->map_index / rmd->data_disks_per_row;
drivers/scsi/smartpqi/smartpqi_init.c
2980
index = rmd->map_index - (group * rmd->data_disks_per_row);
drivers/scsi/smartpqi/smartpqi_init.c
2998
u32 group;
drivers/scsi/smartpqi/smartpqi_init.c
3027
group = device->next_bypass_group[rmd.map_index];
drivers/scsi/smartpqi/smartpqi_init.c
3028
next_bypass_group = group + 1;
drivers/scsi/smartpqi/smartpqi_init.c
3032
rmd.map_index += group * rmd.data_disks_per_row;
drivers/soc/samsung/exynos-asv.h
44
unsigned int group;
drivers/soc/samsung/exynos-asv.h
60
unsigned int level, unsigned int group)
drivers/soc/samsung/exynos-asv.h
62
return __asv_get_table_entry(&subsys->table, level, group + 1);
drivers/soc/samsung/exynos5422-asv.c
428
subsys->asv->group);
drivers/soc/samsung/exynos5422-asv.c
470
asv->group = exynos5422_asv_get_group(asv);
drivers/soc/tegra/pmc.c
2244
unsigned int group)
drivers/soc/tegra/pmc.c
2248
return pmc->soc->io_pads[group].name;
drivers/soc/tegra/pmc.c
2252
unsigned int group,
drivers/soc/tegra/pmc.c
2258
*pins = &pmc->soc->io_pads[group].id;
drivers/soundwire/generic_bandwidth_allocation.c
228
struct sdw_group *group)
drivers/soundwire/generic_bandwidth_allocation.c
237
for (i = 0; i < group->count; i++) {
drivers/soundwire/generic_bandwidth_allocation.c
238
params[i].rate = group->rates[i];
drivers/soundwire/generic_bandwidth_allocation.c
239
params[i].lane = group->lanes[i];
drivers/soundwire/generic_bandwidth_allocation.c
263
for (i = 0; i < group->count; i++) {
drivers/soundwire/generic_bandwidth_allocation.c
275
for (i = 0; i < group->count; i++) {
drivers/soundwire/generic_bandwidth_allocation.c
296
static int sdw_add_element_group_count(struct sdw_group *group,
drivers/soundwire/generic_bandwidth_allocation.c
299
int num = group->count;
drivers/soundwire/generic_bandwidth_allocation.c
303
if (rate == group->rates[i] && lane == group->lanes[i])
drivers/soundwire/generic_bandwidth_allocation.c
309
if (group->count >= group->max_size) {
drivers/soundwire/generic_bandwidth_allocation.c
313
group->max_size += 1;
drivers/soundwire/generic_bandwidth_allocation.c
314
rates = krealloc(group->rates,
drivers/soundwire/generic_bandwidth_allocation.c
315
(sizeof(int) * group->max_size),
drivers/soundwire/generic_bandwidth_allocation.c
320
group->rates = rates;
drivers/soundwire/generic_bandwidth_allocation.c
322
lanes = krealloc(group->lanes,
drivers/soundwire/generic_bandwidth_allocation.c
323
(sizeof(int) * group->max_size),
drivers/soundwire/generic_bandwidth_allocation.c
328
group->lanes = lanes;
drivers/soundwire/generic_bandwidth_allocation.c
331
group->rates[group->count] = rate;
drivers/soundwire/generic_bandwidth_allocation.c
332
group->lanes[group->count++] = lane;
drivers/soundwire/generic_bandwidth_allocation.c
339
struct sdw_group *group)
drivers/soundwire/generic_bandwidth_allocation.c
346
group->count = 0;
drivers/soundwire/generic_bandwidth_allocation.c
347
group->max_size = SDW_STRM_RATE_GROUPING;
drivers/soundwire/generic_bandwidth_allocation.c
348
group->rates = kcalloc(group->max_size, sizeof(int), GFP_KERNEL);
drivers/soundwire/generic_bandwidth_allocation.c
349
if (!group->rates)
drivers/soundwire/generic_bandwidth_allocation.c
352
group->lanes = kcalloc(group->max_size, sizeof(int), GFP_KERNEL);
drivers/soundwire/generic_bandwidth_allocation.c
353
if (!group->lanes) {
drivers/soundwire/generic_bandwidth_allocation.c
354
kfree(group->rates);
drivers/soundwire/generic_bandwidth_allocation.c
355
group->rates = NULL;
drivers/soundwire/generic_bandwidth_allocation.c
367
group->rates[group->count++] = rate;
drivers/soundwire/generic_bandwidth_allocation.c
374
ret = sdw_add_element_group_count(group, rate, p_rt->lane);
drivers/soundwire/generic_bandwidth_allocation.c
376
kfree(group->rates);
drivers/soundwire/generic_bandwidth_allocation.c
377
kfree(group->lanes);
drivers/soundwire/generic_bandwidth_allocation.c
395
struct sdw_group group;
drivers/soundwire/generic_bandwidth_allocation.c
398
ret = sdw_get_group_count(bus, &group);
drivers/soundwire/generic_bandwidth_allocation.c
402
if (group.count == 0)
drivers/soundwire/generic_bandwidth_allocation.c
405
params = kzalloc_objs(*params, group.count);
drivers/soundwire/generic_bandwidth_allocation.c
412
ret = sdw_compute_group_params(bus, stream, params, &group);
drivers/soundwire/generic_bandwidth_allocation.c
416
_sdw_compute_port_params(bus, params, group.count);
drivers/soundwire/generic_bandwidth_allocation.c
421
kfree(group.rates);
drivers/soundwire/generic_bandwidth_allocation.c
422
kfree(group.lanes);
drivers/staging/media/atomisp/pci/ia_css_acc_types.h
308
u32 group; /** Per pipeline data, loaded by dma */
drivers/staging/media/atomisp/pci/runtime/pipeline/src/pipeline.c
438
HIVE_ADDR_sp_group = fw->info.sp.group;
drivers/staging/media/ipu3/ipu3-css-fw.h
108
u32 group; /* Per pipeline data, loaded by dma */
drivers/staging/octeon/ethernet-rx.c
204
BIT(rx_group->group));
drivers/staging/octeon/ethernet-rx.c
210
BIT(rx_group->group));
drivers/staging/octeon/ethernet-rx.c
235
BIT(rx_group->group));
drivers/staging/octeon/ethernet-rx.c
237
BIT(rx_group->group));
drivers/staging/octeon/ethernet-rx.c
242
wq_int.s.iq_dis = BIT(rx_group->group);
drivers/staging/octeon/ethernet-rx.c
243
wq_int.s.wq_int = BIT(rx_group->group);
drivers/staging/octeon/ethernet-rx.c
36
int group;
drivers/staging/octeon/ethernet-rx.c
477
oct_rx_group[i].group = i;
drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c
1166
static void hal_get_chnl_group_8723b(u8 channel, u8 *group)
drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c
1169
*group = 0;
drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c
1171
*group = 1;
drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c
1173
*group = 2;
drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c
1175
*group = 3;
drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c
1177
*group = 4;
drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c
1220
u32 rfPath, eeAddr = EEPROM_TX_PWR_INX_8723B, group, TxCount = 0;
drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c
1230
for (group = 0; group < MAX_CHNL_GROUP_24G; group++) {
drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c
1231
pwrInfo24G->IndexCCK_Base[rfPath][group] = EEPROM_DEFAULT_24G_INDEX;
drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c
1232
pwrInfo24G->IndexBW40_Base[rfPath][group] = EEPROM_DEFAULT_24G_INDEX;
drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c
1255
for (group = 0; group < MAX_CHNL_GROUP_24G; group++) {
drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c
1256
pwrInfo24G->IndexCCK_Base[rfPath][group] = PROMContent[eeAddr++];
drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c
1257
if (pwrInfo24G->IndexCCK_Base[rfPath][group] == 0xFF)
drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c
1258
pwrInfo24G->IndexCCK_Base[rfPath][group] = EEPROM_DEFAULT_24G_INDEX;
drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c
1261
for (group = 0; group < MAX_CHNL_GROUP_24G-1; group++) {
drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c
1262
pwrInfo24G->IndexBW40_Base[rfPath][group] = PROMContent[eeAddr++];
drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c
1263
if (pwrInfo24G->IndexBW40_Base[rfPath][group] == 0xFF)
drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c
1264
pwrInfo24G->IndexBW40_Base[rfPath][group] = EEPROM_DEFAULT_24G_INDEX;
drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c
1338
u8 group = 0;
drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c
1340
hal_get_chnl_group_8723b(ch + 1, &group);
drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c
1344
pHalData->Index24G_BW40_Base[rfPath][ch] = pwrInfo24G.IndexBW40_Base[rfPath][group];
drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c
1346
pHalData->Index24G_CCK_Base[rfPath][ch] = pwrInfo24G.IndexCCK_Base[rfPath][group];
drivers/staging/rtl8723bs/hal/rtl8723b_hal_init.c
1347
pHalData->Index24G_BW40_Base[rfPath][ch] = pwrInfo24G.IndexBW40_Base[rfPath][group];
drivers/target/iscsi/iscsi_target_configfs.c
1187
struct config_group *group,
drivers/target/iscsi/iscsi_target_configfs.c
156
struct config_group *group,
drivers/target/loopback/tcm_loop.c
1029
struct config_group *group,
drivers/target/sbp/sbp_target.c
2024
struct config_group *group,
drivers/target/target_core_configfs.c
183
struct config_group *group,
drivers/target/target_core_configfs.c
190
" %s\n", group, name);
drivers/target/target_core_configfs.c
268
struct config_group *group,
drivers/target/target_core_configfs.c
2875
struct config_group *group,
drivers/target/target_core_configfs.c
2901
struct config_group *group,
drivers/target/target_core_configfs.c
3305
struct config_group *group,
drivers/target/target_core_configfs.c
3308
struct t10_alua *alua = container_of(group, struct t10_alua,
drivers/target/target_core_configfs.c
3332
struct config_group *group,
drivers/target/target_core_configfs.c
3376
struct config_group *group,
drivers/target/target_core_configfs.c
3383
struct config_group *group,
drivers/target/target_core_configfs.c
3401
struct config_group *group,
drivers/target/target_core_configfs.c
3405
struct config_item *hba_ci = &group->cg_item;
drivers/target/target_core_configfs.c
3481
struct config_group *group,
drivers/target/target_core_configfs.c
3607
struct config_group *group,
drivers/target/target_core_configfs.c
3666
struct config_group *group,
drivers/target/target_core_fabric_configfs.c
1102
struct config_group *group,
drivers/target/target_core_fabric_configfs.c
1105
struct target_fabric_configfs *tf = container_of(group,
drivers/target/target_core_fabric_configfs.c
1114
wwn = tf->tf_ops->fabric_make_wwn(tf, group, name);
drivers/target/target_core_fabric_configfs.c
1137
struct config_group *group,
drivers/target/target_core_fabric_configfs.c
236
struct config_group *group,
drivers/target/target_core_fabric_configfs.c
243
struct config_group *group,
drivers/target/target_core_fabric_configfs.c
266
struct config_group *group,
drivers/target/target_core_fabric_configfs.c
269
struct se_node_acl *se_nacl = container_of(group,
drivers/target/target_core_fabric_configfs.c
327
struct config_group *group,
drivers/target/target_core_fabric_configfs.c
374
struct config_group *group,
drivers/target/target_core_fabric_configfs.c
377
struct se_portal_group *se_tpg = container_of(group,
drivers/target/target_core_fabric_configfs.c
422
struct config_group *group,
drivers/target/target_core_fabric_configfs.c
468
struct config_group *group,
drivers/target/target_core_fabric_configfs.c
471
struct se_portal_group *se_tpg = container_of(group,
drivers/target/target_core_fabric_configfs.c
481
se_tpg_np = tf->tf_ops->fabric_make_np(se_tpg, group, name);
drivers/target/target_core_fabric_configfs.c
493
struct config_group *group,
drivers/target/target_core_fabric_configfs.c
716
struct config_group *group,
drivers/target/target_core_fabric_configfs.c
723
struct config_group *group,
drivers/target/target_core_fabric_configfs.c
741
struct config_group *group,
drivers/target/target_core_fabric_configfs.c
745
struct se_portal_group *se_tpg = container_of(group,
drivers/target/target_core_fabric_configfs.c
778
struct config_group *group,
drivers/target/target_core_fabric_configfs.c
927
struct config_group *group,
drivers/target/target_core_fabric_configfs.c
930
struct se_wwn *wwn = container_of(group, struct se_wwn, wwn_group);
drivers/target/target_core_fabric_configfs.c
980
struct config_group *group,
drivers/target/tcm_fc/tfc_conf.c
327
struct config_group *group,
drivers/target/tcm_remote/tcm_remote.c
152
struct config_group *group,
drivers/thermal/tegra/soctherm-fuse.c
126
sensor_group = sensor->group;
drivers/thermal/tegra/soctherm.h
105
const struct tegra_tsensor_group *group;
drivers/thermal/tegra/tegra114-soctherm.c
126
.group = &tegra114_tsensor_group_cpu,
drivers/thermal/tegra/tegra114-soctherm.c
134
.group = &tegra114_tsensor_group_cpu,
drivers/thermal/tegra/tegra114-soctherm.c
142
.group = &tegra114_tsensor_group_cpu,
drivers/thermal/tegra/tegra114-soctherm.c
150
.group = &tegra114_tsensor_group_cpu,
drivers/thermal/tegra/tegra114-soctherm.c
158
.group = &tegra114_tsensor_group_mem,
drivers/thermal/tegra/tegra114-soctherm.c
166
.group = &tegra114_tsensor_group_mem,
drivers/thermal/tegra/tegra114-soctherm.c
174
.group = &tegra114_tsensor_group_gpu,
drivers/thermal/tegra/tegra114-soctherm.c
182
.group = &tegra114_tsensor_group_pll,
drivers/thermal/tegra/tegra124-soctherm.c
135
.group = &tegra124_tsensor_group_cpu,
drivers/thermal/tegra/tegra124-soctherm.c
143
.group = &tegra124_tsensor_group_cpu,
drivers/thermal/tegra/tegra124-soctherm.c
151
.group = &tegra124_tsensor_group_cpu,
drivers/thermal/tegra/tegra124-soctherm.c
159
.group = &tegra124_tsensor_group_cpu,
drivers/thermal/tegra/tegra124-soctherm.c
167
.group = &tegra124_tsensor_group_mem,
drivers/thermal/tegra/tegra124-soctherm.c
175
.group = &tegra124_tsensor_group_mem,
drivers/thermal/tegra/tegra124-soctherm.c
183
.group = &tegra124_tsensor_group_gpu,
drivers/thermal/tegra/tegra124-soctherm.c
191
.group = &tegra124_tsensor_group_pll,
drivers/thermal/tegra/tegra132-soctherm.c
135
.group = &tegra132_tsensor_group_cpu,
drivers/thermal/tegra/tegra132-soctherm.c
143
.group = &tegra132_tsensor_group_cpu,
drivers/thermal/tegra/tegra132-soctherm.c
151
.group = &tegra132_tsensor_group_cpu,
drivers/thermal/tegra/tegra132-soctherm.c
159
.group = &tegra132_tsensor_group_cpu,
drivers/thermal/tegra/tegra132-soctherm.c
167
.group = &tegra132_tsensor_group_mem,
drivers/thermal/tegra/tegra132-soctherm.c
175
.group = &tegra132_tsensor_group_mem,
drivers/thermal/tegra/tegra132-soctherm.c
183
.group = &tegra132_tsensor_group_gpu,
drivers/thermal/tegra/tegra132-soctherm.c
191
.group = &tegra132_tsensor_group_pll,
drivers/thermal/tegra/tegra210-soctherm.c
136
.group = &tegra210_tsensor_group_cpu,
drivers/thermal/tegra/tegra210-soctherm.c
144
.group = &tegra210_tsensor_group_cpu,
drivers/thermal/tegra/tegra210-soctherm.c
152
.group = &tegra210_tsensor_group_cpu,
drivers/thermal/tegra/tegra210-soctherm.c
160
.group = &tegra210_tsensor_group_cpu,
drivers/thermal/tegra/tegra210-soctherm.c
168
.group = &tegra210_tsensor_group_mem,
drivers/thermal/tegra/tegra210-soctherm.c
176
.group = &tegra210_tsensor_group_mem,
drivers/thermal/tegra/tegra210-soctherm.c
184
.group = &tegra210_tsensor_group_gpu,
drivers/thermal/tegra/tegra210-soctherm.c
192
.group = &tegra210_tsensor_group_pll,
drivers/thermal/thermal_netlink.c
84
static int thermal_group_has_listeners(enum thermal_genl_multicast_groups group)
drivers/thermal/thermal_netlink.c
86
return genl_has_listeners(&thermal_genl_family, &init_net, group);
drivers/thunderbolt/tb.c
1432
tb_recalc_estimated_bandwidth_for_group(struct tb_bandwidth_group *group)
drivers/thunderbolt/tb.c
1435
struct tb *tb = group->tb;
drivers/thunderbolt/tb.c
1440
group->index);
drivers/thunderbolt/tb.c
1443
list_for_each_entry(in, &group->ports, group_list) {
drivers/thunderbolt/tb.c
1499
estimated_bw, group->reserved,
drivers/thunderbolt/tb.c
1500
estimated_bw + group->reserved);
drivers/thunderbolt/tb.c
1503
estimated_bw + group->reserved))
drivers/thunderbolt/tb.c
1512
tb_dbg(tb, "bandwidth estimation for group %u done\n", group->index);
drivers/thunderbolt/tb.c
1523
struct tb_bandwidth_group *group = &tcm->groups[i];
drivers/thunderbolt/tb.c
1525
if (!list_empty(&group->ports))
drivers/thunderbolt/tb.c
1526
tb_recalc_estimated_bandwidth_for_group(group);
drivers/thunderbolt/tb.c
1532
static bool __release_group_bandwidth(struct tb_bandwidth_group *group)
drivers/thunderbolt/tb.c
1534
if (group->reserved) {
drivers/thunderbolt/tb.c
1535
tb_dbg(group->tb, "group %d released total %d Mb/s\n", group->index,
drivers/thunderbolt/tb.c
1536
group->reserved);
drivers/thunderbolt/tb.c
1537
group->reserved = 0;
drivers/thunderbolt/tb.c
1543
static void __configure_group_sym(struct tb_bandwidth_group *group)
drivers/thunderbolt/tb.c
1548
if (list_empty(&group->ports))
drivers/thunderbolt/tb.c
1561
in = list_first_entry(&group->ports, struct tb_port, group_list);
drivers/thunderbolt/tb.c
1562
tunnel = tb_find_tunnel(group->tb, TB_TUNNEL_DP, in, NULL);
drivers/thunderbolt/tb.c
1564
tb_configure_sym(group->tb, in, tunnel->dst_port, true);
drivers/thunderbolt/tb.c
1569
struct tb_bandwidth_group *group =
drivers/thunderbolt/tb.c
1570
container_of(work, typeof(*group), release_work.work);
drivers/thunderbolt/tb.c
1571
struct tb *tb = group->tb;
drivers/thunderbolt/tb.c
1574
if (__release_group_bandwidth(group))
drivers/thunderbolt/tb.c
1576
__configure_group_sym(group);
drivers/thunderbolt/tb.c
1585
struct tb_bandwidth_group *group = &tcm->groups[i];
drivers/thunderbolt/tb.c
1587
group->tb = tcm_to_tb(tcm);
drivers/thunderbolt/tb.c
1588
group->index = i + 1;
drivers/thunderbolt/tb.c
1589
INIT_LIST_HEAD(&group->ports);
drivers/thunderbolt/tb.c
1590
INIT_DELAYED_WORK(&group->release_work,
drivers/thunderbolt/tb.c
1595
static void tb_bandwidth_group_attach_port(struct tb_bandwidth_group *group,
drivers/thunderbolt/tb.c
1598
if (!group || WARN_ON(in->group))
drivers/thunderbolt/tb.c
1601
in->group = group;
drivers/thunderbolt/tb.c
1602
list_add_tail(&in->group_list, &group->ports);
drivers/thunderbolt/tb.c
1604
tb_port_dbg(in, "attached to bandwidth group %d\n", group->index);
drivers/thunderbolt/tb.c
1612
struct tb_bandwidth_group *group = &tcm->groups[i];
drivers/thunderbolt/tb.c
1614
if (list_empty(&group->ports))
drivers/thunderbolt/tb.c
1615
return group;
drivers/thunderbolt/tb.c
1625
struct tb_bandwidth_group *group;
drivers/thunderbolt/tb.c
1640
group = tunnel->src_port->group;
drivers/thunderbolt/tb.c
1641
if (group) {
drivers/thunderbolt/tb.c
1642
tb_bandwidth_group_attach_port(group, in);
drivers/thunderbolt/tb.c
1643
return group;
drivers/thunderbolt/tb.c
1649
group = tb_find_free_bandwidth_group(tcm);
drivers/thunderbolt/tb.c
1650
if (group)
drivers/thunderbolt/tb.c
1651
tb_bandwidth_group_attach_port(group, in);
drivers/thunderbolt/tb.c
1655
return group;
drivers/thunderbolt/tb.c
1678
struct tb_bandwidth_group *group = in->group;
drivers/thunderbolt/tb.c
1680
if (group) {
drivers/thunderbolt/tb.c
1681
in->group = NULL;
drivers/thunderbolt/tb.c
1684
tb_port_dbg(in, "detached from bandwidth group %d\n", group->index);
drivers/thunderbolt/tb.c
1687
if (list_empty(&group->ports)) {
drivers/thunderbolt/tb.c
1688
cancel_delayed_work(&group->release_work);
drivers/thunderbolt/tb.c
1689
__release_group_bandwidth(group);
drivers/thunderbolt/tb.c
2539
struct tb_bandwidth_group *group;
drivers/thunderbolt/tb.c
2607
group = in->group;
drivers/thunderbolt/tb.c
2627
group->reserved += reserved;
drivers/thunderbolt/tb.c
2629
group->index, reserved, group->reserved);
drivers/thunderbolt/tb.c
2639
mod_delayed_work(system_percpu_wq, &group->release_work,
drivers/thunderbolt/tb.c
2671
available_up, available_down, group->reserved);
drivers/thunderbolt/tb.c
2674
available_up + group->reserved >= requested_up_corrected) ||
drivers/thunderbolt/tb.c
2676
available_down + group->reserved >= requested_down_corrected)) {
drivers/thunderbolt/tb.c
2705
group->reserved -= released;
drivers/thunderbolt/tb.c
2707
group->index, released, group->reserved);
drivers/thunderbolt/tb.c
625
const struct tb_bandwidth_group *group;
drivers/thunderbolt/tb.c
642
group = tunnel->src_port->group;
drivers/thunderbolt/tb.c
643
if (group && group->reserved && !group_reserved[group->index])
drivers/thunderbolt/tb.c
644
group_reserved[group->index] = group->reserved;
drivers/thunderbolt/tb.h
301
struct tb_bandwidth_group *group;
drivers/thunderbolt/tunnel.c
901
ret = usb4_dp_port_set_group_id(in, in->group->index);
drivers/usb/atm/ueagle-atm.c
1173
u16 function, u16 group, u16 address, u16 offset, u32 data)
drivers/usb/atm/ueagle-atm.c
1184
group, address, offset, data);
drivers/usb/atm/ueagle-atm.c
1190
sc->cmv_dsc.e4.group = group;
drivers/usb/atm/ueagle-atm.c
1193
cmv.wGroup = cpu_to_be16(group);
drivers/usb/atm/ueagle-atm.c
1222
u8 size, u16 group, u16 address, u16 offset, u32 *data)
drivers/usb/atm/ueagle-atm.c
1226
group, address, offset, 0);
drivers/usb/atm/ueagle-atm.c
1252
u8 size, u16 group, u16 address, u16 offset, u32 data)
drivers/usb/atm/ueagle-atm.c
1256
group, address, offset, data);
drivers/usb/atm/ueagle-atm.c
1774
get_unaligned_le32(&cmvs_v2[i].group),
drivers/usb/atm/ueagle-atm.c
2049
be16_to_cpu(cmv->wGroup) != dsc->group ||
drivers/usb/atm/ueagle-atm.c
79
u32 group;
drivers/usb/atm/ueagle-atm.c
97
u16 group;
drivers/usb/gadget/configfs.c
107
struct config_group group;
drivers/usb/gadget/configfs.c
1399
struct config_group *group,
drivers/usb/gadget/configfs.c
1418
desc = container_of(group, struct usb_os_desc, group);
drivers/usb/gadget/configfs.c
1443
static void ext_prop_drop(struct config_group *group, struct config_item *item)
drivers/usb/gadget/configfs.c
1446
struct usb_os_desc *desc = to_usb_os_desc(&group->cg_item);
drivers/usb/gadget/configfs.c
1561
config_group_init_type_name(&d->group, "", interface_type);
drivers/usb/gadget/configfs.c
1562
config_item_set_name(&d->group.cg_item, "interface.%s",
drivers/usb/gadget/configfs.c
1564
configfs_add_default_group(&d->group, os_desc_group);
drivers/usb/gadget/configfs.c
1989
struct config_group *group,
drivers/usb/gadget/configfs.c
1998
config_group_init_type_name(&gi->group, name, &gadget_root_type);
drivers/usb/gadget/configfs.c
2002
configfs_add_default_group(&gi->functions_group, &gi->group);
drivers/usb/gadget/configfs.c
2006
configfs_add_default_group(&gi->configs_group, &gi->group);
drivers/usb/gadget/configfs.c
2010
configfs_add_default_group(&gi->strings_group, &gi->group);
drivers/usb/gadget/configfs.c
2014
configfs_add_default_group(&gi->os_desc_group, &gi->group);
drivers/usb/gadget/configfs.c
2018
configfs_add_default_group(&gi->webusb_group, &gi->group);
drivers/usb/gadget/configfs.c
2049
return &gi->group;
drivers/usb/gadget/configfs.c
2058
static void gadgets_drop(struct config_group *group, struct config_item *item)
drivers/usb/gadget/configfs.c
38
struct config_group group;
drivers/usb/gadget/configfs.c
383
group);
drivers/usb/gadget/configfs.c
390
group);
drivers/usb/gadget/configfs.c
397
struct usb_function_instance, group);
drivers/usb/gadget/configfs.c
602
struct config_group *group,
drivers/usb/gadget/configfs.c
630
ret = config_item_set_name(&fi->group.cg_item, "%s", name);
drivers/usb/gadget/configfs.c
643
gi = container_of(group, struct gadget_info, functions_group);
drivers/usb/gadget/configfs.c
648
return &fi->group;
drivers/usb/gadget/configfs.c
652
struct config_group *group,
drivers/usb/gadget/configfs.c
658
gi = container_of(group, struct gadget_info, functions_group);
drivers/usb/gadget/configfs.c
66
return container_of(to_config_group(item), struct gadget_info, group);
drivers/usb/gadget/configfs.c
697
struct config_group *group,
drivers/usb/gadget/configfs.c
70
struct config_group group;
drivers/usb/gadget/configfs.c
707
gi = container_of(group, struct gadget_info, configs_group);
drivers/usb/gadget/configfs.c
744
config_group_init_type_name(&cfg->group, name,
drivers/usb/gadget/configfs.c
749
configfs_add_default_group(&cfg->strings_group, &cfg->group);
drivers/usb/gadget/configfs.c
755
return &cfg->group;
drivers/usb/gadget/configfs.c
763
struct config_group *group,
drivers/usb/gadget/configfs.c
81
group);
drivers/usb/gadget/configfs.c
865
static struct config_item *gadget_language_string_make(struct config_group *group,
drivers/usb/gadget/configfs.c
871
language = to_gadget_language(&group->cg_item);
drivers/usb/gadget/configfs.c
886
static void gadget_language_string_drop(struct config_group *group,
drivers/usb/gadget/configfs.c
893
language = to_gadget_language(&group->cg_item);
drivers/usb/gadget/configfs.c
916
static struct config_group *gadget_language_make(struct config_group *group,
drivers/usb/gadget/configfs.c
932
config_group_init_type_name(&new->group, name,
drivers/usb/gadget/configfs.c
935
gi = container_of(group, struct gadget_info, strings_group);
drivers/usb/gadget/configfs.c
951
return &new->group;
drivers/usb/gadget/configfs.c
957
static void gadget_language_drop(struct config_group *group,
drivers/usb/gadget/configfs.c
96
struct config_group group;
drivers/usb/gadget/configfs.h
18
return container_of(to_config_group(item), struct usb_os_desc, group);
drivers/usb/gadget/function/f_acm.c
786
func_inst.group);
drivers/usb/gadget/function/f_acm.c
896
config_group_init_type_name(&opts->func_inst.group, "",
drivers/usb/gadget/function/f_ecm.c
806
func_inst.group);
drivers/usb/gadget/function/f_ecm.c
866
config_group_init_type_name(&opts->func_inst.group, "", &ecm_func_type);
drivers/usb/gadget/function/f_eem.c
558
func_inst.group);
drivers/usb/gadget/function/f_eem.c
618
config_group_init_type_name(&opts->func_inst.group, "", &eem_func_type);
drivers/usb/gadget/function/f_fs.c
3968
func_inst.group);
drivers/usb/gadget/function/f_fs.c
4050
config_group_init_type_name(&opts->func_inst.group, "",
drivers/usb/gadget/function/f_fs.c
4303
if (opts->no_configfs || !opts->func_inst.group.cg_item.ci_parent
drivers/usb/gadget/function/f_fs.c
4304
|| !kref_read(&opts->func_inst.group.cg_item.ci_kref))
drivers/usb/gadget/function/f_fs.c
4307
ci = opts->func_inst.group.cg_item.ci_parent->ci_parent;
drivers/usb/gadget/function/f_hid.c
1316
func_inst.group);
drivers/usb/gadget/function/f_hid.c
1558
config_group_init_type_name(&opts->func_inst.group, "", &hid_func_type);
drivers/usb/gadget/function/f_loopback.c
457
func_inst.group);
drivers/usb/gadget/function/f_loopback.c
579
config_group_init_type_name(&lb_opts->func_inst.group, "",
drivers/usb/gadget/function/f_mass_storage.c
3147
return container_of(to_config_group(item), struct fsg_lun_opts, group);
drivers/usb/gadget/function/f_mass_storage.c
3153
func_inst.group);
drivers/usb/gadget/function/f_mass_storage.c
3171
struct fsg_opts *fsg_opts = to_fsg_opts(opts->group.cg_item.ci_parent);
drivers/usb/gadget/function/f_mass_storage.c
3180
struct fsg_opts *fsg_opts = to_fsg_opts(opts->group.cg_item.ci_parent);
drivers/usb/gadget/function/f_mass_storage.c
3196
struct fsg_opts *fsg_opts = to_fsg_opts(opts->group.cg_item.ci_parent);
drivers/usb/gadget/function/f_mass_storage.c
3226
struct fsg_opts *fsg_opts = to_fsg_opts(opts->group.cg_item.ci_parent);
drivers/usb/gadget/function/f_mass_storage.c
3265
struct fsg_opts *fsg_opts = to_fsg_opts(opts->group.cg_item.ci_parent);
drivers/usb/gadget/function/f_mass_storage.c
3290
static struct config_group *fsg_lun_make(struct config_group *group,
drivers/usb/gadget/function/f_mass_storage.c
3311
fsg_opts = to_fsg_opts(&group->cg_item);
drivers/usb/gadget/function/f_mass_storage.c
3332
(const char **)&group->cg_item.ci_name);
drivers/usb/gadget/function/f_mass_storage.c
3341
config_group_init_type_name(&opts->group, name, &fsg_lun_type);
drivers/usb/gadget/function/f_mass_storage.c
3343
return &opts->group;
drivers/usb/gadget/function/f_mass_storage.c
3349
static void fsg_lun_drop(struct config_group *group, struct config_item *item)
drivers/usb/gadget/function/f_mass_storage.c
3355
fsg_opts = to_fsg_opts(&group->cg_item);
drivers/usb/gadget/function/f_mass_storage.c
3361
gadget = group->cg_item.ci_parent->ci_parent;
drivers/usb/gadget/function/f_mass_storage.c
3521
(const char **)&opts->func_inst.group.cg_item.ci_name);
drivers/usb/gadget/function/f_mass_storage.c
3528
config_group_init_type_name(&opts->func_inst.group, "", &fsg_func_type);
drivers/usb/gadget/function/f_mass_storage.c
3530
config_group_init_type_name(&opts->lun0.group, "lun.0", &fsg_lun_type);
drivers/usb/gadget/function/f_mass_storage.c
3531
configfs_add_default_group(&opts->lun0.group, &opts->func_inst.group);
drivers/usb/gadget/function/f_mass_storage.h
65
struct config_group group;
drivers/usb/gadget/function/f_midi.c
1083
func_inst.group);
drivers/usb/gadget/function/f_midi.c
1296
config_group_init_type_name(&opts->func_inst.group, "",
drivers/usb/gadget/function/f_midi2.c
1012
&group);
drivers/usb/gadget/function/f_midi2.c
1015
cable = ep->in_group_to_cable[group];
drivers/usb/gadget/function/f_midi2.c
104
unsigned char group;
drivers/usb/gadget/function/f_midi2.c
1060
unsigned int group, cable, bytes, c, len;
drivers/usb/gadget/function/f_midi2.c
1076
group = midi2->out_cable_mapping[cable].group;
drivers/usb/gadget/function/f_midi2.c
1079
snd_ump_convert_to_ump(cvt, group,
drivers/usb/gadget/function/f_midi2.c
2083
func_inst.group);
drivers/usb/gadget/function/f_midi2.c
2090
group);
drivers/usb/gadget/function/f_midi2.c
2097
group);
drivers/usb/gadget/function/f_midi2.c
2368
f_midi2_opts_block_make(struct config_group *group, const char *name)
drivers/usb/gadget/function/f_midi2.c
2381
ep_opts = to_f_midi2_ep_opts(&group->cg_item);
drivers/usb/gadget/function/f_midi2.c
2391
config_group_init_type_name(&block_opts->group, name,
drivers/usb/gadget/function/f_midi2.c
2393
return &block_opts->group;
drivers/usb/gadget/function/f_midi2.c
2398
f_midi2_opts_block_drop(struct config_group *group, struct config_item *item)
drivers/usb/gadget/function/f_midi2.c
2522
f_midi2_opts_ep_make(struct config_group *group, const char *name)
drivers/usb/gadget/function/f_midi2.c
2535
opts = to_f_midi2_opts(&group->cg_item);
drivers/usb/gadget/function/f_midi2.c
2544
config_group_init_type_name(&ep_opts->group, name, &f_midi2_ep_type);
drivers/usb/gadget/function/f_midi2.c
2545
return &ep_opts->group;
drivers/usb/gadget/function/f_midi2.c
2550
f_midi2_opts_ep_drop(struct config_group *group, struct config_item *item)
drivers/usb/gadget/function/f_midi2.c
2683
config_group_init_type_name(&opts->func_inst.group, "",
drivers/usb/gadget/function/f_midi2.c
2686
config_group_init_type_name(&ep_opts->group, "ep.0",
drivers/usb/gadget/function/f_midi2.c
2688
configfs_add_default_group(&ep_opts->group, &opts->func_inst.group);
drivers/usb/gadget/function/f_midi2.c
2690
config_group_init_type_name(&block_opts->group, "block.0",
drivers/usb/gadget/function/f_midi2.c
2692
configfs_add_default_group(&block_opts->group, &ep_opts->group);
drivers/usb/gadget/function/f_midi2.c
2773
int i, group;
drivers/usb/gadget/function/f_midi2.c
2778
group = binfo->midi1_first_group;
drivers/usb/gadget/function/f_midi2.c
2780
for (i = 0; i < binfo->midi1_num_groups; i++, group++, map++) {
drivers/usb/gadget/function/f_midi2.c
2785
map->group = group;
drivers/usb/gadget/function/f_midi2.c
2788
ep->in_group_to_cable[group] = midi2->num_midi1_in;
drivers/usb/gadget/function/f_midi2.c
2793
group = binfo->midi1_first_group;
drivers/usb/gadget/function/f_midi2.c
2795
for (i = 0; i < binfo->midi1_num_groups; i++, group++, map++) {
drivers/usb/gadget/function/f_midi2.c
2800
map->group = group;
drivers/usb/gadget/function/f_midi2.c
989
unsigned char group, cable;
drivers/usb/gadget/function/f_ncm.c
1581
func_inst.group);
drivers/usb/gadget/function/f_ncm.c
1697
config_group_init_type_name(&opts->func_inst.group, "", &ncm_func_type);
drivers/usb/gadget/function/f_ncm.c
1699
usb_os_desc_prepare_interf_dir(&opts->func_inst.group, 1, descs,
drivers/usb/gadget/function/f_obex.c
383
func_inst.group);
drivers/usb/gadget/function/f_obex.c
439
config_group_init_type_name(&opts->func_inst.group, "",
drivers/usb/gadget/function/f_phonet.c
578
func_inst.group);
drivers/usb/gadget/function/f_phonet.c
638
config_group_init_type_name(&opts->func_inst.group, "",
drivers/usb/gadget/function/f_printer.c
1213
func_inst.group);
drivers/usb/gadget/function/f_printer.c
1407
config_group_init_type_name(&opts->func_inst.group, "",
drivers/usb/gadget/function/f_rndis.c
828
func_inst.group);
drivers/usb/gadget/function/f_rndis.c
916
config_group_init_type_name(&opts->func_inst.group, "",
drivers/usb/gadget/function/f_rndis.c
919
usb_os_desc_prepare_interf_dir(&opts->func_inst.group, 1, descs,
drivers/usb/gadget/function/f_serial.c
253
func_inst.group);
drivers/usb/gadget/function/f_serial.c
330
config_group_init_type_name(&opts->func_inst.group, "",
drivers/usb/gadget/function/f_sourcesink.c
1311
config_group_init_type_name(&ss_opts->func_inst.group, "",
drivers/usb/gadget/function/f_sourcesink.c
883
func_inst.group);
drivers/usb/gadget/function/f_subset.c
377
func_inst.group);
drivers/usb/gadget/function/f_subset.c
437
config_group_init_type_name(&opts->func_inst.group, "",
drivers/usb/gadget/function/f_tcm.c
1688
&opts->func_inst.group.cg_item);
drivers/usb/gadget/function/f_tcm.c
1729
configfs_undepend_item_unlocked(&opts->func_inst.group.cg_item);
drivers/usb/gadget/function/f_tcm.c
1764
&opts->func_inst.group.cg_item);
drivers/usb/gadget/function/f_tcm.c
1774
struct config_group *group,
drivers/usb/gadget/function/f_tcm.c
2454
func_inst.group);
drivers/usb/gadget/function/f_tcm.c
2508
func_inst.group.cg_item.ci_parent->ci_parent);
drivers/usb/gadget/function/f_tcm.c
2575
config_group_init_type_name(&opts->func_inst.group, "",
drivers/usb/gadget/function/f_uac1.c
1505
func_inst.group);
drivers/usb/gadget/function/f_uac1.c
1758
config_group_init_type_name(&opts->func_inst.group, "",
drivers/usb/gadget/function/f_uac1_legacy.c
832
func_inst.group);
drivers/usb/gadget/function/f_uac1_legacy.c
979
config_group_init_type_name(&opts->func_inst.group, "",
drivers/usb/gadget/function/f_uac2.c
1867
func_inst.group);
drivers/usb/gadget/function/f_uac2.c
2199
config_group_init_type_name(&opts->func_inst.group, "",
drivers/usb/gadget/function/f_uvc.c
1113
streaming = config_group_find_item(&opts->func_inst.group, "streaming");
drivers/usb/gadget/function/f_uvc.c
435
return sprintf(buf, "%s\n", uvc->func.fi->group.cg_item.ci_name);
drivers/usb/gadget/function/u_midi2.h
53
struct config_group group;
drivers/usb/gadget/function/u_midi2.h
60
struct config_group group;
drivers/usb/gadget/function/uvc_configfs.c
1010
struct config_group *group = to_config_group(item->ci_parent);
drivers/usb/gadget/function/uvc_configfs.c
1011
struct mutex *su_mutex = &group->cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
1038
struct config_group *group = to_config_group(item->ci_parent);
drivers/usb/gadget/function/uvc_configfs.c
1039
struct mutex *su_mutex = &group->cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
1089
struct config_group *group = to_config_group(item->ci_parent);
drivers/usb/gadget/function/uvc_configfs.c
1090
struct mutex *su_mutex = &group->cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
1117
struct config_group *group = to_config_group(item->ci_parent);
drivers/usb/gadget/function/uvc_configfs.c
1118
struct mutex *su_mutex = &group->cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
121
int (*create_children)(struct config_group *group);
drivers/usb/gadget/function/uvc_configfs.c
1248
static void uvcg_extension_drop(struct config_group *group, struct config_item *item)
drivers/usb/gadget/function/uvc_configfs.c
1254
opts_item = group->cg_item.ci_parent->ci_parent;
drivers/usb/gadget/function/uvc_configfs.c
126
struct config_group *group = to_config_group(item);
drivers/usb/gadget/function/uvc_configfs.c
1267
static struct config_item *uvcg_extension_make(struct config_group *group, const char *name)
drivers/usb/gadget/function/uvc_configfs.c
1273
opts_item = group->cg_item.ci_parent->ci_parent;
drivers/usb/gadget/function/uvc_configfs.c
128
kfree(group);
drivers/usb/gadget/function/uvc_configfs.c
1320
struct config_group group;
drivers/usb/gadget/function/uvc_configfs.c
1327
struct uvcg_control_class_group *group =
drivers/usb/gadget/function/uvc_configfs.c
1329
group.cg_item);
drivers/usb/gadget/function/uvc_configfs.c
1331
if (!strcmp(group->name, "fs"))
drivers/usb/gadget/function/uvc_configfs.c
1334
if (!strcmp(group->name, "ss"))
drivers/usb/gadget/function/uvc_configfs.c
138
static int uvcg_config_create_children(struct config_group *group,
drivers/usb/gadget/function/uvc_configfs.c
1438
struct uvcg_control_class_group *group;
drivers/usb/gadget/function/uvc_configfs.c
1440
group = kzalloc_obj(*group);
drivers/usb/gadget/function/uvc_configfs.c
1441
if (!group)
drivers/usb/gadget/function/uvc_configfs.c
1444
group->name = names[i];
drivers/usb/gadget/function/uvc_configfs.c
1446
config_group_init_type_name(&group->group, group->name,
drivers/usb/gadget/function/uvc_configfs.c
1448
configfs_add_default_group(&group->group, parent);
drivers/usb/gadget/function/uvc_configfs.c
145
return type->create_children(group);
drivers/usb/gadget/function/uvc_configfs.c
1470
struct config_group *group = to_config_group(item);
drivers/usb/gadget/function/uvc_configfs.c
1471
struct mutex *su_mutex = &group->cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
148
ret = uvcg_config_create_group(group, *child);
drivers/usb/gadget/function/uvc_configfs.c
1495
struct config_group *group = to_config_group(item);
drivers/usb/gadget/function/uvc_configfs.c
1496
struct mutex *su_mutex = &group->cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
1518
struct config_group *group = to_config_group(item);
drivers/usb/gadget/function/uvc_configfs.c
1519
struct mutex *su_mutex = &group->cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
159
struct config_group *group;
drivers/usb/gadget/function/uvc_configfs.c
161
group = kzalloc_obj(*group);
drivers/usb/gadget/function/uvc_configfs.c
162
if (!group)
drivers/usb/gadget/function/uvc_configfs.c
165
config_group_init_type_name(group, type->name, &type->type);
drivers/usb/gadget/function/uvc_configfs.c
166
configfs_add_default_group(group, parent);
drivers/usb/gadget/function/uvc_configfs.c
1677
struct mutex *su_mutex = &f->group.cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
168
return uvcg_config_create_children(group, type);
drivers/usb/gadget/function/uvc_configfs.c
1683
opts_item = f->group.cg_item.ci_parent->ci_parent->ci_parent;
drivers/usb/gadget/function/uvc_configfs.c
1704
struct mutex *su_mutex = &ch->group.cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
1709
opts_item = ch->group.cg_item.ci_parent->ci_parent->ci_parent;
drivers/usb/gadget/function/uvc_configfs.c
171
static void uvcg_config_remove_children(struct config_group *group)
drivers/usb/gadget/function/uvc_configfs.c
175
list_for_each_entry_safe(child, n, &group->default_groups, group_entry) {
drivers/usb/gadget/function/uvc_configfs.c
1781
group);
drivers/usb/gadget/function/uvc_configfs.c
1823
group);
drivers/usb/gadget/function/uvc_configfs.c
1898
*uvcg_streaming_header_make(struct config_group *group, const char *name)
drivers/usb/gadget/function/uvc_configfs.c
2157
static struct config_item *uvcg_frame_make(struct config_group *group,
drivers/usb/gadget/function/uvc_configfs.c
2180
opts_item = group->cg_item.ci_parent->ci_parent->ci_parent;
drivers/usb/gadget/function/uvc_configfs.c
2184
fmt = to_uvcg_format(&group->cg_item);
drivers/usb/gadget/function/uvc_configfs.c
2220
static void uvcg_frame_drop(struct config_group *group, struct config_item *item)
drivers/usb/gadget/function/uvc_configfs.c
2228
opts_item = group->cg_item.ci_parent->ci_parent->ci_parent;
drivers/usb/gadget/function/uvc_configfs.c
2233
fmt = to_uvcg_format(&group->cg_item);
drivers/usb/gadget/function/uvc_configfs.c
2275
struct mutex *su_mutex = &ch->fmt.group.cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
2279
opts_item = ch->fmt.group.cg_item.ci_parent->ci_parent->ci_parent;
drivers/usb/gadget/function/uvc_configfs.c
2297
struct mutex *su_mutex = &ch->fmt.group.cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
2304
opts_item = ch->fmt.group.cg_item.ci_parent->ci_parent->ci_parent;
drivers/usb/gadget/function/uvc_configfs.c
2341
struct mutex *su_mutex = &u->fmt.group.cg_subsys->su_mutex; \
drivers/usb/gadget/function/uvc_configfs.c
2346
opts_item = u->fmt.group.cg_item.ci_parent->ci_parent->ci_parent;\
drivers/usb/gadget/function/uvc_configfs.c
2366
struct mutex *su_mutex = &u->fmt.group.cg_subsys->su_mutex; \
drivers/usb/gadget/function/uvc_configfs.c
2371
opts_item = u->fmt.group.cg_item.ci_parent->ci_parent->ci_parent;\
drivers/usb/gadget/function/uvc_configfs.c
2389
struct mutex *su_mutex = &u->fmt.group.cg_subsys->su_mutex; \
drivers/usb/gadget/function/uvc_configfs.c
2395
opts_item = u->fmt.group.cg_item.ci_parent->ci_parent->ci_parent;\
drivers/usb/gadget/function/uvc_configfs.c
2470
static struct config_group *uvcg_uncompressed_make(struct config_group *group,
drivers/usb/gadget/function/uvc_configfs.c
2481
streaming = group->cg_item.ci_parent;
drivers/usb/gadget/function/uvc_configfs.c
2505
config_group_init_type_name(&h->fmt.group, name,
drivers/usb/gadget/function/uvc_configfs.c
2508
return &h->fmt.group;
drivers/usb/gadget/function/uvc_configfs.c
2539
struct mutex *su_mutex = &u->fmt.group.cg_subsys->su_mutex; \
drivers/usb/gadget/function/uvc_configfs.c
2544
opts_item = u->fmt.group.cg_item.ci_parent->ci_parent->ci_parent;\
drivers/usb/gadget/function/uvc_configfs.c
2563
struct mutex *su_mutex = &u->fmt.group.cg_subsys->su_mutex; \
drivers/usb/gadget/function/uvc_configfs.c
2568
opts_item = u->fmt.group.cg_item.ci_parent->ci_parent->ci_parent;\
drivers/usb/gadget/function/uvc_configfs.c
2586
struct mutex *su_mutex = &u->fmt.group.cg_subsys->su_mutex; \
drivers/usb/gadget/function/uvc_configfs.c
2592
opts_item = u->fmt.group.cg_item.ci_parent->ci_parent->ci_parent;\
drivers/usb/gadget/function/uvc_configfs.c
2666
static struct config_group *uvcg_mjpeg_make(struct config_group *group,
drivers/usb/gadget/function/uvc_configfs.c
2673
streaming = group->cg_item.ci_parent;
drivers/usb/gadget/function/uvc_configfs.c
268
static struct config_item *uvcg_control_header_make(struct config_group *group,
drivers/usb/gadget/function/uvc_configfs.c
2695
config_group_init_type_name(&h->fmt.group, name,
drivers/usb/gadget/function/uvc_configfs.c
2698
return &h->fmt.group;
drivers/usb/gadget/function/uvc_configfs.c
2730
struct mutex *su_mutex = &u->fmt.group.cg_subsys->su_mutex; \
drivers/usb/gadget/function/uvc_configfs.c
2735
opts_item = u->fmt.group.cg_item.ci_parent->ci_parent->ci_parent;\
drivers/usb/gadget/function/uvc_configfs.c
2755
struct mutex *su_mutex = &u->fmt.group.cg_subsys->su_mutex; \
drivers/usb/gadget/function/uvc_configfs.c
2760
opts_item = u->fmt.group.cg_item.ci_parent->ci_parent->ci_parent;\
drivers/usb/gadget/function/uvc_configfs.c
2778
struct mutex *su_mutex = &u->fmt.group.cg_subsys->su_mutex; \
drivers/usb/gadget/function/uvc_configfs.c
2784
opts_item = u->fmt.group.cg_item.ci_parent->ci_parent->ci_parent;\
drivers/usb/gadget/function/uvc_configfs.c
2827
struct mutex *su_mutex = &ch->fmt.group.cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
2831
opts_item = ch->fmt.group.cg_item.ci_parent->ci_parent->ci_parent;
drivers/usb/gadget/function/uvc_configfs.c
2849
struct mutex *su_mutex = &ch->fmt.group.cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
2854
opts_item = ch->fmt.group.cg_item.ci_parent->ci_parent->ci_parent;
drivers/usb/gadget/function/uvc_configfs.c
2913
static struct config_group *uvcg_framebased_make(struct config_group *group,
drivers/usb/gadget/function/uvc_configfs.c
2924
streaming = group->cg_item.ci_parent;
drivers/usb/gadget/function/uvc_configfs.c
2950
config_group_init_type_name(&h->fmt.group, name,
drivers/usb/gadget/function/uvc_configfs.c
2953
return &h->fmt.group;
drivers/usb/gadget/function/uvc_configfs.c
2977
struct config_group *group = to_config_group(item); \
drivers/usb/gadget/function/uvc_configfs.c
2979
to_uvcg_color_matching(group); \
drivers/usb/gadget/function/uvc_configfs.c
2982
struct mutex *su_mutex = &group->cg_subsys->su_mutex; \
drivers/usb/gadget/function/uvc_configfs.c
2987
opts_item = group->cg_item.ci_parent->ci_parent->ci_parent; \
drivers/usb/gadget/function/uvc_configfs.c
3002
struct config_group *group = to_config_group(item); \
drivers/usb/gadget/function/uvc_configfs.c
3003
struct mutex *su_mutex = &group->cg_subsys->su_mutex; \
drivers/usb/gadget/function/uvc_configfs.c
3005
to_uvcg_color_matching(group); \
drivers/usb/gadget/function/uvc_configfs.c
3022
opts_item = group->cg_item.ci_parent->ci_parent->ci_parent; \
drivers/usb/gadget/function/uvc_configfs.c
3073
static struct config_group *uvcg_color_matching_make(struct config_group *group,
drivers/usb/gadget/function/uvc_configfs.c
3086
config_group_init_type_name(&color_match->group, name,
drivers/usb/gadget/function/uvc_configfs.c
3089
return &color_match->group;
drivers/usb/gadget/function/uvc_configfs.c
309
struct config_group *group = to_config_group(item); \
drivers/usb/gadget/function/uvc_configfs.c
3111
config_group_init_type_name(&color_match->group, "default",
drivers/usb/gadget/function/uvc_configfs.c
3113
configfs_add_default_group(&color_match->group, parent);
drivers/usb/gadget/function/uvc_configfs.c
312
struct mutex *su_mutex = &group->cg_subsys->su_mutex; \
drivers/usb/gadget/function/uvc_configfs.c
3133
struct config_group group;
drivers/usb/gadget/function/uvc_configfs.c
3140
struct uvcg_streaming_class_group *group =
drivers/usb/gadget/function/uvc_configfs.c
3142
group.cg_item);
drivers/usb/gadget/function/uvc_configfs.c
3144
if (!strcmp(group->name, "fs"))
drivers/usb/gadget/function/uvc_configfs.c
3147
if (!strcmp(group->name, "hs"))
drivers/usb/gadget/function/uvc_configfs.c
3150
if (!strcmp(group->name, "ss"))
drivers/usb/gadget/function/uvc_configfs.c
318
opts_item = group->cg_item.ci_parent->ci_parent->ci_parent; \
drivers/usb/gadget/function/uvc_configfs.c
3203
grp = &f->fmt->group;
drivers/usb/gadget/function/uvc_configfs.c
342
struct config_group *group = to_config_group(item);
drivers/usb/gadget/function/uvc_configfs.c
343
struct mutex *su_mutex = &group->cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
353
opts_item = group->cg_item.ci_parent->ci_parent->ci_parent;
drivers/usb/gadget/function/uvc_configfs.c
3554
struct uvcg_streaming_class_group *group;
drivers/usb/gadget/function/uvc_configfs.c
3556
group = kzalloc_obj(*group);
drivers/usb/gadget/function/uvc_configfs.c
3557
if (!group)
drivers/usb/gadget/function/uvc_configfs.c
3560
group->name = names[i];
drivers/usb/gadget/function/uvc_configfs.c
3562
config_group_init_type_name(&group->group, group->name,
drivers/usb/gadget/function/uvc_configfs.c
3564
configfs_add_default_group(&group->group, parent);
drivers/usb/gadget/function/uvc_configfs.c
3586
struct config_group *group = to_config_group(item);
drivers/usb/gadget/function/uvc_configfs.c
3587
struct mutex *su_mutex = &group->cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
3826
config_group_init_type_name(&opts->func_inst.group, uvc_func_type.name,
drivers/usb/gadget/function/uvc_configfs.c
3829
ret = uvcg_config_create_children(&opts->func_inst.group,
drivers/usb/gadget/function/uvc_configfs.c
3832
config_group_put(&opts->func_inst.group);
drivers/usb/gadget/function/uvc_configfs.c
400
struct config_group *group = to_config_group(item);
drivers/usb/gadget/function/uvc_configfs.c
403
struct mutex *su_mutex = &group->cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
410
opts_item = group->cg_item.ci_parent->ci_parent->ci_parent;
drivers/usb/gadget/function/uvc_configfs.c
470
struct config_group *group = to_config_group(item); \
drivers/usb/gadget/function/uvc_configfs.c
473
struct mutex *su_mutex = &group->cg_subsys->su_mutex; \
drivers/usb/gadget/function/uvc_configfs.c
479
opts_item = group->cg_item.ci_parent->ci_parent->ci_parent-> \
drivers/usb/gadget/function/uvc_configfs.c
511
struct config_group *group = to_config_group(item);
drivers/usb/gadget/function/uvc_configfs.c
512
struct mutex *su_mutex = &group->cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
522
opts_item = group->cg_item.ci_parent->ci_parent->ci_parent->
drivers/usb/gadget/function/uvc_configfs.c
570
struct config_group *group = to_config_group(item);
drivers/usb/gadget/function/uvc_configfs.c
573
struct mutex *su_mutex = &group->cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
580
opts_item = group->cg_item.ci_parent->ci_parent->ci_parent->
drivers/usb/gadget/function/uvc_configfs.c
643
struct config_group *group = to_config_group(item); \
drivers/usb/gadget/function/uvc_configfs.c
646
struct mutex *su_mutex = &group->cg_subsys->su_mutex; \
drivers/usb/gadget/function/uvc_configfs.c
652
opts_item = group->cg_item.ci_parent->ci_parent-> \
drivers/usb/gadget/function/uvc_configfs.c
678
struct config_group *group = to_config_group(item);
drivers/usb/gadget/function/uvc_configfs.c
681
struct mutex *su_mutex = &group->cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
687
opts_item = group->cg_item.ci_parent->ci_parent->
drivers/usb/gadget/function/uvc_configfs.c
704
struct config_group *group = to_config_group(item);
drivers/usb/gadget/function/uvc_configfs.c
707
struct mutex *su_mutex = &group->cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
718
opts_item = group->cg_item.ci_parent->ci_parent->
drivers/usb/gadget/function/uvc_configfs.c
792
struct config_group *group = to_config_group(item->ci_parent); \
drivers/usb/gadget/function/uvc_configfs.c
793
struct mutex *su_mutex = &group->cg_subsys->su_mutex; \
drivers/usb/gadget/function/uvc_configfs.c
821
struct config_group *group = to_config_group(item->ci_parent);
drivers/usb/gadget/function/uvc_configfs.c
822
struct mutex *su_mutex = &group->cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
855
struct config_group *group = to_config_group(item->ci_parent);
drivers/usb/gadget/function/uvc_configfs.c
856
struct mutex *su_mutex = &group->cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
908
struct config_group *group = to_config_group(item->ci_parent);
drivers/usb/gadget/function/uvc_configfs.c
909
struct mutex *su_mutex = &group->cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
958
struct config_group *group = to_config_group(item->ci_parent);
drivers/usb/gadget/function/uvc_configfs.c
959
struct mutex *su_mutex = &group->cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.c
981
struct config_group *group = to_config_group(item->ci_parent);
drivers/usb/gadget/function/uvc_configfs.c
982
struct mutex *su_mutex = &group->cg_subsys->su_mutex;
drivers/usb/gadget/function/uvc_configfs.h
22
func_inst.group);
drivers/usb/gadget/function/uvc_configfs.h
41
struct config_group group;
drivers/usb/gadget/function/uvc_configfs.h
47
container_of(group_ptr, struct uvcg_color_matching, group)
drivers/usb/gadget/function/uvc_configfs.h
56
struct config_group group;
drivers/usb/gadget/function/uvc_configfs.h
72
return container_of(to_config_group(item), struct uvcg_format, group);
drivers/usb/gadget/function/uvc_v4l2.c
278
to_uvcg_uncompressed(&uformat->group.cg_item);
drivers/usb/gadget/function/uvc_v4l2.c
38
unc = to_uvcg_uncompressed(&uformat->group.cg_item);
drivers/usb/gadget/function/uvc_v4l2.c
46
unc = to_uvcg_framebased(&uformat->group.cg_item);
drivers/usb/gadget/function/uvc_v4l2.c
66
u = to_uvcg_uncompressed(&uformat->group.cg_item);
drivers/usb/typec/bus.h
21
struct attribute_group group;
drivers/usb/typec/class.c
672
alt->group.name = alt->group_name;
drivers/usb/typec/class.c
673
alt->group.attrs = alt->attrs;
drivers/usb/typec/class.c
674
alt->groups[0] = &alt->group;
drivers/vdpa/mlx5/net/mlx5_vnet.c
3648
static int mlx5_set_group_asid(struct vdpa_device *vdev, u32 group,
drivers/vdpa/mlx5/net/mlx5_vnet.c
3654
mvdev->mres.group2asid[group] = asid;
drivers/vdpa/mlx5/net/mlx5_vnet.c
3657
if (group == MLX5_VDPA_CVQ_GROUP && mvdev->mres.mr[asid])
drivers/vdpa/vdpa_sim/vdpa_sim.c
600
static int vdpasim_set_group_asid(struct vdpa_device *vdpa, unsigned int group,
drivers/vdpa/vdpa_sim/vdpa_sim.c
612
if (vdpasim_get_vq_group(vdpa, i) == group)
drivers/vdpa/vdpa_user/vduse_dev.c
1004
if (!token.group)
drivers/vdpa/vdpa_user/vduse_dev.c
1010
guard(vq_group_as_read_lock)(token.group);
drivers/vdpa/vdpa_user/vduse_dev.c
1011
domain = token.group->as->domain;
drivers/vdpa/vdpa_user/vduse_dev.c
1020
if (!token.group)
drivers/vdpa/vdpa_user/vduse_dev.c
1023
guard(vq_group_as_read_lock)(token.group);
drivers/vdpa/vdpa_user/vduse_dev.c
1024
return dma_addr < token.group->as->domain->bounce_size;
drivers/vdpa/vdpa_user/vduse_dev.c
1036
if (!token.group)
drivers/vdpa/vdpa_user/vduse_dev.c
1039
guard(vq_group_as_read_lock)(token.group);
drivers/vdpa/vdpa_user/vduse_dev.c
1040
return token.group->as->domain->bounce_size;
drivers/vdpa/vdpa_user/vduse_dev.c
1438
if (config.group)
drivers/vdpa/vdpa_user/vduse_dev.c
1441
if (config.group >= dev->ngroups)
drivers/vdpa/vdpa_user/vduse_dev.c
1454
dev->vqs[index]->group = config.group;
drivers/vdpa/vdpa_user/vduse_dev.c
638
return dev->vqs[idx]->group;
drivers/vdpa/vdpa_user/vduse_dev.c
646
.group = &dev->groups[vq_group],
drivers/vdpa/vdpa_user/vduse_dev.c
664
static int vduse_set_group_asid(struct vdpa_device *vdpa, unsigned int group,
drivers/vdpa/vdpa_user/vduse_dev.c
675
msg.req.vq_group_asid.group = group;
drivers/vdpa/vdpa_user/vduse_dev.c
682
guard(vq_group_as_write_lock)(&dev->groups[group]);
drivers/vdpa/vdpa_user/vduse_dev.c
683
dev->groups[group].as = &dev->as[asid];
drivers/vdpa/vdpa_user/vduse_dev.c
74
u32 group;
drivers/vdpa/vdpa_user/vduse_dev.c
919
if (!token.group)
drivers/vdpa/vdpa_user/vduse_dev.c
922
guard(vq_group_as_read_lock)(token.group);
drivers/vdpa/vdpa_user/vduse_dev.c
923
domain = token.group->as->domain;
drivers/vdpa/vdpa_user/vduse_dev.c
933
if (!token.group)
drivers/vdpa/vdpa_user/vduse_dev.c
936
guard(vq_group_as_read_lock)(token.group);
drivers/vdpa/vdpa_user/vduse_dev.c
937
domain = token.group->as->domain;
drivers/vdpa/vdpa_user/vduse_dev.c
948
if (!token.group)
drivers/vdpa/vdpa_user/vduse_dev.c
951
guard(vq_group_as_read_lock)(token.group);
drivers/vdpa/vdpa_user/vduse_dev.c
952
domain = token.group->as->domain;
drivers/vdpa/vdpa_user/vduse_dev.c
962
if (!token.group)
drivers/vdpa/vdpa_user/vduse_dev.c
965
guard(vq_group_as_read_lock)(token.group);
drivers/vdpa/vdpa_user/vduse_dev.c
966
domain = token.group->as->domain;
drivers/vdpa/vdpa_user/vduse_dev.c
976
if (!token.group)
drivers/vdpa/vdpa_user/vduse_dev.c
986
guard(vq_group_as_read_lock)(token.group);
drivers/vdpa/vdpa_user/vduse_dev.c
987
domain = token.group->as->domain;
drivers/vfio/container.c
167
device->group->container->iommu_driver;
drivers/vfio/container.c
171
device->group->container->iommu_data, device);
drivers/vfio/container.c
177
device->group->container->iommu_driver;
drivers/vfio/container.c
181
device->group->container->iommu_data, device);
drivers/vfio/container.c
239
struct vfio_group *group;
drivers/vfio/container.c
242
list_for_each_entry(group, &container->group_list, container_next) {
drivers/vfio/container.c
243
ret = driver->ops->attach_group(data, group->iommu_group,
drivers/vfio/container.c
244
group->type);
drivers/vfio/container.c
252
list_for_each_entry_continue_reverse(group, &container->group_list,
drivers/vfio/container.c
254
driver->ops->detach_group(data, group->iommu_group);
drivers/vfio/container.c
417
struct vfio_group *group)
drivers/vfio/container.c
422
lockdep_assert_held(&group->group_lock);
drivers/vfio/container.c
424
if (group->type == VFIO_NO_IOMMU && !capable(CAP_SYS_RAWIO))
drivers/vfio/container.c
431
container->noiommu != (group->type == VFIO_NO_IOMMU)) {
drivers/vfio/container.c
436
if (group->type == VFIO_IOMMU) {
drivers/vfio/container.c
437
ret = iommu_group_claim_dma_owner(group->iommu_group, group);
drivers/vfio/container.c
445
group->iommu_group,
drivers/vfio/container.c
446
group->type);
drivers/vfio/container.c
448
if (group->type == VFIO_IOMMU)
drivers/vfio/container.c
450
group->iommu_group);
drivers/vfio/container.c
455
group->container = container;
drivers/vfio/container.c
456
group->container_users = 1;
drivers/vfio/container.c
457
container->noiommu = (group->type == VFIO_NO_IOMMU);
drivers/vfio/container.c
458
list_add(&group->container_next, &container->group_list);
drivers/vfio/container.c
468
void vfio_group_detach_container(struct vfio_group *group)
drivers/vfio/container.c
470
struct vfio_container *container = group->container;
drivers/vfio/container.c
473
lockdep_assert_held(&group->group_lock);
drivers/vfio/container.c
474
WARN_ON(group->container_users != 1);
drivers/vfio/container.c
481
group->iommu_group);
drivers/vfio/container.c
483
if (group->type == VFIO_IOMMU)
drivers/vfio/container.c
484
iommu_group_release_dma_owner(group->iommu_group);
drivers/vfio/container.c
486
group->container = NULL;
drivers/vfio/container.c
487
group->container_users = 0;
drivers/vfio/container.c
488
list_del(&group->container_next);
drivers/vfio/container.c
503
int vfio_group_use_container(struct vfio_group *group)
drivers/vfio/container.c
505
lockdep_assert_held(&group->group_lock);
drivers/vfio/container.c
511
if (!group->container->iommu_driver)
drivers/vfio/container.c
514
if (group->type == VFIO_NO_IOMMU && !capable(CAP_SYS_RAWIO))
drivers/vfio/container.c
517
get_file(group->opened_file);
drivers/vfio/container.c
518
group->container_users++;
drivers/vfio/container.c
522
void vfio_group_unuse_container(struct vfio_group *group)
drivers/vfio/container.c
524
lockdep_assert_held(&group->group_lock);
drivers/vfio/container.c
526
WARN_ON(group->container_users <= 1);
drivers/vfio/container.c
527
group->container_users--;
drivers/vfio/container.c
528
fput(group->opened_file);
drivers/vfio/container.c
535
struct vfio_container *container = device->group->container;
drivers/vfio/container.c
536
struct iommu_group *iommu_group = device->group->iommu_group;
drivers/vfio/container.c
551
struct vfio_container *container = device->group->container;
drivers/vfio/container.c
564
struct vfio_container *container = device->group->container;
drivers/vfio/device_cdev.c
110
if (df->group)
drivers/vfio/group.c
102
static int vfio_group_ioctl_set_container(struct vfio_group *group,
drivers/vfio/group.c
117
mutex_lock(&group->group_lock);
drivers/vfio/group.c
118
if (vfio_group_has_iommu(group)) {
drivers/vfio/group.c
122
if (!group->iommu_group) {
drivers/vfio/group.c
129
ret = vfio_container_attach_group(container, group);
drivers/vfio/group.c
136
group->type == VFIO_NO_IOMMU)
drivers/vfio/group.c
146
group->iommufd = iommufd;
drivers/vfio/group.c
154
mutex_unlock(&group->group_lock);
drivers/vfio/group.c
160
spin_lock(&device->group->kvm_ref_lock);
drivers/vfio/group.c
161
vfio_device_get_kvm_safe(device, device->group->kvm);
drivers/vfio/group.c
162
spin_unlock(&device->group->kvm_ref_lock);
drivers/vfio/group.c
170
mutex_lock(&device->group->group_lock);
drivers/vfio/group.c
171
if (!vfio_group_has_iommu(device->group)) {
drivers/vfio/group.c
187
df->iommufd = device->group->iommufd;
drivers/vfio/group.c
218
mutex_unlock(&device->group->group_lock);
drivers/vfio/group.c
229
mutex_unlock(&device->group->group_lock);
drivers/vfio/group.c
237
mutex_lock(&device->group->group_lock);
drivers/vfio/group.c
247
mutex_unlock(&device->group->group_lock);
drivers/vfio/group.c
26
static struct vfio_device *vfio_device_get_from_name(struct vfio_group *group,
drivers/vfio/group.c
262
df->group = device->group;
drivers/vfio/group.c
281
if (device->group->type == VFIO_NO_IOMMU)
drivers/vfio/group.c
298
static int vfio_group_ioctl_get_device_fd(struct vfio_group *group,
drivers/vfio/group.c
309
device = vfio_device_get_from_name(group, buf);
drivers/vfio/group.c
31
mutex_lock(&group->device_lock);
drivers/vfio/group.c
32
list_for_each_entry(it, &group->device_list, group_next) {
drivers/vfio/group.c
320
static int vfio_group_ioctl_get_status(struct vfio_group *group,
drivers/vfio/group.c
334
mutex_lock(&group->group_lock);
drivers/vfio/group.c
335
if (!group->iommu_group) {
drivers/vfio/group.c
336
mutex_unlock(&group->group_lock);
drivers/vfio/group.c
347
if (vfio_group_has_iommu(group))
drivers/vfio/group.c
350
else if (!iommu_group_dma_owner_claimed(group->iommu_group))
drivers/vfio/group.c
352
mutex_unlock(&group->group_lock);
drivers/vfio/group.c
362
struct vfio_group *group = filep->private_data;
drivers/vfio/group.c
367
return vfio_group_ioctl_get_device_fd(group, uarg);
drivers/vfio/group.c
369
return vfio_group_ioctl_get_status(group, uarg);
drivers/vfio/group.c
371
return vfio_group_ioctl_set_container(group, uarg);
drivers/vfio/group.c
373
return vfio_group_ioctl_unset_container(group);
drivers/vfio/group.c
381
struct vfio_group *group = device->group;
drivers/vfio/group.c
384
mutex_lock(&group->group_lock);
drivers/vfio/group.c
385
if (group->opened_file) {
drivers/vfio/group.c
390
group->cdev_device_open_cnt++;
drivers/vfio/group.c
393
mutex_unlock(&group->group_lock);
drivers/vfio/group.c
399
struct vfio_group *group = device->group;
drivers/vfio/group.c
401
mutex_lock(&group->group_lock);
drivers/vfio/group.c
402
group->cdev_device_open_cnt--;
drivers/vfio/group.c
403
mutex_unlock(&group->group_lock);
drivers/vfio/group.c
408
struct vfio_group *group =
drivers/vfio/group.c
412
mutex_lock(&group->group_lock);
drivers/vfio/group.c
418
if (refcount_read(&group->drivers) == 0) {
drivers/vfio/group.c
423
if (group->type == VFIO_NO_IOMMU && !capable(CAP_SYS_RAWIO)) {
drivers/vfio/group.c
428
if (group->cdev_device_open_cnt) {
drivers/vfio/group.c
436
if (group->opened_file) {
drivers/vfio/group.c
440
group->opened_file = filep;
drivers/vfio/group.c
441
filep->private_data = group;
drivers/vfio/group.c
444
mutex_unlock(&group->group_lock);
drivers/vfio/group.c
450
struct vfio_group *group = filep->private_data;
drivers/vfio/group.c
454
mutex_lock(&group->group_lock);
drivers/vfio/group.c
459
WARN_ON(group->notifier.head);
drivers/vfio/group.c
460
if (group->container)
drivers/vfio/group.c
461
vfio_group_detach_container(group);
drivers/vfio/group.c
462
if (group->iommufd) {
drivers/vfio/group.c
463
iommufd_ctx_put(group->iommufd);
drivers/vfio/group.c
464
group->iommufd = NULL;
drivers/vfio/group.c
466
group->opened_file = NULL;
drivers/vfio/group.c
467
mutex_unlock(&group->group_lock);
drivers/vfio/group.c
485
struct vfio_group *group;
drivers/vfio/group.c
493
list_for_each_entry(group, &vfio.group_list, vfio_next) {
drivers/vfio/group.c
494
if (group->iommu_group == iommu_group)
drivers/vfio/group.c
495
return group;
drivers/vfio/group.c
50
mutex_unlock(&group->device_lock);
drivers/vfio/group.c
502
struct vfio_group *group = container_of(dev, struct vfio_group, dev);
drivers/vfio/group.c
504
mutex_destroy(&group->device_lock);
drivers/vfio/group.c
505
mutex_destroy(&group->group_lock);
drivers/vfio/group.c
506
WARN_ON(group->iommu_group);
drivers/vfio/group.c
507
WARN_ON(group->cdev_device_open_cnt);
drivers/vfio/group.c
508
ida_free(&vfio.group_ida, MINOR(group->dev.devt));
drivers/vfio/group.c
509
kfree(group);
drivers/vfio/group.c
515
struct vfio_group *group;
drivers/vfio/group.c
518
group = kzalloc_obj(*group);
drivers/vfio/group.c
519
if (!group)
drivers/vfio/group.c
524
kfree(group);
drivers/vfio/group.c
528
device_initialize(&group->dev);
drivers/vfio/group.c
529
group->dev.devt = MKDEV(MAJOR(vfio.group_devt), minor);
drivers/vfio/group.c
530
group->dev.class = vfio.class;
drivers/vfio/group.c
531
group->dev.release = vfio_group_release;
drivers/vfio/group.c
532
cdev_init(&group->cdev, &vfio_group_fops);
drivers/vfio/group.c
533
group->cdev.owner = THIS_MODULE;
drivers/vfio/group.c
535
refcount_set(&group->drivers, 1);
drivers/vfio/group.c
536
mutex_init(&group->group_lock);
drivers/vfio/group.c
537
spin_lock_init(&group->kvm_ref_lock);
drivers/vfio/group.c
538
INIT_LIST_HEAD(&group->device_list);
drivers/vfio/group.c
539
mutex_init(&group->device_lock);
drivers/vfio/group.c
540
group->iommu_group = iommu_group;
drivers/vfio/group.c
543
group->type = type;
drivers/vfio/group.c
544
BLOCKING_INIT_NOTIFIER_HEAD(&group->notifier);
drivers/vfio/group.c
546
return group;
drivers/vfio/group.c
552
struct vfio_group *group;
drivers/vfio/group.c
558
group = vfio_group_alloc(iommu_group, type);
drivers/vfio/group.c
559
if (IS_ERR(group))
drivers/vfio/group.c
560
return group;
drivers/vfio/group.c
562
err = dev_set_name(&group->dev, "%s%d",
drivers/vfio/group.c
563
group->type == VFIO_NO_IOMMU ? "noiommu-" : "",
drivers/vfio/group.c
570
err = cdev_device_add(&group->cdev, &group->dev);
drivers/vfio/group.c
576
list_add(&group->vfio_next, &vfio.group_list);
drivers/vfio/group.c
578
return group;
drivers/vfio/group.c
58
static bool vfio_group_has_iommu(struct vfio_group *group)
drivers/vfio/group.c
581
put_device(&group->dev);
drivers/vfio/group.c
589
struct vfio_group *group;
drivers/vfio/group.c
60
lockdep_assert_held(&group->group_lock);
drivers/vfio/group.c
604
group = vfio_create_group(iommu_group, type);
drivers/vfio/group.c
606
if (IS_ERR(group)) {
drivers/vfio/group.c
607
ret = PTR_ERR(group);
drivers/vfio/group.c
611
return group;
drivers/vfio/group.c
620
static bool vfio_group_has_device(struct vfio_group *group, struct device *dev)
drivers/vfio/group.c
624
mutex_lock(&group->device_lock);
drivers/vfio/group.c
625
list_for_each_entry(device, &group->device_list, group_next) {
drivers/vfio/group.c
627
mutex_unlock(&group->device_lock);
drivers/vfio/group.c
631
mutex_unlock(&group->device_lock);
drivers/vfio/group.c
638
struct vfio_group *group;
drivers/vfio/group.c
648
group = vfio_noiommu_group_alloc(dev, VFIO_NO_IOMMU);
drivers/vfio/group.c
649
if (!IS_ERR(group)) {
drivers/vfio/group.c
65
WARN_ON(!group->container != !group->container_users);
drivers/vfio/group.c
653
return group;
drivers/vfio/group.c
660
group = vfio_group_find_from_iommu(iommu_group);
drivers/vfio/group.c
661
if (group) {
drivers/vfio/group.c
662
if (WARN_ON(vfio_group_has_device(group, dev)))
drivers/vfio/group.c
663
group = ERR_PTR(-EINVAL);
drivers/vfio/group.c
665
refcount_inc(&group->drivers);
drivers/vfio/group.c
667
group = vfio_create_group(iommu_group, VFIO_IOMMU);
drivers/vfio/group.c
67
return group->container || group->iommufd;
drivers/vfio/group.c
673
return group;
drivers/vfio/group.c
679
struct vfio_group *group;
drivers/vfio/group.c
682
group = vfio_group_find_or_alloc(device->dev);
drivers/vfio/group.c
684
group = vfio_noiommu_group_alloc(device->dev, type);
drivers/vfio/group.c
686
if (IS_ERR(group))
drivers/vfio/group.c
687
return PTR_ERR(group);
drivers/vfio/group.c
690
device->group = group;
drivers/vfio/group.c
696
struct vfio_group *group = device->group;
drivers/vfio/group.c
699
if (group->type == VFIO_NO_IOMMU || group->type == VFIO_EMULATED_IOMMU)
drivers/vfio/group.c
703
if (!refcount_dec_and_mutex_lock(&group->drivers, &vfio.group_lock))
drivers/vfio/group.c
705
list_del(&group->vfio_next);
drivers/vfio/group.c
713
cdev_device_del(&group->cdev, &group->dev);
drivers/vfio/group.c
715
mutex_lock(&group->group_lock);
drivers/vfio/group.c
722
WARN_ON(!list_empty(&group->device_list));
drivers/vfio/group.c
723
WARN_ON(group->notifier.head);
drivers/vfio/group.c
730
if (group->container)
drivers/vfio/group.c
731
vfio_group_detach_container(group);
drivers/vfio/group.c
732
iommu_group = group->iommu_group;
drivers/vfio/group.c
733
group->iommu_group = NULL;
drivers/vfio/group.c
734
mutex_unlock(&group->group_lock);
drivers/vfio/group.c
738
put_device(&group->dev);
drivers/vfio/group.c
743
mutex_lock(&device->group->device_lock);
drivers/vfio/group.c
744
list_add(&device->group_next, &device->group->device_list);
drivers/vfio/group.c
745
mutex_unlock(&device->group->device_lock);
drivers/vfio/group.c
750
mutex_lock(&device->group->device_lock);
drivers/vfio/group.c
752
mutex_unlock(&device->group->device_lock);
drivers/vfio/group.c
757
struct vfio_group *group = device->group;
drivers/vfio/group.c
76
static int vfio_group_ioctl_unset_container(struct vfio_group *group)
drivers/vfio/group.c
760
lockdep_assert_held(&group->group_lock);
drivers/vfio/group.c
762
if (WARN_ON(!group->container))
drivers/vfio/group.c
765
ret = vfio_group_use_container(group);
drivers/vfio/group.c
774
struct vfio_group *group = device->group;
drivers/vfio/group.c
776
lockdep_assert_held(&group->group_lock);
drivers/vfio/group.c
778
if (WARN_ON(!group->container))
drivers/vfio/group.c
782
vfio_group_unuse_container(group);
drivers/vfio/group.c
787
return device->group->container;
drivers/vfio/group.c
792
struct vfio_group *group = file->private_data;
drivers/vfio/group.c
796
return group;
drivers/vfio/group.c
80
mutex_lock(&group->group_lock);
drivers/vfio/group.c
809
struct vfio_group *group = vfio_group_from_file(file);
drivers/vfio/group.c
81
if (!vfio_group_has_iommu(group)) {
drivers/vfio/group.c
815
if (!group)
drivers/vfio/group.c
818
mutex_lock(&group->group_lock);
drivers/vfio/group.c
819
if (group->iommu_group) {
drivers/vfio/group.c
820
iommu_group = group->iommu_group;
drivers/vfio/group.c
823
mutex_unlock(&group->group_lock);
drivers/vfio/group.c
838
bool vfio_group_enforced_coherent(struct vfio_group *group)
drivers/vfio/group.c
849
mutex_lock(&group->device_lock);
drivers/vfio/group.c
85
if (group->container) {
drivers/vfio/group.c
850
list_for_each_entry(device, &group->device_list, group_next) {
drivers/vfio/group.c
857
mutex_unlock(&group->device_lock);
drivers/vfio/group.c
86
if (group->container_users != 1) {
drivers/vfio/group.c
861
void vfio_group_set_kvm(struct vfio_group *group, struct kvm *kvm)
drivers/vfio/group.c
863
spin_lock(&group->kvm_ref_lock);
drivers/vfio/group.c
864
group->kvm = kvm;
drivers/vfio/group.c
865
spin_unlock(&group->kvm_ref_lock);
drivers/vfio/group.c
877
struct vfio_group *group = vfio_group_from_file(file);
drivers/vfio/group.c
879
if (!group)
drivers/vfio/group.c
882
return group == device->group;
drivers/vfio/group.c
90
vfio_group_detach_container(group);
drivers/vfio/group.c
92
if (group->iommufd) {
drivers/vfio/group.c
93
iommufd_ctx_put(group->iommufd);
drivers/vfio/group.c
94
group->iommufd = NULL;
drivers/vfio/group.c
98
mutex_unlock(&group->group_lock);
drivers/vfio/iommufd.c
101
if (iommufd_ctx_has_group(ictx, group))
drivers/vfio/iommufd.c
106
iommu_group_put(group);
drivers/vfio/iommufd.c
91
struct iommu_group *group;
drivers/vfio/iommufd.c
97
group = iommu_group_get(vdev->dev);
drivers/vfio/iommufd.c
98
if (!group)
drivers/vfio/vfio.h
110
bool vfio_group_enforced_coherent(struct vfio_group *group);
drivers/vfio/vfio.h
111
void vfio_group_set_kvm(struct vfio_group *group, struct kvm *kvm);
drivers/vfio/vfio.h
119
vdev->group->type == VFIO_NO_IOMMU;
drivers/vfio/vfio.h
169
static inline bool vfio_group_enforced_coherent(struct vfio_group *group)
drivers/vfio/vfio.h
174
static inline void vfio_group_set_kvm(struct vfio_group *group, struct kvm *kvm)
drivers/vfio/vfio.h
21
struct vfio_group *group;
drivers/vfio/vfio.h
210
struct iommu_group *group,
drivers/vfio/vfio.h
213
struct iommu_group *group);
drivers/vfio/vfio.h
215
struct iommu_group *group,
drivers/vfio/vfio.h
228
struct iommu_group *group);
drivers/vfio/vfio.h
240
int vfio_group_use_container(struct vfio_group *group);
drivers/vfio/vfio.h
241
void vfio_group_unuse_container(struct vfio_group *group);
drivers/vfio/vfio.h
243
struct vfio_group *group);
drivers/vfio/vfio.h
244
void vfio_group_detach_container(struct vfio_group *group);
drivers/vfio/vfio.h
265
static inline int vfio_group_use_container(struct vfio_group *group)
drivers/vfio/vfio.h
270
static inline void vfio_group_unuse_container(struct vfio_group *group)
drivers/vfio/vfio.h
275
struct vfio_group *group)
drivers/vfio/vfio.h
280
static inline void vfio_group_detach_container(struct vfio_group *group)
drivers/vfio/vfio_iommu_spapr_tce.c
777
static long vfio_spapr_ioctl_eeh_pe_op(struct iommu_group *group,
drivers/vfio/vfio_iommu_spapr_tce.c
784
pe = eeh_iommu_group_to_pe(group);
drivers/vfio/vfio_iommu_type1.c
1965
struct vfio_iommu_group *group;
drivers/vfio/vfio_iommu_type1.c
1969
group = find_iommu_group(domain, iommu_group);
drivers/vfio/vfio_iommu_type1.c
1970
if (group)
drivers/vfio/vfio_iommu_type1.c
1971
return group;
drivers/vfio/vfio_iommu_type1.c
1974
list_for_each_entry(group, &iommu->emulated_iommu_groups, next)
drivers/vfio/vfio_iommu_type1.c
1975
if (group->iommu_group == iommu_group)
drivers/vfio/vfio_iommu_type1.c
1976
return group;
drivers/vfio/vfio_iommu_type1.c
2241
struct vfio_iommu_group *group;
drivers/vfio/vfio_iommu_type1.c
2262
group = kzalloc_obj(*group);
drivers/vfio/vfio_iommu_type1.c
2263
if (!group)
drivers/vfio/vfio_iommu_type1.c
2265
group->iommu_group = iommu_group;
drivers/vfio/vfio_iommu_type1.c
2268
list_add(&group->next, &iommu->emulated_iommu_groups);
drivers/vfio/vfio_iommu_type1.c
2275
group->pinned_page_dirty_scope = true;
drivers/vfio/vfio_iommu_type1.c
2297
ret = iommu_attach_group(domain->domain, group->iommu_group);
drivers/vfio/vfio_iommu_type1.c
2339
list_add(&group->next, &domain->group_list);
drivers/vfio/vfio_iommu_type1.c
2370
iommu_detach_group(domain->domain, group->iommu_group);
drivers/vfio/vfio_iommu_type1.c
2372
group->iommu_group)) {
drivers/vfio/vfio_iommu_type1.c
2373
list_add(&group->next, &d->group_list);
drivers/vfio/vfio_iommu_type1.c
2380
group->iommu_group);
drivers/vfio/vfio_iommu_type1.c
2415
iommu_detach_group(domain->domain, group->iommu_group);
drivers/vfio/vfio_iommu_type1.c
2423
kfree(group);
drivers/vfio/vfio_iommu_type1.c
2544
struct vfio_iommu_group *group;
drivers/vfio/vfio_iommu_type1.c
2549
list_for_each_entry(group, &iommu->emulated_iommu_groups, next) {
drivers/vfio/vfio_iommu_type1.c
2550
if (group->iommu_group != iommu_group)
drivers/vfio/vfio_iommu_type1.c
2552
update_dirty_scope = !group->pinned_page_dirty_scope;
drivers/vfio/vfio_iommu_type1.c
2553
list_del(&group->next);
drivers/vfio/vfio_iommu_type1.c
2554
kfree(group);
drivers/vfio/vfio_iommu_type1.c
2572
group = find_iommu_group(domain, iommu_group);
drivers/vfio/vfio_iommu_type1.c
2573
if (!group)
drivers/vfio/vfio_iommu_type1.c
2576
iommu_detach_group(domain->domain, group->iommu_group);
drivers/vfio/vfio_iommu_type1.c
2577
update_dirty_scope = !group->pinned_page_dirty_scope;
drivers/vfio/vfio_iommu_type1.c
2578
list_del(&group->next);
drivers/vfio/vfio_iommu_type1.c
2579
kfree(group);
drivers/vfio/vfio_iommu_type1.c
2659
struct vfio_iommu_group *group, *group_tmp;
drivers/vfio/vfio_iommu_type1.c
2661
list_for_each_entry_safe(group, group_tmp,
drivers/vfio/vfio_iommu_type1.c
2663
iommu_detach_group(domain->domain, group->iommu_group);
drivers/vfio/vfio_iommu_type1.c
2664
list_del(&group->next);
drivers/vfio/vfio_iommu_type1.c
2665
kfree(group);
drivers/vfio/vfio_iommu_type1.c
2675
struct vfio_iommu_group *group, *next_group;
drivers/vfio/vfio_iommu_type1.c
2677
list_for_each_entry_safe(group, next_group,
drivers/vfio/vfio_iommu_type1.c
2679
list_del(&group->next);
drivers/vfio/vfio_iommu_type1.c
2680
kfree(group);
drivers/vfio/vfio_iommu_type1.c
901
struct vfio_iommu_group *group;
drivers/vfio/vfio_iommu_type1.c
997
group = vfio_iommu_find_iommu_group(iommu, iommu_group);
drivers/vfio/vfio_iommu_type1.c
998
if (!group->pinned_page_dirty_scope) {
drivers/vfio/vfio_iommu_type1.c
999
group->pinned_page_dirty_scope = true;
drivers/vfio/vfio_main.c
1328
if (IS_ENABLED(CONFIG_VFIO_DEVICE_CDEV) && !df->group) {
drivers/vfio/vfio_main.c
1471
struct vfio_group *group;
drivers/vfio/vfio_main.c
1473
group = vfio_group_from_file(file);
drivers/vfio/vfio_main.c
1474
if (group)
drivers/vfio/vfio_main.c
1475
return vfio_group_enforced_coherent(group);
drivers/vfio/vfio_main.c
1510
struct vfio_group *group;
drivers/vfio/vfio_main.c
1512
group = vfio_group_from_file(file);
drivers/vfio/vfio_main.c
1513
if (group)
drivers/vfio/vfio_main.c
1514
vfio_group_set_kvm(group, kvm);
drivers/vfio/vfio_main.c
570
if (device->open_count != 0 && !df->group)
drivers/vfio/vfio_main.c
637
if (df->group)
drivers/vhost/scsi.c
2836
struct config_group *group,
drivers/virt/coco/guest/report.c
422
static struct config_item *tsm_report_make_item(struct config_group *group,
drivers/virt/coco/guest/report.c
440
static void tsm_report_drop_item(struct config_group *group, struct config_item *item)
drivers/watchdog/imx_sc_wdt.c
139
unsigned long event, void *group)
drivers/watchdog/imx_sc_wdt.c
147
*(u8 *)group == SC_IRQ_GROUP_WDOG)
drivers/xen/xen-scsiback.c
1345
struct config_group *group,
fs/afs/afs.h
136
s64 group; /* group ID */
fs/afs/fsclient.c
106
status->group = ntohl(xdr->group);
fs/afs/fsclient.c
178
u32 mask = 0, mtime = 0, owner = 0, group = 0, mode = 0;
fs/afs/fsclient.c
193
group = from_kgid(&init_user_ns, attr->ia_gid);
fs/afs/fsclient.c
204
*bp++ = htonl(group);
fs/afs/inode.c
183
inode->i_gid = make_kgid(&init_user_ns, status->group);
fs/afs/inode.c
285
if (status->group != vnode->status.group)
fs/afs/inode.c
286
inode->i_gid = make_kgid(&init_user_ns, status->group);
fs/afs/protocol_yfs.h
119
struct yfs_xdr_u64 group;
fs/afs/protocol_yfs.h
96
struct yfs_xdr_u64 group;
fs/afs/xdr_fs.h
28
__be32 group;
fs/afs/yfsclient.c
100
x->group = u64_to_xdr(0);
fs/afs/yfsclient.c
206
status->group = xdr_to_u64(xdr->group);
fs/afs/yfsclient.c
268
s64 mtime = 0, owner = 0, group = 0;
fs/afs/yfsclient.c
284
group = from_kgid(&init_user_ns, attr->ia_gid);
fs/afs/yfsclient.c
296
x->group = u64_to_xdr(group);
fs/btrfs/qgroup.c
1500
qgroup_iterator_add(&qgroup_list, glist->group);
fs/btrfs/qgroup.c
1576
if (list->group == parent) {
fs/btrfs/qgroup.c
1633
if (list->group == parent) {
fs/btrfs/qgroup.c
1827
list->group->qgroupid);
fs/btrfs/qgroup.c
2777
qgroup_iterator_nested_add(qgroups, glist->group);
fs/btrfs/qgroup.c
2778
qgroup_iterator_add(&tmp, glist->group);
fs/btrfs/qgroup.c
280
prealloc->group = parent;
fs/btrfs/qgroup.c
324
if (list->group == parent) {
fs/btrfs/qgroup.c
3243
qgids[i++] = qg_list->group->qgroupid;
fs/btrfs/qgroup.c
3288
if (list->group->qgroupid != parentid)
fs/btrfs/qgroup.c
3316
qgroup_iterator_add(&qgroup_list, list->group);
fs/btrfs/qgroup.c
3633
qgroup_iterator_add(&qgroup_list, glist->group);
fs/btrfs/qgroup.c
3697
qgroup_iterator_add(&qgroup_list, glist->group);
fs/btrfs/qgroup.c
4603
qgroup_iterator_add(&qgroup_list, glist->group);
fs/btrfs/qgroup.c
4954
qgroup_iterator_add(&qgroup_list, glist->group);
fs/btrfs/qgroup.h
294
struct btrfs_qgroup *group;
fs/btrfs/relocation.c
3724
const struct btrfs_block_group *group)
fs/btrfs/relocation.c
3726
struct btrfs_fs_info *fs_info = group->fs_info;
fs/btrfs/relocation.c
3755
inode->reloc_block_group_start = group->start;
fs/configfs/dir.c
1293
struct config_group *group = NULL;
fs/configfs/dir.c
1363
group = type->ct_group_ops->make_group(to_config_group(parent_item), name);
fs/configfs/dir.c
1364
if (!group)
fs/configfs/dir.c
1365
group = ERR_PTR(-ENOMEM);
fs/configfs/dir.c
1366
if (!IS_ERR(group)) {
fs/configfs/dir.c
1367
link_group(to_config_group(parent_item), group);
fs/configfs/dir.c
1368
item = &group->cg_item;
fs/configfs/dir.c
1370
ret = PTR_ERR(group);
fs/configfs/dir.c
1426
if (group)
fs/configfs/dir.c
1443
if (group)
fs/configfs/dir.c
1444
unlink_group(group);
fs/configfs/dir.c
1755
struct config_group *group)
fs/configfs/dir.c
1767
link_group(parent_group, group);
fs/configfs/dir.c
1773
ret = create_default_group(parent_group, group, frag);
fs/configfs/dir.c
1778
configfs_dir_set_ready(group->cg_item.ci_dentry->d_fsdata);
fs/configfs/dir.c
1786
unlink_group(group);
fs/configfs/dir.c
1799
void configfs_unregister_group(struct config_group *group)
fs/configfs/dir.c
1801
struct configfs_subsystem *subsys = group->cg_subsys;
fs/configfs/dir.c
1802
struct dentry *dentry = group->cg_item.ci_dentry;
fs/configfs/dir.c
1803
struct dentry *parent = group->cg_item.ci_parent->ci_dentry;
fs/configfs/dir.c
1816
configfs_detach_group(&group->cg_item);
fs/configfs/dir.c
1826
unlink_group(group);
fs/configfs/dir.c
1848
struct config_group *group;
fs/configfs/dir.c
1850
group = kzalloc_obj(*group);
fs/configfs/dir.c
1851
if (!group)
fs/configfs/dir.c
1853
config_group_init_type_name(group, name, item_type);
fs/configfs/dir.c
1855
ret = configfs_register_group(parent_group, group);
fs/configfs/dir.c
1857
kfree(group);
fs/configfs/dir.c
1860
return group;
fs/configfs/dir.c
1868
void configfs_unregister_default_group(struct config_group *group)
fs/configfs/dir.c
1870
configfs_unregister_group(group);
fs/configfs/dir.c
1871
kfree(group);
fs/configfs/dir.c
1878
struct config_group *group = &subsys->su_group;
fs/configfs/dir.c
1894
if (!group->cg_item.ci_name)
fs/configfs/dir.c
1895
group->cg_item.ci_name = group->cg_item.ci_namebuf;
fs/configfs/dir.c
1899
link_group(to_config_group(sd->s_element), group);
fs/configfs/dir.c
1905
dentry = d_alloc_name(root, group->cg_item.ci_name);
fs/configfs/dir.c
1912
&group->cg_item,
fs/configfs/dir.c
1929
unlink_group(group);
fs/configfs/dir.c
1940
struct config_group *group = &subsys->su_group;
fs/configfs/dir.c
1941
struct dentry *dentry = group->cg_item.ci_dentry;
fs/configfs/dir.c
1965
configfs_detach_group(&group->cg_item);
fs/configfs/dir.c
1978
unlink_group(group);
fs/configfs/dir.c
650
static void detach_groups(struct config_group *group)
fs/configfs/dir.c
652
struct dentry * dentry = dget(group->cg_item.ci_dentry);
fs/configfs/dir.c
695
struct config_group *group,
fs/configfs/dir.c
703
if (!group->cg_item.ci_name)
fs/configfs/dir.c
704
group->cg_item.ci_name = group->cg_item.ci_namebuf;
fs/configfs/dir.c
707
child = d_alloc_name(parent, group->cg_item.ci_name);
fs/configfs/dir.c
712
&group->cg_item, child, frag);
fs/configfs/dir.c
726
static int populate_groups(struct config_group *group,
fs/configfs/dir.c
732
list_for_each_entry(new_group, &group->default_groups, group_entry) {
fs/configfs/dir.c
733
ret = create_default_group(group, new_group, frag);
fs/configfs/dir.c
735
detach_groups(group);
fs/configfs/dir.c
743
void configfs_remove_default_groups(struct config_group *group)
fs/configfs/dir.c
747
list_for_each_entry_safe(g, n, &group->default_groups, group_entry) {
fs/configfs/dir.c
761
struct config_group *group;
fs/configfs/dir.c
763
group = item->ci_group;
fs/configfs/dir.c
764
if (group) {
fs/configfs/dir.c
774
config_group_put(group);
fs/configfs/dir.c
800
static void unlink_group(struct config_group *group)
fs/configfs/dir.c
804
list_for_each_entry(new_group, &group->default_groups, group_entry)
fs/configfs/dir.c
807
group->cg_subsys = NULL;
fs/configfs/dir.c
808
unlink_obj(&group->cg_item);
fs/configfs/dir.c
811
static void link_group(struct config_group *parent_group, struct config_group *group)
fs/configfs/dir.c
816
link_obj(&parent_group->cg_item, &group->cg_item);
fs/configfs/dir.c
821
subsys = to_configfs_subsystem(group);
fs/configfs/dir.c
824
group->cg_subsys = subsys;
fs/configfs/dir.c
826
list_for_each_entry(new_group, &group->default_groups, group_entry)
fs/configfs/dir.c
827
link_group(group, new_group);
fs/configfs/item.c
157
void config_group_init(struct config_group *group)
fs/configfs/item.c
159
config_item_init(&group->cg_item);
fs/configfs/item.c
160
INIT_LIST_HEAD(&group->cg_children);
fs/configfs/item.c
161
INIT_LIST_HEAD(&group->default_groups);
fs/configfs/item.c
174
struct config_item *config_group_find_item(struct config_group *group,
fs/configfs/item.c
180
list_for_each(entry, &group->cg_children) {
fs/configfs/item.c
92
void config_group_init_type_name(struct config_group *group, const char *name,
fs/configfs/item.c
95
config_item_set_name(&group->cg_item, "%s", name);
fs/configfs/item.c
96
group->cg_item.ci_type = type;
fs/configfs/item.c
97
config_group_init(group);
fs/dlm/config.c
283
struct config_group group;
fs/dlm/config.c
408
return i ? container_of(to_config_group(i), struct dlm_space, group) :
fs/dlm/config.c
439
config_group_init_type_name(&cl->group, name, &cluster_type);
fs/dlm/config.c
443
configfs_add_default_group(&sps->ss_group, &cl->group);
fs/dlm/config.c
444
configfs_add_default_group(&cms->cs_group, &cl->group);
fs/dlm/config.c
448
return &cl->group;
fs/dlm/config.c
461
configfs_remove_default_groups(&cl->group);
fs/dlm/config.c
489
config_group_init_type_name(&sp->group, name, &space_type);
fs/dlm/config.c
492
configfs_add_default_group(&nds->ns_group, &sp->group);
fs/dlm/config.c
499
return &sp->group;
fs/dlm/config.c
513
configfs_remove_default_groups(&sp->group);
fs/dlm/config.c
76
struct config_group group;
fs/dlm/config.c
83
return i ? container_of(to_config_group(i), struct dlm_cluster, group) :
fs/dlm/config.c
892
config_item_put(&sp->group.cg_item);
fs/ext2/balloc.c
1064
ext2_try_to_allocate_with_rsv(struct super_block *sb, unsigned int group,
fs/ext2/balloc.c
1080
return ext2_try_to_allocate(sb, group, bitmap_bh,
fs/ext2/balloc.c
1089
group_first_block = ext2_group_first_block_no(sb, group);
fs/ext2/balloc.c
1090
group_last_block = ext2_group_last_block_no(sb, group);
fs/ext2/balloc.c
1110
grp_goal, group, sb)) {
fs/ext2/balloc.c
1114
group, bitmap_bh);
fs/ext2/balloc.c
1119
grp_goal, group, sb))
fs/ext2/balloc.c
1134
group, grp_goal, group_first_block,
fs/ext2/balloc.c
1140
ret = ext2_try_to_allocate(sb, group, bitmap_bh, grp_goal,
fs/ext2/balloc.c
1498
static int ext2_group_sparse(int group)
fs/ext2/balloc.c
1500
if (group <= 1)
fs/ext2/balloc.c
1502
return (test_root(group, 3) || test_root(group, 5) ||
fs/ext2/balloc.c
1503
test_root(group, 7));
fs/ext2/balloc.c
1514
int ext2_bg_has_super(struct super_block *sb, int group)
fs/ext2/balloc.c
1517
!ext2_group_sparse(group))
fs/ext2/balloc.c
1531
unsigned long ext2_bg_num_gdb(struct super_block *sb, int group)
fs/ext2/balloc.c
1533
return ext2_bg_has_super(sb, group) ? EXT2_SB(sb)->s_gdb_count : 0;
fs/ext2/balloc.c
271
unsigned int group, struct super_block * sb)
fs/ext2/balloc.c
275
group_first_block = ext2_group_first_block_no(sb, group);
fs/ext2/balloc.c
276
group_last_block = ext2_group_last_block_no(sb, group);
fs/ext2/balloc.c
668
ext2_try_to_allocate(struct super_block *sb, int group,
fs/ext2/balloc.c
673
ext2_fsblk_t group_first_block = ext2_group_first_block_no(sb, group);
fs/ext2/balloc.c
674
ext2_fsblk_t group_last_block = ext2_group_last_block_no(sb, group);
fs/ext2/balloc.c
707
if (ext2_set_bit_atomic(sb_bgl_lock(EXT2_SB(sb), group),
fs/ext2/balloc.c
872
unsigned int group, struct buffer_head *bitmap_bh)
fs/ext2/balloc.c
882
group_first_block = ext2_group_first_block_no(sb, group);
fs/ext2/balloc.c
883
group_end_block = ext2_group_last_block_no(sb, group);
fs/ext2/ext2.h
702
extern int ext2_bg_has_super(struct super_block *sb, int group);
fs/ext2/ext2.h
703
extern unsigned long ext2_bg_num_gdb(struct super_block *sb, int group);
fs/ext2/ialloc.c
204
int group, best_group = -1;
fs/ext2/ialloc.c
206
for (group = 0; group < ngroups; group++) {
fs/ext2/ialloc.c
207
desc = ext2_get_group_desc (sb, group, NULL);
fs/ext2/ialloc.c
215
best_group = group;
fs/ext2/ialloc.c
265
int group = -1, i;
fs/ext2/ialloc.c
281
group = (parent_group + i) % ngroups;
fs/ext2/ialloc.c
282
desc = ext2_get_group_desc (sb, group, NULL);
fs/ext2/ialloc.c
291
best_group = group;
fs/ext2/ialloc.c
295
group = best_group;
fs/ext2/ialloc.c
319
group = (parent_group + i) % ngroups;
fs/ext2/ialloc.c
320
desc = ext2_get_group_desc (sb, group, NULL);
fs/ext2/ialloc.c
323
if (sbi->s_debts[group] >= max_debt)
fs/ext2/ialloc.c
336
group = (parent_group + i) % ngroups;
fs/ext2/ialloc.c
337
desc = ext2_get_group_desc (sb, group, NULL);
fs/ext2/ialloc.c
356
return group;
fs/ext2/ialloc.c
364
int group, i;
fs/ext2/ialloc.c
369
group = parent_group;
fs/ext2/ialloc.c
370
desc = ext2_get_group_desc (sb, group, NULL);
fs/ext2/ialloc.c
384
group = (group + parent->i_ino) % ngroups;
fs/ext2/ialloc.c
391
group += i;
fs/ext2/ialloc.c
392
if (group >= ngroups)
fs/ext2/ialloc.c
393
group -= ngroups;
fs/ext2/ialloc.c
394
desc = ext2_get_group_desc (sb, group, NULL);
fs/ext2/ialloc.c
404
group = parent_group;
fs/ext2/ialloc.c
406
if (++group >= ngroups)
fs/ext2/ialloc.c
407
group = 0;
fs/ext2/ialloc.c
408
desc = ext2_get_group_desc (sb, group, NULL);
fs/ext2/ialloc.c
416
return group;
fs/ext2/ialloc.c
425
int group, i;
fs/ext2/ialloc.c
444
group = find_group_dir(sb, dir);
fs/ext2/ialloc.c
446
group = find_group_orlov(sb, dir);
fs/ext2/ialloc.c
448
group = find_group_other(sb, dir);
fs/ext2/ialloc.c
450
if (group == -1) {
fs/ext2/ialloc.c
456
gdp = ext2_get_group_desc(sb, group, &bh2);
fs/ext2/ialloc.c
458
if (++group == sbi->s_groups_count)
fs/ext2/ialloc.c
459
group = 0;
fs/ext2/ialloc.c
463
bitmap_bh = read_inode_bitmap(sb, group);
fs/ext2/ialloc.c
482
if (++group == sbi->s_groups_count)
fs/ext2/ialloc.c
483
group = 0;
fs/ext2/ialloc.c
486
if (ext2_set_bit_atomic(sb_bgl_lock(sbi, group),
fs/ext2/ialloc.c
491
if (++group == sbi->s_groups_count)
fs/ext2/ialloc.c
492
group = 0;
fs/ext2/ialloc.c
513
ino += group * EXT2_INODES_PER_GROUP(sb) + 1;
fs/ext2/ialloc.c
517
"block_group = %d,inode=%lu", group,
fs/ext2/ialloc.c
527
spin_lock(sb_bgl_lock(sbi, group));
fs/ext2/ialloc.c
530
if (sbi->s_debts[group] < 255)
fs/ext2/ialloc.c
531
sbi->s_debts[group]++;
fs/ext2/ialloc.c
534
if (sbi->s_debts[group])
fs/ext2/ialloc.c
535
sbi->s_debts[group]--;
fs/ext2/ialloc.c
537
spin_unlock(sb_bgl_lock(sbi, group));
fs/ext2/ialloc.c
560
ei->i_block_group = group;
fs/ext2/ialloc.c
66
static void ext2_release_inode(struct super_block *sb, int group, int dir)
fs/ext2/ialloc.c
71
desc = ext2_get_group_desc(sb, group, &bh);
fs/ext2/ialloc.c
74
"can't get descriptor for group %d", group);
fs/ext2/ialloc.c
78
spin_lock(sb_bgl_lock(EXT2_SB(sb), group));
fs/ext2/ialloc.c
82
spin_unlock(sb_bgl_lock(EXT2_SB(sb), group));
fs/ext4/balloc.c
327
ext4_group_t group)
fs/ext4/balloc.c
332
if (unlikely(group >= EXT4_SB(sb)->s_groups_count))
fs/ext4/balloc.c
334
indexv = group >> (EXT4_DESC_PER_BLOCK_BITS(sb));
fs/ext4/balloc.c
335
indexh = group & ((EXT4_DESC_PER_BLOCK(sb)) - 1);
fs/ext4/balloc.c
39
ext4_group_t group;
fs/ext4/balloc.c
42
group = (block -
fs/ext4/balloc.c
46
ext4_get_group_no_and_offset(sb, block, &group, NULL);
fs/ext4/balloc.c
47
return group;
fs/ext4/balloc.c
854
int ext4_bg_has_super(struct super_block *sb, ext4_group_t group)
fs/ext4/balloc.c
858
if (group == 0)
fs/ext4/balloc.c
861
if (group == le32_to_cpu(es->s_backup_bgs[0]) ||
fs/ext4/balloc.c
862
group == le32_to_cpu(es->s_backup_bgs[1]))
fs/ext4/balloc.c
866
if ((group <= 1) || !ext4_has_feature_sparse_super(sb))
fs/ext4/balloc.c
868
if (!(group & 1))
fs/ext4/balloc.c
870
if (test_root(group, 3) || (test_root(group, 5)) ||
fs/ext4/balloc.c
871
test_root(group, 7))
fs/ext4/balloc.c
878
ext4_group_t group)
fs/ext4/balloc.c
880
unsigned long metagroup = group / EXT4_DESC_PER_BLOCK(sb);
fs/ext4/balloc.c
884
if (group == first || group == first + 1 || group == last)
fs/ext4/balloc.c
890
ext4_group_t group)
fs/ext4/balloc.c
892
if (!ext4_bg_has_super(sb, group))
fs/ext4/balloc.c
910
unsigned long ext4_bg_num_gdb(struct super_block *sb, ext4_group_t group)
fs/ext4/balloc.c
914
unsigned long metagroup = group / EXT4_DESC_PER_BLOCK(sb);
fs/ext4/balloc.c
917
return ext4_bg_num_gdb_nometa(sb, group);
fs/ext4/balloc.c
919
return ext4_bg_num_gdb_meta(sb,group);
fs/ext4/ext4.h
2788
extern int ext4_bg_has_super(struct super_block *sb, ext4_group_t group);
fs/ext4/ext4.h
2790
ext4_group_t group);
fs/ext4/ext4.h
2803
ext4_group_t group);
fs/ext4/ext4.h
2964
ext4_group_t group, int barrier);
fs/ext4/ext4.h
3006
ext4_group_t group,
fs/ext4/ext4.h
3008
extern void ext4_mb_prefetch_fini(struct super_block *sb, ext4_group_t group,
fs/ext4/ext4.h
3360
extern int ext4_group_desc_csum_verify(struct super_block *sb, __u32 group,
fs/ext4/ext4.h
3362
extern void ext4_group_desc_csum_set(struct super_block *sb, __u32 group,
fs/ext4/ext4.h
3562
ext4_group_t group)
fs/ext4/ext4.h
3564
return bgl_lock_ptr(EXT4_SB(sb)->s_blockgroup_lock, group);
fs/ext4/ext4.h
3576
static inline bool ext4_try_lock_group(struct super_block *sb, ext4_group_t group)
fs/ext4/ext4.h
3578
if (!spin_trylock(ext4_group_lock_ptr(sb, group)))
fs/ext4/ext4.h
3588
static inline void ext4_lock_group(struct super_block *sb, ext4_group_t group)
fs/ext4/ext4.h
3590
if (!ext4_try_lock_group(sb, group)) {
fs/ext4/ext4.h
3597
spin_lock(ext4_group_lock_ptr(sb, group));
fs/ext4/ext4.h
3602
ext4_group_t group)
fs/ext4/ext4.h
3604
spin_unlock(ext4_group_lock_ptr(sb, group));
fs/ext4/ext4.h
662
u32 group;
fs/ext4/ext4.h
674
__u32 group;
fs/ext4/ialloc.c
1017
group = (goal - 1) / EXT4_INODES_PER_GROUP(sb);
fs/ext4/ialloc.c
1024
ret2 = find_group_orlov(sb, dir, &group, mode, qstr);
fs/ext4/ialloc.c
1026
ret2 = find_group_other(sb, dir, &group, mode);
fs/ext4/ialloc.c
1029
EXT4_I(dir)->i_last_alloc_group = group;
fs/ext4/ialloc.c
1042
gdp = ext4_get_group_desc(sb, group, &group_desc_bh);
fs/ext4/ialloc.c
1053
grp = ext4_get_group_info(sb, group);
fs/ext4/ialloc.c
1063
inode_bitmap_bh = ext4_read_inode_bitmap(sb, group);
fs/ext4/ialloc.c
1073
ret2 = find_inode_bit(sb, group, inode_bitmap_bh, &ino);
fs/ext4/ialloc.c
1077
if (group == 0 && (ino + 1) < EXT4_FIRST_INO(sb)) {
fs/ext4/ialloc.c
1080
ext4_mark_group_bitmap_corrupted(sb, group,
fs/ext4/ialloc.c
1103
ext4_lock_group(sb, group);
fs/ext4/ialloc.c
1109
ret2 = find_inode_bit(sb, group, inode_bitmap_bh, &ino);
fs/ext4/ialloc.c
1117
ext4_unlock_group(sb, group);
fs/ext4/ialloc.c
1123
if (++group == ngroups)
fs/ext4/ialloc.c
1124
group = 0;
fs/ext4/ialloc.c
1150
block_bitmap_bh = ext4_read_block_bitmap(sb, group);
fs/ext4/ialloc.c
1168
ext4_lock_group(sb, group);
fs/ext4/ialloc.c
1173
ext4_free_clusters_after_init(sb, group, gdp));
fs/ext4/ialloc.c
1175
ext4_group_desc_csum_set(sb, group, gdp);
fs/ext4/ialloc.c
1177
ext4_unlock_group(sb, group);
fs/ext4/ialloc.c
1192
grp = ext4_get_group_info(sb, group);
fs/ext4/ialloc.c
1202
ext4_lock_group(sb, group); /* while we modify the bg desc */
fs/ext4/ialloc.c
1220
ext4_lock_group(sb, group);
fs/ext4/ialloc.c
1227
ext4_group_t f = ext4_flex_group(sbi, group);
fs/ext4/ialloc.c
1235
ext4_group_desc_csum_set(sb, group, gdp);
fs/ext4/ialloc.c
1237
ext4_unlock_group(sb, group);
fs/ext4/ialloc.c
1251
flex_group = ext4_flex_group(sbi, group);
fs/ext4/ialloc.c
1256
inode->i_ino = ino + group * EXT4_INODES_PER_GROUP(sb);
fs/ext4/ialloc.c
1272
ei->i_block_group = group;
fs/ext4/ialloc.c
1286
ext4_mark_group_bitmap_corrupted(sb, group,
fs/ext4/ialloc.c
1517
int ext4_init_inode_table(struct super_block *sb, ext4_group_t group,
fs/ext4/ialloc.c
1520
struct ext4_group_info *grp = ext4_get_group_info(sb, group);
fs/ext4/ialloc.c
1529
gdp = ext4_get_group_desc(sb, group, &group_desc_bh);
fs/ext4/ialloc.c
1562
group, used_blks,
fs/ext4/ialloc.c
1568
used_inos += group * EXT4_INODES_PER_GROUP(sb);
fs/ext4/ialloc.c
1578
group, ext4_itable_unused_count(sb, gdp),
fs/ext4/ialloc.c
1603
group);
fs/ext4/ialloc.c
1611
ext4_lock_group(sb, group);
fs/ext4/ialloc.c
1613
ext4_group_desc_csum_set(sb, group, gdp);
fs/ext4/ialloc.c
1614
ext4_unlock_group(sb, group);
fs/ext4/ialloc.c
423
ext4_group_t *group, umode_t mode,
fs/ext4/ialloc.c
487
*group = grp;
fs/ext4/ialloc.c
504
*group = grp+i;
fs/ext4/ialloc.c
552
*group = grp;
fs/ext4/ialloc.c
571
ext4_group_t *group, umode_t mode)
fs/ext4/ialloc.c
596
*group = i;
fs/ext4/ialloc.c
610
*group = parent_group + flex_size;
fs/ext4/ialloc.c
611
if (*group > ngroups)
fs/ext4/ialloc.c
612
*group = 0;
fs/ext4/ialloc.c
613
return find_group_orlov(sb, parent, group, mode, NULL);
fs/ext4/ialloc.c
619
*group = parent_group;
fs/ext4/ialloc.c
620
desc = ext4_get_group_desc(sb, *group, NULL);
fs/ext4/ialloc.c
634
*group = (*group + parent->i_ino) % ngroups;
fs/ext4/ialloc.c
641
*group += i;
fs/ext4/ialloc.c
642
if (*group >= ngroups)
fs/ext4/ialloc.c
643
*group -= ngroups;
fs/ext4/ialloc.c
644
desc = ext4_get_group_desc(sb, *group, NULL);
fs/ext4/ialloc.c
654
*group = parent_group;
fs/ext4/ialloc.c
656
if (++*group >= ngroups)
fs/ext4/ialloc.c
657
*group = 0;
fs/ext4/ialloc.c
658
desc = ext4_get_group_desc(sb, *group, NULL);
fs/ext4/ialloc.c
675
static int recently_deleted(struct super_block *sb, ext4_group_t group, int ino)
fs/ext4/ialloc.c
685
gdp = ext4_get_group_desc(sb, group, NULL);
fs/ext4/ialloc.c
725
static int find_inode_bit(struct super_block *sb, ext4_group_t group,
fs/ext4/ialloc.c
738
if (check_recently_deleted && recently_deleted(sb, group, *ino)) {
fs/ext4/ialloc.c
764
ext4_group_t group;
fs/ext4/ialloc.c
771
group = (ino - 1) / EXT4_INODES_PER_GROUP(sb);
fs/ext4/ialloc.c
773
inode_bitmap_bh = ext4_read_inode_bitmap(sb, group);
fs/ext4/ialloc.c
782
gdp = ext4_get_group_desc(sb, group, &group_desc_bh);
fs/ext4/ialloc.c
807
block_bitmap_bh = ext4_read_block_bitmap(sb, group);
fs/ext4/ialloc.c
818
ext4_lock_group(sb, group);
fs/ext4/ialloc.c
823
ext4_free_clusters_after_init(sb, group, gdp));
fs/ext4/ialloc.c
825
ext4_group_desc_csum_set(sb, group, gdp);
fs/ext4/ialloc.c
827
ext4_unlock_group(sb, group);
fs/ext4/ialloc.c
840
ext4_lock_group(sb, group); /* while we modify the bg desc */
fs/ext4/ialloc.c
857
ext4_lock_group(sb, group);
fs/ext4/ialloc.c
863
ext4_group_desc_csum_set(sb, group, gdp);
fs/ext4/ialloc.c
866
ext4_unlock_group(sb, group);
fs/ext4/ialloc.c
941
ext4_group_t ngroups, group = 0;
fs/ext4/ioctl.c
1957
err = get_user(input.group, &uinput->group);
fs/ext4/ioctl.c
981
err = ext4_register_li_request(sb, input->group);
fs/ext4/mballoc-test.c
235
static void mbt_ctx_mark_used(struct super_block *sb, ext4_group_t group,
fs/ext4/mballoc-test.c
238
struct mbt_grp_ctx *grp_ctx = MBT_GRP_CTX(sb, group);
fs/ext4/mballoc-test.c
243
static void *mbt_ctx_bitmap(struct super_block *sb, ext4_group_t group)
fs/ext4/mballoc-test.c
245
struct mbt_grp_ctx *grp_ctx = MBT_GRP_CTX(sb, group);
fs/ext4/mballoc-test.c
330
ext4_group_t group, ext4_grpblk_t blkoff,
fs/ext4/mballoc-test.c
334
struct mbt_grp_ctx *grp_ctx = MBT_GRP_CTX(sb, group);
fs/ext4/mballoc.c
1012
ext4_group_t group)
fs/ext4/mballoc.c
1018
start = group;
fs/ext4/mballoc.c
1062
ext4_group_t group)
fs/ext4/mballoc.c
1100
start = group;
fs/ext4/mballoc.c
1157
static void next_linear_group(ext4_group_t *group, ext4_group_t ngroups)
fs/ext4/mballoc.c
1163
*group = *group + 1 >= ngroups ? 0 : *group + 1;
fs/ext4/mballoc.c
1173
ext4_group_t group = *start;
fs/ext4/mballoc.c
1175
for (i = 0; i < count; i++, next_linear_group(&group, ngroups)) {
fs/ext4/mballoc.c
1176
ret = ext4_mb_scan_group(ac, group);
fs/ext4/mballoc.c
1182
*start = group;
fs/ext4/mballoc.c
1282
void *buddy, void *bitmap, ext4_group_t group,
fs/ext4/mballoc.c
1314
ext4_grp_locked_error(sb, group, 0, 0,
fs/ext4/mballoc.c
1323
ext4_mark_group_bitmap_corrupted(sb, group,
fs/ext4/mballoc.c
1381
ext4_group_t first_group, group;
fs/ext4/mballoc.c
1412
for (i = 0, group = first_group; i < groups_per_folio; i++, group++) {
fs/ext4/mballoc.c
1413
if (group >= ngroups)
fs/ext4/mballoc.c
1416
grinfo = ext4_get_group_info(sb, group);
fs/ext4/mballoc.c
1430
bh[i] = ext4_read_block_bitmap_nowait(sb, group, false);
fs/ext4/mballoc.c
1436
mb_debug(sb, "read bitmap for group %u\n", group);
fs/ext4/mballoc.c
1440
for (i = 0, group = first_group; i < groups_per_folio; i++, group++) {
fs/ext4/mballoc.c
1445
err2 = ext4_wait_block_bitmap(sb, group, bh[i]);
fs/ext4/mballoc.c
1452
group = (first_block + i) >> 1;
fs/ext4/mballoc.c
1453
if (group >= ngroups)
fs/ext4/mballoc.c
1456
if (!bh[group - first_group])
fs/ext4/mballoc.c
1460
if (!buffer_verified(bh[group - first_group]))
fs/ext4/mballoc.c
1472
bitmap = bh[group - first_group]->b_data;
fs/ext4/mballoc.c
1478
grinfo = ext4_get_group_info(sb, group);
fs/ext4/mballoc.c
1487
group, folio->index, i * blocksize);
fs/ext4/mballoc.c
1488
trace_ext4_mb_buddy_bitmap_load(sb, group);
fs/ext4/mballoc.c
1496
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
1499
ext4_mb_generate_buddy(sb, data, incore, group, grinfo);
fs/ext4/mballoc.c
1500
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
1506
group, folio->index, i * blocksize);
fs/ext4/mballoc.c
1507
trace_ext4_mb_bitmap_load(sb, group);
fs/ext4/mballoc.c
1510
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
1514
ext4_mb_generate_from_pa(sb, data, group);
fs/ext4/mballoc.c
1516
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
1543
ext4_group_t group, struct ext4_buddy *e4b, gfp_t gfp)
fs/ext4/mballoc.c
1557
block = group * 2;
fs/ext4/mballoc.c
1605
int ext4_mb_init_group(struct super_block *sb, ext4_group_t group, gfp_t gfp)
fs/ext4/mballoc.c
1614
mb_debug(sb, "init group %u\n", group);
fs/ext4/mballoc.c
1615
this_grp = ext4_get_group_info(sb, group);
fs/ext4/mballoc.c
1628
ret = ext4_mb_get_buddy_folio_lock(sb, group, &e4b, gfp);
fs/ext4/mballoc.c
1675
ext4_mb_load_buddy_gfp(struct super_block *sb, ext4_group_t group,
fs/ext4/mballoc.c
1687
mb_debug(sb, "load group %u\n", group);
fs/ext4/mballoc.c
1689
grp = ext4_get_group_info(sb, group);
fs/ext4/mballoc.c
1696
e4b->bd_group = group;
fs/ext4/mballoc.c
1705
ret = ext4_mb_init_group(sb, group, gfp);
fs/ext4/mballoc.c
1715
block = group * 2;
fs/ext4/mballoc.c
1831
static int ext4_mb_load_buddy(struct super_block *sb, ext4_group_t group,
fs/ext4/mballoc.c
1834
return ext4_mb_load_buddy_gfp(sb, group, e4b, GFP_NOFS);
fs/ext4/mballoc.c
2404
ext4_group_t group = ex.fe_group;
fs/ext4/mballoc.c
2409
err = ext4_mb_load_buddy(ac->ac_sb, group, e4b);
fs/ext4/mballoc.c
2413
ext4_lock_group(ac->ac_sb, group);
fs/ext4/mballoc.c
2425
ext4_unlock_group(ac->ac_sb, group);
fs/ext4/mballoc.c
2433
ext4_group_t group = ac->ac_g_ex.fe_group;
fs/ext4/mballoc.c
2437
struct ext4_group_info *grp = ext4_get_group_info(ac->ac_sb, group);
fs/ext4/mballoc.c
2447
err = ext4_mb_load_buddy(ac->ac_sb, group, e4b);
fs/ext4/mballoc.c
2455
ext4_lock_group(ac->ac_sb, group);
fs/ext4/mballoc.c
2492
ext4_unlock_group(ac->ac_sb, group);
fs/ext4/mballoc.c
2709
ext4_group_t group, enum criteria cr)
fs/ext4/mballoc.c
2713
struct ext4_group_info *grp = ext4_get_group_info(ac->ac_sb, group);
fs/ext4/mballoc.c
2735
((group % flex_size) == 0))
fs/ext4/mballoc.c
2778
ext4_group_t group, enum criteria cr)
fs/ext4/mballoc.c
2780
struct ext4_group_info *grp = ext4_get_group_info(ac->ac_sb, group);
fs/ext4/mballoc.c
2792
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
2793
__release(ext4_group_lock_ptr(sb, group));
fs/ext4/mballoc.c
2808
__acquire(ext4_group_lock_ptr(sb, group));
fs/ext4/mballoc.c
2809
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
2815
ext4_get_group_desc(sb, group, NULL);
fs/ext4/mballoc.c
2829
((group & ((1 << sbi->s_log_groups_per_flex) - 1)) != 0)) &&
fs/ext4/mballoc.c
2833
ret = ext4_mb_init_group(sb, group, GFP_NOFS);
fs/ext4/mballoc.c
2839
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
2840
__release(ext4_group_lock_ptr(sb, group));
fs/ext4/mballoc.c
2842
ret = ext4_mb_good_group(ac, group, cr);
fs/ext4/mballoc.c
2845
__acquire(ext4_group_lock_ptr(sb, group));
fs/ext4/mballoc.c
2846
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
2855
ext4_group_t ext4_mb_prefetch(struct super_block *sb, ext4_group_t group,
fs/ext4/mballoc.c
2864
struct ext4_group_desc *gdp = ext4_get_group_desc(sb, group,
fs/ext4/mballoc.c
2866
struct ext4_group_info *grp = ext4_get_group_info(sb, group);
fs/ext4/mballoc.c
2878
bh = ext4_read_block_bitmap_nowait(sb, group, true);
fs/ext4/mballoc.c
2885
if (++group >= ngroups)
fs/ext4/mballoc.c
2886
group = 0;
fs/ext4/mballoc.c
2889
return group;
fs/ext4/mballoc.c
2899
ext4_group_t group)
fs/ext4/mballoc.c
2903
if (ac->ac_prefetch_grp != group)
fs/ext4/mballoc.c
2913
nr -= group & (nr - 1);
fs/ext4/mballoc.c
2918
ac->ac_prefetch_grp = ext4_mb_prefetch(ac->ac_sb, group, nr,
fs/ext4/mballoc.c
2935
void ext4_mb_prefetch_fini(struct super_block *sb, ext4_group_t group,
fs/ext4/mballoc.c
2942
if (!group)
fs/ext4/mballoc.c
2943
group = ext4_get_groups_count(sb);
fs/ext4/mballoc.c
2944
group--;
fs/ext4/mballoc.c
2945
gdp = ext4_get_group_desc(sb, group, NULL);
fs/ext4/mballoc.c
2946
grp = ext4_get_group_info(sb, group);
fs/ext4/mballoc.c
2950
if (ext4_mb_init_group(sb, group, GFP_NOFS))
fs/ext4/mballoc.c
2957
ext4_group_t group)
fs/ext4/mballoc.c
2963
ext4_mb_might_prefetch(ac, group);
fs/ext4/mballoc.c
2966
if (cr < CR_ANY_FREE && spin_is_locked(ext4_group_lock_ptr(sb, group)))
fs/ext4/mballoc.c
2970
ret = ext4_mb_good_group_nolock(ac, group, cr);
fs/ext4/mballoc.c
2977
ret = ext4_mb_load_buddy(sb, group, ac->ac_e4b);
fs/ext4/mballoc.c
2983
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
2984
else if (!ext4_try_lock_group(sb, group))
fs/ext4/mballoc.c
2988
if (unlikely(!ext4_mb_good_group(ac, group, cr)))
fs/ext4/mballoc.c
2994
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
3123
ext4_group_t group;
fs/ext4/mballoc.c
3127
group = *pos + 1;
fs/ext4/mballoc.c
3128
return (void *) ((unsigned long) group);
fs/ext4/mballoc.c
3134
ext4_group_t group;
fs/ext4/mballoc.c
3139
group = *pos + 1;
fs/ext4/mballoc.c
3140
return (void *) ((unsigned long) group);
fs/ext4/mballoc.c
3146
ext4_group_t group = (ext4_group_t) ((unsigned long) v);
fs/ext4/mballoc.c
3157
group--;
fs/ext4/mballoc.c
3158
if (group == 0)
fs/ext4/mballoc.c
3166
grinfo = ext4_get_group_info(sb, group);
fs/ext4/mballoc.c
3171
err = ext4_mb_load_buddy(sb, group, &e4b);
fs/ext4/mballoc.c
3173
seq_printf(seq, "#%-5u: %s\n", group, ext4_decode_error(NULL, err, nbuf));
fs/ext4/mballoc.c
3184
seq_printf(seq, "#%-5u: %-5u %-5u %-5u [", group, sg->bb_free,
fs/ext4/mballoc.c
3427
int ext4_mb_add_groupinfo(struct super_block *sb, ext4_group_t group,
fs/ext4/mballoc.c
3432
int idx = group >> EXT4_DESC_PER_BLOCK_BITS(sb);
fs/ext4/mballoc.c
3442
if (group % EXT4_DESC_PER_BLOCK(sb) == 0) {
fs/ext4/mballoc.c
3457
i = group & (EXT4_DESC_PER_BLOCK(sb) - 1);
fs/ext4/mballoc.c
3474
ext4_free_clusters_after_init(sb, group, desc);
fs/ext4/mballoc.c
3485
meta_group_info[i]->bb_group = group;
fs/ext4/mballoc.c
3487
mb_group_bb_bitmap_alloc(sb, meta_group_info[i], group);
fs/ext4/mballoc.c
3492
if (group % EXT4_DESC_PER_BLOCK(sb) == 0) {
fs/ext4/mballoc.c
4089
ext4_group_t group, ext4_grpblk_t blkoff,
fs/ext4/mballoc.c
4100
handle, sb, state, group, blkoff, len,
fs/ext4/mballoc.c
4105
bitmap_bh = ext4_read_block_bitmap(sb, group);
fs/ext4/mballoc.c
4118
gdp = ext4_get_group_desc(sb, group, &gdp_bh);
fs/ext4/mballoc.c
4130
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
4135
ext4_free_clusters_after_init(sb, group, gdp));
fs/ext4/mballoc.c
4158
ext4_group_desc_csum_set(sb, group, gdp);
fs/ext4/mballoc.c
4159
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
4164
ext4_group_t flex_group = ext4_flex_group(sbi, group);
fs/ext4/mballoc.c
425
ext4_group_t group);
fs/ext4/mballoc.c
4263
ext4_group_t group;
fs/ext4/mballoc.c
4269
ext4_get_group_no_and_offset(sb, block, &group, &blkoff);
fs/ext4/mballoc.c
429
ext4_group_t group);
fs/ext4/mballoc.c
4290
group, blkoff, clen,
fs/ext4/mballoc.c
5088
ext4_group_t group)
fs/ext4/mballoc.c
5090
struct ext4_group_info *grp = ext4_get_group_info(sb, group);
fs/ext4/mballoc.c
5118
BUG_ON(groupnr != group);
fs/ext4/mballoc.c
5122
mb_debug(sb, "preallocated %d for group %u\n", preallocated, group);
fs/ext4/mballoc.c
5430
ext4_group_t group;
fs/ext4/mballoc.c
5436
ext4_get_group_no_and_offset(sb, pa->pa_pstart, &group, &bit);
fs/ext4/mballoc.c
5438
BUG_ON(group != e4b->bd_group && pa->pa_len != 0);
fs/ext4/mballoc.c
5447
(unsigned) ext4_group_first_block_no(sb, group) + bit,
fs/ext4/mballoc.c
5448
(unsigned) next - bit, (unsigned) group);
fs/ext4/mballoc.c
5451
trace_ext4_mballoc_discard(sb, NULL, group, bit, next - bit);
fs/ext4/mballoc.c
5464
ext4_grp_locked_error(sb, group, 0, 0, "free %u, pa_free %u",
fs/ext4/mballoc.c
5479
ext4_group_t group;
fs/ext4/mballoc.c
5484
ext4_get_group_no_and_offset(sb, pa->pa_pstart, &group, &bit);
fs/ext4/mballoc.c
5485
if (unlikely(group != e4b->bd_group && pa->pa_len != 0)) {
fs/ext4/mballoc.c
5487
e4b->bd_group, group, pa->pa_pstart);
fs/ext4/mballoc.c
5492
trace_ext4_mballoc_discard(sb, NULL, group, bit, pa->pa_len);
fs/ext4/mballoc.c
5506
ext4_group_t group, int *busy)
fs/ext4/mballoc.c
5508
struct ext4_group_info *grp = ext4_get_group_info(sb, group);
fs/ext4/mballoc.c
5519
mb_debug(sb, "discard preallocation for group %u\n", group);
fs/ext4/mballoc.c
5523
bitmap_bh = ext4_read_block_bitmap(sb, group);
fs/ext4/mballoc.c
5528
err, group);
fs/ext4/mballoc.c
5532
err = ext4_mb_load_buddy(sb, group, &e4b);
fs/ext4/mballoc.c
5535
err, group);
fs/ext4/mballoc.c
5540
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
5595
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
5600
free, group, grp->bb_free);
fs/ext4/mballoc.c
5619
ext4_group_t group = 0;
fs/ext4/mballoc.c
5689
group = ext4_get_group_number(sb, pa->pa_pstart);
fs/ext4/mballoc.c
5691
err = ext4_mb_load_buddy_gfp(sb, group, &e4b,
fs/ext4/mballoc.c
5695
err, group);
fs/ext4/mballoc.c
5699
bitmap_bh = ext4_read_block_bitmap(sb, group);
fs/ext4/mballoc.c
5703
err, group);
fs/ext4/mballoc.c
5708
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
5711
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
5895
ext4_group_t group;
fs/ext4/mballoc.c
5912
ext4_get_group_no_and_offset(sb, goal, &group, &block);
fs/ext4/mballoc.c
5920
ac->ac_o_ex.fe_group = group;
fs/ext4/mballoc.c
5945
ext4_group_t group = 0;
fs/ext4/mballoc.c
5996
group = ext4_get_group_number(sb, pa->pa_pstart);
fs/ext4/mballoc.c
5997
err = ext4_mb_load_buddy_gfp(sb, group, &e4b,
fs/ext4/mballoc.c
6001
err, group);
fs/ext4/mballoc.c
6004
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
6007
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
6173
ext4_group_t group, nr;
fs/ext4/mballoc.c
6186
ext4_get_group_no_and_offset(sb, goal, &group, &blkoff);
fs/ext4/mballoc.c
6188
bitmap_bh = ext4_read_block_bitmap(sb, group);
fs/ext4/mballoc.c
6201
ext4_group_first_block_no(sb, group) +
fs/ext4/mballoc.c
621
struct ext4_group_info *grp, ext4_group_t group)
fs/ext4/mballoc.c
6211
if (++group >= ext4_get_groups_count(sb))
fs/ext4/mballoc.c
6212
group = 0;
fs/ext4/mballoc.c
6222
block = ext4_group_first_block_no(sb, group) + EXT4_C2B(sbi, i);
fs/ext4/mballoc.c
629
bh = ext4_read_block_bitmap(sb, group);
fs/ext4/mballoc.c
6442
ext4_group_t group = e4b->bd_group;
fs/ext4/mballoc.c
6477
ext4_grp_locked_error(sb, group, 0,
fs/ext4/mballoc.c
6478
ext4_group_first_block_no(sb, group) +
fs/ext4/mballoc.c
6517
ext4_group_t group;
fs/ext4/mballoc.c
6520
ext4_get_group_no_and_offset(sb, block, &group, &blkoff);
fs/ext4/mballoc.c
6521
ext4_mb_mark_context(NULL, sb, false, group, blkoff, count,
fs/ext4/mballoc.c
662
struct ext4_group_info *grp, ext4_group_t group)
fs/ext4/mballoc.c
6881
ext4_group_t group = e4b->bd_group;
fs/ext4/mballoc.c
6884
trace_ext4_trim_extent(sb, group, start, count);
fs/ext4/mballoc.c
6886
assert_spin_locked(ext4_group_lock_ptr(sb, group));
fs/ext4/mballoc.c
6889
ex.fe_group = group;
fs/ext4/mballoc.c
6897
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
6898
ret = ext4_issue_discard(sb, group, start, count);
fs/ext4/mballoc.c
6899
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
6998
ext4_trim_all_free(struct super_block *sb, ext4_group_t group,
fs/ext4/mballoc.c
7005
trace_ext4_trim_all_free(sb, group, start, max);
fs/ext4/mballoc.c
7007
ret = ext4_mb_load_buddy(sb, group, &e4b);
fs/ext4/mballoc.c
7010
ret, group);
fs/ext4/mballoc.c
7014
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
7022
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
7026
ret, group);
fs/ext4/mballoc.c
7047
ext4_group_t group, first_group, last_group;
fs/ext4/mballoc.c
7087
for (group = first_group; group <= last_group; group++) {
fs/ext4/mballoc.c
7090
grp = ext4_get_group_info(sb, group);
fs/ext4/mballoc.c
7095
ret = ext4_mb_init_group(sb, group, GFP_NOFS);
fs/ext4/mballoc.c
7106
if (group == last_group)
fs/ext4/mballoc.c
7109
cnt = ext4_trim_all_free(sb, group, first_cluster,
fs/ext4/mballoc.c
7137
ext4_group_t group,
fs/ext4/mballoc.c
7149
error = ext4_mb_load_buddy(sb, group, &e4b);
fs/ext4/mballoc.c
7154
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
7162
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
7163
error = meta_formatter(sb, group, first, start - first,
fs/ext4/mballoc.c
7167
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
7175
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
7176
error = formatter(sb, group, start, next - start, priv);
fs/ext4/mballoc.c
7179
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
7184
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
7244
void *bitmap, ext4_group_t group,
fs/ext4/mballoc.c
7247
ext4_mb_generate_buddy(sb, buddy, bitmap, group, grp);
fs/ext4/mballoc.c
7251
int ext4_mb_load_buddy_test(struct super_block *sb, ext4_group_t group,
fs/ext4/mballoc.c
7254
return ext4_mb_load_buddy(sb, group, e4b);
fs/ext4/mballoc.c
918
unsigned long group = start;
fs/ext4/mballoc.c
924
xa_for_each_range(xa, group, grp, start, end - 1) {
fs/ext4/mballoc.c
961
ext4_group_t group)
fs/ext4/mballoc.c
968
start = group;
fs/ext4/mballoc.h
275
ext4_group_t group, ext4_grpblk_t blkoff,
fs/ext4/mballoc.h
293
void *buddy, void *bitmap, ext4_group_t group,
fs/ext4/mballoc.h
296
ext4_group_t group, struct ext4_buddy *e4b);
fs/ext4/resize.c
1019
ext4_group_t group)
fs/ext4/resize.c
1066
gdbackups = verify_reserved_gdb(sb, group, primary[res]);
fs/ext4/resize.c
108
ext4_group_t group) {
fs/ext4/resize.c
1090
blk = group * EXT4_BLOCKS_PER_GROUP(sb);
fs/ext4/resize.c
110
overhead = ext4_bg_num_gdb(sb, group);
fs/ext4/resize.c
111
if (ext4_bg_has_super(sb, group))
fs/ext4/resize.c
1115
ext4_group_t group)
fs/ext4/resize.c
1119
es->s_block_group_nr = cpu_to_le16(group);
fs/ext4/resize.c
1149
ext4_group_t group = 0;
fs/ext4/resize.c
1156
group = 1;
fs/ext4/resize.c
1162
group = ext4_list_backups(sb, &three, &five, &seven);
fs/ext4/resize.c
1165
group = ext4_get_group_number(sb, blk_off) + 1;
fs/ext4/resize.c
1166
last = (ext4_group_t)(group + EXT4_DESC_PER_BLOCK(sb) - 2);
fs/ext4/resize.c
1169
while (group < sbi->s_groups_count) {
fs/ext4/resize.c
1172
int has_super = ext4_bg_has_super(sb, group);
fs/ext4/resize.c
1173
ext4_fsblk_t first_block = ext4_group_first_block_no(sb, group);
fs/ext4/resize.c
1181
backup_block = ((ext4_fsblk_t)group) * bpg + blk_off;
fs/ext4/resize.c
1192
ext4_group_first_block_no(sb, group));
fs/ext4/resize.c
1204
ext4_set_block_group_nr(sb, bh->b_data, group);
fs/ext4/resize.c
1213
group = ext4_list_backups(sb, &three, &five, &seven);
fs/ext4/resize.c
1214
else if (group == last)
fs/ext4/resize.c
1217
group = last;
fs/ext4/resize.c
1235
"forcing fsck on next reboot", group, err);
fs/ext4/resize.c
1253
ext4_group_t group, struct inode *resize_inode,
fs/ext4/resize.c
1263
for (i = 0; i < count; i++, group++) {
fs/ext4/resize.c
1264
int reserved_gdb = ext4_bg_has_super(sb, group) ?
fs/ext4/resize.c
1267
gdb_off = group % EXT4_DESC_PER_BLOCK(sb);
fs/ext4/resize.c
1268
gdb_num = group / EXT4_DESC_PER_BLOCK(sb);
fs/ext4/resize.c
127
ext4_group_t group = input->group;
fs/ext4/resize.c
1283
if (!err && reserved_gdb && ext4_bg_num_gdb(sb, group))
fs/ext4/resize.c
1284
err = reserve_backup_gdb(handle, resize_inode, group);
fs/ext4/resize.c
1286
err = add_new_gdb_meta_bg(sb, handle, group);
fs/ext4/resize.c
1288
err = add_new_gdb(handle, resize_inode, group);
fs/ext4/resize.c
1345
ext4_group_t group;
fs/ext4/resize.c
135
if (group != sbi->s_groups_count) {
fs/ext4/resize.c
1351
group = group_data->group;
fs/ext4/resize.c
1353
gdb_off = group % EXT4_DESC_PER_BLOCK(sb);
fs/ext4/resize.c
1354
gdb_num = group / EXT4_DESC_PER_BLOCK(sb);
fs/ext4/resize.c
137
input->group, sbi->s_groups_count);
fs/ext4/resize.c
1381
ext4_group_desc_csum_set(sb, group, gdp);
fs/ext4/resize.c
1393
err = ext4_mb_add_groupinfo(sb, group, gdp);
fs/ext4/resize.c
141
overhead = ext4_group_overhead_blocks(sb, group);
fs/ext4/resize.c
150
ext4_bg_has_super(sb, input->group) ? "normal" :
fs/ext4/resize.c
1501
flex_group = ext4_flex_group(sbi, group_data[0].group);
fs/ext4/resize.c
151
"no-super", input->group, input->blocks_count,
fs/ext4/resize.c
1543
ext4_group_t group;
fs/ext4/resize.c
1552
ext4_get_group_no_and_offset(sb, o_blocks_count, &group, &last);
fs/ext4/resize.c
1581
group = flex_gd->groups[0].group;
fs/ext4/resize.c
1582
BUG_ON(group != sbi->s_groups_count);
fs/ext4/resize.c
1583
err = ext4_add_new_descs(handle, sb, group,
fs/ext4/resize.c
1602
int gdb_num = group / EXT4_DESC_PER_BLOCK(sb);
fs/ext4/resize.c
1603
int gdb_num_end = ((group + flex_gd->count - 1) /
fs/ext4/resize.c
1634
ext4_group_t group;
fs/ext4/resize.c
1647
ext4_get_group_no_and_offset(sb, o_blocks_count, &group, &last);
fs/ext4/resize.c
1651
last_group = group | (flex_gd->resize_bg - 1);
fs/ext4/resize.c
1655
flex_gd->count = last_group - group + 1;
fs/ext4/resize.c
1660
group_data[i].group = group + i;
fs/ext4/resize.c
1662
overhead = ext4_group_overhead_blocks(sb, group + i);
fs/ext4/resize.c
1705
int reserved_gdb = ext4_bg_has_super(sb, input->group) ?
fs/ext4/resize.c
1712
gdb_off = input->group % EXT4_DESC_PER_BLOCK(sb);
fs/ext4/resize.c
1750
err = ext4_alloc_flex_bg_array(sb, input->group + 1);
fs/ext4/resize.c
1754
err = ext4_mb_alloc_groupinfo(sb, input->group + 1);
fs/ext4/resize.c
1841
ext4_group_t group;
fs/ext4/resize.c
1866
ext4_get_group_no_and_offset(sb, o_blocks_count, &group, &last);
fs/ext4/resize.c
314
ext4_group_t group;
fs/ext4/resize.c
322
src_group = group_data[0].group;
fs/ext4/resize.c
328
group = group_data[0].group;
fs/ext4/resize.c
329
if (src_group >= group_data[0].group + flex_gd->count)
fs/ext4/resize.c
332
last_blk = start_blk + group_data[src_group - group].blocks_count;
fs/ext4/resize.c
343
last_blk += group_data[src_group - group].blocks_count;
fs/ext4/resize.c
353
group = ext4_get_group_number(sb, start_blk - 1);
fs/ext4/resize.c
354
group -= group_data[0].group;
fs/ext4/resize.c
355
group_data[group].mdata_blocks++;
fs/ext4/resize.c
356
flex_gd->bg_flags[group] &= uninit_mask;
fs/ext4/resize.c
364
group = ext4_get_group_number(sb, start_blk - 1);
fs/ext4/resize.c
365
group -= group_data[0].group;
fs/ext4/resize.c
366
group_data[group].mdata_blocks++;
fs/ext4/resize.c
367
flex_gd->bg_flags[group] &= uninit_mask;
fs/ext4/resize.c
378
group = ext4_get_group_number(sb, start_blk);
fs/ext4/resize.c
379
next_group_start = ext4_group_first_block_no(sb, group + 1);
fs/ext4/resize.c
380
group -= group_data[0].group;
fs/ext4/resize.c
383
flex_gd->bg_flags[group + 1] &= uninit_mask;
fs/ext4/resize.c
385
group_data[group + 1].mdata_blocks += overhead;
fs/ext4/resize.c
389
group_data[group].mdata_blocks += itb;
fs/ext4/resize.c
390
flex_gd->bg_flags[group] &= uninit_mask;
fs/ext4/resize.c
403
group = group_data[0].group;
fs/ext4/resize.c
412
ext4_bg_has_super(sb, group + i) ? "normal" :
fs/ext4/resize.c
413
"no-super", group + i,
fs/ext4/resize.c
472
ext4_group_t group;
fs/ext4/resize.c
475
group = ext4_get_group_number(sb, EXT4_C2B(sbi, first_cluster));
fs/ext4/resize.c
476
start = EXT4_B2C(sbi, ext4_group_first_block_no(sb, group));
fs/ext4/resize.c
477
group -= flex_gd->groups[0].group;
fs/ext4/resize.c
483
if (flex_gd->bg_flags[group] & EXT4_BG_BLOCK_UNINIT) {
fs/ext4/resize.c
492
bh = sb_getblk(sb, flex_gd->groups[group].block_bitmap);
fs/ext4/resize.c
541
ext4_group_t group, count;
fs/ext4/resize.c
547
group_data[0].group != sbi->s_groups_count);
fs/ext4/resize.c
557
group = group_data[0].group;
fs/ext4/resize.c
558
for (i = 0; i < flex_gd->count; i++, group++) {
fs/ext4/resize.c
562
gdblocks = ext4_bg_num_gdb(sb, group);
fs/ext4/resize.c
563
start = ext4_group_first_block_no(sb, group);
fs/ext4/resize.c
565
if (meta_bg == 0 && !ext4_bg_has_super(sb, group))
fs/ext4/resize.c
571
block = start + ext4_bg_has_super(sb, group);
fs/ext4/resize.c
609
if (ext4_bg_has_super(sb, group)) {
fs/ext4/resize.c
645
overhead = ext4_group_overhead_blocks(sb, group);
fs/ext4/resize.c
820
ext4_group_t group)
fs/ext4/resize.c
824
unsigned long gdb_num = group / EXT4_DESC_PER_BLOCK(sb);
fs/ext4/resize.c
843
gdbackups = verify_reserved_gdb(sb, group, gdb_bh);
fs/ext4/resize.c
860
group, gdblock);
fs/ext4/resize.c
962
handle_t *handle, ext4_group_t group) {
fs/ext4/resize.c
966
unsigned long gdb_num = group / EXT4_DESC_PER_BLOCK(sb);
fs/ext4/resize.c
969
gdblock = ext4_group_first_block_no(sb, group) +
fs/ext4/resize.c
970
ext4_bg_has_super(sb, group);
fs/ext4/super.c
1101
ext4_group_t group,
fs/ext4/super.c
1105
struct ext4_group_info *grp = ext4_get_group_info(sb, group);
fs/ext4/super.c
1106
struct ext4_group_desc *gdp = ext4_get_group_desc(sb, group, NULL);
fs/ext4/super.c
3729
ext4_group_t group = elr->lr_next_group;
fs/ext4/super.c
3736
elr->lr_next_group = ext4_mb_prefetch(sb, group, nr, &prefetch_ios);
fs/ext4/super.c
3738
trace_ext4_prefetch_bitmaps(sb, group, elr->lr_next_group, nr);
fs/ext4/super.c
3739
if (group >= elr->lr_next_group) {
fs/ext4/super.c
3752
for (; group < ngroups; group++) {
fs/ext4/super.c
3753
gdp = ext4_get_group_desc(sb, group, NULL);
fs/ext4/super.c
3763
if (group >= ngroups)
fs/ext4/super.c
3768
ret = ext4_init_inode_table(sb, group,
fs/ext4/super.c
3770
trace_ext4_lazy_itable_init(sb, group);
fs/ext4/super.c
3776
elr->lr_next_group = group + 1;
fs/ext4/super.c
3967
ext4_group_t group, ngroups = EXT4_SB(sb)->s_groups_count;
fs/ext4/super.c
3973
for (group = 0; group < ngroups; group++) {
fs/ext4/super.c
3974
gdp = ext4_get_group_desc(sb, group, NULL);
fs/ext4/super.c
3982
return group;
fs/hfsplus/catalog.c
95
perms->group = cpu_to_be32(i_gid_read(inode));
fs/hfsplus/inode.c
211
i_gid_write(inode, be32_to_cpu(perms->group));
fs/init.c
110
error = chown_common(&path, user, group);
fs/init.c
99
int __init init_chown(const char *filename, uid_t user, gid_t group, int flags)
fs/internal.h
204
int do_fchownat(int dfd, const char __user *filename, uid_t user, gid_t group,
fs/internal.h
206
int chown_common(const struct path *path, uid_t user, gid_t group);
fs/netfs/buffered_write.c
154
struct netfs_group *group;
fs/netfs/buffered_write.c
213
group = netfs_folio_group(folio);
fs/netfs/buffered_write.c
215
if (unlikely(group != netfs_group) &&
fs/netfs/buffered_write.c
216
group != NETFS_FOLIO_COPY_TO_CACHE)
fs/netfs/buffered_write.c
507
struct netfs_group *group;
fs/netfs/buffered_write.c
533
group = netfs_folio_group(folio);
fs/netfs/buffered_write.c
534
if (group != netfs_group && group != NETFS_FOLIO_COPY_TO_CACHE) {
fs/netfs/write_collect.c
338
netfs_put_group_many(wreq->group, wreq->nr_group_rel);
fs/netfs/write_collect.c
62
struct netfs_group *group = NULL;
fs/netfs/write_collect.c
76
group = finfo->netfs_group;
fs/netfs/write_collect.c
83
if ((group = netfs_folio_group(folio))) {
fs/netfs/write_collect.c
84
if (group == NETFS_FOLIO_COPY_TO_CACHE) {
fs/netfs/write_issue.c
351
netfs_put_group_many(wreq->group, wreq->nr_group_rel);
fs/netfs/write_issue.c
401
} else if (fgroup != wreq->group) {
fs/netfs/write_issue.c
51
struct netfs_group *group = NULL;
fs/netfs/write_issue.c
60
group = finfo->netfs_group;
fs/netfs/write_issue.c
63
group = priv;
fs/netfs/write_issue.c
64
if (group == NETFS_FOLIO_COPY_TO_CACHE) {
fs/netfs/write_issue.c
67
group = NULL;
fs/netfs/write_issue.c
81
netfs_put_group(group);
fs/nfs/nfs4idmap.c
125
struct nfs4_string *group = fattr->group_name;
fs/nfs/nfs4idmap.c
130
if (nfs_map_group_to_gid(server, group->data, group->len, &gid) == 0) {
fs/nfsd/nfs4acl.c
182
unsigned short group;
fs/nfsd/nfs4acl.c
207
pas->group = pa->e_perm;
fs/nfsd/nfs4acl.c
225
pas->group &= pas->mask;
fs/nfsd/nfs4acl.c
253
deny &= pas.users | pas.group | pas.groups | pas.other;
fs/nfsd/nfs4acl.c
273
deny &= pas.groups | pas.group | pas.other;
fs/nfsd/nfs4acl.c
303
ace->access_mask = mask_from_posix(pas.group, flags);
fs/nfsd/nfs4acl.c
325
deny = ~pas.group & pas.other;
fs/nfsd/nfs4acl.c
453
struct posix_ace_state group;
fs/nfsd/nfs4acl.c
540
low_mode_from_nfs4(state->group.allow, &pace->e_perm, flags);
fs/nfsd/nfs4acl.c
541
add_to_mask(state, &state->group);
fs/nfsd/nfs4acl.c
656
allow_bits(&state->group, mask);
fs/nfsd/nfs4acl.c
658
deny_bits(&state->group, mask);
fs/nfsd/nfs4acl.c
659
mask = state->group.deny;
fs/nfsd/nfs4acl.c
674
deny_bits(&state->group, mask);
fs/nfsd/nfs4acl.c
683
allow_bits(&state->group, mask);
fs/nfsd/nfs4acl.c
690
deny_bits(&state->group, mask);
fs/nfsd/nfs4acl.c
755
default_acl_state.group = effective_acl_state.group;
fs/nilfs2/alloc.c
109
nilfs_palloc_desc_blkoff(const struct inode *inode, unsigned long group)
fs/nilfs2/alloc.c
112
group / nilfs_palloc_groups_per_desc_block(inode);
fs/nilfs2/alloc.c
128
nilfs_palloc_bitmap_blkoff(const struct inode *inode, unsigned long group)
fs/nilfs2/alloc.c
131
group % nilfs_palloc_groups_per_desc_block(inode);
fs/nilfs2/alloc.c
132
return nilfs_palloc_desc_blkoff(inode, group) + 1 +
fs/nilfs2/alloc.c
188
unsigned long group, group_offset;
fs/nilfs2/alloc.c
190
group = nilfs_palloc_group(inode, nr, &group_offset);
fs/nilfs2/alloc.c
192
return nilfs_palloc_bitmap_blkoff(inode, group) + 1 +
fs/nilfs2/alloc.c
291
unsigned long group,
fs/nilfs2/alloc.c
297
nilfs_palloc_desc_blkoff(inode, group),
fs/nilfs2/alloc.c
312
unsigned long group,
fs/nilfs2/alloc.c
318
nilfs_palloc_bitmap_blkoff(inode, group),
fs/nilfs2/alloc.c
331
unsigned long group)
fs/nilfs2/alloc.c
337
group),
fs/nilfs2/alloc.c
387
unsigned long group,
fs/nilfs2/alloc.c
392
(group % nilfs_palloc_groups_per_desc_block(inode));
fs/nilfs2/alloc.c
588
unsigned long group, maxgroup, ngroups;
fs/nilfs2/alloc.c
597
group = nilfs_palloc_group(inode, req->pr_entry_nr, &group_offset);
fs/nilfs2/alloc.c
601
if (group >= ngroups && wrap) {
fs/nilfs2/alloc.c
603
group = 0;
fs/nilfs2/alloc.c
607
ret = nilfs_palloc_get_desc_block(inode, group, 1, &desc_bh);
fs/nilfs2/alloc.c
611
doff = nilfs_palloc_group_desc_offset(inode, group, desc_bh);
fs/nilfs2/alloc.c
613
n = nilfs_palloc_rest_groups_in_desc_block(inode, group,
fs/nilfs2/alloc.c
615
for (j = 0; j < n; j++, group++, group_offset = 0) {
fs/nilfs2/alloc.c
616
lock = nilfs_mdt_bgl_lock(inode, group);
fs/nilfs2/alloc.c
621
ret = nilfs_palloc_get_bitmap_block(inode, group, 1,
fs/nilfs2/alloc.c
662
req->pr_entry_nr = entries_per_group * group + pos;
fs/nilfs2/alloc.c
694
unsigned long group, group_offset;
fs/nilfs2/alloc.c
700
group = nilfs_palloc_group(inode, req->pr_entry_nr, &group_offset);
fs/nilfs2/alloc.c
701
doff = nilfs_palloc_group_desc_offset(inode, group, req->pr_desc_bh);
fs/nilfs2/alloc.c
706
lock = nilfs_mdt_bgl_lock(inode, group);
fs/nilfs2/alloc.c
738
unsigned long group, group_offset;
fs/nilfs2/alloc.c
741
group = nilfs_palloc_group(inode, req->pr_entry_nr, &group_offset);
fs/nilfs2/alloc.c
742
doff = nilfs_palloc_group_desc_offset(inode, group, req->pr_desc_bh);
fs/nilfs2/alloc.c
747
lock = nilfs_mdt_bgl_lock(inode, group);
fs/nilfs2/alloc.c
779
unsigned long group, group_offset;
fs/nilfs2/alloc.c
782
group = nilfs_palloc_group(inode, req->pr_entry_nr, &group_offset);
fs/nilfs2/alloc.c
783
ret = nilfs_palloc_get_desc_block(inode, group, 1, &desc_bh);
fs/nilfs2/alloc.c
786
ret = nilfs_palloc_get_bitmap_block(inode, group, 1, &bitmap_bh);
fs/nilfs2/alloc.c
827
unsigned long group, group_offset;
fs/nilfs2/alloc.c
840
group = nilfs_palloc_group(inode, entry_nrs[i], &group_offset);
fs/nilfs2/alloc.c
841
ret = nilfs_palloc_get_desc_block(inode, group, 0, &desc_bh);
fs/nilfs2/alloc.c
844
ret = nilfs_palloc_get_bitmap_block(inode, group, 0,
fs/nilfs2/alloc.c
852
group_min_nr = (__u64)group * epg;
fs/nilfs2/alloc.c
856
lock = nilfs_mdt_bgl_lock(inode, group);
fs/nilfs2/alloc.c
914
doff = nilfs_palloc_group_desc_offset(inode, group, desc_bh);
fs/nilfs2/alloc.c
923
ret = nilfs_palloc_delete_bitmap_block(inode, group);
fs/nilfs2/alloc.c
927
ret, group, inode->i_ino);
fs/nilfs2/alloc.c
94
__u64 group = nr;
fs/nilfs2/alloc.c
96
*offset = do_div(group, nilfs_palloc_entries_per_group(inode));
fs/nilfs2/alloc.c
97
return group;
fs/nilfs2/bmap.c
458
unsigned long group = bmap->b_inode->i_ino / entries_per_group;
fs/nilfs2/bmap.c
460
return group * entries_per_group +
fs/notify/fanotify/fanotify.c
1002
fsnotify_destroy_event(group, fsn_event);
fs/notify/fanotify/fanotify.c
1006
ret = fanotify_get_response(group, FANOTIFY_PERM(event),
fs/notify/fanotify/fanotify.c
1016
static void fanotify_free_group_priv(struct fsnotify_group *group)
fs/notify/fanotify/fanotify.c
1018
put_user_ns(group->user_ns);
fs/notify/fanotify/fanotify.c
1019
kfree(group->fanotify_data.merge_hash);
fs/notify/fanotify/fanotify.c
1020
if (group->fanotify_data.ucounts)
fs/notify/fanotify/fanotify.c
1021
dec_ucount(group->fanotify_data.ucounts,
fs/notify/fanotify/fanotify.c
1024
if (mempool_initialized(&group->fanotify_data.error_events_pool))
fs/notify/fanotify/fanotify.c
1025
mempool_exit(&group->fanotify_data.error_events_pool);
fs/notify/fanotify/fanotify.c
1054
static void fanotify_free_error_event(struct fsnotify_group *group,
fs/notify/fanotify/fanotify.c
1059
mempool_free(fee, &group->fanotify_data.error_events_pool);
fs/notify/fanotify/fanotify.c
1067
static void fanotify_free_event(struct fsnotify_group *group,
fs/notify/fanotify/fanotify.c
1091
fanotify_free_error_event(group, event);
fs/notify/fanotify/fanotify.c
1102
struct fsnotify_group *group)
fs/notify/fanotify/fanotify.c
1104
if (!FAN_GROUP_FLAG(group, FAN_UNLIMITED_MARKS))
fs/notify/fanotify/fanotify.c
1105
dec_ucount(group->fanotify_data.ucounts, UCOUNT_FANOTIFY_MARKS);
fs/notify/fanotify/fanotify.c
182
static int fanotify_merge(struct fsnotify_group *group,
fs/notify/fanotify/fanotify.c
186
unsigned int bucket = fanotify_event_hash_bucket(group, new);
fs/notify/fanotify/fanotify.c
187
struct hlist_head *hlist = &group->fanotify_data.merge_hash[bucket];
fs/notify/fanotify/fanotify.c
191
group, event, bucket);
fs/notify/fanotify/fanotify.c
224
static int fanotify_get_response(struct fsnotify_group *group,
fs/notify/fanotify/fanotify.c
230
pr_debug("%s: group=%p event=%p\n", __func__, group, event);
fs/notify/fanotify/fanotify.c
232
ret = wait_event_state(group->fanotify_data.access_waitq,
fs/notify/fanotify/fanotify.c
238
spin_lock(&group->notification_lock);
fs/notify/fanotify/fanotify.c
243
spin_unlock(&group->notification_lock);
fs/notify/fanotify/fanotify.c
248
fsnotify_remove_queued_event(group, &event->fae.fse);
fs/notify/fanotify/fanotify.c
257
spin_unlock(&group->notification_lock);
fs/notify/fanotify/fanotify.c
286
group, event, ret);
fs/notify/fanotify/fanotify.c
288
fsnotify_destroy_event(group, &event->fae.fse);
fs/notify/fanotify/fanotify.c
299
static u32 fanotify_group_event_mask(struct fsnotify_group *group,
fs/notify/fanotify/fanotify.c
309
unsigned int fid_mode = FAN_GROUP_FLAG(group, FANOTIFY_FID_BITS);
fs/notify/fanotify/fanotify.c
317
if (FAN_GROUP_FLAG(group, FAN_REPORT_MNT)) {
fs/notify/fanotify/fanotify.c
709
struct fsnotify_group *group,
fs/notify/fanotify/fanotify.c
723
fee = mempool_alloc(&group->fanotify_data.error_events_pool, GFP_NOFS);
fs/notify/fanotify/fanotify.c
747
struct fsnotify_group *group,
fs/notify/fanotify/fanotify.c
754
unsigned int fid_mode = FAN_GROUP_FLAG(group, FANOTIFY_FID_BITS);
fs/notify/fanotify/fanotify.c
837
if (group->max_events == UINT_MAX)
fs/notify/fanotify/fanotify.c
843
old_memcg = set_active_memcg(group->memcg);
fs/notify/fanotify/fanotify.c
848
event = fanotify_alloc_error_event(group, fsid, data,
fs/notify/fanotify/fanotify.c
866
if (FAN_GROUP_FLAG(group, FAN_REPORT_TID))
fs/notify/fanotify/fanotify.c
907
static void fanotify_insert_event(struct fsnotify_group *group,
fs/notify/fanotify/fanotify.c
911
unsigned int bucket = fanotify_event_hash_bucket(group, event);
fs/notify/fanotify/fanotify.c
912
struct hlist_head *hlist = &group->fanotify_data.merge_hash[bucket];
fs/notify/fanotify/fanotify.c
914
assert_spin_locked(&group->notification_lock);
fs/notify/fanotify/fanotify.c
920
group, event, bucket);
fs/notify/fanotify/fanotify.c
925
static int fanotify_handle_event(struct fsnotify_group *group, u32 mask,
fs/notify/fanotify/fanotify.c
962
mask = fanotify_group_event_mask(group, iter_info, &match_mask,
fs/notify/fanotify/fanotify.c
968
group, mask, match_mask);
fs/notify/fanotify/fanotify.c
979
if (FAN_GROUP_FLAG(group, FANOTIFY_FID_BITS))
fs/notify/fanotify/fanotify.c
982
event = fanotify_alloc_event(group, mask, data, data_type, dir,
fs/notify/fanotify/fanotify.c
991
fsnotify_queue_overflow(group);
fs/notify/fanotify/fanotify.c
996
ret = fsnotify_insert_event(group, fsn_event, fanotify_merge,
fs/notify/fanotify/fanotify.h
515
struct fsnotify_group *group,
fs/notify/fanotify/fanotify_user.c
1001
pr_debug("%s: group=%p\n", __func__, group);
fs/notify/fanotify/fanotify_user.c
1003
add_wait_queue(&group->notification_waitq, &wait);
fs/notify/fanotify/fanotify_user.c
1010
event = get_one_event(group, count);
fs/notify/fanotify/fanotify_user.c
1032
ret = copy_event_to_user(group, event, buf, count);
fs/notify/fanotify/fanotify_user.c
1039
fsnotify_destroy_event(group, &event->fse);
fs/notify/fanotify/fanotify_user.c
1042
spin_lock(&group->notification_lock);
fs/notify/fanotify/fanotify_user.c
1043
finish_permission_event(group,
fs/notify/fanotify/fanotify_user.c
1045
wake_up(&group->fanotify_data.access_waitq);
fs/notify/fanotify/fanotify_user.c
1047
spin_lock(&group->notification_lock);
fs/notify/fanotify/fanotify_user.c
1049
&group->fanotify_data.access_list);
fs/notify/fanotify/fanotify_user.c
1051
spin_unlock(&group->notification_lock);
fs/notify/fanotify/fanotify_user.c
1059
remove_wait_queue(&group->notification_waitq, &wait);
fs/notify/fanotify/fanotify_user.c
1069
struct fsnotify_group *group;
fs/notify/fanotify/fanotify_user.c
1077
group = file->private_data;
fs/notify/fanotify/fanotify_user.c
1079
pr_debug("%s: group=%p count=%zu\n", __func__, group, count);
fs/notify/fanotify/fanotify_user.c
1089
ret = process_access_response(group, &response, info_buf, info_len);
fs/notify/fanotify/fanotify_user.c
1100
struct fsnotify_group *group = file->private_data;
fs/notify/fanotify/fanotify_user.c
1108
fsnotify_group_stop_queueing(group);
fs/notify/fanotify/fanotify_user.c
1110
fanotify_perm_watchdog_group_remove(group);
fs/notify/fanotify/fanotify_user.c
1116
spin_lock(&group->notification_lock);
fs/notify/fanotify/fanotify_user.c
1117
while (!list_empty(&group->fanotify_data.access_list)) {
fs/notify/fanotify/fanotify_user.c
1120
event = list_first_entry(&group->fanotify_data.access_list,
fs/notify/fanotify/fanotify_user.c
1123
finish_permission_event(group, event, FAN_ALLOW, NULL);
fs/notify/fanotify/fanotify_user.c
1124
spin_lock(&group->notification_lock);
fs/notify/fanotify/fanotify_user.c
1132
while ((fsn_event = fsnotify_remove_first_event(group))) {
fs/notify/fanotify/fanotify_user.c
1136
spin_unlock(&group->notification_lock);
fs/notify/fanotify/fanotify_user.c
1137
fsnotify_destroy_event(group, fsn_event);
fs/notify/fanotify/fanotify_user.c
1139
finish_permission_event(group, FANOTIFY_PERM(event),
fs/notify/fanotify/fanotify_user.c
1142
spin_lock(&group->notification_lock);
fs/notify/fanotify/fanotify_user.c
1144
spin_unlock(&group->notification_lock);
fs/notify/fanotify/fanotify_user.c
1147
wake_up(&group->fanotify_data.access_waitq);
fs/notify/fanotify/fanotify_user.c
1150
fsnotify_destroy_group(group);
fs/notify/fanotify/fanotify_user.c
1157
struct fsnotify_group *group;
fs/notify/fanotify/fanotify_user.c
1163
group = file->private_data;
fs/notify/fanotify/fanotify_user.c
1169
spin_lock(&group->notification_lock);
fs/notify/fanotify/fanotify_user.c
1170
list_for_each_entry(fsn_event, &group->notification_list, list)
fs/notify/fanotify/fanotify_user.c
1172
spin_unlock(&group->notification_lock);
fs/notify/fanotify/fanotify_user.c
119
struct fsnotify_group *group;
fs/notify/fanotify/fanotify_user.c
1269
static int fanotify_remove_mark(struct fsnotify_group *group,
fs/notify/fanotify/fanotify_user.c
1277
fsnotify_group_lock(group);
fs/notify/fanotify/fanotify_user.c
1278
fsn_mark = fsnotify_find_mark(obj, obj_type, group);
fs/notify/fanotify/fanotify_user.c
128
list_for_each_entry(group, &perm_group_list,
fs/notify/fanotify/fanotify_user.c
1280
fsnotify_group_unlock(group);
fs/notify/fanotify/fanotify_user.c
1290
fsnotify_group_unlock(group);
fs/notify/fanotify/fanotify_user.c
134
if (list_empty(&group->fanotify_data.access_list))
fs/notify/fanotify/fanotify_user.c
1367
static int fanotify_set_mark_fsid(struct fsnotify_group *group,
fs/notify/fanotify/fanotify_user.c
137
spin_lock(&group->notification_lock);
fs/notify/fanotify/fanotify_user.c
138
list_for_each_entry(event, &group->fanotify_data.access_list,
fs/notify/fanotify/fanotify_user.c
1381
if (list_empty(&group->marks_list))
fs/notify/fanotify/fanotify_user.c
1385
list_for_each_entry(old, &group->marks_list, g_list) {
fs/notify/fanotify/fanotify_user.c
1418
static struct fsnotify_mark *fanotify_add_new_mark(struct fsnotify_group *group,
fs/notify/fanotify/fanotify_user.c
1424
struct ucounts *ucounts = group->fanotify_data.ucounts;
fs/notify/fanotify/fanotify_user.c
1435
if (!FAN_GROUP_FLAG(group, FAN_UNLIMITED_MARKS) &&
fs/notify/fanotify/fanotify_user.c
1446
fsnotify_init_mark(mark, group);
fs/notify/fanotify/fanotify_user.c
1452
ret = fanotify_set_mark_fsid(group, mark, fsid);
fs/notify/fanotify/fanotify_user.c
1468
if (!FAN_GROUP_FLAG(group, FAN_UNLIMITED_MARKS))
fs/notify/fanotify/fanotify_user.c
1473
static int fanotify_group_init_error_pool(struct fsnotify_group *group)
fs/notify/fanotify/fanotify_user.c
1475
if (mempool_initialized(&group->fanotify_data.error_events_pool))
fs/notify/fanotify/fanotify_user.c
1478
return mempool_init_kmalloc_pool(&group->fanotify_data.error_events_pool,
fs/notify/fanotify/fanotify_user.c
1519
static int fanotify_add_mark(struct fsnotify_group *group,
fs/notify/fanotify/fanotify_user.c
1528
fsnotify_group_lock(group);
fs/notify/fanotify/fanotify_user.c
1529
fsn_mark = fsnotify_find_mark(obj, obj_type, group);
fs/notify/fanotify/fanotify_user.c
1531
fsn_mark = fanotify_add_new_mark(group, obj, obj_type,
fs/notify/fanotify/fanotify_user.c
1534
fsnotify_group_unlock(group);
fs/notify/fanotify/fanotify_user.c
1552
ret = fanotify_group_init_error_pool(group);
fs/notify/fanotify/fanotify_user.c
1562
fsnotify_group_unlock(group);
fs/notify/fanotify/fanotify_user.c
1567
fanotify_perm_watchdog_group_add(group);
fs/notify/fanotify/fanotify_user.c
162
spin_unlock(&group->notification_lock);
fs/notify/fanotify/fanotify_user.c
167
static void fanotify_perm_watchdog_group_remove(struct fsnotify_group *group)
fs/notify/fanotify/fanotify_user.c
169
if (!list_empty(&group->fanotify_data.perm_grp_list)) {
fs/notify/fanotify/fanotify_user.c
1697
CLASS(fsnotify_group, group)(&fanotify_fsnotify_ops,
fs/notify/fanotify/fanotify_user.c
1700
if (IS_ERR(group))
fs/notify/fanotify/fanotify_user.c
1701
return PTR_ERR(group);
fs/notify/fanotify/fanotify_user.c
1704
group->fanotify_data.ucounts = inc_ucount(user_ns, current_euid(),
fs/notify/fanotify/fanotify_user.c
1706
if (!group->fanotify_data.ucounts)
fs/notify/fanotify/fanotify_user.c
1709
group->fanotify_data.flags = flags | internal_flags;
fs/notify/fanotify/fanotify_user.c
1710
group->memcg = get_mem_cgroup_from_mm(current->mm);
fs/notify/fanotify/fanotify_user.c
1711
group->user_ns = get_user_ns(user_ns);
fs/notify/fanotify/fanotify_user.c
1713
group->fanotify_data.merge_hash = fanotify_alloc_merge_hash();
fs/notify/fanotify/fanotify_user.c
1714
if (!group->fanotify_data.merge_hash)
fs/notify/fanotify/fanotify_user.c
1717
group->overflow_event = fanotify_alloc_overflow_event();
fs/notify/fanotify/fanotify_user.c
1718
if (unlikely(!group->overflow_event))
fs/notify/fanotify/fanotify_user.c
172
list_del_init(&group->fanotify_data.perm_grp_list);
fs/notify/fanotify/fanotify_user.c
1723
group->fanotify_data.f_flags = event_f_flags;
fs/notify/fanotify/fanotify_user.c
1724
init_waitqueue_head(&group->fanotify_data.access_waitq);
fs/notify/fanotify/fanotify_user.c
1725
INIT_LIST_HEAD(&group->fanotify_data.access_list);
fs/notify/fanotify/fanotify_user.c
1726
INIT_LIST_HEAD(&group->fanotify_data.perm_grp_list);
fs/notify/fanotify/fanotify_user.c
1729
group->priority = FSNOTIFY_PRIO_NORMAL;
fs/notify/fanotify/fanotify_user.c
1732
group->priority = FSNOTIFY_PRIO_CONTENT;
fs/notify/fanotify/fanotify_user.c
1735
group->priority = FSNOTIFY_PRIO_PRE_CONTENT;
fs/notify/fanotify/fanotify_user.c
1743
group->max_events = UINT_MAX;
fs/notify/fanotify/fanotify_user.c
1745
group->max_events = fanotify_max_queued_events;
fs/notify/fanotify/fanotify_user.c
1755
group, f_flags, FMODE_NONOTIFY));
fs/notify/fanotify/fanotify_user.c
1757
retain_and_null_ptr(group);
fs/notify/fanotify/fanotify_user.c
177
static void fanotify_perm_watchdog_group_add(struct fsnotify_group *group)
fs/notify/fanotify/fanotify_user.c
1828
static int fanotify_events_supported(struct fsnotify_group *group,
fs/notify/fanotify/fanotify_user.c
183
if (list_empty(&group->fanotify_data.perm_grp_list)) {
fs/notify/fanotify/fanotify_user.c
1835
bool strict_dir_events = FAN_GROUP_FLAG(group, FAN_REPORT_TARGET_FID) ||
fs/notify/fanotify/fanotify_user.c
187
list_add_tail(&group->fanotify_data.perm_grp_list, &perm_group_list);
fs/notify/fanotify/fanotify_user.c
1892
struct fsnotify_group *group;
fs/notify/fanotify/fanotify_user.c
1974
group = fd_file(f)->private_data;
fs/notify/fanotify/fanotify_user.c
1977
if (FAN_GROUP_FLAG(group, FAN_REPORT_MNT)) {
fs/notify/fanotify/fanotify_user.c
1993
if (!ns_capable(group->user_ns, CAP_SYS_ADMIN) &&
fs/notify/fanotify/fanotify_user.c
2002
group->priority == FSNOTIFY_PRIO_NORMAL)
fs/notify/fanotify/fanotify_user.c
2005
group->priority == FSNOTIFY_PRIO_CONTENT)
fs/notify/fanotify/fanotify_user.c
2027
fid_mode = FAN_GROUP_FLAG(group, FANOTIFY_FID_BITS);
fs/notify/fanotify/fanotify_user.c
2045
fsnotify_clear_marks_by_group(group, obj_type);
fs/notify/fanotify/fanotify_user.c
2055
ret = fanotify_events_supported(group, &path, mask, flags);
fs/notify/fanotify/fanotify_user.c
2138
ret = fanotify_add_mark(group, obj, obj_type, mask, flags,
fs/notify/fanotify/fanotify_user.c
2142
ret = fanotify_remove_mark(group, obj, obj_type, mask, flags,
fs/notify/fanotify/fanotify_user.c
293
static void fanotify_unhash_event(struct fsnotify_group *group,
fs/notify/fanotify/fanotify_user.c
296
assert_spin_locked(&group->notification_lock);
fs/notify/fanotify/fanotify_user.c
299
group, event, fanotify_event_hash_bucket(group, event));
fs/notify/fanotify/fanotify_user.c
313
static struct fanotify_event *get_one_event(struct fsnotify_group *group,
fs/notify/fanotify/fanotify_user.c
319
unsigned int info_mode = FAN_GROUP_FLAG(group, FANOTIFY_INFO_MODES);
fs/notify/fanotify/fanotify_user.c
321
pr_debug("%s: group=%p count=%zd\n", __func__, group, count);
fs/notify/fanotify/fanotify_user.c
323
spin_lock(&group->notification_lock);
fs/notify/fanotify/fanotify_user.c
324
fsn_event = fsnotify_peek_first_event(group);
fs/notify/fanotify/fanotify_user.c
340
fsnotify_remove_first_event(group);
fs/notify/fanotify/fanotify_user.c
344
fanotify_unhash_event(group, event);
fs/notify/fanotify/fanotify_user.c
346
spin_unlock(&group->notification_lock);
fs/notify/fanotify/fanotify_user.c
350
static int create_fd(struct fsnotify_group *group, const struct path *path,
fs/notify/fanotify/fanotify_user.c
356
client_fd = get_unused_fd_flags(group->fanotify_data.f_flags);
fs/notify/fanotify/fanotify_user.c
364
new_file = dentry_open_nonotify(path, group->fanotify_data.f_flags,
fs/notify/fanotify/fanotify_user.c
400
static void finish_permission_event(struct fsnotify_group *group,
fs/notify/fanotify/fanotify_user.c
403
__releases(&group->notification_lock)
fs/notify/fanotify/fanotify_user.c
407
assert_spin_locked(&group->notification_lock);
fs/notify/fanotify/fanotify_user.c
416
spin_unlock(&group->notification_lock);
fs/notify/fanotify/fanotify_user.c
418
fsnotify_destroy_event(group, &event->fae.fse);
fs/notify/fanotify/fanotify_user.c
421
static int process_access_response(struct fsnotify_group *group,
fs/notify/fanotify/fanotify_user.c
434
__func__, group, fd, response, errno, info, info_len);
fs/notify/fanotify/fanotify_user.c
450
if (errno && group->priority != FSNOTIFY_PRIO_PRE_CONTENT)
fs/notify/fanotify/fanotify_user.c
475
if ((response & FAN_AUDIT) && !FAN_GROUP_FLAG(group, FAN_ENABLE_AUDIT))
fs/notify/fanotify/fanotify_user.c
491
spin_lock(&group->notification_lock);
fs/notify/fanotify/fanotify_user.c
492
list_for_each_entry(event, &group->fanotify_data.access_list,
fs/notify/fanotify/fanotify_user.c
498
finish_permission_event(group, event, response, &friar);
fs/notify/fanotify/fanotify_user.c
499
wake_up(&group->fanotify_data.access_waitq);
fs/notify/fanotify/fanotify_user.c
502
spin_unlock(&group->notification_lock);
fs/notify/fanotify/fanotify_user.c
833
static ssize_t copy_event_to_user(struct fsnotify_group *group,
fs/notify/fanotify/fanotify_user.c
840
unsigned int info_mode = FAN_GROUP_FLAG(group, FANOTIFY_INFO_MODES);
fs/notify/fanotify/fanotify_user.c
845
pr_debug("%s: group=%p event=%p\n", __func__, group, event);
fs/notify/fanotify/fanotify_user.c
858
if (FAN_GROUP_FLAG(group, FANOTIFY_UNPRIV) &&
fs/notify/fanotify/fanotify_user.c
868
if (!FAN_GROUP_FLAG(group, FANOTIFY_UNPRIV) &&
fs/notify/fanotify/fanotify_user.c
870
fd = create_fd(group, path, &f);
fs/notify/fanotify/fanotify_user.c
887
if (!FAN_GROUP_FLAG(group, FAN_REPORT_FD_ERROR)) {
fs/notify/fanotify/fanotify_user.c
900
if (FAN_GROUP_FLAG(group, FAN_REPORT_FD_ERROR))
fs/notify/fanotify/fanotify_user.c
912
WARN_ON_ONCE(FAN_GROUP_FLAG(group, FAN_REPORT_TID));
fs/notify/fanotify/fanotify_user.c
926
if (!FAN_GROUP_FLAG(group, FAN_REPORT_FD_ERROR) && pidfd < 0)
fs/notify/fanotify/fanotify_user.c
977
struct fsnotify_group *group = file->private_data;
fs/notify/fanotify/fanotify_user.c
980
poll_wait(file, &group->notification_waitq, wait);
fs/notify/fanotify/fanotify_user.c
981
spin_lock(&group->notification_lock);
fs/notify/fanotify/fanotify_user.c
982
if (!fsnotify_notify_queue_is_empty(group))
fs/notify/fanotify/fanotify_user.c
984
spin_unlock(&group->notification_lock);
fs/notify/fanotify/fanotify_user.c
992
struct fsnotify_group *group;
fs/notify/fanotify/fanotify_user.c
999
group = file->private_data;
fs/notify/fdinfo.c
140
struct fsnotify_group *group = f->private_data;
fs/notify/fdinfo.c
143
group->fanotify_data.flags & FANOTIFY_INIT_FLAGS,
fs/notify/fdinfo.c
144
group->fanotify_data.f_flags);
fs/notify/fdinfo.c
30
struct fsnotify_group *group = f->private_data;
fs/notify/fdinfo.c
33
fsnotify_group_lock(group);
fs/notify/fdinfo.c
34
list_for_each_entry(mark, &group->marks_list, g_list) {
fs/notify/fdinfo.c
39
fsnotify_group_unlock(group);
fs/notify/fsnotify.c
248
static int fsnotify_handle_inode_event(struct fsnotify_group *group,
fs/notify/fsnotify.c
256
const struct fsnotify_ops *ops = group->ops;
fs/notify/fsnotify.c
275
static int fsnotify_handle_event(struct fsnotify_group *group, __u32 mask,
fs/notify/fsnotify.c
301
ret = fsnotify_handle_inode_event(group, parent_mark, mask,
fs/notify/fsnotify.c
327
return fsnotify_handle_inode_event(group, inode_mark, mask, data, data_type,
fs/notify/fsnotify.c
335
struct fsnotify_group *group = NULL;
fs/notify/fsnotify.c
357
group = mark->group;
fs/notify/fsnotify.c
364
__func__, group, mask, marks_mask, marks_ignore_mask,
fs/notify/fsnotify.c
370
if (group->ops->handle_event) {
fs/notify/fsnotify.c
371
return group->ops->handle_event(group, mask, data, data_type, dir,
fs/notify/fsnotify.c
375
return fsnotify_handle_event(group, mask, data, data_type, dir,
fs/notify/fsnotify.c
419
fsnotify_compare_groups(max_prio_group, mark->group) > 0)
fs/notify/fsnotify.c
420
max_prio_group = mark->group;
fs/notify/fsnotify.c
431
if (mark && mark->group == iter_info->current_group) {
fs/notify/fsnotify.c
466
if (mark && mark->group == iter_info->current_group)
fs/notify/fsnotify.h
71
extern void fsnotify_flush_notify(struct fsnotify_group *group);
fs/notify/group.c
101
refcount_inc(&group->refcnt);
fs/notify/group.c
107
void fsnotify_put_group(struct fsnotify_group *group)
fs/notify/group.c
109
if (refcount_dec_and_test(&group->refcnt))
fs/notify/group.c
110
fsnotify_final_destroy_group(group);
fs/notify/group.c
118
struct fsnotify_group *group;
fs/notify/group.c
120
group = kzalloc_obj(struct fsnotify_group, gfp);
fs/notify/group.c
121
if (!group)
fs/notify/group.c
125
refcount_set(&group->refcnt, 1);
fs/notify/group.c
126
atomic_set(&group->user_waits, 0);
fs/notify/group.c
128
spin_lock_init(&group->notification_lock);
fs/notify/group.c
129
INIT_LIST_HEAD(&group->notification_list);
fs/notify/group.c
130
init_waitqueue_head(&group->notification_waitq);
fs/notify/group.c
131
group->max_events = UINT_MAX;
fs/notify/group.c
133
mutex_init(&group->mark_mutex);
fs/notify/group.c
134
INIT_LIST_HEAD(&group->marks_list);
fs/notify/group.c
136
group->ops = ops;
fs/notify/group.c
137
group->flags = flags;
fs/notify/group.c
139
return group;
fs/notify/group.c
157
struct fsnotify_group *group = file->private_data;
fs/notify/group.c
159
return fasync_helper(fd, file, on, &group->fsn_fa) >= 0 ? 0 : -EIO;
fs/notify/group.c
22
static void fsnotify_final_destroy_group(struct fsnotify_group *group)
fs/notify/group.c
24
if (group->ops->free_group_priv)
fs/notify/group.c
25
group->ops->free_group_priv(group);
fs/notify/group.c
27
mem_cgroup_put(group->memcg);
fs/notify/group.c
28
mutex_destroy(&group->mark_mutex);
fs/notify/group.c
30
kfree(group);
fs/notify/group.c
37
void fsnotify_group_stop_queueing(struct fsnotify_group *group)
fs/notify/group.c
39
spin_lock(&group->notification_lock);
fs/notify/group.c
40
group->shutdown = true;
fs/notify/group.c
41
spin_unlock(&group->notification_lock);
fs/notify/group.c
50
void fsnotify_destroy_group(struct fsnotify_group *group)
fs/notify/group.c
58
fsnotify_group_stop_queueing(group);
fs/notify/group.c
61
fsnotify_clear_marks_by_group(group, FSNOTIFY_OBJ_TYPE_ANY);
fs/notify/group.c
68
wait_event(group->notification_waitq, !atomic_read(&group->user_waits));
fs/notify/group.c
84
fsnotify_flush_notify(group);
fs/notify/group.c
90
if (group->overflow_event)
fs/notify/group.c
91
group->ops->free_event(group, group->overflow_event);
fs/notify/group.c
93
fsnotify_put_group(group);
fs/notify/group.c
99
void fsnotify_get_group(struct fsnotify_group *group)
fs/notify/inotify/inotify.h
45
struct fsnotify_group *group);
fs/notify/inotify/inotify_fsnotify.c
104
fsnotify_queue_overflow(group);
fs/notify/inotify/inotify_fsnotify.c
126
ret = fsnotify_add_event(group, fsn_event, inotify_merge);
fs/notify/inotify/inotify_fsnotify.c
129
fsnotify_destroy_event(group, fsn_event);
fs/notify/inotify/inotify_fsnotify.c
133
fsnotify_destroy_mark(inode_mark, group);
fs/notify/inotify/inotify_fsnotify.c
138
static void inotify_freeing_mark(struct fsnotify_mark *fsn_mark, struct fsnotify_group *group)
fs/notify/inotify/inotify_fsnotify.c
140
inotify_ignored_and_remove_idr(fsn_mark, group);
fs/notify/inotify/inotify_fsnotify.c
174
fsn_mark->group, i_mark->wd);
fs/notify/inotify/inotify_fsnotify.c
178
static void inotify_free_group_priv(struct fsnotify_group *group)
fs/notify/inotify/inotify_fsnotify.c
181
idr_for_each(&group->inotify_data.idr, idr_callback, group);
fs/notify/inotify/inotify_fsnotify.c
182
idr_destroy(&group->inotify_data.idr);
fs/notify/inotify/inotify_fsnotify.c
183
if (group->inotify_data.ucounts)
fs/notify/inotify/inotify_fsnotify.c
184
dec_inotify_instances(group->inotify_data.ucounts);
fs/notify/inotify/inotify_fsnotify.c
187
static void inotify_free_event(struct fsnotify_group *group,
fs/notify/inotify/inotify_fsnotify.c
49
static int inotify_merge(struct fsnotify_group *group,
fs/notify/inotify/inotify_fsnotify.c
52
struct list_head *list = &group->notification_list;
fs/notify/inotify/inotify_fsnotify.c
66
struct fsnotify_group *group = inode_mark->group;
fs/notify/inotify/inotify_fsnotify.c
77
pr_debug("%s: group=%p mark=%p mask=%x\n", __func__, group, inode_mark,
fs/notify/inotify/inotify_fsnotify.c
95
old_memcg = set_active_memcg(group->memcg);
fs/notify/inotify/inotify_user.c
141
struct fsnotify_group *group = file->private_data;
fs/notify/inotify/inotify_user.c
144
poll_wait(file, &group->notification_waitq, wait);
fs/notify/inotify/inotify_user.c
145
spin_lock(&group->notification_lock);
fs/notify/inotify/inotify_user.c
146
if (!fsnotify_notify_queue_is_empty(group))
fs/notify/inotify/inotify_user.c
148
spin_unlock(&group->notification_lock);
fs/notify/inotify/inotify_user.c
170
static struct fsnotify_event *get_one_event(struct fsnotify_group *group,
fs/notify/inotify/inotify_user.c
176
event = fsnotify_peek_first_event(group);
fs/notify/inotify/inotify_user.c
180
pr_debug("%s: group=%p event=%p\n", __func__, group, event);
fs/notify/inotify/inotify_user.c
188
fsnotify_remove_first_event(group);
fs/notify/inotify/inotify_user.c
199
static ssize_t copy_event_to_user(struct fsnotify_group *group,
fs/notify/inotify/inotify_user.c
209
pr_debug("%s: group=%p event=%p\n", __func__, group, fsn_event);
fs/notify/inotify/inotify_user.c
252
struct fsnotify_group *group;
fs/notify/inotify/inotify_user.c
259
group = file->private_data;
fs/notify/inotify/inotify_user.c
261
add_wait_queue(&group->notification_waitq, &wait);
fs/notify/inotify/inotify_user.c
263
spin_lock(&group->notification_lock);
fs/notify/inotify/inotify_user.c
264
kevent = get_one_event(group, count);
fs/notify/inotify/inotify_user.c
265
spin_unlock(&group->notification_lock);
fs/notify/inotify/inotify_user.c
267
pr_debug("%s: group=%p kevent=%p\n", __func__, group, kevent);
fs/notify/inotify/inotify_user.c
273
ret = copy_event_to_user(group, kevent, buf);
fs/notify/inotify/inotify_user.c
274
fsnotify_destroy_event(group, kevent);
fs/notify/inotify/inotify_user.c
294
remove_wait_queue(&group->notification_waitq, &wait);
fs/notify/inotify/inotify_user.c
303
struct fsnotify_group *group = file->private_data;
fs/notify/inotify/inotify_user.c
305
pr_debug("%s: group=%p\n", __func__, group);
fs/notify/inotify/inotify_user.c
308
fsnotify_destroy_group(group);
fs/notify/inotify/inotify_user.c
316
struct fsnotify_group *group;
fs/notify/inotify/inotify_user.c
322
group = file->private_data;
fs/notify/inotify/inotify_user.c
325
pr_debug("%s: group=%p cmd=%u\n", __func__, group, cmd);
fs/notify/inotify/inotify_user.c
329
spin_lock(&group->notification_lock);
fs/notify/inotify/inotify_user.c
330
list_for_each_entry(fsn_event, &group->notification_list,
fs/notify/inotify/inotify_user.c
335
spin_unlock(&group->notification_lock);
fs/notify/inotify/inotify_user.c
344
data = &group->inotify_data;
fs/notify/inotify/inotify_user.c
414
static struct inotify_inode_mark *inotify_idr_find_locked(struct fsnotify_group *group,
fs/notify/inotify/inotify_user.c
417
struct idr *idr = &group->inotify_data.idr;
fs/notify/inotify/inotify_user.c
418
spinlock_t *idr_lock = &group->inotify_data.idr_lock;
fs/notify/inotify/inotify_user.c
435
static struct inotify_inode_mark *inotify_idr_find(struct fsnotify_group *group,
fs/notify/inotify/inotify_user.c
439
spinlock_t *idr_lock = &group->inotify_data.idr_lock;
fs/notify/inotify/inotify_user.c
442
i_mark = inotify_idr_find_locked(group, wd);
fs/notify/inotify/inotify_user.c
452
static void inotify_remove_from_idr(struct fsnotify_group *group,
fs/notify/inotify/inotify_user.c
455
struct idr *idr = &group->inotify_data.idr;
fs/notify/inotify/inotify_user.c
456
spinlock_t *idr_lock = &group->inotify_data.idr_lock;
fs/notify/inotify/inotify_user.c
469
__func__, i_mark, i_mark->wd, i_mark->fsn_mark.group);
fs/notify/inotify/inotify_user.c
474
found_i_mark = inotify_idr_find_locked(group, wd);
fs/notify/inotify/inotify_user.c
477
__func__, i_mark, i_mark->wd, i_mark->fsn_mark.group);
fs/notify/inotify/inotify_user.c
490
i_mark->wd, i_mark->fsn_mark.group, found_i_mark,
fs/notify/inotify/inotify_user.c
491
found_i_mark->wd, found_i_mark->fsn_mark.group);
fs/notify/inotify/inotify_user.c
501
__func__, i_mark, i_mark->wd, i_mark->fsn_mark.group);
fs/notify/inotify/inotify_user.c
521
struct fsnotify_group *group)
fs/notify/inotify/inotify_user.c
531
inotify_remove_from_idr(group, i_mark);
fs/notify/inotify/inotify_user.c
533
dec_inotify_watches(group->inotify_data.ucounts);
fs/notify/inotify/inotify_user.c
536
static int inotify_update_existing_watch(struct fsnotify_group *group,
fs/notify/inotify/inotify_user.c
547
fsn_mark = fsnotify_find_inode_mark(inode, group);
fs/notify/inotify/inotify_user.c
590
static int inotify_new_watch(struct fsnotify_group *group,
fs/notify/inotify/inotify_user.c
596
struct idr *idr = &group->inotify_data.idr;
fs/notify/inotify/inotify_user.c
597
spinlock_t *idr_lock = &group->inotify_data.idr_lock;
fs/notify/inotify/inotify_user.c
603
fsnotify_init_mark(&tmp_i_mark->fsn_mark, group);
fs/notify/inotify/inotify_user.c
613
if (!inc_inotify_watches(group->inotify_data.ucounts)) {
fs/notify/inotify/inotify_user.c
614
inotify_remove_from_idr(group, tmp_i_mark);
fs/notify/inotify/inotify_user.c
623
inotify_remove_from_idr(group, tmp_i_mark);
fs/notify/inotify/inotify_user.c
638
static int inotify_update_watch(struct fsnotify_group *group, struct inode *inode, u32 arg)
fs/notify/inotify/inotify_user.c
642
fsnotify_group_lock(group);
fs/notify/inotify/inotify_user.c
644
ret = inotify_update_existing_watch(group, inode, arg);
fs/notify/inotify/inotify_user.c
647
ret = inotify_new_watch(group, inode, arg);
fs/notify/inotify/inotify_user.c
648
fsnotify_group_unlock(group);
fs/notify/inotify/inotify_user.c
655
struct fsnotify_group *group;
fs/notify/inotify/inotify_user.c
658
group = fsnotify_alloc_group(&inotify_fsnotify_ops,
fs/notify/inotify/inotify_user.c
660
if (IS_ERR(group))
fs/notify/inotify/inotify_user.c
661
return group;
fs/notify/inotify/inotify_user.c
665
fsnotify_destroy_group(group);
fs/notify/inotify/inotify_user.c
668
group->overflow_event = &oevent->fse;
fs/notify/inotify/inotify_user.c
669
fsnotify_init_event(group->overflow_event);
fs/notify/inotify/inotify_user.c
675
group->max_events = max_events;
fs/notify/inotify/inotify_user.c
676
group->memcg = get_mem_cgroup_from_mm(current->mm);
fs/notify/inotify/inotify_user.c
678
spin_lock_init(&group->inotify_data.idr_lock);
fs/notify/inotify/inotify_user.c
679
idr_init(&group->inotify_data.idr);
fs/notify/inotify/inotify_user.c
680
group->inotify_data.ucounts = inc_ucount(current_user_ns(),
fs/notify/inotify/inotify_user.c
684
if (!group->inotify_data.ucounts) {
fs/notify/inotify/inotify_user.c
685
fsnotify_destroy_group(group);
fs/notify/inotify/inotify_user.c
689
return group;
fs/notify/inotify/inotify_user.c
696
struct fsnotify_group *group;
fs/notify/inotify/inotify_user.c
707
group = inotify_new_group(inotify_max_queued_events);
fs/notify/inotify/inotify_user.c
708
if (IS_ERR(group))
fs/notify/inotify/inotify_user.c
709
return PTR_ERR(group);
fs/notify/inotify/inotify_user.c
711
ret = anon_inode_getfd("inotify", &inotify_fops, group,
fs/notify/inotify/inotify_user.c
714
fsnotify_destroy_group(group);
fs/notify/inotify/inotify_user.c
732
struct fsnotify_group *group;
fs/notify/inotify/inotify_user.c
778
group = fd_file(f)->private_data;
fs/notify/inotify/inotify_user.c
781
ret = inotify_update_watch(group, inode, mask);
fs/notify/inotify/inotify_user.c
788
struct fsnotify_group *group;
fs/notify/inotify/inotify_user.c
799
group = fd_file(f)->private_data;
fs/notify/inotify/inotify_user.c
801
i_mark = inotify_idr_find(group, wd);
fs/notify/inotify/inotify_user.c
805
fsnotify_destroy_mark(&i_mark->fsn_mark, group);
fs/notify/mark.c
1003
head = &group->marks_list;
fs/notify/mark.c
1015
fsnotify_group_lock(group);
fs/notify/mark.c
1016
list_for_each_entry_safe(mark, lmark, &group->marks_list, g_list) {
fs/notify/mark.c
1020
fsnotify_group_unlock(group);
fs/notify/mark.c
1024
fsnotify_group_lock(group);
fs/notify/mark.c
1026
fsnotify_group_unlock(group);
fs/notify/mark.c
1032
fsnotify_group_unlock(group);
fs/notify/mark.c
1062
fsnotify_destroy_mark(mark, mark->group);
fs/notify/mark.c
1081
struct fsnotify_group *group)
fs/notify/mark.c
1086
fsnotify_get_group(group);
fs/notify/mark.c
1087
mark->group = group;
fs/notify/mark.c
184
highest_prio = first_mark->group->priority;
fs/notify/mark.c
372
struct fsnotify_group *group = mark->group;
fs/notify/mark.c
374
if (WARN_ON_ONCE(!group))
fs/notify/mark.c
376
group->ops->free_mark(mark);
fs/notify/mark.c
377
fsnotify_put_group(group);
fs/notify/mark.c
467
atomic_inc(&mark->group->user_waits);
fs/notify/mark.c
485
struct fsnotify_group *group = mark->group;
fs/notify/mark.c
492
if (atomic_dec_and_test(&group->user_waits) && group->shutdown)
fs/notify/mark.c
493
wake_up(&group->notification_waitq);
fs/notify/mark.c
547
fsnotify_group_assert_locked(mark->group);
fs/notify/mark.c
576
struct fsnotify_group *group = mark->group;
fs/notify/mark.c
592
if (group->ops->freeing_mark)
fs/notify/mark.c
593
group->ops->freeing_mark(mark, group);
fs/notify/mark.c
597
struct fsnotify_group *group)
fs/notify/mark.c
599
fsnotify_group_lock(group);
fs/notify/mark.c
601
fsnotify_group_unlock(group);
fs/notify/mark.c
872
if ((lmark->group == mark->group) &&
fs/notify/mark.c
874
!(mark->group->flags & FSNOTIFY_GROUP_DUPS)) {
fs/notify/mark.c
879
cmp = fsnotify_compare_groups(lmark->group, mark->group);
fs/notify/mark.c
913
struct fsnotify_group *group = mark->group;
fs/notify/mark.c
916
fsnotify_group_assert_locked(group);
fs/notify/mark.c
927
list_add(&mark->g_list, &group->marks_list);
fs/notify/mark.c
953
struct fsnotify_group *group = mark->group;
fs/notify/mark.c
955
fsnotify_group_lock(group);
fs/notify/mark.c
957
fsnotify_group_unlock(group);
fs/notify/mark.c
967
struct fsnotify_group *group)
fs/notify/mark.c
981
if (mark->group == group &&
fs/notify/mark.c
994
void fsnotify_clear_marks_by_group(struct fsnotify_group *group,
fs/notify/notification.c
100
if (event == group->overflow_event ||
fs/notify/notification.c
101
group->q_len >= group->max_events) {
fs/notify/notification.c
104
if (!list_empty(&group->overflow_event->list)) {
fs/notify/notification.c
105
spin_unlock(&group->notification_lock);
fs/notify/notification.c
108
event = group->overflow_event;
fs/notify/notification.c
113
ret = merge(group, event);
fs/notify/notification.c
115
spin_unlock(&group->notification_lock);
fs/notify/notification.c
121
group->q_len++;
fs/notify/notification.c
124
insert(group, event);
fs/notify/notification.c
125
spin_unlock(&group->notification_lock);
fs/notify/notification.c
127
wake_up(&group->notification_waitq);
fs/notify/notification.c
128
kill_fasync(&group->fsn_fa, SIGIO, POLL_IN);
fs/notify/notification.c
132
void fsnotify_remove_queued_event(struct fsnotify_group *group,
fs/notify/notification.c
135
assert_spin_locked(&group->notification_lock);
fs/notify/notification.c
141
group->q_len--;
fs/notify/notification.c
148
struct fsnotify_event *fsnotify_peek_first_event(struct fsnotify_group *group)
fs/notify/notification.c
150
assert_spin_locked(&group->notification_lock);
fs/notify/notification.c
152
if (fsnotify_notify_queue_is_empty(group))
fs/notify/notification.c
155
return list_first_entry(&group->notification_list,
fs/notify/notification.c
163
struct fsnotify_event *fsnotify_remove_first_event(struct fsnotify_group *group)
fs/notify/notification.c
165
struct fsnotify_event *event = fsnotify_peek_first_event(group);
fs/notify/notification.c
170
pr_debug("%s: group=%p event=%p\n", __func__, group, event);
fs/notify/notification.c
172
fsnotify_remove_queued_event(group, event);
fs/notify/notification.c
181
void fsnotify_flush_notify(struct fsnotify_group *group)
fs/notify/notification.c
185
spin_lock(&group->notification_lock);
fs/notify/notification.c
186
while (!fsnotify_notify_queue_is_empty(group)) {
fs/notify/notification.c
187
event = fsnotify_remove_first_event(group);
fs/notify/notification.c
188
spin_unlock(&group->notification_lock);
fs/notify/notification.c
189
fsnotify_destroy_event(group, event);
fs/notify/notification.c
190
spin_lock(&group->notification_lock);
fs/notify/notification.c
192
spin_unlock(&group->notification_lock);
fs/notify/notification.c
50
void fsnotify_destroy_event(struct fsnotify_group *group,
fs/notify/notification.c
54
if (!event || event == group->overflow_event)
fs/notify/notification.c
63
spin_lock(&group->notification_lock);
fs/notify/notification.c
65
spin_unlock(&group->notification_lock);
fs/notify/notification.c
67
group->ops->free_event(group, event);
fs/notify/notification.c
81
int fsnotify_insert_event(struct fsnotify_group *group,
fs/notify/notification.c
89
struct list_head *list = &group->notification_list;
fs/notify/notification.c
91
pr_debug("%s: group=%p event=%p\n", __func__, group, event);
fs/notify/notification.c
93
spin_lock(&group->notification_lock);
fs/notify/notification.c
95
if (group->shutdown) {
fs/notify/notification.c
96
spin_unlock(&group->notification_lock);
fs/ntfs3/fsntfs.c
1821
const struct SID *group = Add2Ptr(sd, sd_group);
fs/ntfs3/fsntfs.c
1826
if (group->Revision != 1)
fs/ntfs3/fsntfs.c
1829
if (sd_group + sid_length(group) > len)
fs/ocfs2/alloc.c
7484
u64 group, u32 start, u32 count)
fs/ocfs2/alloc.c
7500
if (group != osb->first_cluster_group_blkno)
fs/ocfs2/alloc.c
7509
struct ocfs2_group_desc *gd, u64 group,
fs/ocfs2/alloc.c
7528
ret = ocfs2_trim_extent(sb, gd, group,
fs/ocfs2/alloc.c
7557
u64 start, len, trimmed = 0, first_group, last_group = 0, group = 0;
fs/ocfs2/alloc.c
7597
if (!group) {
fs/ocfs2/alloc.c
7618
group = first_group;
fs/ocfs2/alloc.c
7628
main_bm, group,
fs/ocfs2/alloc.c
7636
cnt = ocfs2_trim_group(sb, gd, group,
fs/ocfs2/alloc.c
7649
if (group == osb->first_cluster_group_blkno)
fs/ocfs2/alloc.c
7650
group = ocfs2_clusters_to_blocks(sb, osb->bitmap_cpg);
fs/ocfs2/alloc.c
7652
group += ocfs2_clusters_to_blocks(sb, osb->bitmap_cpg);
fs/ocfs2/alloc.c
7668
if (ret >= 0 && group <= last_group) {
fs/ocfs2/cluster/heartbeat.c
1960
static struct o2hb_heartbeat_group *to_o2hb_heartbeat_group(struct config_group *group)
fs/ocfs2/cluster/heartbeat.c
1962
return group ?
fs/ocfs2/cluster/heartbeat.c
1963
container_of(group, struct o2hb_heartbeat_group, hs_group)
fs/ocfs2/cluster/heartbeat.c
1996
static struct config_item *o2hb_heartbeat_group_make_item(struct config_group *group,
fs/ocfs2/cluster/heartbeat.c
2066
static void o2hb_heartbeat_group_drop_item(struct config_group *group,
fs/ocfs2/cluster/heartbeat.c
2226
void o2hb_free_hb_set(struct config_group *group)
fs/ocfs2/cluster/heartbeat.c
2228
struct o2hb_heartbeat_group *hs = to_o2hb_heartbeat_group(group);
fs/ocfs2/cluster/heartbeat.h
50
void o2hb_free_hb_set(struct config_group *group);
fs/ocfs2/cluster/nodemanager.c
417
static struct o2nm_node_group *to_o2nm_node_group(struct config_group *group)
fs/ocfs2/cluster/nodemanager.c
419
return group ?
fs/ocfs2/cluster/nodemanager.c
420
container_of(group, struct o2nm_node_group, ns_group)
fs/ocfs2/cluster/nodemanager.c
582
static struct config_item *o2nm_node_group_make_item(struct config_group *group,
fs/ocfs2/cluster/nodemanager.c
603
static void o2nm_node_group_drop_item(struct config_group *group,
fs/ocfs2/cluster/nodemanager.c
607
struct o2nm_cluster *cluster = to_o2nm_cluster(group->cg_item.ci_parent);
fs/ocfs2/cluster/nodemanager.c
678
static struct o2nm_cluster_group *to_o2nm_cluster_group(struct config_group *group)
fs/ocfs2/cluster/nodemanager.c
680
return group ?
fs/ocfs2/cluster/nodemanager.c
681
container_of(to_configfs_subsystem(group), struct o2nm_cluster_group, cs_subsys)
fs/ocfs2/cluster/nodemanager.c
686
static struct config_group *o2nm_cluster_group_make_group(struct config_group *group,
fs/ocfs2/cluster/nodemanager.c
733
static void o2nm_cluster_group_drop_item(struct config_group *group, struct config_item *item)
fs/ocfs2/ocfs2_ioctl.h
41
__u64 group; /* Group descriptor's blkno. */
fs/ocfs2/ocfs2_trace.h
837
TP_PROTO(unsigned long long inode, unsigned long long group,
fs/ocfs2/ocfs2_trace.h
839
TP_ARGS(inode, group, start_bit, count),
fs/ocfs2/ocfs2_trace.h
842
__field(unsigned long long, group)
fs/ocfs2/ocfs2_trace.h
848
__entry->group = group;
fs/ocfs2/ocfs2_trace.h
852
TP_printk("%llu %llu %u %u", __entry->inode, __entry->group,
fs/ocfs2/resize.c
107
group = (struct ocfs2_group_desc *)group_bh->b_data;
fs/ocfs2/resize.c
109
old_bg_clusters = le16_to_cpu(group->bg_bits) / cl_bpc;
fs/ocfs2/resize.c
112
le16_add_cpu(&group->bg_bits, num_bits);
fs/ocfs2/resize.c
113
le16_add_cpu(&group->bg_free_bits_count, num_bits);
fs/ocfs2/resize.c
122
group,
fs/ocfs2/resize.c
124
le16_add_cpu(&group->bg_free_bits_count, -1 * backups);
fs/ocfs2/resize.c
127
contig_bits = ocfs2_find_max_contig_free_bits(group->bg_bitmap,
fs/ocfs2/resize.c
128
le16_to_cpu(group->bg_bits), 0);
fs/ocfs2/resize.c
129
old_bg_contig_free_bits = group->bg_contig_free_bits;
fs/ocfs2/resize.c
130
group->bg_contig_free_bits = cpu_to_le16(contig_bits);
fs/ocfs2/resize.c
142
chain = le16_to_cpu(group->bg_chain);
fs/ocfs2/resize.c
165
group,
fs/ocfs2/resize.c
167
le16_add_cpu(&group->bg_free_bits_count, backups);
fs/ocfs2/resize.c
168
le16_add_cpu(&group->bg_bits, -1 * num_bits);
fs/ocfs2/resize.c
169
le16_add_cpu(&group->bg_free_bits_count, -1 * num_bits);
fs/ocfs2/resize.c
170
group->bg_contig_free_bits = old_bg_contig_free_bits;
fs/ocfs2/resize.c
273
struct ocfs2_group_desc *group = NULL;
fs/ocfs2/resize.c
329
group = (struct ocfs2_group_desc *)group_bh->b_data;
fs/ocfs2/resize.c
332
if (le16_to_cpu(group->bg_bits) / cl_bpc + new_clusters >
fs/ocfs2/resize.c
340
(unsigned long long)le64_to_cpu(group->bg_blkno), new_clusters);
fs/ocfs2/resize.c
423
u32 cluster = ocfs2_blocks_to_clusters(inode->i_sb, input->group);
fs/ocfs2/resize.c
443
else if (input->group != ocfs2_which_cluster_group(inode, cluster))
fs/ocfs2/resize.c
463
struct ocfs2_group_desc *group = NULL;
fs/ocfs2/resize.c
500
ret = ocfs2_read_blocks_sync(osb, input->group, 1, &group_bh);
fs/ocfs2/resize.c
503
"from the device.", (unsigned long long)input->group);
fs/ocfs2/resize.c
515
trace_ocfs2_group_add((unsigned long long)input->group,
fs/ocfs2/resize.c
536
group = (struct ocfs2_group_desc *)group_bh->b_data;
fs/ocfs2/resize.c
537
bg_ptr = le64_to_cpu(group->bg_next_group);
fs/ocfs2/resize.c
538
group->bg_next_group = cr->c_blkno;
fs/ocfs2/resize.c
544
group->bg_next_group = cpu_to_le64(bg_ptr);
fs/ocfs2/resize.c
554
cr->c_blkno = cpu_to_le64(input->group);
fs/ocfs2/resize.c
89
struct ocfs2_group_desc *group;
fs/ocfs2/stackglue.c
305
const char *group,
fs/ocfs2/stackglue.c
316
BUG_ON(group == NULL);
fs/ocfs2/stackglue.c
337
strscpy(new_conn->cc_name, group, GROUP_NAME_MAX + 1);
fs/ocfs2/stackglue.c
373
int ocfs2_cluster_connect_agnostic(const char *group,
fs/ocfs2/stackglue.c
385
return ocfs2_cluster_connect(stack_name, NULL, 0, group, grouplen,
fs/ocfs2/stackglue.c
416
static void ocfs2_leave_group(const char *group)
fs/ocfs2/stackglue.c
424
argv[3] = (char *)group;
fs/ocfs2/stackglue.c
448
void ocfs2_cluster_hangup(const char *group, int grouplen)
fs/ocfs2/stackglue.c
450
BUG_ON(group == NULL);
fs/ocfs2/stackglue.c
451
BUG_ON(group[grouplen] != '\0');
fs/ocfs2/stackglue.c
453
ocfs2_leave_group(group);
fs/ocfs2/stackglue.h
240
const char *group,
fs/ocfs2/stackglue.h
251
int ocfs2_cluster_connect_agnostic(const char *group,
fs/ocfs2/stackglue.h
260
void ocfs2_cluster_hangup(const char *group, int grouplen);
fs/ocfs2/suballoc.c
2667
struct ocfs2_group_desc *group = (struct ocfs2_group_desc *) group_bh->b_data;
fs/ocfs2/suballoc.c
2669
idx = le16_to_cpu(group->bg_chain);
fs/ocfs2/suballoc.c
2748
memset(group, 0, sizeof(struct ocfs2_group_desc));
fs/ocfs2/suballoc.c
2777
group = (struct ocfs2_group_desc *) group_bh->b_data;
fs/ocfs2/suballoc.c
2779
if ((count + start_bit) > le16_to_cpu(group->bg_bits)) {
fs/ocfs2/suballoc.c
2782
count + start_bit, le16_to_cpu(group->bg_bits));
fs/ocfs2/suballoc.c
2786
old_bg_contig_free_bits = group->bg_contig_free_bits;
fs/ocfs2/suballoc.c
2788
group, group_bh,
fs/ocfs2/suballoc.c
2800
ocfs2_block_group_set_bits(handle, main_bm_inode, group, group_bh,
fs/ocfs2/suballoc.c
2806
idx = le16_to_cpu(group->bg_chain);
fs/ocfs2/suballoc.c
2846
struct ocfs2_group_desc *group;
fs/ocfs2/suballoc.c
2869
group = (struct ocfs2_group_desc *) group_bh->b_data;
fs/ocfs2/suballoc.c
2871
BUG_ON((count + start_bit) > le16_to_cpu(group->bg_bits));
fs/ocfs2/suballoc.c
2874
old_bg_contig_free_bits = group->bg_contig_free_bits;
fs/ocfs2/suballoc.c
2876
group, group_bh,
fs/ocfs2/suballoc.c
2887
ocfs2_block_group_set_bits(handle, alloc_inode, group, group_bh,
fs/ocfs2/suballoc.c
2893
idx = le16_to_cpu(group->bg_chain);
fs/ocfs2/suballoc.c
3163
struct ocfs2_group_desc *group;
fs/ocfs2/suballoc.c
3194
group = (struct ocfs2_group_desc *) group_bh->b_data;
fs/ocfs2/suballoc.c
3195
*res = ocfs2_test_bit(bit, (unsigned long *)group->bg_bitmap);
fs/ocfs2/suballoc.h
147
u64 group = block - (u64) bit;
fs/ocfs2/suballoc.h
149
return group;
fs/open.c
741
int chown_common(const struct path *path, uid_t user, gid_t group)
fs/open.c
753
gid = make_kgid(current_user_ns(), group);
fs/open.c
764
if ((group != (gid_t)-1) && !setattr_vfsgid(&newattrs, gid))
fs/open.c
789
int do_fchownat(int dfd, const char __user *filename, uid_t user, gid_t group,
fs/open.c
806
error = chown_common(&path, user, group);
fs/open.c
819
gid_t, group, int, flag)
fs/open.c
821
return do_fchownat(dfd, filename, user, group, flag);
fs/open.c
824
SYSCALL_DEFINE3(chown, const char __user *, filename, uid_t, user, gid_t, group)
fs/open.c
826
return do_fchownat(AT_FDCWD, filename, user, group, 0);
fs/open.c
829
SYSCALL_DEFINE3(lchown, const char __user *, filename, uid_t, user, gid_t, group)
fs/open.c
831
return do_fchownat(AT_FDCWD, filename, user, group,
fs/open.c
835
int vfs_fchown(struct file *file, uid_t user, gid_t group)
fs/open.c
843
error = chown_common(&file->f_path, user, group);
fs/open.c
848
int ksys_fchown(unsigned int fd, uid_t user, gid_t group)
fs/open.c
855
return vfs_fchown(fd_file(f), user, group);
fs/open.c
858
SYSCALL_DEFINE3(fchown, unsigned int, fd, uid_t, user, gid_t, group)
fs/open.c
860
return ksys_fchown(fd, user, group);
fs/orangefs/orangefs-kernel.h
469
sys_attr.group = from_kgid(&init_user_ns, current_fsgid()); \
fs/orangefs/orangefs-utils.c
150
attrs->group = from_kgid(&init_user_ns, inode->i_gid);
fs/orangefs/orangefs-utils.c
152
gossip_debug(GOSSIP_UTILS_DEBUG, "(GID) %d\n", attrs->group);
fs/orangefs/orangefs-utils.c
359
downcall.resp.getattr.attributes.group);
fs/orangefs/protocol.h
228
__u32 group;
fs/smb/client/cifsacl.h
93
struct owner_sid group;
fs/smb/client/readdir.c
285
sid_to_id(cifs_sb, &parsed.group, fattr, SIDGROUP);
fs/smb/client/smb2pdu.c
2380
memcpy(&posix->group, sid, sid_len);
fs/smb/client/smb2pdu.c
2672
sids->group.Revision = 1;
fs/smb/client/smb2pdu.c
2673
sids->group.NumAuth = 3;
fs/smb/client/smb2pdu.c
2674
sids->group.Authority[5] = 5;
fs/smb/client/smb2pdu.c
2675
sids->group.SubAuthorities[0] = cpu_to_le32(88);
fs/smb/client/smb2pdu.c
2676
sids->group.SubAuthorities[1] = cpu_to_le32(2);
fs/smb/client/smb2pdu.c
2677
sids->group.SubAuthorities[2] = cpu_to_le32(current_fsgid().val);
fs/smb/client/smb2pdu.c
2709
group_offset = owner_offset + offsetof(struct owner_group_sids, group);
fs/smb/client/smb2pdu.c
5434
memcpy(&out->group, group_sid, group_len);
fs/smb/client/smb2pdu.h
263
struct smb_sid group; /* var-sized on the wire */
fs/smb/client/smb2pdu.h
307
struct smb_sid group;
fs/smb/server/smbacl.c
321
pace->e_perm = state->group.allow;
fs/smb/server/smbacl.c
538
acl_state.group.allow = (mode & 0070) >> 3;
fs/smb/server/smbacl.c
543
default_acl_state.group.allow = (mode & 0070) >> 3;
fs/smb/server/smbacl.h
76
struct posix_ace_state group;
fs/smb/server/vfs.c
1820
acl_state.group.allow = (inode->i_mode & 0070) >> 3;
fs/smb/server/vfs.c
1827
acl_state.group.allow;
fs/sysfs/file.c
399
const struct attribute *attr, const char *group)
fs/sysfs/file.c
406
if (group) {
fs/sysfs/file.c
407
parent = kernfs_find_and_get(kobj->sd, group);
fs/sysfs/file.c
556
const struct attribute *attr, const char *group)
fs/sysfs/file.c
560
if (group) {
fs/sysfs/file.c
561
parent = kernfs_find_and_get(kobj->sd, group);
fs/xfs/libxfs/xfs_fs.h
1129
struct xfs_health_monitor_group group;
fs/xfs/scrub/health.c
215
switch (type_to_health_flag[sc->sm->sm_type].group) {
fs/xfs/scrub/health.c
295
type_to_health_flag[sc->sm->sm_type].group == XHG_AG)
fs/xfs/scrub/health.c
82
enum xchk_health_group group;
fs/xfs/xfs_discard.c
122
struct xfs_group *xg = busyp->group;
fs/xfs/xfs_extent_busy.c
45
new->group = xfs_group_hold(xg);
fs/xfs/xfs_extent_busy.c
536
struct xfs_extent_busy_tree *eb = busyp->group->xg_busy_extents;
fs/xfs/xfs_extent_busy.c
544
trace_xfs_extent_busy_clear(busyp->group, busyp->bno,
fs/xfs/xfs_extent_busy.c
550
xfs_group_put(busyp->group);
fs/xfs/xfs_extent_busy.c
572
struct xfs_group *xg = xfs_group_hold(busyp->group);
fs/xfs/xfs_extent_busy.c
583
busyp->group == xg);
fs/xfs/xfs_extent_busy.c
693
diff = b1->group->xg_gno - b2->group->xg_gno;
fs/xfs/xfs_extent_busy.h
22
struct xfs_group *group;
fs/xfs/xfs_fsmap.c
1024
info->group = NULL;
fs/xfs/xfs_fsmap.c
1036
info->group = NULL;
fs/xfs/xfs_fsmap.c
1277
info.group = NULL;
fs/xfs/xfs_fsmap.c
164
struct xfs_group *group; /* group info, if applicable */
fs/xfs/xfs_fsmap.c
220
if (!xfs_has_reflink(mp) || !info->group)
fs/xfs/xfs_fsmap.c
223
if (info->group->xg_type == XG_TYPE_RTG)
fs/xfs/xfs_fsmap.c
224
cur = xfs_rtrefcountbt_init_cursor(tp, to_rtg(info->group));
fs/xfs/xfs_fsmap.c
227
to_perag(info->group));
fs/xfs/xfs_fsmap.c
343
info->group ? info->group->xg_gno : NULLAGNUMBER,
fs/xfs/xfs_fsmap.c
535
info->group = pag_group(pag);
fs/xfs/xfs_fsmap.c
585
info->group = NULL;
fs/xfs/xfs_fsmap.c
595
if (info->group) {
fs/xfs/xfs_fsmap.c
597
info->group = NULL;
fs/xfs/xfs_fsmap.c
620
to_perag(info->group));
fs/xfs/xfs_fsmap.c
653
to_perag(info->group));
fs/xfs/xfs_fsmap.c
795
info->group = rtg_group(rtg);
fs/xfs/xfs_fsmap.c
821
info->group = NULL;
fs/xfs/xfs_fsmap.c
827
if (info->group) {
fs/xfs/xfs_fsmap.c
829
info->group = NULL;
fs/xfs/xfs_fsmap.c
863
struct xfs_rtgroup *rtg = to_rtg(info->group);
fs/xfs/xfs_fsmap.c
976
info->group = rtg_group(rtg);
fs/xfs/xfs_healthmon.c
206
if (existing->group == new->group){
fs/xfs/xfs_healthmon.c
492
.group = xg->xg_gno,
fs/xfs/xfs_healthmon.c
778
hme.e.group.mask = xfs_healthmon_rtgroup_mask(event->grpmask);
fs/xfs/xfs_healthmon.c
779
hme.e.group.gno = event->group;
fs/xfs/xfs_healthmon.c
782
hme.e.group.mask = xfs_healthmon_perag_mask(event->grpmask);
fs/xfs/xfs_healthmon.c
783
hme.e.group.gno = event->group;
fs/xfs/xfs_healthmon.h
134
unsigned int group;
fs/xfs/xfs_trace.h
6072
__field(unsigned int, group)
fs/xfs/xfs_trace.h
6082
__entry->group = 0;
fs/xfs/xfs_trace.h
6105
__entry->group = event->group;
fs/xfs/xfs_trace.h
6135
__entry->group,
fs/xfs/xfs_trace.h
6193
__field(unsigned int, group)
fs/xfs/xfs_trace.h
6202
__entry->group = event->group;
fs/xfs/xfs_trace.h
6211
__entry->group)
include/acpi/actbl3.h
382
u8 group;
include/kvm/arm_vgic.h
163
u8 group; /* 0 == group 0, 1 == group 1 */
include/linux/configfs.h
100
static inline void config_group_put(struct config_group *group)
include/linux/configfs.h
102
config_item_put(&group->cg_item);
include/linux/configfs.h
110
struct config_group *group)
include/linux/configfs.h
112
list_add_tail(&new_group->group_entry, &group->default_groups);
include/linux/configfs.h
219
struct config_item *(*make_item)(struct config_group *group, const char *name);
include/linux/configfs.h
220
struct config_group *(*make_group)(struct config_group *group, const char *name);
include/linux/configfs.h
221
void (*disconnect_notify)(struct config_group *group, struct config_item *item);
include/linux/configfs.h
222
void (*drop_item)(struct config_group *group, struct config_item *item);
include/linux/configfs.h
233
static inline struct configfs_subsystem *to_configfs_subsystem(struct config_group *group)
include/linux/configfs.h
235
return group ?
include/linux/configfs.h
236
container_of(group, struct configfs_subsystem, su_group) :
include/linux/configfs.h
244
struct config_group *group);
include/linux/configfs.h
245
void configfs_unregister_group(struct config_group *group);
include/linux/configfs.h
247
void configfs_remove_default_groups(struct config_group *group);
include/linux/configfs.h
253
void configfs_unregister_default_group(struct config_group *group);
include/linux/configfs.h
85
extern void config_group_init(struct config_group *group);
include/linux/configfs.h
86
extern void config_group_init_type_name(struct config_group *group,
include/linux/configfs.h
95
static inline struct config_group *config_group_get(struct config_group *group)
include/linux/configfs.h
97
return group ? to_config_group(config_item_get(&group->cg_item)) : NULL;
include/linux/connector.h
102
u32 group, gfp_t gfp_mask,
include/linux/connector.h
127
int cn_netlink_send(struct cn_msg *msg, u32 portid, u32 group, gfp_t gfp_mask);
include/linux/connector.h
45
u32 seq, group;
include/linux/cs5535.h
176
int cs5535_gpio_set_irq(unsigned group, unsigned irq);
include/linux/cs5535.h
49
static inline int cs5535_pic_unreqz_select_high(unsigned int group,
include/linux/cs5535.h
55
lo &= ~(0xF << (group * 4));
include/linux/cs5535.h
56
lo |= (irq & 0xF) << (group * 4);
include/linux/fanotify.h
8
#define FAN_GROUP_FLAG(group, flag) \
include/linux/fanotify.h
9
((group)->fanotify_data.flags & (flag))
include/linux/fault-inject.h
99
struct config_group group;
include/linux/firewire.h
169
struct attribute_group group;
include/linux/firmware/imx/sci.h
23
int imx_scu_irq_group_enable(u8 group, u32 mask, u8 enable);
include/linux/firmware/imx/sci.h
24
int imx_scu_irq_get_status(u8 group, u32 *irq_status);
include/linux/firmware/imx/sci.h
47
static inline int imx_scu_irq_group_enable(u8 group, u32 mask, u8 enable)
include/linux/firmware/imx/sci.h
52
static inline int imx_scu_irq_get_status(u8 group, u32 *irq_status)
include/linux/fs.h
1823
int vfs_fchown(struct file *file, uid_t user, gid_t group);
include/linux/fsnotify_backend.h
173
int (*handle_event)(struct fsnotify_group *group, u32 mask,
include/linux/fsnotify_backend.h
180
void (*free_group_priv)(struct fsnotify_group *group);
include/linux/fsnotify_backend.h
181
void (*freeing_mark)(struct fsnotify_mark *mark, struct fsnotify_group *group);
include/linux/fsnotify_backend.h
182
void (*free_event)(struct fsnotify_group *group, struct fsnotify_event *event);
include/linux/fsnotify_backend.h
287
static inline void fsnotify_group_lock(struct fsnotify_group *group)
include/linux/fsnotify_backend.h
289
mutex_lock(&group->mark_mutex);
include/linux/fsnotify_backend.h
290
group->owner_flags = memalloc_nofs_save();
include/linux/fsnotify_backend.h
293
static inline void fsnotify_group_unlock(struct fsnotify_group *group)
include/linux/fsnotify_backend.h
295
memalloc_nofs_restore(group->owner_flags);
include/linux/fsnotify_backend.h
296
mutex_unlock(&group->mark_mutex);
include/linux/fsnotify_backend.h
299
static inline void fsnotify_group_assert_locked(struct fsnotify_group *group)
include/linux/fsnotify_backend.h
301
WARN_ON_ONCE(!mutex_is_locked(&group->mark_mutex));
include/linux/fsnotify_backend.h
615
struct fsnotify_group *group;
include/linux/fsnotify_backend.h
714
extern void fsnotify_get_group(struct fsnotify_group *group);
include/linux/fsnotify_backend.h
716
extern void fsnotify_put_group(struct fsnotify_group *group);
include/linux/fsnotify_backend.h
718
extern void fsnotify_group_stop_queueing(struct fsnotify_group *group);
include/linux/fsnotify_backend.h
720
extern void fsnotify_destroy_group(struct fsnotify_group *group);
include/linux/fsnotify_backend.h
724
extern void fsnotify_destroy_event(struct fsnotify_group *group,
include/linux/fsnotify_backend.h
727
extern int fsnotify_insert_event(struct fsnotify_group *group,
include/linux/fsnotify_backend.h
734
static inline int fsnotify_add_event(struct fsnotify_group *group,
include/linux/fsnotify_backend.h
739
return fsnotify_insert_event(group, event, merge, NULL);
include/linux/fsnotify_backend.h
743
static inline void fsnotify_queue_overflow(struct fsnotify_group *group)
include/linux/fsnotify_backend.h
745
fsnotify_add_event(group, group->overflow_event, NULL);
include/linux/fsnotify_backend.h
753
static inline bool fsnotify_notify_queue_is_empty(struct fsnotify_group *group)
include/linux/fsnotify_backend.h
755
assert_spin_locked(&group->notification_lock);
include/linux/fsnotify_backend.h
757
return list_empty(&group->notification_list);
include/linux/fsnotify_backend.h
760
extern bool fsnotify_notify_queue_is_empty(struct fsnotify_group *group);
include/linux/fsnotify_backend.h
762
extern struct fsnotify_event *fsnotify_peek_first_event(struct fsnotify_group *group);
include/linux/fsnotify_backend.h
764
extern struct fsnotify_event *fsnotify_remove_first_event(struct fsnotify_group *group);
include/linux/fsnotify_backend.h
766
extern void fsnotify_remove_queued_event(struct fsnotify_group *group,
include/linux/fsnotify_backend.h
871
struct fsnotify_group *group);
include/linux/fsnotify_backend.h
874
struct fsnotify_group *group);
include/linux/fsnotify_backend.h
899
struct fsnotify_group *group)
include/linux/fsnotify_backend.h
901
return fsnotify_find_mark(inode, FSNOTIFY_OBJ_TYPE_INODE, group);
include/linux/fsnotify_backend.h
906
struct fsnotify_group *group);
include/linux/fsnotify_backend.h
914
extern void fsnotify_clear_marks_by_group(struct fsnotify_group *group,
include/linux/genl_magic_func.h
15
#define GENL_mc_group(group)
include/linux/genl_magic_func.h
261
#define GENL_mc_group(group) { .name = #group, },
include/linux/genl_magic_func.h
267
#define GENL_mc_group(group) CONCATENATE(GENL_MAGIC_FAMILY, _group_ ## group),
include/linux/genl_magic_func.h
272
#define GENL_mc_group(group) \
include/linux/genl_magic_func.h
273
static int CONCATENATE(GENL_MAGIC_FAMILY, _genl_multicast_ ## group)( \
include/linux/genl_magic_func.h
277
CONCATENATE(GENL_MAGIC_FAMILY, _group_ ## group); \
include/linux/genl_magic_func.h
285
#define GENL_mc_group(group)
include/linux/genl_magic_struct.h
124
#define GENL_mc_group(group)
include/linux/hfs_common.h
452
__be32 group;
include/linux/hid.h
652
__u16 group; /* Report group */
include/linux/hid.h
783
.bus = (b), .group = (g), .vendor = (ven), .product = (prod)
include/linux/host1x.h
101
struct iommu_group *group;
include/linux/iio/sw_device.h
27
struct config_group *group;
include/linux/iio/sw_device.h
33
struct config_group group;
include/linux/iio/sw_device.h
45
group);
include/linux/iio/sw_device.h
60
config_group_init_type_name(&d->group, name, type);
include/linux/iio/sw_trigger.h
27
struct config_group *group;
include/linux/iio/sw_trigger.h
33
struct config_group group;
include/linux/iio/sw_trigger.h
45
group);
include/linux/iio/sw_trigger.h
60
config_group_init_type_name(&t->group, name, type);
include/linux/init_syscalls.h
8
int __init init_chown(const char *filename, uid_t user, gid_t group, int flags);
include/linux/iommu.h
1178
int iommu_group_claim_dma_owner(struct iommu_group *group, void *owner);
include/linux/iommu.h
1179
void iommu_group_release_dma_owner(struct iommu_group *group);
include/linux/iommu.h
1180
bool iommu_group_dma_owner_claimed(struct iommu_group *group);
include/linux/iommu.h
1298
static inline int iommu_get_group_resv_regions(struct iommu_group *group,
include/linux/iommu.h
1318
struct iommu_group *group)
include/linux/iommu.h
1324
struct iommu_group *group)
include/linux/iommu.h
1333
static inline void *iommu_group_get_iommudata(struct iommu_group *group)
include/linux/iommu.h
1338
static inline void iommu_group_set_iommudata(struct iommu_group *group,
include/linux/iommu.h
1344
static inline int iommu_group_set_name(struct iommu_group *group,
include/linux/iommu.h
1350
static inline int iommu_group_add_device(struct iommu_group *group,
include/linux/iommu.h
1360
static inline int iommu_group_for_each_dev(struct iommu_group *group,
include/linux/iommu.h
1372
static inline void iommu_group_put(struct iommu_group *group)
include/linux/iommu.h
1376
static inline int iommu_group_id(struct iommu_group *group)
include/linux/iommu.h
1478
iommu_group_claim_dma_owner(struct iommu_group *group, void *owner)
include/linux/iommu.h
1483
static inline void iommu_group_release_dma_owner(struct iommu_group *group)
include/linux/iommu.h
1487
static inline bool iommu_group_dma_owner_claimed(struct iommu_group *group)
include/linux/iommu.h
1671
void iopf_free_group(struct iopf_group *group);
include/linux/iommu.h
1673
void iopf_group_response(struct iopf_group *group,
include/linux/iommu.h
1706
static inline void iopf_free_group(struct iopf_group *group)
include/linux/iommu.h
1716
static inline void iopf_group_response(struct iopf_group *group,
include/linux/iommu.h
231
int (*iopf_handler)(struct iopf_group *group);
include/linux/iommu.h
901
extern bool iommu_group_has_isolated_msi(struct iommu_group *group);
include/linux/iommu.h
941
extern int iommu_get_group_resv_regions(struct iommu_group *group,
include/linux/iommu.h
945
struct iommu_group *group);
include/linux/iommu.h
947
struct iommu_group *group);
include/linux/iommu.h
949
extern void *iommu_group_get_iommudata(struct iommu_group *group);
include/linux/iommu.h
950
extern void iommu_group_set_iommudata(struct iommu_group *group,
include/linux/iommu.h
953
extern int iommu_group_set_name(struct iommu_group *group, const char *name);
include/linux/iommu.h
954
extern int iommu_group_add_device(struct iommu_group *group,
include/linux/iommu.h
957
extern int iommu_group_for_each_dev(struct iommu_group *group, void *data,
include/linux/iommu.h
960
extern struct iommu_group *iommu_group_ref_get(struct iommu_group *group);
include/linux/iommu.h
961
extern void iommu_group_put(struct iommu_group *group);
include/linux/iommu.h
963
extern int iommu_group_id(struct iommu_group *group);
include/linux/iommufd.h
204
bool iommufd_ctx_has_group(struct iommufd_ctx *ictx, struct iommu_group *group);
include/linux/irqchip/arm-gic-v4.h
136
bool group;
include/linux/irqchip/arm-gic-v4.h
151
int its_prop_update_vsgi(int irq, u8 priority, bool group);
include/linux/irqchip/arm-gic-v4.h
67
bool group;
include/linux/mcb.h
69
int group;
include/linux/memory.h
156
struct memory_group *group);
include/linux/memory.h
91
struct memory_group *group; /* group (if any) for this block */
include/linux/memory_hotplug.h
106
struct memory_group *group,
include/linux/memory_hotplug.h
113
struct zone *zone, struct memory_group *group);
include/linux/memory_hotplug.h
265
struct zone *zone, struct memory_group *group);
include/linux/memory_hotplug.h
274
struct zone *zone, struct memory_group *group)
include/linux/memory_hotplug.h
314
struct memory_group *group, unsigned long start_pfn,
include/linux/mlx5/mlx5_ifc.h
11964
u8 group[0x4];
include/linux/mmc/host.h
753
int mmc_sd_switch(struct mmc_card *card, bool mode, int group,
include/linux/mod_devicetable.h
175
__u16 group;
include/linux/mtd/mtd.h
145
int group;
include/linux/netdevice.h
2235
int group;
include/linux/netfilter/nfnetlink.h
54
int nfnetlink_has_listeners(struct net *net, unsigned int group);
include/linux/netfilter/nfnetlink.h
56
unsigned int group, int echo, gfp_t flags);
include/linux/netfilter/nfnetlink.h
57
int nfnetlink_set_err(struct net *net, u32 portid, u32 group, int error);
include/linux/netfilter/nfnetlink.h
60
__u32 group, gfp_t allocation);
include/linux/netfs.h
237
struct netfs_group *group; /* Writeback group being written back */
include/linux/netlink.h
223
void __netlink_clear_multicast_users(struct sock *sk, unsigned int group);
include/linux/netlink.h
226
int netlink_has_listeners(struct sock *sk, unsigned int group);
include/linux/netlink.h
231
__u32 group, gfp_t allocation);
include/linux/netlink.h
236
__u32 portid, __u32 group, gfp_t allocation,
include/linux/netlink.h
239
int netlink_set_err(struct sock *ssk, __u32 portid, __u32 group, int code);
include/linux/netlink.h
51
int (*bind)(struct net *net, int group);
include/linux/netlink.h
52
void (*unbind)(struct net *net, int group);
include/linux/pci-ep-cfs.h
16
void pci_ep_cfs_remove_epc_group(struct config_group *group);
include/linux/pci-ep-cfs.h
18
void pci_ep_cfs_remove_epf_group(struct config_group *group);
include/linux/pci-ep-cfs.h
25
static inline void pci_ep_cfs_remove_epc_group(struct config_group *group)
include/linux/pci-ep-cfs.h
34
static inline void pci_ep_cfs_remove_epf_group(struct config_group *group)
include/linux/pci-epc.h
179
struct config_group *group;
include/linux/pci-epf.h
218
struct config_group *group;
include/linux/pci-epf.h
69
struct config_group *group);
include/linux/pinctrl/machine.h
34
const char *group;
include/linux/pinctrl/machine.h
94
.group = grp, \
include/linux/platform_data/x86/simatic-ipc.h
48
u8 group;
include/linux/platform_data/x86/simatic-ipc.h
57
data_entry->group == SIMATIC_IPC_DMI_GROUP &&
include/linux/psi.h
25
int psi_show(struct seq_file *s, struct psi_group *group, enum psi_res res);
include/linux/psi.h
26
struct psi_trigger *psi_trigger_create(struct psi_group *group, char *buf,
include/linux/psi.h
43
void psi_cgroup_restart(struct psi_group *group);
include/linux/psi.h
65
static inline void psi_cgroup_restart(struct psi_group *group) {}
include/linux/psi_types.h
133
struct psi_group *group;
include/linux/rtnetlink.h
12
extern int rtnetlink_send(struct sk_buff *skb, struct net *net, u32 pid, u32 group, int echo);
include/linux/rtnetlink.h
15
u32 pid, u32 group, int echo)
include/linux/rtnetlink.h
17
return !skb ? 0 : rtnetlink_send(skb, net, pid, group, echo);
include/linux/rtnetlink.h
22
u32 group, const struct nlmsghdr *nlh, gfp_t flags);
include/linux/rtnetlink.h
221
static inline int rtnl_has_listeners(const struct net *net, u32 group)
include/linux/rtnetlink.h
225
return netlink_has_listeners(rtnl, group);
include/linux/rtnetlink.h
23
extern void rtnl_set_sk_err(struct net *net, u32 group, int error);
include/linux/rtnetlink.h
238
rtnl_notify_needed(const struct net *net, u16 nlflags, u32 group)
include/linux/rtnetlink.h
240
return (nlflags & NLM_F_ECHO) || rtnl_has_listeners(net, group);
include/linux/scc.h
41
unsigned char group; /* Group ID for AX.25 TX interlocking */
include/linux/scmi_protocol.h
802
u32 group);
include/linux/sock_diag.h
91
const enum sknetlink_groups group = sock_diag_destroy_group(sk);
include/linux/sock_diag.h
93
return group != SKNLGRP_NONE && n->diag_nlsk &&
include/linux/sock_diag.h
94
netlink_has_listeners(n->diag_nlsk, group);
include/linux/syscalls.h
1045
uid_t user, gid_t group);
include/linux/syscalls.h
1049
uid_t user, gid_t group);
include/linux/syscalls.h
1158
old_uid_t user, old_gid_t group);
include/linux/syscalls.h
1160
old_uid_t user, old_gid_t group);
include/linux/syscalls.h
1161
asmlinkage long sys_fchown16(unsigned int fd, old_uid_t user, old_gid_t group);
include/linux/syscalls.h
1236
int ksys_fchown(unsigned int fd, uid_t user, gid_t group);
include/linux/syscalls.h
1271
gid_t group, int flag);
include/linux/syscalls.h
1274
gid_t group)
include/linux/syscalls.h
1276
return do_fchownat(AT_FDCWD, filename, user, group, 0);
include/linux/syscalls.h
1280
gid_t group)
include/linux/syscalls.h
1282
return do_fchownat(AT_FDCWD, filename, user, group,
include/linux/syscalls.h
462
gid_t group, int flag);
include/linux/syscalls.h
463
asmlinkage long sys_fchown(unsigned int fd, uid_t user, gid_t group);
include/linux/sysfs.h
458
const struct attribute *attr, const char *group);
include/linux/sysfs.h
460
const struct attribute *attr, const char *group);
include/linux/sysfs.h
660
const struct attribute *attr, const char *group)
include/linux/sysfs.h
666
const struct attribute *attr, const char *group)
include/linux/usb/composite.h
580
struct config_group group;
include/linux/usb/composite.h
94
struct config_group group;
include/linux/usb/gadget_configfs.h
45
struct config_group *group, \
include/linux/usb/gadget_configfs.h
61
config_group_init_type_name(&new->group, name, \
include/linux/usb/gadget_configfs.h
64
gi = container_of(group, struct struct_member, strings_group); \
include/linux/usb/gadget_configfs.h
76
return &new->group; \
include/linux/usb/gadget_configfs.h
83
struct config_group *group, \
include/linux/user_namespace.h
83
kgid_t group;
include/linux/vdpa.h
442
int (*set_group_asid)(struct vdpa_device *vdev, unsigned int group,
include/linux/vfio.h
48
struct vfio_group *group;
include/linux/virtio.h
52
struct vduse_vq_group *group;
include/media/ipu-bridge.h
132
const struct software_node *group[SWNODE_COUNT + 1];
include/net/addrconf.h
274
bool ipv6_chk_mcast_addr(struct net_device *dev, const struct in6_addr *group,
include/net/devlink.h
1467
const struct devlink_trap_group *group);
include/net/devlink.h
1475
const struct devlink_trap_group *group,
include/net/devlink.h
1485
const struct devlink_trap_group *group,
include/net/genetlink.h
334
struct genl_info *info, u32 group, gfp_t flags);
include/net/genetlink.h
488
u32 portid, unsigned int group, gfp_t flags,
include/net/genetlink.h
492
if (WARN_ON_ONCE(group >= family->n_mcgrps))
include/net/genetlink.h
494
group = family->mcgrp_offset + group;
include/net/genetlink.h
495
return nlmsg_multicast_filtered(net->genl_sock, skb, portid, group,
include/net/genetlink.h
510
u32 portid, unsigned int group, gfp_t flags)
include/net/genetlink.h
513
group, flags, NULL, NULL);
include/net/genetlink.h
526
unsigned int group, gfp_t flags)
include/net/genetlink.h
529
portid, group, flags);
include/net/genetlink.h
543
unsigned int group);
include/net/genetlink.h
628
u32 group, int code)
include/net/genetlink.h
630
if (WARN_ON_ONCE(group >= family->n_mcgrps))
include/net/genetlink.h
632
group = family->mcgrp_offset + group;
include/net/genetlink.h
633
return netlink_set_err(net->genl_sock, portid, group, code);
include/net/genetlink.h
637
struct net *net, unsigned int group)
include/net/genetlink.h
639
if (WARN_ON_ONCE(group >= family->n_mcgrps))
include/net/genetlink.h
641
group = family->mcgrp_offset + group;
include/net/genetlink.h
642
return netlink_has_listeners(net->genl_sock, group);
include/net/net_shaper.h
86
int (*group)(struct net_shaper_binding *binding, int leaves_count,
include/net/netfilter/nf_log.h
28
u_int16_t group;
include/net/netlink.h
1156
u32 portid, unsigned int group,
include/net/netlink.h
1163
NETLINK_CB(skb).dst_group = group;
include/net/netlink.h
1165
err = netlink_broadcast_filtered(sk, skb, portid, group, flags,
include/net/netlink.h
1182
u32 portid, unsigned int group, gfp_t flags)
include/net/netlink.h
1184
return nlmsg_multicast_filtered(sk, skb, portid, group, flags,
include/net/netlink.h
539
unsigned int group, int report, gfp_t flags);
include/net/psample.h
34
void psample_group_take(struct psample_group *group);
include/net/psample.h
35
void psample_group_put(struct psample_group *group);
include/net/psample.h
41
void psample_sample_packet(struct psample_group *group,
include/net/psample.h
47
static inline void psample_sample_packet(struct psample_group *group,
include/rdma/rdma_netlink.h
106
unsigned int group, gfp_t flags);
include/rdma/rdma_netlink.h
113
bool rdma_nl_chk_listeners(unsigned int group);
include/sound/core.h
298
const struct attribute_group *group);
include/sound/pcm.h
488
struct snd_pcm_group *group; /* pointer to current group */
include/sound/pcm.h
634
return substream->group != &substream->self_group;
include/sound/pcm.h
697
list_for_each_entry(s, &substream->group->substreams, link_list)
include/sound/sdca_function.h
1190
struct sdca_entity *group;
include/sound/ump.h
17
int group; /* group index (0-based) */
include/sound/ump.h
243
static inline u32 ump_compose(unsigned char type, unsigned char group,
include/sound/ump.h
246
return ((u32)type << 28) | ((u32)group << 24) | ((u32)status << 20) |
include/sound/ump_convert.h
37
void snd_ump_convert_to_ump(struct ump_cvt_to_ump *cvt, unsigned char group,
include/sound/ump_msg.h
132
u32 group:4;
include/sound/ump_msg.h
142
u32 group:4;
include/sound/ump_msg.h
151
u32 group:4;
include/sound/ump_msg.h
161
u32 group:4;
include/sound/ump_msg.h
170
u32 group:4;
include/sound/ump_msg.h
180
u32 group:4;
include/sound/ump_msg.h
189
u32 group:4;
include/sound/ump_msg.h
199
u32 group:4;
include/sound/ump_msg.h
208
u32 group:4;
include/sound/ump_msg.h
218
u32 group:4;
include/sound/ump_msg.h
227
u32 group:4;
include/sound/ump_msg.h
237
u32 group:4;
include/sound/ump_msg.h
246
u32 group:4;
include/sound/ump_msg.h
254
u32 group:4;
include/sound/ump_msg.h
276
u32 group:4;
include/sound/ump_msg.h
290
u32 group:4;
include/sound/ump_msg.h
303
u32 group:4;
include/sound/ump_msg.h
316
u32 group:4;
include/sound/ump_msg.h
328
u32 group:4;
include/sound/ump_msg.h
341
u32 group:4;
include/sound/ump_msg.h
353
u32 group:4;
include/sound/ump_msg.h
366
u32 group:4;
include/sound/ump_msg.h
378
u32 group:4;
include/sound/ump_msg.h
391
u32 group:4;
include/sound/ump_msg.h
403
u32 group:4;
include/sound/ump_msg.h
416
u32 group:4;
include/sound/ump_msg.h
428
u32 group:4;
include/sound/ump_msg.h
444
u32 group:4;
include/sound/ump_msg.h
459
u32 group:4;
include/sound/ump_msg.h
470
u32 group:4;
include/sound/ump_msg.h
482
u32 group:4;
include/sound/ump_msg.h
493
u32 group:4;
include/sound/ump_msg.h
505
u32 group:4;
include/sound/ump_msg.h
518
u32 group:4;
include/trace/events/bridge.h
128
const struct br_ip *group),
include/trace/events/bridge.h
130
TP_ARGS(dev, group),
include/trace/events/bridge.h
145
__entry->vid = group->vid;
include/trace/events/bridge.h
147
if (!group->proto) {
include/trace/events/bridge.h
152
memcpy(__entry->grpmac, group->dst.mac_addr, ETH_ALEN);
include/trace/events/bridge.h
153
} else if (group->proto == htons(ETH_P_IP)) {
include/trace/events/bridge.h
157
ipv6_addr_set_v4mapped(group->src.ip4, in6);
include/trace/events/bridge.h
160
ipv6_addr_set_v4mapped(group->dst.ip4, in6);
include/trace/events/bridge.h
169
*in6 = group->src.ip6;
include/trace/events/bridge.h
172
*in6 = group->dst.ip6;
include/trace/events/ext4.h
1169
ext4_group_t group,
include/trace/events/ext4.h
1173
TP_ARGS(sb, inode, group, start, len),
include/trace/events/ext4.h
1187
__entry->result_group = group;
include/trace/events/ext4.h
1202
ext4_group_t group,
include/trace/events/ext4.h
1206
TP_ARGS(sb, inode, group, start, len)
include/trace/events/ext4.h
1213
ext4_group_t group,
include/trace/events/ext4.h
1217
TP_ARGS(sb, inode, group, start, len)
include/trace/events/ext4.h
1345
TP_PROTO(struct super_block *sb, unsigned long group),
include/trace/events/ext4.h
1347
TP_ARGS(sb, group),
include/trace/events/ext4.h
1351
__field( __u32, group )
include/trace/events/ext4.h
1357
__entry->group = group;
include/trace/events/ext4.h
1362
__entry->group)
include/trace/events/ext4.h
1367
TP_PROTO(struct super_block *sb, unsigned long group),
include/trace/events/ext4.h
1369
TP_ARGS(sb, group)
include/trace/events/ext4.h
1374
TP_PROTO(struct super_block *sb, unsigned long group),
include/trace/events/ext4.h
1376
TP_ARGS(sb, group)
include/trace/events/ext4.h
1381
TP_PROTO(struct super_block *sb, unsigned long group),
include/trace/events/ext4.h
1383
TP_ARGS(sb, group)
include/trace/events/ext4.h
1387
TP_PROTO(struct super_block *sb, unsigned long group, bool prefetch),
include/trace/events/ext4.h
1389
TP_ARGS(sb, group, prefetch),
include/trace/events/ext4.h
1393
__field( __u32, group )
include/trace/events/ext4.h
1400
__entry->group = group;
include/trace/events/ext4.h
1406
__entry->group, __entry->prefetch)
include/trace/events/ext4.h
1876
ext4_group_t group,
include/trace/events/ext4.h
1880
TP_ARGS(sb, group, start, len),
include/trace/events/ext4.h
1885
__field( __u32, group )
include/trace/events/ext4.h
1893
__entry->group = group;
include/trace/events/ext4.h
1900
__entry->group, __entry->start, __entry->len)
include/trace/events/ext4.h
1906
ext4_group_t group,
include/trace/events/ext4.h
1910
TP_ARGS(sb, group, start, len)
include/trace/events/ext4.h
1916
ext4_group_t group,
include/trace/events/ext4.h
1920
TP_ARGS(sb, group, start, len)
include/trace/events/ext4.h
2678
TP_PROTO(struct super_block *sb, ext4_group_t group,
include/trace/events/ext4.h
2681
TP_ARGS(sb, group, next, prefetch_ios),
include/trace/events/ext4.h
2685
__field( __u32, group )
include/trace/events/ext4.h
2692
__entry->group = group;
include/trace/events/ext4.h
2699
__entry->group, __entry->next, __entry->ios)
include/trace/events/ext4.h
2703
TP_PROTO(struct super_block *sb, ext4_group_t group),
include/trace/events/ext4.h
2705
TP_ARGS(sb, group),
include/trace/events/ext4.h
2709
__field( __u32, group )
include/trace/events/ext4.h
2714
__entry->group = group;
include/trace/events/ext4.h
2718
MAJOR(__entry->dev), MINOR(__entry->dev), __entry->group)
include/trace/events/signal.h
53
int group, int result),
include/trace/events/signal.h
55
TP_ARGS(sig, info, task, group, result),
include/trace/events/signal.h
63
__field( int, group )
include/trace/events/signal.h
72
__entry->group = group;
include/trace/events/signal.h
78
__entry->comm, __entry->pid, __entry->group,
include/trace/events/timer_migration.h
109
__entry->group = group;
include/trace/events/timer_migration.h
110
__entry->parent = group->parent;
include/trace/events/timer_migration.h
111
__entry->lvl = group->level;
include/trace/events/timer_migration.h
112
__entry->numa_node = group->numa_node;
include/trace/events/timer_migration.h
120
__entry->group, __entry->lvl, __entry->numa_node,
include/trace/events/timer_migration.h
127
TP_PROTO(struct tmigr_group *group, union tmigr_state state, u32 childmask),
include/trace/events/timer_migration.h
129
TP_ARGS(group, state, childmask)
include/trace/events/timer_migration.h
134
TP_PROTO(struct tmigr_group *group, union tmigr_state state, u32 childmask),
include/trace/events/timer_migration.h
136
TP_ARGS(group, state, childmask)
include/trace/events/timer_migration.h
14
TP_PROTO(struct tmigr_group *group),
include/trace/events/timer_migration.h
16
TP_ARGS(group),
include/trace/events/timer_migration.h
19
__field( void *, group )
include/trace/events/timer_migration.h
237
TP_PROTO(struct tmigr_group *child, struct tmigr_group *group,
include/trace/events/timer_migration.h
241
TP_ARGS(child, group, childstate, groupstate, nextevt),
include/trace/events/timer_migration.h
245
__field( void *, group )
include/trace/events/timer_migration.h
25
__entry->group = group;
include/trace/events/timer_migration.h
257
__entry->group = group;
include/trace/events/timer_migration.h
259
__entry->group_next_expiry = group->next_expiry;
include/trace/events/timer_migration.h
26
__entry->lvl = group->level;
include/trace/events/timer_migration.h
261
__entry->group_lvl = group->level;
include/trace/events/timer_migration.h
269
__entry->child, __entry->group, __entry->group_lvl, __entry->child_active,
include/trace/events/timer_migration.h
27
__entry->numa_node = group->numa_node;
include/trace/events/timer_migration.h
277
TP_PROTO(struct tmigr_group *group),
include/trace/events/timer_migration.h
279
TP_ARGS(group),
include/trace/events/timer_migration.h
282
__field( void * , group )
include/trace/events/timer_migration.h
287
__entry->group = group;
include/trace/events/timer_migration.h
288
__entry->lvl = group->level;
include/trace/events/timer_migration.h
292
__entry->group, __entry->lvl)
include/trace/events/timer_migration.h
31
__entry->group, __entry->lvl, __entry->numa_node)
include/trace/events/timer_migration.h
94
TP_PROTO(struct tmigr_group *group, union tmigr_state state, u32 childmask),
include/trace/events/timer_migration.h
96
TP_ARGS(group, state, childmask),
include/trace/events/timer_migration.h
99
__field( void *, group )
include/uapi/linux/batadv_packet.h
201
__be16 group; /* group id */
include/uapi/linux/ext4.h
104
__u32 group; /* Group number for this data */
include/uapi/linux/igmp.h
35
__be32 group;
include/uapi/linux/igmp.h
67
__be32 group;
include/uapi/linux/kvm.h
1181
__u32 group; /* device-defined */
include/uapi/linux/netfilter/xt_NFLOG.h
18
__u16 group;
include/uapi/linux/netfilter_bridge/ebt_nflog.h
17
__u16 group;
include/uapi/linux/netlink.h
179
__u32 group;
include/uapi/linux/vduse.h
139
__u32 group;
include/uapi/linux/vduse.h
177
__u32 group;
kernel/audit.c
1638
static void audit_log_multicast(int group, const char *op, int err)
kernel/audit.c
1665
audit_log_format(ab, " nl-mcgrp=%d op=%s res=%d", group, op, !err);
kernel/audit.c
1670
static int audit_multicast_bind(struct net *net, int group)
kernel/audit.c
1676
audit_log_multicast(group, "connect", err);
kernel/audit.c
1680
static void audit_multicast_unbind(struct net *net, int group)
kernel/audit.c
1682
audit_log_multicast(group, "disconnect", 0);
kernel/audit_fsnotify.c
165
if (WARN_ON_ONCE(inode_mark->group != audit_fsnotify_group))
kernel/audit_tree.c
1050
struct fsnotify_group *group)
kernel/audit_tree.c
1054
fsnotify_group_lock(mark->group);
kernel/audit_tree.c
1059
fsnotify_group_unlock(mark->group);
kernel/audit_watch.c
479
if (WARN_ON_ONCE(inode_mark->group != audit_watch_group))
kernel/crash_dump_dm_crypt.c
245
static struct config_item *config_keys_make_item(struct config_group *group,
kernel/events/core.c
13890
CLASS(fd, group)(group_fd); // group_fd == -1 => empty
kernel/events/core.c
13892
if (!is_perf_file(group)) {
kernel/events/core.c
13896
group_leader = fd_file(group)->private_data;
kernel/events/core.c
4775
bool group;
kernel/events/core.c
4832
if (data->group)
kernel/events/core.c
4838
if (!data->group) {
kernel/events/core.c
4956
static int perf_event_read(struct perf_event *event, bool group)
kernel/events/core.c
4983
.group = group,
kernel/events/core.c
5022
if (group)
kernel/sched/fair.c
10134
struct sched_group *group, *sdg = sd->groups;
kernel/sched/fair.c
10170
group = child->groups;
kernel/sched/fair.c
10172
struct sched_group_capacity *sgc = group->sgc;
kernel/sched/fair.c
10177
group = group->next;
kernel/sched/fair.c
10178
} while (group != child->groups);
kernel/sched/fair.c
10233
static inline int sg_imbalanced(struct sched_group *group)
kernel/sched/fair.c
10235
return group->sgc->imbalance;
kernel/sched/fair.c
10294
struct sched_group *group,
kernel/sched/fair.c
10300
if (sg_imbalanced(group))
kernel/sched/fair.c
10363
sched_group_asym(struct lb_env *env, struct sg_lb_stats *sgs, struct sched_group *group)
kernel/sched/fair.c
10369
if ((group->flags & SD_SHARE_CPUCAPACITY) &&
kernel/sched/fair.c
10373
return sched_asym(env->sd, env->dst_cpu, READ_ONCE(group->asym_prefer_cpu));
kernel/sched/fair.c
10388
struct sched_group *group)
kernel/sched/fair.c
10399
if (group->flags & SD_SHARE_CPUCAPACITY &&
kernel/sched/fair.c
10465
struct sched_group *group,
kernel/sched/fair.c
10475
local_group = group == sds->local;
kernel/sched/fair.c
10477
for_each_cpu_and(i, sched_group_span(group), env->cpus) {
kernel/sched/fair.c
10528
sgs->group_capacity = group->sgc->capacity;
kernel/sched/fair.c
10530
sgs->group_weight = group->group_weight;
kernel/sched/fair.c
10534
sched_group_asym(env, sgs, group))
kernel/sched/fair.c
10538
if (!local_group && smt_balance(env, sgs, group))
kernel/sched/fair.c
10541
sgs->group_type = group_classify(env->sd->imbalance_pct, group, sgs);
kernel/sched/fair.c
10784
struct sched_group *group,
kernel/sched/fair.c
10796
for_each_cpu_and(i, sched_group_span(group), p->cpus_ptr) {
kernel/sched/fair.c
10823
sgs->group_capacity = group->sgc->capacity;
kernel/sched/fair.c
10825
sgs->group_weight = group->group_weight;
kernel/sched/fair.c
10827
sgs->group_type = group_classify(sd->imbalance_pct, group, sgs);
kernel/sched/fair.c
10841
struct sched_group *group,
kernel/sched/fair.c
10871
if (idlest->sgc->max_capacity >= group->sgc->max_capacity)
kernel/sched/fair.c
10900
struct sched_group *idlest = NULL, *local = NULL, *group = sd->groups;
kernel/sched/fair.c
10913
if (!cpumask_intersects(sched_group_span(group),
kernel/sched/fair.c
10918
if (!sched_group_cookie_match(cpu_rq(this_cpu), p, group))
kernel/sched/fair.c
10922
sched_group_span(group));
kernel/sched/fair.c
10926
local = group;
kernel/sched/fair.c
10931
update_sg_wakeup_stats(sd, group, sgs, p);
kernel/sched/fair.c
10933
if (!local_group && update_pick_idlest(idlest, &idlest_sgs, group, sgs)) {
kernel/sched/fair.c
10934
idlest = group;
kernel/sched/fair.c
10938
} while (group = group->next, group != sd->groups);
kernel/sched/fair.c
11548
struct sched_group *group)
kernel/sched/fair.c
11555
for_each_cpu_and(i, sched_group_span(group), env->cpus) {
kernel/sched/fair.c
11871
struct sched_group *group;
kernel/sched/fair.c
11906
group = sched_balance_find_src_group(&env);
kernel/sched/fair.c
11907
if (!group) {
kernel/sched/fair.c
11912
busiest = sched_balance_find_src_rq(&env, group);
kernel/sched/fair.c
1715
static inline unsigned long group_faults_cpu(struct numa_group *group, int nid)
kernel/sched/fair.c
1717
return group->faults[task_faults_idx(NUMA_CPU, nid, 0)] +
kernel/sched/fair.c
1718
group->faults[task_faults_idx(NUMA_CPU, nid, 1)];
kernel/sched/fair.c
7446
sched_balance_find_dst_group_cpu(struct sched_group *group, struct task_struct *p, int this_cpu)
kernel/sched/fair.c
7456
if (group->group_weight == 1)
kernel/sched/fair.c
7457
return cpumask_first(sched_group_span(group));
kernel/sched/fair.c
7460
for_each_cpu_and(i, sched_group_span(group), p->cpus_ptr) {
kernel/sched/fair.c
7518
struct sched_group *group;
kernel/sched/fair.c
7527
group = sched_balance_find_dst_group(sd, p, cpu);
kernel/sched/fair.c
7528
if (!group) {
kernel/sched/fair.c
7533
new_cpu = sched_balance_find_dst_group_cpu(group, p, cpu);
kernel/sched/psi.c
1033
for_each_group(group, task_psi_group(curr)) {
kernel/sched/psi.c
1034
if (!group->enabled)
kernel/sched/psi.c
1037
groupc = per_cpu_ptr(group->pcpu, cpu);
kernel/sched/psi.c
1042
if (group->rtpoll_states & (1 << PSI_IRQ_FULL))
kernel/sched/psi.c
1043
psi_schedule_rtpoll_work(group, 1, false);
kernel/sched/psi.c
1210
void psi_cgroup_restart(struct psi_group *group)
kernel/sched/psi.c
1229
if (!group->enabled)
kernel/sched/psi.c
1239
psi_group_change(group, cpu, 0, 0, now, true);
kernel/sched/psi.c
1245
int psi_show(struct seq_file *m, struct psi_group *group, enum psi_res res)
kernel/sched/psi.c
1260
mutex_lock(&group->avgs_lock);
kernel/sched/psi.c
1262
collect_percpu_times(group, PSI_AVGS, NULL);
kernel/sched/psi.c
1263
if (now >= group->avg_next_update)
kernel/sched/psi.c
1264
group->avg_next_update = update_averages(group, now);
kernel/sched/psi.c
1265
mutex_unlock(&group->avgs_lock);
kernel/sched/psi.c
1277
if (!(group == &psi_system && res == PSI_CPU && full)) {
kernel/sched/psi.c
1279
avg[w] = group->avg[res * 2 + full][w];
kernel/sched/psi.c
1280
total = div_u64(group->total[PSI_AVGS][res * 2 + full],
kernel/sched/psi.c
1295
struct psi_trigger *psi_trigger_create(struct psi_group *group, char *buf,
kernel/sched/psi.c
1347
t->group = group;
kernel/sched/psi.c
1352
group->total[PSI_POLL][t->state], 0);
kernel/sched/psi.c
1363
mutex_lock(&group->rtpoll_trigger_lock);
kernel/sched/psi.c
1365
if (!rcu_access_pointer(group->rtpoll_task)) {
kernel/sched/psi.c
1368
task = kthread_create(psi_rtpoll_worker, group, "psimon");
kernel/sched/psi.c
1371
mutex_unlock(&group->rtpoll_trigger_lock);
kernel/sched/psi.c
1374
atomic_set(&group->rtpoll_wakeup, 0);
kernel/sched/psi.c
1376
rcu_assign_pointer(group->rtpoll_task, task);
kernel/sched/psi.c
1379
list_add(&t->node, &group->rtpoll_triggers);
kernel/sched/psi.c
1380
group->rtpoll_min_period = min(group->rtpoll_min_period,
kernel/sched/psi.c
1382
group->rtpoll_nr_triggers[t->state]++;
kernel/sched/psi.c
1383
group->rtpoll_states |= (1 << t->state);
kernel/sched/psi.c
1385
mutex_unlock(&group->rtpoll_trigger_lock);
kernel/sched/psi.c
1387
mutex_lock(&group->avgs_lock);
kernel/sched/psi.c
1389
list_add(&t->node, &group->avg_triggers);
kernel/sched/psi.c
1390
group->avg_nr_triggers[t->state]++;
kernel/sched/psi.c
1392
mutex_unlock(&group->avgs_lock);
kernel/sched/psi.c
1399
struct psi_group *group;
kernel/sched/psi.c
1409
group = t->group;
kernel/sched/psi.c
1421
mutex_lock(&group->avgs_lock);
kernel/sched/psi.c
1424
group->avg_nr_triggers[t->state]--;
kernel/sched/psi.c
1426
mutex_unlock(&group->avgs_lock);
kernel/sched/psi.c
1428
mutex_lock(&group->rtpoll_trigger_lock);
kernel/sched/psi.c
1434
group->rtpoll_nr_triggers[t->state]--;
kernel/sched/psi.c
1435
if (!group->rtpoll_nr_triggers[t->state])
kernel/sched/psi.c
1436
group->rtpoll_states &= ~(1 << t->state);
kernel/sched/psi.c
1441
if (group->rtpoll_min_period == div_u64(t->win.size, UPDATES_PER_WINDOW)) {
kernel/sched/psi.c
1442
list_for_each_entry(tmp, &group->rtpoll_triggers, node)
kernel/sched/psi.c
1445
group->rtpoll_min_period = period;
kernel/sched/psi.c
1448
if (group->rtpoll_states == 0) {
kernel/sched/psi.c
1449
group->rtpoll_until = 0;
kernel/sched/psi.c
1451
group->rtpoll_task,
kernel/sched/psi.c
1452
lockdep_is_held(&group->rtpoll_trigger_lock));
kernel/sched/psi.c
1453
rcu_assign_pointer(group->rtpoll_task, NULL);
kernel/sched/psi.c
1454
timer_delete(&group->rtpoll_timer);
kernel/sched/psi.c
1457
mutex_unlock(&group->rtpoll_trigger_lock);
kernel/sched/psi.c
1477
atomic_set(&group->rtpoll_scheduled, 0);
kernel/sched/psi.c
205
static void group_init(struct psi_group *group)
kernel/sched/psi.c
207
group->enabled = true;
kernel/sched/psi.c
208
group->avg_last_update = sched_clock();
kernel/sched/psi.c
209
group->avg_next_update = group->avg_last_update + psi_period;
kernel/sched/psi.c
210
mutex_init(&group->avgs_lock);
kernel/sched/psi.c
213
INIT_LIST_HEAD(&group->avg_triggers);
kernel/sched/psi.c
214
memset(group->avg_nr_triggers, 0, sizeof(group->avg_nr_triggers));
kernel/sched/psi.c
215
INIT_DELAYED_WORK(&group->avgs_work, psi_avgs_work);
kernel/sched/psi.c
218
atomic_set(&group->rtpoll_scheduled, 0);
kernel/sched/psi.c
219
mutex_init(&group->rtpoll_trigger_lock);
kernel/sched/psi.c
220
INIT_LIST_HEAD(&group->rtpoll_triggers);
kernel/sched/psi.c
221
group->rtpoll_min_period = U32_MAX;
kernel/sched/psi.c
222
group->rtpoll_next_update = ULLONG_MAX;
kernel/sched/psi.c
223
init_waitqueue_head(&group->rtpoll_wait);
kernel/sched/psi.c
224
timer_setup(&group->rtpoll_timer, poll_timer_fn, 0);
kernel/sched/psi.c
225
rcu_assign_pointer(group->rtpoll_task, NULL);
kernel/sched/psi.c
271
static void get_recent_times(struct psi_group *group, int cpu,
kernel/sched/psi.c
275
struct psi_group_cpu *groupc = per_cpu_ptr(group->pcpu, cpu);
kernel/sched/psi.c
327
if (current_work() == &group->avgs_work.work) {
kernel/sched/psi.c
362
static void collect_percpu_times(struct psi_group *group,
kernel/sched/psi.c
385
get_recent_times(group, cpu, aggregator, times,
kernel/sched/psi.c
410
group->total[aggregator][s] +=
kernel/sched/psi.c
463
static void update_triggers(struct psi_group *group, u64 now,
kernel/sched/psi.c
467
u64 *total = group->total[aggregator];
kernel/sched/psi.c
472
triggers = &group->avg_triggers;
kernel/sched/psi.c
473
aggregator_total = group->avg_total;
kernel/sched/psi.c
475
triggers = &group->rtpoll_triggers;
kernel/sched/psi.c
476
aggregator_total = group->rtpoll_total;
kernel/sched/psi.c
525
static u64 update_averages(struct psi_group *group, u64 now)
kernel/sched/psi.c
533
expires = group->avg_next_update;
kernel/sched/psi.c
545
period = now - (group->avg_last_update + (missed_periods * psi_period));
kernel/sched/psi.c
546
group->avg_last_update = now;
kernel/sched/psi.c
551
sample = group->total[PSI_AVGS][s] - group->avg_total[s];
kernel/sched/psi.c
571
group->avg_total[s] += sample;
kernel/sched/psi.c
572
calc_avgs(group->avg[s], missed_periods, sample, period);
kernel/sched/psi.c
581
struct psi_group *group;
kernel/sched/psi.c
586
group = container_of(dwork, struct psi_group, avgs_work);
kernel/sched/psi.c
588
mutex_lock(&group->avgs_lock);
kernel/sched/psi.c
592
collect_percpu_times(group, PSI_AVGS, &changed_states);
kernel/sched/psi.c
600
if (now >= group->avg_next_update) {
kernel/sched/psi.c
601
update_triggers(group, now, PSI_AVGS);
kernel/sched/psi.c
602
group->avg_next_update = update_averages(group, now);
kernel/sched/psi.c
607
group->avg_next_update - now) + 1);
kernel/sched/psi.c
610
mutex_unlock(&group->avgs_lock);
kernel/sched/psi.c
613
static void init_rtpoll_triggers(struct psi_group *group, u64 now)
kernel/sched/psi.c
617
list_for_each_entry(t, &group->rtpoll_triggers, node)
kernel/sched/psi.c
619
group->total[PSI_POLL][t->state], 0);
kernel/sched/psi.c
620
memcpy(group->rtpoll_total, group->total[PSI_POLL],
kernel/sched/psi.c
621
sizeof(group->rtpoll_total));
kernel/sched/psi.c
622
group->rtpoll_next_update = now + group->rtpoll_min_period;
kernel/sched/psi.c
626
static void psi_schedule_rtpoll_work(struct psi_group *group, unsigned long delay,
kernel/sched/psi.c
635
if (atomic_xchg(&group->rtpoll_scheduled, 1) && !force)
kernel/sched/psi.c
640
task = rcu_dereference(group->rtpoll_task);
kernel/sched/psi.c
646
mod_timer(&group->rtpoll_timer, jiffies + delay);
kernel/sched/psi.c
648
atomic_set(&group->rtpoll_scheduled, 0);
kernel/sched/psi.c
653
static void psi_rtpoll_work(struct psi_group *group)
kernel/sched/psi.c
659
mutex_lock(&group->rtpoll_trigger_lock);
kernel/sched/psi.c
663
if (now > group->rtpoll_until) {
kernel/sched/psi.c
673
atomic_set(&group->rtpoll_scheduled, 0);
kernel/sched/psi.c
700
collect_percpu_times(group, PSI_POLL, &changed_states);
kernel/sched/psi.c
702
if (changed_states & group->rtpoll_states) {
kernel/sched/psi.c
704
if (now > group->rtpoll_until)
kernel/sched/psi.c
705
init_rtpoll_triggers(group, now);
kernel/sched/psi.c
712
group->rtpoll_until = now +
kernel/sched/psi.c
713
group->rtpoll_min_period * UPDATES_PER_WINDOW;
kernel/sched/psi.c
716
if (now > group->rtpoll_until) {
kernel/sched/psi.c
717
group->rtpoll_next_update = ULLONG_MAX;
kernel/sched/psi.c
721
if (now >= group->rtpoll_next_update) {
kernel/sched/psi.c
722
if (changed_states & group->rtpoll_states) {
kernel/sched/psi.c
723
update_triggers(group, now, PSI_POLL);
kernel/sched/psi.c
724
memcpy(group->rtpoll_total, group->total[PSI_POLL],
kernel/sched/psi.c
725
sizeof(group->rtpoll_total));
kernel/sched/psi.c
727
group->rtpoll_next_update = now + group->rtpoll_min_period;
kernel/sched/psi.c
730
psi_schedule_rtpoll_work(group,
kernel/sched/psi.c
731
nsecs_to_jiffies(group->rtpoll_next_update - now) + 1,
kernel/sched/psi.c
735
mutex_unlock(&group->rtpoll_trigger_lock);
kernel/sched/psi.c
740
struct psi_group *group = (struct psi_group *)data;
kernel/sched/psi.c
745
wait_event_interruptible(group->rtpoll_wait,
kernel/sched/psi.c
746
atomic_cmpxchg(&group->rtpoll_wakeup, 1, 0) ||
kernel/sched/psi.c
751
psi_rtpoll_work(group);
kernel/sched/psi.c
758
struct psi_group *group = timer_container_of(group, t, rtpoll_timer);
kernel/sched/psi.c
760
atomic_set(&group->rtpoll_wakeup, 1);
kernel/sched/psi.c
761
wake_up_interruptible(&group->rtpoll_wait);
kernel/sched/psi.c
793
#define for_each_group(iter, group) \
kernel/sched/psi.c
794
for (typeof(group) iter = group; iter; iter = iter->parent)
kernel/sched/psi.c
796
static void psi_group_change(struct psi_group *group, int cpu,
kernel/sched/psi.c
805
groupc = per_cpu_ptr(group->pcpu, cpu);
kernel/sched/psi.c
845
if (!group->enabled) {
kernel/sched/psi.c
878
if (state_mask & group->rtpoll_states)
kernel/sched/psi.c
879
psi_schedule_rtpoll_work(group, 1, false);
kernel/sched/psi.c
881
if (wake_clock && !delayed_work_pending(&group->avgs_work))
kernel/sched/psi.c
882
schedule_delayed_work(&group->avgs_work, PSI_FREQ);
kernel/sched/psi.c
921
for_each_group(group, task_psi_group(task))
kernel/sched/psi.c
922
psi_group_change(group, cpu, clear, set, now, true);
kernel/sched/psi.c
943
for_each_group(group, task_psi_group(next)) {
kernel/sched/psi.c
944
struct psi_group_cpu *groupc = per_cpu_ptr(group->pcpu, cpu);
kernel/sched/psi.c
947
common = group;
kernel/sched/psi.c
950
psi_group_change(group, cpu, 0, TSK_ONCPU, now, true);
kernel/sched/psi.c
984
for_each_group(group, task_psi_group(prev)) {
kernel/sched/psi.c
985
if (group == common)
kernel/sched/psi.c
987
psi_group_change(group, cpu, clear, set, now, wake_clock);
kernel/sched/psi.c
998
for_each_group(group, common)
kernel/sched/psi.c
999
psi_group_change(group, cpu, clear, set, now, wake_clock);
kernel/sched/sched.h
1514
struct sched_group *group)
kernel/sched/sched.h
1522
for_each_cpu_and(cpu, sched_group_span(group), p->cpus_ptr) {
kernel/sched/sched.h
1575
struct sched_group *group)
kernel/sched/topology.c
102
group->sgc->id,
kernel/sched/topology.c
103
cpumask_pr_args(sched_group_span(group)));
kernel/sched/topology.c
106
!cpumask_equal(group_balance_mask(group), sched_group_span(group))) {
kernel/sched/topology.c
108
cpumask_pr_args(group_balance_mask(group)));
kernel/sched/topology.c
111
if (group->sgc->capacity != SCHED_CAPACITY_SCALE)
kernel/sched/topology.c
112
printk(KERN_CONT " cap=%lu", group->sgc->capacity);
kernel/sched/topology.c
114
if (group == sd->groups && sd->child &&
kernel/sched/topology.c
116
sched_group_span(group))) {
kernel/sched/topology.c
122
group = group->next;
kernel/sched/topology.c
124
if (group != sd->groups)
kernel/sched/topology.c
127
} while (group != sd->groups);
kernel/sched/topology.c
46
struct sched_group *group = sd->groups;
kernel/sched/topology.c
59
if (group && !cpumask_test_cpu(cpu, sched_group_span(group))) {
kernel/sched/topology.c
80
if (!group) {
kernel/sched/topology.c
86
if (cpumask_empty(sched_group_span(group))) {
kernel/sched/topology.c
93
cpumask_intersects(groupmask, sched_group_span(group))) {
kernel/sched/topology.c
99
cpumask_or(groupmask, groupmask, sched_group_span(group));
kernel/time/timer_migration.c
1040
static bool tmigr_handle_remote_up(struct tmigr_group *group,
kernel/time/timer_migration.c
1054
trace_tmigr_handle_remote(group);
kernel/time/timer_migration.c
1061
if (!tmigr_check_migrator(group, childmask))
kernel/time/timer_migration.c
1064
raw_spin_lock_irq(&group->lock);
kernel/time/timer_migration.c
1066
evt = tmigr_next_expired_groupevt(group, now);
kernel/time/timer_migration.c
1071
raw_spin_unlock_irq(&group->lock);
kernel/time/timer_migration.c
1084
data->firstexp = group->next_expiry;
kernel/time/timer_migration.c
1086
raw_spin_unlock_irq(&group->lock);
kernel/time/timer_migration.c
1139
static bool tmigr_requires_handle_remote_up(struct tmigr_group *group,
kernel/time/timer_migration.c
1152
if (!tmigr_check_migrator(group, childmask))
kernel/time/timer_migration.c
1161
data->firstexp = READ_ONCE(group->next_expiry);
kernel/time/timer_migration.c
1167
raw_spin_lock(&group->lock);
kernel/time/timer_migration.c
1168
data->firstexp = group->next_expiry;
kernel/time/timer_migration.c
1169
if (data->now >= group->next_expiry) {
kernel/time/timer_migration.c
1171
raw_spin_unlock(&group->lock);
kernel/time/timer_migration.c
1174
raw_spin_unlock(&group->lock);
kernel/time/timer_migration.c
1272
static bool tmigr_inactive_up(struct tmigr_group *group,
kernel/time/timer_migration.c
1289
curstate.state = atomic_read_acquire(&group->migr_state);
kernel/time/timer_migration.c
1327
if (atomic_try_cmpxchg(&group->migr_state, &curstate.state, newstate.state)) {
kernel/time/timer_migration.c
1328
trace_tmigr_group_set_cpu_inactive(group, newstate, childmask);
kernel/time/timer_migration.c
1344
tmigr_update_events(group, child, data);
kernel/time/timer_migration.c
1424
struct tmigr_group *group = tmc->tmgroup;
kernel/time/timer_migration.c
1436
if (!tmigr_check_lonely(group))
kernel/time/timer_migration.c
1445
nextevt = min_t(u64, nextevt, READ_ONCE(group->next_expiry));
kernel/time/timer_migration.c
1446
group = group->parent;
kernel/time/timer_migration.c
1447
} while (group);
kernel/time/timer_migration.c
1632
static void tmigr_init_group(struct tmigr_group *group, unsigned int lvl,
kernel/time/timer_migration.c
1637
raw_spin_lock_init(&group->lock);
kernel/time/timer_migration.c
1639
group->level = lvl;
kernel/time/timer_migration.c
1640
group->numa_node = lvl < tmigr_crossnode_level ? node : NUMA_NO_NODE;
kernel/time/timer_migration.c
1642
group->num_children = 0;
kernel/time/timer_migration.c
1647
atomic_set(&group->migr_state, s.state);
kernel/time/timer_migration.c
1649
timerqueue_init_head(&group->events);
kernel/time/timer_migration.c
1650
timerqueue_init(&group->groupevt.nextevt);
kernel/time/timer_migration.c
1651
group->groupevt.nextevt.expires = KTIME_MAX;
kernel/time/timer_migration.c
1652
WRITE_ONCE(group->next_expiry, KTIME_MAX);
kernel/time/timer_migration.c
1653
group->groupevt.ignore = true;
kernel/time/timer_migration.c
1658
struct tmigr_group *tmp, *group = NULL;
kernel/time/timer_migration.c
1682
group = tmp;
kernel/time/timer_migration.c
1686
if (group)
kernel/time/timer_migration.c
1687
return group;
kernel/time/timer_migration.c
1690
group = kzalloc_node(sizeof(*group), GFP_KERNEL, node);
kernel/time/timer_migration.c
1691
if (!group)
kernel/time/timer_migration.c
1694
tmigr_init_group(group, lvl, node);
kernel/time/timer_migration.c
1697
list_add(&group->list, &tmigr_level_list[lvl]);
kernel/time/timer_migration.c
1698
trace_tmigr_group_set(group);
kernel/time/timer_migration.c
1699
return group;
kernel/time/timer_migration.c
1702
static bool tmigr_init_root(struct tmigr_group *group, bool activate)
kernel/time/timer_migration.c
1704
if (!group->parent && group != tmigr_root) {
kernel/time/timer_migration.c
1710
group->groupmask = BIT(0);
kernel/time/timer_migration.c
1763
struct tmigr_group *group, *child, **stack;
kernel/time/timer_migration.c
1780
group = tmigr_get_group(node, i);
kernel/time/timer_migration.c
1781
if (IS_ERR(group)) {
kernel/time/timer_migration.c
1782
err = PTR_ERR(group);
kernel/time/timer_migration.c
1788
stack[i] = group;
kernel/time/timer_migration.c
1799
if (group->parent)
kernel/time/timer_migration.c
1811
group = stack[i];
kernel/time/timer_migration.c
1814
list_del(&group->list);
kernel/time/timer_migration.c
1815
kfree(group);
kernel/time/timer_migration.c
1819
WARN_ON_ONCE(i != group->level);
kernel/time/timer_migration.c
1827
tmc->tmgroup = group;
kernel/time/timer_migration.c
1828
tmc->groupmask = BIT(group->num_children++);
kernel/time/timer_migration.c
1830
tmigr_init_root(group, activate);
kernel/time/timer_migration.c
1838
tmigr_connect_child_parent(child, group, activate);
kernel/time/timer_migration.c
1880
group = list_first_entry(&tmigr_level_list[top],
kernel/time/timer_migration.c
1881
typeof(*group), list);
kernel/time/timer_migration.c
1882
WARN_ON_ONCE(group->parent);
kernel/time/timer_migration.c
1887
tmigr_root = group;
kernel/time/timer_migration.c
477
static bool tmigr_check_migrator(struct tmigr_group *group, u8 childmask)
kernel/time/timer_migration.c
481
s.state = atomic_read(&group->migr_state);
kernel/time/timer_migration.c
489
static bool tmigr_check_migrator_and_lonely(struct tmigr_group *group, u8 childmask)
kernel/time/timer_migration.c
495
s.state = atomic_read(&group->migr_state);
kernel/time/timer_migration.c
506
static bool tmigr_check_lonely(struct tmigr_group *group)
kernel/time/timer_migration.c
511
s.state = atomic_read(&group->migr_state);
kernel/time/timer_migration.c
560
struct tmigr_group *child, struct tmigr_group *group)
kernel/time/timer_migration.c
563
WARN_ON_ONCE(group->level >= tmigr_hierarchy_levels);
kernel/time/timer_migration.c
565
if (up(group, child, data))
kernel/time/timer_migration.c
568
child = group;
kernel/time/timer_migration.c
573
group = READ_ONCE(group->parent);
kernel/time/timer_migration.c
576
} while (group);
kernel/time/timer_migration.c
598
static struct tmigr_event *tmigr_next_groupevt(struct tmigr_group *group)
kernel/time/timer_migration.c
603
lockdep_assert_held(&group->lock);
kernel/time/timer_migration.c
605
WRITE_ONCE(group->next_expiry, KTIME_MAX);
kernel/time/timer_migration.c
607
while ((node = timerqueue_getnext(&group->events))) {
kernel/time/timer_migration.c
611
WRITE_ONCE(group->next_expiry, evt->nextevt.expires);
kernel/time/timer_migration.c
619
if (!timerqueue_del(&group->events, node))
kernel/time/timer_migration.c
631
static struct tmigr_event *tmigr_next_expired_groupevt(struct tmigr_group *group,
kernel/time/timer_migration.c
634
struct tmigr_event *evt = tmigr_next_groupevt(group);
kernel/time/timer_migration.c
642
timerqueue_del(&group->events, &evt->nextevt);
kernel/time/timer_migration.c
643
tmigr_next_groupevt(group);
kernel/time/timer_migration.c
648
static u64 tmigr_next_groupevt_expires(struct tmigr_group *group)
kernel/time/timer_migration.c
652
evt = tmigr_next_groupevt(group);
kernel/time/timer_migration.c
660
static bool tmigr_active_up(struct tmigr_group *group,
kernel/time/timer_migration.c
674
curstate.state = atomic_read(&group->migr_state);
kernel/time/timer_migration.c
690
} while (!atomic_try_cmpxchg(&group->migr_state, &curstate.state, newstate.state));
kernel/time/timer_migration.c
692
trace_tmigr_group_set_cpu_active(group, newstate, childmask);
kernel/time/timer_migration.c
706
WRITE_ONCE(group->groupevt.ignore, true);
kernel/time/timer_migration.c
760
bool tmigr_update_events(struct tmigr_group *group, struct tmigr_group *child,
kernel/time/timer_migration.c
772
raw_spin_lock_nested(&group->lock, SINGLE_DEPTH_NESTING);
kernel/time/timer_migration.c
775
groupstate.state = atomic_read(&group->migr_state);
kernel/time/timer_migration.c
824
if (ignore && !remote && group->parent)
kernel/time/timer_migration.c
827
raw_spin_lock(&group->lock);
kernel/time/timer_migration.c
830
groupstate.state = atomic_read(&group->migr_state);
kernel/time/timer_migration.c
844
if (!timerqueue_del(&group->events, &evt->nextevt))
kernel/time/timer_migration.c
845
WRITE_ONCE(group->next_expiry, KTIME_MAX);
kernel/time/timer_migration.c
868
if (timerqueue_add(&group->events, &evt->nextevt))
kernel/time/timer_migration.c
869
WRITE_ONCE(group->next_expiry, nextexp);
kernel/time/timer_migration.c
873
if (!group->parent && (groupstate.migrator == TMIGR_NONE)) {
kernel/time/timer_migration.c
892
data->firstexp = tmigr_next_groupevt_expires(group);
kernel/time/timer_migration.c
895
trace_tmigr_update_events(child, group, childstate, groupstate,
kernel/time/timer_migration.c
899
raw_spin_unlock(&group->lock);
kernel/time/timer_migration.c
907
static bool tmigr_new_timer_up(struct tmigr_group *group,
kernel/time/timer_migration.c
911
return tmigr_update_events(group, child, data);
kernel/trace/trace_boot.c
468
const char *p, *group, *event;
kernel/trace/trace_boot.c
470
group = xbc_node_get_data(gnode);
kernel/trace/trace_boot.c
473
if (!strcmp(group, "kprobes"))
kernel/trace/trace_boot.c
476
if (!strcmp(group, "synthetic"))
kernel/trace/trace_boot.c
481
file = find_event_file(tr, group, event);
kernel/trace/trace_boot.c
483
pr_err("Failed to find event: %s:%s\n", group, event);
kernel/trace/trace_boot.c
511
group, event);
kernel/trace/trace_eprobe.c
198
static struct trace_eprobe *alloc_event_probe(const char *group,
kernel/trace/trace_eprobe.c
227
ret = trace_probe_init(&ep->tp, this_event, group, false, nargs);
kernel/trace/trace_eprobe.c
847
const char *event = NULL, *group = EPROBE_EVENT_SYSTEM;
kernel/trace/trace_eprobe.c
867
ret = traceprobe_parse_event_name(&event, &group, gbuf,
kernel/trace/trace_eprobe.c
910
ep = alloc_event_probe(group, event, event_call, argc - 2);
kernel/trace/trace_events_user.c
1010
return user_event_parse(group, name, args, flags, newuser, reg_flags);
kernel/trace/trace_events_user.c
1510
static struct user_event *find_user_event(struct user_event_group *group,
kernel/trace/trace_events_user.c
1519
hash_for_each_possible(group->register_table, user, node, key) {
kernel/trace/trace_events_user.c
164
struct user_event_group *group;
kernel/trace/trace_events_user.c
1764
struct user_event_group *group;
kernel/trace/trace_events_user.c
1780
group = current_user_event_group();
kernel/trace/trace_events_user.c
1782
if (!group) {
kernel/trace/trace_events_user.c
1787
mutex_lock(&group->reg_mutex);
kernel/trace/trace_events_user.c
1790
ret = user_event_parse_cmd(group, name, &user, USER_EVENT_REG_PERSIST);
kernel/trace/trace_events_user.c
1795
mutex_unlock(&group->reg_mutex);
kernel/trace/trace_events_user.c
1920
match = strcmp(system, user->group->system_name) == 0 ||
kernel/trace/trace_events_user.c
1921
strcmp(system, user->group->system_multi_name) == 0;
kernel/trace/trace_events_user.c
1957
lockdep_assert_held(&user->group->reg_mutex);
kernel/trace/trace_events_user.c
1963
user->reg_name, user->group->multi_id);
kernel/trace/trace_events_user.c
1972
user->group->multi_id++;
kernel/trace/trace_events_user.c
204
static int user_event_parse(struct user_event_group *group, char *name,
kernel/trace/trace_events_user.c
2061
static int user_event_parse(struct user_event_group *group, char *name,
kernel/trace/trace_events_user.c
2087
user = find_user_event(group, name, argc, (const char **)argv,
kernel/trace/trace_events_user.c
2117
user->group = group;
kernel/trace/trace_events_user.c
2143
user->class.system = group->system_multi_name;
kernel/trace/trace_events_user.c
2145
user->class.system = group->system_name;
kernel/trace/trace_events_user.c
2177
hash_add(group->register_table, &user->node, key);
kernel/trace/trace_events_user.c
2202
static int delete_user_event(struct user_event_group *group, char *name)
kernel/trace/trace_events_user.c
2210
hash_for_each_possible_safe(group->register_table, user, tmp, node, key) {
kernel/trace/trace_events_user.c
2310
struct user_event_group *group;
kernel/trace/trace_events_user.c
2313
group = current_user_event_group();
kernel/trace/trace_events_user.c
2315
if (!group)
kernel/trace/trace_events_user.c
2323
info->group = group;
kernel/trace/trace_events_user.c
2352
struct user_event_group *group = info->group;
kernel/trace/trace_events_user.c
2357
lockdep_is_held(&group->reg_mutex));
kernel/trace/trace_events_user.c
2483
ret = user_event_parse_cmd(info->group, name, &user, reg.flags);
kernel/trace/trace_events_user.c
2542
ret = delete_user_event(info->group, name);
kernel/trace/trace_events_user.c
2672
struct user_event_group *group = info->group;
kernel/trace/trace_events_user.c
2677
mutex_lock(&group->reg_mutex);
kernel/trace/trace_events_user.c
2679
mutex_unlock(&group->reg_mutex);
kernel/trace/trace_events_user.c
2683
mutex_lock(&group->reg_mutex);
kernel/trace/trace_events_user.c
2685
mutex_unlock(&group->reg_mutex);
kernel/trace/trace_events_user.c
2689
mutex_lock(&group->reg_mutex);
kernel/trace/trace_events_user.c
2691
mutex_unlock(&group->reg_mutex);
kernel/trace/trace_events_user.c
2704
struct user_event_group *group;
kernel/trace/trace_events_user.c
2711
group = info->group;
kernel/trace/trace_events_user.c
2717
mutex_lock(&group->reg_mutex);
kernel/trace/trace_events_user.c
2735
mutex_unlock(&group->reg_mutex);
kernel/trace/trace_events_user.c
2771
struct user_event_group *group = m->private;
kernel/trace/trace_events_user.c
2776
if (!group)
kernel/trace/trace_events_user.c
2779
mutex_lock(&group->reg_mutex);
kernel/trace/trace_events_user.c
2781
hash_for_each(group->register_table, i, user, node) {
kernel/trace/trace_events_user.c
2801
mutex_unlock(&group->reg_mutex);
kernel/trace/trace_events_user.c
2819
struct user_event_group *group;
kernel/trace/trace_events_user.c
2822
group = current_user_event_group();
kernel/trace/trace_events_user.c
2824
if (!group)
kernel/trace/trace_events_user.c
2833
m->private = group;
kernel/trace/trace_events_user.c
337
static void user_event_group_destroy(struct user_event_group *group)
kernel/trace/trace_events_user.c
339
kfree(group->system_name);
kernel/trace/trace_events_user.c
340
kfree(group->system_multi_name);
kernel/trace/trace_events_user.c
341
kfree(group);
kernel/trace/trace_events_user.c
371
struct user_event_group *group;
kernel/trace/trace_events_user.c
373
group = kzalloc_obj(*group);
kernel/trace/trace_events_user.c
375
if (!group)
kernel/trace/trace_events_user.c
378
group->system_name = user_event_group_system_name();
kernel/trace/trace_events_user.c
380
if (!group->system_name)
kernel/trace/trace_events_user.c
383
group->system_multi_name = user_event_group_system_multi_name();
kernel/trace/trace_events_user.c
385
if (!group->system_multi_name)
kernel/trace/trace_events_user.c
388
mutex_init(&group->reg_mutex);
kernel/trace/trace_events_user.c
389
hash_init(group->register_table);
kernel/trace/trace_events_user.c
391
return group;
kernel/trace/trace_events_user.c
393
if (group)
kernel/trace/trace_events_user.c
394
user_event_group_destroy(group);
kernel/trace/trace_events_user.c
84
struct user_event_group *group;
kernel/trace/trace_events_user.c
994
static int user_event_parse_cmd(struct user_event_group *group,
kernel/trace/trace_fprobe.c
1241
const char *event = NULL, *group = FPROBE_EVENT_SYSTEM;
kernel/trace/trace_fprobe.c
1259
group = TRACEPOINT_EVENT_SYSTEM;
kernel/trace/trace_fprobe.c
1283
ret = traceprobe_parse_event_name(&event, &group, gbuf,
kernel/trace/trace_fprobe.c
1351
tf = alloc_trace_fprobe(group, event, symbol, argc, is_return, is_tracepoint);
kernel/trace/trace_fprobe.c
573
static struct trace_fprobe *alloc_trace_fprobe(const char *group,
kernel/trace/trace_fprobe.c
597
ret = trace_probe_init(&tf->tp, event, group, false, nargs);
kernel/trace/trace_fprobe.c
606
const char *group)
kernel/trace/trace_fprobe.c
613
strcmp(trace_probe_group_name(&tf->tp), group) == 0)
kernel/trace/trace_kprobe.c
1032
tk = alloc_trace_kprobe(group, event, addr, symbol, offset, maxactive,
kernel/trace/trace_kprobe.c
1762
const char *group = event->tp_event->class->system;
kernel/trace/trace_kprobe.c
1766
tk = find_trace_kprobe(pevent, group);
kernel/trace/trace_kprobe.c
267
static struct trace_kprobe *alloc_trace_kprobe(const char *group,
kernel/trace/trace_kprobe.c
304
ret = trace_probe_init(&tk->tp, event, group, false, nargs);
kernel/trace/trace_kprobe.c
313
const char *group)
kernel/trace/trace_kprobe.c
320
strcmp(trace_probe_group_name(&tk->tp), group) == 0)
kernel/trace/trace_kprobe.c
865
const char *event = NULL, *group = KPROBE_EVENT_SYSTEM;
kernel/trace/trace_kprobe.c
984
ret = traceprobe_parse_event_name(&event, &group, gbuf,
kernel/trace/trace_probe.c
2058
const char *group, bool alloc_filter, int nargs)
kernel/trace/trace_probe.c
2064
if (!event || !group)
kernel/trace/trace_probe.c
2088
tp->event->class.system = kstrdup(group, GFP_KERNEL);
kernel/trace/trace_probe.h
356
const char *group, bool alloc_filter, int nargs);
kernel/trace/trace_uprobe.c
1473
const char *group = event->tp_event->class->system;
kernel/trace/trace_uprobe.c
1477
tu = find_probe_event(pevent, group);
kernel/trace/trace_uprobe.c
336
alloc_trace_uprobe(const char *group, const char *event, int nargs, bool is_ret)
kernel/trace/trace_uprobe.c
351
ret = trace_probe_init(&tu->tp, event, group, true, nargs);
kernel/trace/trace_uprobe.c
381
static struct trace_uprobe *find_probe_event(const char *event, const char *group)
kernel/trace/trace_uprobe.c
388
strcmp(trace_probe_group_name(&tu->tp), group) == 0)
kernel/trace/trace_uprobe.c
547
const char *event = NULL, *group = UPROBE_EVENT_SYSTEM;
kernel/trace/trace_uprobe.c
659
ret = traceprobe_parse_event_name(&event, &group, gbuf,
kernel/trace/trace_uprobe.c
688
tu = alloc_trace_uprobe(group, event, argc, is_return);
kernel/uid16.c
119
old_gid_t group;
kernel/uid16.c
124
group = high2lowgid(from_kgid_munged(user_ns, kgid));
kernel/uid16.c
125
if (put_user(group, grouplist+i))
kernel/uid16.c
137
old_gid_t group;
kernel/uid16.c
141
if (get_user(group, grouplist+i))
kernel/uid16.c
144
kgid = make_kgid(user_ns, low2highgid(group));
kernel/uid16.c
23
SYSCALL_DEFINE3(chown16, const char __user *, filename, old_uid_t, user, old_gid_t, group)
kernel/uid16.c
25
return ksys_chown(filename, low2highuid(user), low2highgid(group));
kernel/uid16.c
28
SYSCALL_DEFINE3(lchown16, const char __user *, filename, old_uid_t, user, old_gid_t, group)
kernel/uid16.c
30
return ksys_lchown(filename, low2highuid(user), low2highgid(group));
kernel/uid16.c
33
SYSCALL_DEFINE3(fchown16, unsigned int, fd, old_uid_t, user, old_gid_t, group)
kernel/uid16.c
35
return ksys_fchown(fd, low2highuid(user), low2highgid(group));
kernel/user.c
70
.group = GLOBAL_ROOT_GID,
kernel/user_namespace.c
115
!kgid_has_mapping(parent_ns, group))
kernel/user_namespace.c
137
ns->group = group;
kernel/user_namespace.c
87
kgid_t group = new->egid;
kernel/watchdog.c
497
int i, group;
kernel/watchdog.c
512
group = (tail + i) % NUM_SAMPLE_PERIODS;
kernel/watchdog.c
515
__this_cpu_read(cpustat_util[group][STATS_SYSTEM]),
kernel/watchdog.c
516
__this_cpu_read(cpustat_util[group][STATS_SOFTIRQ]),
kernel/watchdog.c
517
__this_cpu_read(cpustat_util[group][STATS_HARDIRQ]),
kernel/watchdog.c
518
__this_cpu_read(cpustat_util[group][STATS_IDLE]));
lib/fault-inject.c
358
return container_of(to_config_group(item), struct fault_config, group);
lib/fault-inject.c
458
config_group_init_type_name(&config->group, name, &fault_config_type);
lib/tests/printf_kunit.c
689
const struct software_node *group[] = { &first, &second, &third, NULL };
lib/tests/printf_kunit.c
696
rval = software_node_register_node_group(group);
lib/tests/printf_kunit.c
707
software_node_unregister_node_group(group);
mm/memory_hotplug.c
1011
if (!auto_movable_can_online_movable(NUMA_NO_NODE, group, nr_pages))
mm/memory_hotplug.c
1016
!auto_movable_can_online_movable(nid, group, nr_pages))
mm/memory_hotplug.c
1050
struct memory_group *group, unsigned long start_pfn,
mm/memory_hotplug.c
1060
return auto_movable_zone_for_pfn(nid, group, start_pfn, nr_pages);
mm/memory_hotplug.c
1069
void adjust_present_page_count(struct page *page, struct memory_group *group,
mm/memory_hotplug.c
1084
if (group && movable)
mm/memory_hotplug.c
1085
group->present_movable_pages += nr_pages;
mm/memory_hotplug.c
1086
else if (group && !movable)
mm/memory_hotplug.c
1087
group->present_kernel_pages += nr_pages;
mm/memory_hotplug.c
1145
struct zone *zone, struct memory_group *group)
mm/memory_hotplug.c
1208
adjust_present_page_count(pfn_to_page(pfn), group, nr_pages);
mm/memory_hotplug.c
1444
static int create_altmaps_and_memory_blocks(int nid, struct memory_group *group,
mm/memory_hotplug.c
1477
params.altmap, group);
mm/memory_hotplug.c
1502
struct memory_group *group = NULL;
mm/memory_hotplug.c
1515
group = memory_group_find_by_id(nid);
mm/memory_hotplug.c
1516
if (!group)
mm/memory_hotplug.c
1518
nid = group->nid;
mm/memory_hotplug.c
1554
ret = create_altmaps_and_memory_blocks(nid, group, start, size);
mm/memory_hotplug.c
1563
ret = create_memory_block_devices(start, size, nid, NULL, group);
mm/memory_hotplug.c
1902
struct zone *zone, struct memory_group *group)
mm/memory_hotplug.c
2062
adjust_present_page_count(pfn_to_page(start_pfn), group, -nr_pages);
mm/memory_hotplug.c
816
static int auto_movable_stats_account_group(struct memory_group *group,
mm/memory_hotplug.c
834
pages = group->present_movable_pages * 100 / ratio;
mm/memory_hotplug.c
835
pages -= group->present_kernel_pages;
mm/memory_hotplug.c
839
stats->movable_pages += group->present_movable_pages;
mm/memory_hotplug.c
843
static bool auto_movable_can_online_movable(int nid, struct memory_group *group,
mm/memory_hotplug.c
876
group, &group_stats);
mm/memory_hotplug.c
882
if (group && group->is_dynamic)
mm/memory_hotplug.c
883
kernel_early_pages += group->present_kernel_pages;
mm/memory_hotplug.c
965
struct memory_group *group,
mm/memory_hotplug.c
975
if (group && !group->is_dynamic) {
mm/memory_hotplug.c
976
max_pages = group->s.max_pages;
mm/memory_hotplug.c
977
online_pages = group->present_movable_pages;
mm/memory_hotplug.c
980
if (group->present_kernel_pages)
mm/memory_hotplug.c
982
} else if (!group || group->d.unit_pages == nr_pages) {
mm/memory_hotplug.c
985
max_pages = group->d.unit_pages;
mm/memory_hotplug.c
992
pfn = ALIGN_DOWN(pfn, group->d.unit_pages);
mm/memory_hotplug.c
993
end_pfn = pfn + group->d.unit_pages;
mm/percpu.c
2466
int group, v;
mm/percpu.c
2486
for (group = 0; group < ai->nr_groups; group++) {
mm/percpu.c
2487
const struct pcpu_group_info *gi = &ai->groups[group];
mm/percpu.c
2497
pr_cont("[%0*d] ", group_width, group);
mm/percpu.c
2574
int group, unit, i;
mm/percpu.c
2625
for (group = 0, unit = 0; group < ai->nr_groups; group++, unit += i) {
mm/percpu.c
2626
const struct pcpu_group_info *gi = &ai->groups[group];
mm/percpu.c
2628
group_offsets[group] = gi->base_offset;
mm/percpu.c
2629
group_sizes[group] = gi->nr_units * ai->unit_size;
mm/percpu.c
2817
int last_allocs, group, unit;
mm/percpu.c
2850
for (group = 0; !cpumask_empty(&mask); group++) {
mm/percpu.c
2853
group_map[cpu] = group;
mm/percpu.c
2854
group_cnt[group]++;
mm/percpu.c
2861
group_map[tcpu] = group;
mm/percpu.c
2862
group_cnt[group]++;
mm/percpu.c
2867
nr_groups = group;
mm/percpu.c
2882
for (group = 0; group < nr_groups; group++) {
mm/percpu.c
2883
int this_allocs = DIV_ROUND_UP(group_cnt[group], upa);
mm/percpu.c
2885
wasted += this_allocs * upa - group_cnt[group];
mm/percpu.c
2906
for (group = 0; group < nr_groups; group++)
mm/percpu.c
2907
nr_units += roundup(group_cnt[group], upa);
mm/percpu.c
2914
for (group = 0; group < nr_groups; group++) {
mm/percpu.c
2915
ai->groups[group].cpu_map = cpu_map;
mm/percpu.c
2916
cpu_map += roundup(group_cnt[group], upa);
mm/percpu.c
2926
for (group = 0, unit = 0; group < nr_groups; group++) {
mm/percpu.c
2927
struct pcpu_group_info *gi = &ai->groups[group];
mm/percpu.c
2937
if (group_map[cpu] == group)
mm/percpu.c
3026
int group, i, highest_group, rc = 0;
mm/percpu.c
3044
for (group = 0; group < ai->nr_groups; group++) {
mm/percpu.c
3045
struct pcpu_group_info *gi = &ai->groups[group];
mm/percpu.c
3061
areas[group] = ptr;
mm/percpu.c
3065
highest_group = group;
mm/percpu.c
3086
for (group = 0; group < ai->nr_groups; group++) {
mm/percpu.c
3087
struct pcpu_group_info *gi = &ai->groups[group];
mm/percpu.c
3088
void *ptr = areas[group];
mm/percpu.c
3103
for (group = 0; group < ai->nr_groups; group++) {
mm/percpu.c
3104
ai->groups[group].base_offset = areas[group] - base;
mm/percpu.c
3115
for (group = 0; group < ai->nr_groups; group++)
mm/percpu.c
3116
if (areas[group])
mm/percpu.c
3117
pcpu_fc_free(areas[group],
mm/percpu.c
3118
ai->groups[group].nr_units * ai->unit_size);
net/9p/trans_usbg.c
792
func_inst.group);
net/9p/trans_usbg.c
933
config_group_init_type_name(&usb9pfs_opts->func_inst.group, "",
net/batman-adv/bridge_loop_avoidance.c
1043
if (bla_dst->group == bla_dst_own->group)
net/batman-adv/bridge_loop_avoidance.c
1056
if (ntohs(bla_dst->group) > ntohs(bla_dst_own->group)) {
net/batman-adv/bridge_loop_avoidance.c
1059
ntohs(bla_dst->group));
net/batman-adv/bridge_loop_avoidance.c
1060
bla_dst_own->group = bla_dst->group;
net/batman-adv/bridge_loop_avoidance.c
1335
__be16 group;
net/batman-adv/bridge_loop_avoidance.c
1339
group = htons(crc16(0, primary_if->net_dev->dev_addr, ETH_ALEN));
net/batman-adv/bridge_loop_avoidance.c
1340
bat_priv->bla.claim_dest.group = group;
net/batman-adv/bridge_loop_avoidance.c
1545
bat_priv->bla.claim_dest.group = htons(crc);
net/batman-adv/bridge_loop_avoidance.c
1548
bat_priv->bla.claim_dest.group = 0; /* will be set later */
net/batman-adv/netlink.c
249
ntohs(bat_priv->bla.claim_dest.group)))
net/bridge/br_mdb.c
1010
p = br_multicast_new_port_group(cfg->p, &cfg->group, *pp, flags, NULL,
net/bridge/br_mdb.c
1029
if (br_multicast_should_handle_mode(brmctx, cfg->group.proto) &&
net/bridge/br_mdb.c
1048
struct br_ip group = cfg->group;
net/bridge/br_mdb.c
1055
mp = br_multicast_new_group(br, &group);
net/bridge/br_mdb.c
1075
if (br_multicast_is_star_g(&group))
net/bridge/br_mdb.c
1180
__mdb_entry_to_br_ip(cfg->entry, &cfg->group, mdb_attrs);
net/bridge/br_mdb.c
1187
if (!br_multicast_is_star_g(&cfg->group)) {
net/bridge/br_mdb.c
1201
if (!br_multicast_is_star_g(&cfg->group)) {
net/bridge/br_mdb.c
1287
__mdb_entry_to_br_ip(cfg->entry, &cfg->group, NULL);
net/bridge/br_mdb.c
1311
if (!cfg.p && !br_group_is_l2(&cfg.group)) {
net/bridge/br_mdb.c
1317
if (!br_multicast_is_star_g(&cfg.group)) {
net/bridge/br_mdb.c
1323
if (br_group_is_l2(&cfg.group) && cfg.entry->state != MDB_PERMANENT) {
net/bridge/br_mdb.c
1344
cfg.group.vid = v->vid;
net/bridge/br_mdb.c
1365
struct br_ip ip = cfg->group;
net/bridge/br_mdb.c
1422
cfg.group.vid = v->vid;
net/bridge/br_mdb.c
1573
struct br_ip *group, struct netlink_ext_ack *extack)
net/bridge/br_mdb.c
1580
__mdb_entry_to_br_ip(entry, group, NULL);
net/bridge/br_mdb.c
1595
__mdb_entry_to_br_ip(entry, group, mdbe_attrs);
net/bridge/br_mdb.c
1681
struct br_ip group;
net/bridge/br_mdb.c
1684
err = br_mdb_get_parse(dev, tb, &group, extack);
net/bridge/br_mdb.c
1694
mp = br_mdb_ip_get(br, &group);
net/bridge/br_mdb.c
785
p = br_multicast_new_port_group(cfg->p, &cfg->group, *pp, flags, NULL,
net/bridge/br_mdb.c
799
if (br_multicast_should_handle_mode(brmctx, cfg->group.proto)) {
net/bridge/br_mdb.c
823
sg_ip = cfg->group;
net/bridge/br_mdb.c
840
sg_cfg.group = sg_ip;
net/bridge/br_mdb.c
978
if (br_multicast_should_handle_mode(brmctx, cfg->group.proto))
net/bridge/br_multicast.c
1013
ih->code = (group ? brmctx->multicast_last_member_interval :
net/bridge/br_multicast.c
1016
ih->group = group;
net/bridge/br_multicast.c
1024
ihv3->code = (group ? brmctx->multicast_last_member_interval :
net/bridge/br_multicast.c
1027
ihv3->group = group;
net/bridge/br_multicast.c
1075
const struct in6_addr *group,
net/bridge/br_multicast.c
1168
interval = ipv6_addr_any(group) ?
net/bridge/br_multicast.c
1180
mldq->mld_mca = *group;
net/bridge/br_multicast.c
1196
mld2q->mld2q_mca = *group;
net/bridge/br_multicast.c
1240
struct br_ip *group,
net/bridge/br_multicast.c
1247
switch (group->proto) {
net/bridge/br_multicast.c
1251
ip4_dst, group->dst.ip4,
net/bridge/br_multicast.c
1266
&ip6_dst, &group->dst.ip6,
net/bridge/br_multicast.c
1277
struct br_ip *group)
net/bridge/br_multicast.c
1282
mp = br_mdb_ip_get(br, group);
net/bridge/br_multicast.c
1287
trace_br_mdb_full(br->dev, group);
net/bridge/br_multicast.c
1298
mp->addr = *group;
net/bridge/br_multicast.c
1402
const struct br_ip *group,
net/bridge/br_multicast.c
1413
err = br_multicast_port_ngroups_inc(port, group, extack);
net/bridge/br_multicast.c
1423
p->key.addr = *group;
net/bridge/br_multicast.c
1433
if (!br_multicast_is_star_g(group) &&
net/bridge/br_multicast.c
1455
br_multicast_port_ngroups_dec(port, group->vid);
net/bridge/br_multicast.c
1504
struct br_ip *group,
net/bridge/br_multicast.c
1518
mp = br_multicast_new_group(brmctx->br, group);
net/bridge/br_multicast.c
1536
p = br_multicast_new_port_group(pmctx->port, group, *pp, 0, src,
net/bridge/br_multicast.c
1558
struct br_ip *group,
net/bridge/br_multicast.c
1567
pg = __br_multicast_add_group(brmctx, pmctx, group, src, filter_mode,
net/bridge/br_multicast.c
1578
__be32 group,
net/bridge/br_multicast.c
1586
if (ipv4_is_local_multicast(group))
net/bridge/br_multicast.c
1590
br_group.dst.ip4 = group;
net/bridge/br_multicast.c
1602
const struct in6_addr *group,
net/bridge/br_multicast.c
1610
if (ipv6_addr_is_ll_all_nodes(group))
net/bridge/br_multicast.c
1614
br_group.dst.ip6 = *group;
net/bridge/br_multicast.c
1783
struct br_ip *group,
net/bridge/br_multicast.c
1797
skb = br_multicast_alloc_query(brmctx, pmctx, pg, ip_dst, group,
net/bridge/br_multicast.c
1816
br_multicast_select_own_querier(brmctx, group, skb);
net/bridge/br_multicast.c
2855
__be32 group, *h_addr;
net/bridge/br_multicast.c
2870
group = grec->grec_mca;
net/bridge/br_multicast.c
2897
group, vid, src);
net/bridge/br_multicast.c
2901
err = br_ip4_multicast_add_group(brmctx, pmctx, group,
net/bridge/br_multicast.c
2914
mdst = br_mdb_ip4_get(brmctx->br, group, vid);
net/bridge/br_multicast.c
3498
__be32 group;
net/bridge/br_multicast.c
3504
group = ih->group;
net/bridge/br_multicast.c
3511
group = 0;
net/bridge/br_multicast.c
3516
(brmctx->multicast_igmp_version == 3 && group &&
net/bridge/br_multicast.c
3526
if (!group) {
net/bridge/br_multicast.c
3536
mp = br_mdb_ip4_get(brmctx->br, group, vid);
net/bridge/br_multicast.c
3579
const struct in6_addr *group = NULL;
net/bridge/br_multicast.c
3595
group = &mld->mld_mca;
net/bridge/br_multicast.c
3603
group = &mld2q->mld2q_mca;
net/bridge/br_multicast.c
3612
is_general_query = group && ipv6_addr_any(group);
net/bridge/br_multicast.c
3622
} else if (!group) {
net/bridge/br_multicast.c
3626
mp = br_mdb_ip6_get(brmctx->br, group, vid);
net/bridge/br_multicast.c
3657
struct br_ip *group,
net/bridge/br_multicast.c
3671
mp = br_mdb_ip_get(brmctx->br, group);
net/bridge/br_multicast.c
3758
__be32 group,
net/bridge/br_multicast.c
3765
if (ipv4_is_local_multicast(group))
net/bridge/br_multicast.c
3771
br_group.dst.ip4 = group;
net/bridge/br_multicast.c
3783
const struct in6_addr *group,
net/bridge/br_multicast.c
3790
if (ipv6_addr_is_ll_all_nodes(group))
net/bridge/br_multicast.c
3796
br_group.dst.ip6 = *group;
net/bridge/br_multicast.c
3907
err = br_ip4_multicast_add_group(brmctx, pmctx, ih->group, vid,
net/bridge/br_multicast.c
3917
br_ip4_multicast_leave_group(brmctx, pmctx, ih->group, vid, src);
net/bridge/br_multicast.c
4875
struct net_bridge_port_group *group;
net/bridge/br_multicast.c
4893
hlist_for_each_entry_rcu(group, &port->mglist, mglist) {
net/bridge/br_multicast.c
4898
entry->addr = group->key.addr;
net/bridge/br_multicast.c
59
__be32 group,
net/bridge/br_multicast.c
71
const struct in6_addr *group,
net/bridge/br_multicast.c
723
const struct br_ip *group,
net/bridge/br_multicast.c
735
trace_br_mdb_full(port->dev, group);
net/bridge/br_multicast.c
740
if (!group->vid)
net/bridge/br_multicast.c
743
pmctx = br_multicast_port_vid_to_port_ctx(port, group->vid);
net/bridge/br_multicast.c
749
trace_br_mdb_full(port->dev, group);
net/bridge/br_multicast.c
77
struct br_ip *group,
net/bridge/br_multicast.c
927
__be32 ip_dst, __be32 group,
net/bridge/br_private.h
1001
br_multicast_new_group(struct net_bridge *br, struct br_ip *group);
net/bridge/br_private.h
1004
const struct br_ip *group,
net/bridge/br_private.h
107
struct br_ip group;
net/bridge/br_private.h
1083
static inline bool br_group_is_l2(const struct br_ip *group)
net/bridge/br_private.h
1085
return group->proto == 0;
net/bridge/netfilter/ebt_nflog.c
32
li.u.ulog.group = info->group;
net/core/dev.c
12136
dev->group = INIT_NETDEV_GROUP;
net/core/dev.c
9975
dev->group = new_group;
net/core/net-sysfs.c
621
NETDEVICE_SHOW(group, fmt_dec);
net/core/rtnetlink.c
2090
nla_put_u32(skb, IFLA_GROUP, READ_ONCE(dev->group)) ||
net/core/rtnetlink.c
3504
static int rtnl_group_dellink(const struct net *net, int group)
net/core/rtnetlink.c
3510
if (!group)
net/core/rtnetlink.c
3514
if (dev->group == group) {
net/core/rtnetlink.c
3528
if (dev->group == group) {
net/core/rtnetlink.c
3803
int group, struct ifinfomsg *ifm,
net/core/rtnetlink.c
3811
if (dev->group == group) {
net/core/rtnetlink.c
7010
static int rtnetlink_bind(struct net *net, int group)
net/core/rtnetlink.c
7012
switch (group) {
net/core/rtnetlink.c
944
int rtnetlink_send(struct sk_buff *skb, struct net *net, u32 pid, unsigned int group, int echo)
net/core/rtnetlink.c
948
return nlmsg_notify(rtnl, skb, pid, group, echo, GFP_KERNEL);
net/core/rtnetlink.c
959
void rtnl_notify(struct sk_buff *skb, struct net *net, u32 pid, u32 group,
net/core/rtnetlink.c
964
nlmsg_notify(rtnl, skb, pid, group, nlmsg_report(nlh), flags);
net/core/rtnetlink.c
968
void rtnl_set_sk_err(struct net *net, u32 group, int error)
net/core/rtnetlink.c
972
netlink_set_err(rtnl, 0, group, error);
net/core/sock_diag.c
151
const enum sknetlink_groups group = sock_diag_destroy_group(sk);
net/core/sock_diag.c
154
WARN_ON(group == SKNLGRP_NONE);
net/core/sock_diag.c
167
nlmsg_multicast(sock_net(sk)->diag_nlsk, skb, 0, group,
net/core/sock_diag.c
296
static int sock_diag_bind(struct net *net, int group)
net/core/sock_diag.c
298
switch (group) {
net/devlink/trap.c
1125
devlink_trap_group_generic_verify(const struct devlink_trap_group *group)
net/devlink/trap.c
1127
if (group->id > DEVLINK_TRAP_GROUP_GENERIC_ID_MAX)
net/devlink/trap.c
1130
if (strcmp(group->name, devlink_trap_group_generic[group->id].name))
net/devlink/trap.c
1137
devlink_trap_group_driver_verify(const struct devlink_trap_group *group)
net/devlink/trap.c
1141
if (group->id <= DEVLINK_TRAP_GROUP_GENERIC_ID_MAX)
net/devlink/trap.c
1145
if (!strcmp(group->name, devlink_trap_group_generic[i].name))
net/devlink/trap.c
1152
static int devlink_trap_group_verify(const struct devlink_trap_group *group)
net/devlink/trap.c
1154
if (group->generic)
net/devlink/trap.c
1155
return devlink_trap_group_generic_verify(group);
net/devlink/trap.c
1157
return devlink_trap_group_driver_verify(group);
net/devlink/trap.c
1467
metadata->trap_group_name = trap_item->group_item->group->name;
net/devlink/trap.c
1523
u32 policer_id = group_item->group->init_policer_id;
net/devlink/trap.c
1540
const struct devlink_trap_group *group)
net/devlink/trap.c
1545
if (devlink_trap_group_item_lookup(devlink, group->name))
net/devlink/trap.c
1558
group_item->group = group;
net/devlink/trap.c
1565
err = devlink->ops->trap_group_init(devlink, group);
net/devlink/trap.c
1586
const struct devlink_trap_group *group)
net/devlink/trap.c
1590
group_item = devlink_trap_group_item_lookup(devlink, group->name);
net/devlink/trap.c
1617
const struct devlink_trap_group *group = &groups[i];
net/devlink/trap.c
1619
err = devlink_trap_group_verify(group);
net/devlink/trap.c
1623
err = devlink_trap_group_register(devlink, group);
net/devlink/trap.c
269
group_item->group->name))
net/devlink/trap.c
436
if (!strcmp(group_item->group->name, name))
net/devlink/trap.c
449
if (group_item->group->id == id)
net/devlink/trap.c
45
const struct devlink_trap_group *group;
net/devlink/trap.c
486
group_item->group->name))
net/devlink/trap.c
489
if (group_item->group->generic &&
net/devlink/trap.c
586
const char *group_name = group_item->group->name;
net/devlink/trap.c
591
err = devlink->ops->trap_group_action_set(devlink, group_item->group,
net/devlink/trap.c
597
if (strcmp(trap_item->group_item->group->name, group_name))
net/devlink/trap.c
609
if (strcmp(trap_item->group_item->group->name, group_name))
net/devlink/trap.c
672
err = devlink->ops->trap_group_set(devlink, group_item->group, policer,
net/ieee802154/ieee802154.h
29
int ieee802154_nl_mcast(struct sk_buff *msg, unsigned int group);
net/ieee802154/netlink.c
45
int ieee802154_nl_mcast(struct sk_buff *msg, unsigned int group)
net/ieee802154/netlink.c
52
return genlmsg_multicast(&nl802154_family, msg, 0, group, GFP_ATOMIC);
net/ipv4/fib_rules.c
169
if (rule->suppress_ifgroup != -1 && dev && dev->group == rule->suppress_ifgroup)
net/ipv4/igmp.c
1028
if (!group) { /* general query */
net/ipv4/igmp.c
1052
if (group && group != im->multiaddr)
net/ipv4/igmp.c
1113
dropped = igmp_heard_report(in_dev, ih->group);
net/ipv4/igmp.c
1644
if (!igmp_hdr(skb)->group &&
net/ipv4/igmp.c
2354
__be32 group = imr->imr_multiaddr.s_addr;
net/ipv4/igmp.c
2369
if (iml->multi.imr_multiaddr.s_addr != group)
net/ipv4/igmp.c
2383
ip_mc_dec_group(in_dev, group);
net/ipv4/igmp.c
738
__be32 group = pmc ? pmc->multiaddr : 0;
net/ipv4/igmp.c
746
if (ipv4_is_local_multicast(group) &&
net/ipv4/igmp.c
753
dst = group;
net/ipv4/igmp.c
796
ih->group = group;
net/ipv4/igmp.c
922
static bool igmp_heard_report(struct in_device *in_dev, __be32 group)
net/ipv4/igmp.c
929
if (group == IGMP_ALL_HOSTS)
net/ipv4/igmp.c
931
if (ipv4_is_local_multicast(group) &&
net/ipv4/igmp.c
937
if (im->multiaddr == group) {
net/ipv4/igmp.c
953
__be32 group = ih->group;
net/ipv4/igmp.c
967
group = 0;
net/ipv4/igmp.c
986
group = 0;
net/ipv4/ip_sockglue.c
664
struct sockaddr_storage *group,
net/ipv4/ip_sockglue.c
675
psin = (struct sockaddr_in *)group;
net/ipv4/ipmr.c
2362
pim->group != PIM_V1_VERSION || pim->code != PIM_V1_REGISTER)
net/ipv4/ping.c
254
kgid_t group = current_egid();
net/ipv4/ping.c
264
if (gid_lte(low, group) && gid_lte(group, high))
net/ipv6/fib6_rules.c
301
if (rule->suppress_ifgroup != -1 && dev && dev->group == rule->suppress_ifgroup)
net/ipv6/mcast.c
1038
bool ipv6_chk_mcast_addr(struct net_device *dev, const struct in6_addr *group,
net/ipv6/mcast.c
1050
if (ipv6_addr_equal(&mc->mca_addr, group))
net/ipv6/mcast.c
1427
const struct in6_addr *group;
net/ipv6/mcast.c
1461
group = &mld->mld_mca;
net/ipv6/mcast.c
1462
group_type = ipv6_addr_type(group);
net/ipv6/mcast.c
1512
if (!ipv6_addr_equal(group, &ma->mca_addr))
net/ipv6/mcast.c
173
const struct in6_addr *group,
net/ipv6/mcast.c
181
rt = rt6_lookup(net, group, NULL, 0, NULL, 0);
net/ipv6/mcast.c
314
const struct in6_addr *group,
net/ipv6/mcast.c
320
dev = ip6_mc_find_dev(net, group, ifindex);
net/ipv6/mcast.c
357
struct in6_addr *source, *group;
net/ipv6/mcast.c
367
group = &((struct sockaddr_in6 *)&pgsr->gsr_group)->sin6_addr;
net/ipv6/mcast.c
369
if (!ipv6_addr_is_multicast(group))
net/ipv6/mcast.c
372
idev = ip6_mc_find_idev(net, group, pgsr->gsr_interface);
net/ipv6/mcast.c
388
if (ipv6_addr_equal(&pmc->addr, group))
net/ipv6/mcast.c
403
ip6_mc_add_src(idev, group, omode, 0, NULL, 0);
net/ipv6/mcast.c
404
ip6_mc_del_src(idev, group, pmc->sfmode, 0, NULL, 0);
net/ipv6/mcast.c
428
ip6_mc_del_src(idev, group, omode, 1, source, 1);
net/ipv6/mcast.c
478
ip6_mc_add_src(idev, group, omode, 1, source, 1);
net/ipv6/mcast.c
483
err = ipv6_sock_mc_drop(sk, pgsr->gsr_interface, group);
net/ipv6/mcast.c
493
const struct in6_addr *group;
net/ipv6/mcast.c
499
group = &((struct sockaddr_in6 *)&gsf->gf_group)->sin6_addr;
net/ipv6/mcast.c
501
if (!ipv6_addr_is_multicast(group))
net/ipv6/mcast.c
507
idev = ip6_mc_find_idev(net, group, gsf->gf_interface);
net/ipv6/mcast.c
528
if (ipv6_addr_equal(&pmc->addr, group))
net/ipv6/mcast.c
551
err = ip6_mc_add_src(idev, group, gsf->gf_fmode,
net/ipv6/mcast.c
560
ip6_mc_add_src(idev, group, gsf->gf_fmode, 0, NULL, 0);
net/ipv6/mcast.c
565
ip6_mc_del_src(idev, group, pmc->sfmode,
net/ipv6/mcast.c
570
ip6_mc_del_src(idev, group, pmc->sfmode, 0, NULL, 0);
net/ipv6/mcast.c
581
err = ipv6_sock_mc_drop(sk, gsf->gf_interface, group);
net/ipv6/mcast.c
589
const struct in6_addr *group;
net/ipv6/mcast.c
595
group = &((struct sockaddr_in6 *)&gsf->gf_group)->sin6_addr;
net/ipv6/mcast.c
597
if (!ipv6_addr_is_multicast(group))
net/ipv6/mcast.c
603
if (ipv6_addr_equal(group, &pmc->addr))
net/mac80211/airtime.c
512
int group, idx;
net/mac80211/airtime.c
540
group = VHT_GROUP_IDX(streams, sgi, bw);
net/mac80211/airtime.c
545
group = HT_GROUP_IDX(streams, sgi, bw);
net/mac80211/airtime.c
550
group = HE_GROUP_IDX(streams, status->he_gi, bw);
net/mac80211/airtime.c
555
group = EHT_GROUP_IDX(streams, status->eht.gi, bw);
net/mac80211/airtime.c
577
duration = airtime_mcs_groups[group].duration[idx];
net/mac80211/airtime.c
578
duration <<= airtime_mcs_groups[group].shift;
net/mac80211/rc80211_minstrel_ht.c
1002
mi->sample[MINSTREL_SAMPLE_TYPE_JUMP].sample_group = group;
net/mac80211/rc80211_minstrel_ht.c
1063
int group, i, j, cur_prob;
net/mac80211/rc80211_minstrel_ht.c
1081
group = MINSTREL_CCK_GROUP;
net/mac80211/rc80211_minstrel_ht.c
1083
group = MINSTREL_OFDM_GROUP;
net/mac80211/rc80211_minstrel_ht.c
1085
group = 0;
net/mac80211/rc80211_minstrel_ht.c
1087
index = MI_RATE(group, 0);
net/mac80211/rc80211_minstrel_ht.c
1092
group = MINSTREL_VHT_GROUP_0;
net/mac80211/rc80211_minstrel_ht.c
1094
group = MINSTREL_HT_GROUP_0;
net/mac80211/rc80211_minstrel_ht.c
1096
group = MINSTREL_CCK_GROUP;
net/mac80211/rc80211_minstrel_ht.c
1098
group = MINSTREL_OFDM_GROUP;
net/mac80211/rc80211_minstrel_ht.c
1100
index = MI_RATE(group, 0);
net/mac80211/rc80211_minstrel_ht.c
1106
for (group = 0; group < ARRAY_SIZE(minstrel_mcs_groups); group++) {
net/mac80211/rc80211_minstrel_ht.c
1110
mg = &mi->groups[group];
net/mac80211/rc80211_minstrel_ht.c
1111
if (!mi->supported[group])
net/mac80211/rc80211_minstrel_ht.c
1116
tmp_group_tp_rate[j] = MI_RATE(group, 0);
net/mac80211/rc80211_minstrel_ht.c
1118
if (group == MINSTREL_CCK_GROUP && ht_supported)
net/mac80211/rc80211_minstrel_ht.c
1122
if (!(mi->supported[group] & BIT(i)))
net/mac80211/rc80211_minstrel_ht.c
1125
index = MI_RATE(group, i);
net/mac80211/rc80211_minstrel_ht.c
1137
if (minstrel_ht_get_tp_avg(mi, group, i, cur_prob) == 0)
net/mac80211/rc80211_minstrel_ht.c
1157
for (group = 0; group < ARRAY_SIZE(minstrel_mcs_groups); group++) {
net/mac80211/rc80211_minstrel_ht.c
1158
if (!mi->supported[group])
net/mac80211/rc80211_minstrel_ht.c
1161
mg = &mi->groups[group];
net/mac80211/rc80211_minstrel_ht.c
1162
mg->max_group_prob_rate = MI_RATE(group, 0);
net/mac80211/rc80211_minstrel_ht.c
1165
if (!(mi->supported[group] & BIT(i)))
net/mac80211/rc80211_minstrel_ht.c
1168
index = MI_RATE(group, i);
net/mac80211/rc80211_minstrel_ht.c
1260
int group, orig_group;
net/mac80211/rc80211_minstrel_ht.c
1262
orig_group = group = MI_RATE_GROUP(*idx);
net/mac80211/rc80211_minstrel_ht.c
1263
while (group > 0) {
net/mac80211/rc80211_minstrel_ht.c
1264
group--;
net/mac80211/rc80211_minstrel_ht.c
1266
if (!mi->supported[group])
net/mac80211/rc80211_minstrel_ht.c
1269
if (minstrel_mcs_groups[group].streams >
net/mac80211/rc80211_minstrel_ht.c
1274
*idx = mi->groups[group].max_group_tp_rate[0];
net/mac80211/rc80211_minstrel_ht.c
1276
*idx = mi->groups[group].max_group_tp_rate[1];
net/mac80211/rc80211_minstrel_ht.c
1446
const struct mcs_group *group = &minstrel_mcs_groups[group_idx];
net/mac80211/rc80211_minstrel_ht.c
1449
u16 flags = group->flags;
net/mac80211/rc80211_minstrel_ht.c
1472
idx = ((group->streams - 1) << 4) |
net/mac80211/rc80211_minstrel_ht.c
1475
idx = index + (group->streams - 1) * 8;
net/mac80211/rc80211_minstrel_ht.c
1483
group->streams > 1)) {
net/mac80211/rc80211_minstrel_ht.c
1495
int group = MI_RATE_GROUP(rate);
net/mac80211/rc80211_minstrel_ht.c
1497
return mi->groups[group].rates[rate].prob_avg;
net/mac80211/rc80211_minstrel_ht.c
1503
int group = MI_RATE_GROUP(mi->max_prob_rate);
net/mac80211/rc80211_minstrel_ht.c
1504
const struct mcs_group *g = &minstrel_mcs_groups[group];
net/mac80211/rc80211_minstrel_ht.c
1509
if (mi->groups[group].rates[rate].prob_avg < MINSTREL_FRAC(50, 100))
net/mac80211/rc80211_minstrel_ht.c
320
minstrel_ht_is_legacy_group(int group)
net/mac80211/rc80211_minstrel_ht.c
322
return group == MINSTREL_CCK_GROUP ||
net/mac80211/rc80211_minstrel_ht.c
323
group == MINSTREL_OFDM_GROUP;
net/mac80211/rc80211_minstrel_ht.c
373
int group, idx;
net/mac80211/rc80211_minstrel_ht.c
376
group = minstrel_ht_get_group_idx(rate);
net/mac80211/rc80211_minstrel_ht.c
382
group = minstrel_vht_get_group_idx(rate);
net/mac80211/rc80211_minstrel_ht.c
387
group = MINSTREL_CCK_GROUP;
net/mac80211/rc80211_minstrel_ht.c
389
if (!(mi->supported[group] & BIT(idx)))
net/mac80211/rc80211_minstrel_ht.c
396
if ((mi->supported[group] & BIT(idx + 4)) &&
net/mac80211/rc80211_minstrel_ht.c
402
group = MINSTREL_OFDM_GROUP;
net/mac80211/rc80211_minstrel_ht.c
409
return &mi->groups[group].rates[idx];
net/mac80211/rc80211_minstrel_ht.c
419
int group, idx;
net/mac80211/rc80211_minstrel_ht.c
423
group = minstrel_ht_ri_get_group_idx(rate);
net/mac80211/rc80211_minstrel_ht.c
429
group = minstrel_vht_ri_get_group_idx(rate);
net/mac80211/rc80211_minstrel_ht.c
434
group = MINSTREL_CCK_GROUP;
net/mac80211/rc80211_minstrel_ht.c
440
if ((mi->supported[group] & BIT(idx + 4)) &&
net/mac80211/rc80211_minstrel_ht.c
446
group = MINSTREL_OFDM_GROUP;
net/mac80211/rc80211_minstrel_ht.c
453
return &mi->groups[group].rates[idx];
net/mac80211/rc80211_minstrel_ht.c
464
const struct mcs_group *group = &minstrel_mcs_groups[MI_RATE_GROUP(index)];
net/mac80211/rc80211_minstrel_ht.c
465
unsigned int duration = group->duration[MI_RATE_IDX(index)];
net/mac80211/rc80211_minstrel_ht.c
467
return duration << group->shift;
net/mac80211/rc80211_minstrel_ht.c
500
minstrel_ht_get_tp_avg(struct minstrel_ht_sta *mi, int group, int rate,
net/mac80211/rc80211_minstrel_ht.c
510
if (minstrel_ht_is_legacy_group(group))
net/mac80211/rc80211_minstrel_ht.c
516
nsecs += minstrel_mcs_groups[group].duration[rate] <<
net/mac80211/rc80211_minstrel_ht.c
517
minstrel_mcs_groups[group].shift;
net/mac80211/rc80211_minstrel_ht.c
674
int tmp_max_streams, group, tmp_idx, tmp_prob;
net/mac80211/rc80211_minstrel_ht.c
680
group = MI_RATE_GROUP(mi->max_tp_rate[0]);
net/mac80211/rc80211_minstrel_ht.c
681
tmp_max_streams = minstrel_mcs_groups[group].streams;
net/mac80211/rc80211_minstrel_ht.c
682
for (group = 0; group < ARRAY_SIZE(minstrel_mcs_groups); group++) {
net/mac80211/rc80211_minstrel_ht.c
683
mg = &mi->groups[group];
net/mac80211/rc80211_minstrel_ht.c
684
if (!mi->supported[group] || group == MINSTREL_CCK_GROUP)
net/mac80211/rc80211_minstrel_ht.c
688
tmp_prob = mi->groups[group].rates[tmp_idx].prob_avg;
net/mac80211/rc80211_minstrel_ht.c
690
if (tmp_tp < minstrel_ht_get_tp_avg(mi, group, tmp_idx, tmp_prob) &&
net/mac80211/rc80211_minstrel_ht.c
691
(minstrel_mcs_groups[group].streams < tmp_max_streams)) {
net/mac80211/rc80211_minstrel_ht.c
693
tmp_tp = minstrel_ht_get_tp_avg(mi, group,
net/mac80211/rc80211_minstrel_ht.c
850
minstrel_ht_group_min_rate_offset(struct minstrel_ht_sta *mi, int group,
net/mac80211/rc80211_minstrel_ht.c
853
u16 supported = mi->supported[group];
net/mac80211/rc80211_minstrel_ht.c
860
if (minstrel_get_duration(MI_RATE(group, i)) >= max_duration)
net/mac80211/rc80211_minstrel_ht.c
879
u8 group;
net/mac80211/rc80211_minstrel_ht.c
881
group = mi->sample[type].sample_group;
net/mac80211/rc80211_minstrel_ht.c
883
group = (group + 1) % ARRAY_SIZE(minstrel_mcs_groups);
net/mac80211/rc80211_minstrel_ht.c
885
index = minstrel_ht_group_min_rate_offset(mi, group,
net/mac80211/rc80211_minstrel_ht.c
890
index = MI_RATE(group, index & 0xf);
net/mac80211/rc80211_minstrel_ht.c
897
mi->sample[type].sample_group = group;
net/mac80211/rc80211_minstrel_ht.c
903
minstrel_ht_next_group_sample_rate(struct minstrel_ht_sta *mi, int group,
net/mac80211/rc80211_minstrel_ht.c
906
struct minstrel_mcs_group_data *mg = &mi->groups[group];
net/mac80211/rc80211_minstrel_ht.c
924
return MI_RATE(group, idx);
net/mac80211/rc80211_minstrel_ht.c
946
u8 group;
net/mac80211/rc80211_minstrel_ht.c
952
group = mi->sample[MINSTREL_SAMPLE_TYPE_JUMP].sample_group;
net/mac80211/rc80211_minstrel_ht.c
956
group = (group + 1) % ARRAY_SIZE(minstrel_mcs_groups);
net/mac80211/rc80211_minstrel_ht.c
958
supported = mi->supported[group];
net/mac80211/rc80211_minstrel_ht.c
962
offset = minstrel_ht_group_min_rate_offset(mi, group,
net/mac80211/rc80211_minstrel_ht.c
967
index = minstrel_ht_next_group_sample_rate(mi, group, supported,
net/mac80211/rc80211_minstrel_ht.h
199
int minstrel_ht_get_tp_avg(struct minstrel_ht_sta *mi, int group, int rate,
net/netfilter/nf_conntrack_netlink.c
3093
unsigned int type, group;
net/netfilter/nf_conntrack_netlink.c
3098
group = NFNLGRP_CONNTRACK_EXP_DESTROY;
net/netfilter/nf_conntrack_netlink.c
3102
group = NFNLGRP_CONNTRACK_EXP_NEW;
net/netfilter/nf_conntrack_netlink.c
3106
if (!item->report && !nfnetlink_has_listeners(net, group))
net/netfilter/nf_conntrack_netlink.c
3123
nfnetlink_send(skb, net, item->portid, group, item->report, GFP_ATOMIC);
net/netfilter/nf_conntrack_netlink.c
754
unsigned int flags = 0, group;
net/netfilter/nf_conntrack_netlink.c
759
group = NFNLGRP_CONNTRACK_DESTROY;
net/netfilter/nf_conntrack_netlink.c
763
group = NFNLGRP_CONNTRACK_NEW;
net/netfilter/nf_conntrack_netlink.c
766
group = NFNLGRP_CONNTRACK_UPDATE;
net/netfilter/nf_conntrack_netlink.c
771
if (!item->report && !nfnetlink_has_listeners(net, group))
net/netfilter/nf_conntrack_netlink.c
867
err = nfnetlink_send(skb, net, item->portid, group, item->report,
net/netfilter/nf_conntrack_netlink.c
879
if (nfnetlink_set_err(net, 0, group, -ENOBUFS) > 0)
net/netfilter/nfnetlink.c
168
int nfnetlink_has_listeners(struct net *net, unsigned int group)
net/netfilter/nfnetlink.c
172
return netlink_has_listeners(nfnlnet->nfnl, group);
net/netfilter/nfnetlink.c
177
unsigned int group, int echo, gfp_t flags)
net/netfilter/nfnetlink.c
181
return nlmsg_notify(nfnlnet->nfnl, skb, portid, group, echo, flags);
net/netfilter/nfnetlink.c
185
int nfnetlink_set_err(struct net *net, u32 portid, u32 group, int error)
net/netfilter/nfnetlink.c
189
return netlink_set_err(nfnlnet->nfnl, portid, group, error);
net/netfilter/nfnetlink.c
207
__u32 group, gfp_t allocation)
net/netfilter/nfnetlink.c
211
netlink_broadcast(nfnlnet->nfnl, skb, portid, group, allocation);
net/netfilter/nfnetlink.c
672
static void nfnetlink_bind_event(struct net *net, unsigned int group)
net/netfilter/nfnetlink.c
681
if (group >= 8)
net/netfilter/nfnetlink.c
684
type = nfnl_group2type[group];
net/netfilter/nfnetlink.c
695
group_bit = (1 << group);
net/netfilter/nfnetlink.c
710
static int nfnetlink_bind(struct net *net, int group)
net/netfilter/nfnetlink.c
715
if (group <= NFNLGRP_NONE || group > NFNLGRP_MAX)
net/netfilter/nfnetlink.c
718
type = nfnl_group2type[group];
net/netfilter/nfnetlink.c
726
nfnetlink_bind_event(net, group);
net/netfilter/nfnetlink.c
730
static void nfnetlink_unbind(struct net *net, int group)
net/netfilter/nfnetlink.c
735
if (group <= NFNLGRP_NONE || group > NFNLGRP_MAX)
net/netfilter/nfnetlink.c
738
type = nfnl_group2type[group];
net/netfilter/nfnetlink.c
750
if (group >= 8)
net/netfilter/nfnetlink.c
753
group_bit = (1 << group);
net/netfilter/nfnetlink.c
756
if (!nfnetlink_has_listeners(net, group)) {
net/netfilter/nfnetlink_log.c
672
.group = 0,
net/netfilter/nfnetlink_log.c
705
inst = instance_lookup_get_rcu(log, li->u.ulog.group);
net/netfilter/nft_log.c
141
li->u.ulog.group = ntohs(nla_get_be16(tb[NFTA_LOG_GROUP]));
net/netfilter/nft_log.c
209
if (nla_put_be16(skb, NFTA_LOG_GROUP, htons(li->u.ulog.group)))
net/netfilter/nft_meta.c
226
*dest = dev->group;
net/netfilter/nft_set_pipapo.c
1001
for (group = 0; group < f->groups; group++) {
net/netfilter/nft_set_pipapo.c
1005
v = k[group / (BITS_PER_BYTE / f->bb)];
net/netfilter/nft_set_pipapo.c
1012
if (mask_bits >= (group + 1) * f->bb) {
net/netfilter/nft_set_pipapo.c
1014
pipapo_bucket_set(f, rule, group, v);
net/netfilter/nft_set_pipapo.c
1015
} else if (mask_bits <= group * f->bb) {
net/netfilter/nft_set_pipapo.c
1018
pipapo_bucket_set(f, rule, group, i);
net/netfilter/nft_set_pipapo.c
1022
mask >>= mask_bits - group * f->bb;
net/netfilter/nft_set_pipapo.c
1025
pipapo_bucket_set(f, rule, group, i);
net/netfilter/nft_set_pipapo.c
716
int group, bucket, err;
net/netfilter/nft_set_pipapo.c
747
for (group = 0; group < f->groups; group++) {
net/netfilter/nft_set_pipapo.c
783
static void pipapo_bucket_set(struct nft_pipapo_field *f, int rule, int group,
net/netfilter/nft_set_pipapo.c
789
pos += f->bsize * NFT_PIPAPO_BUCKETS(f->bb) * group;
net/netfilter/nft_set_pipapo.c
993
unsigned int rule = f->rules, group, ret, bit_offset = 0;
net/netfilter/nft_set_pipapo.h
196
int group;
net/netfilter/nft_set_pipapo.h
198
for (group = 0; group < f->groups; group += BITS_PER_BYTE / 4, data++) {
net/netfilter/nft_set_pipapo.h
224
int group;
net/netfilter/nft_set_pipapo.h
226
for (group = 0; group < f->groups; group++, data++) {
net/netfilter/nft_set_pipapo_avx2.c
45
#define NFT_PIPAPO_AVX2_BUCKET_LOAD4(reg, lt, group, v, bsize) \
net/netfilter/nft_set_pipapo_avx2.c
47
lt[((group) * NFT_PIPAPO_BUCKETS(4) + \
net/netfilter/nft_set_pipapo_avx2.c
49
#define NFT_PIPAPO_AVX2_BUCKET_LOAD8(reg, lt, group, v, bsize) \
net/netfilter/nft_set_pipapo_avx2.c
51
lt[((group) * NFT_PIPAPO_BUCKETS(8) + \
net/netfilter/xt_NFLOG.c
29
li.u.ulog.group = info->group;
net/netfilter/xt_devgroup.c
24
(((info->src_group ^ xt_in(par)->group) & info->src_mask ? 1 : 0) ^
net/netfilter/xt_devgroup.c
29
(((info->dst_group ^ xt_out(par)->group) & info->dst_mask ? 1 : 0) ^
net/netfilter/xt_owner.c
110
kgid_t group = gi->gid[i];
net/netfilter/xt_owner.c
112
if (gid_gte(group, gid_min) &&
net/netfilter/xt_owner.c
113
gid_lte(group, gid_max)) {
net/netlink/af_netlink.c
1010
int group;
net/netlink/af_netlink.c
1013
for (group = 0; group < BITS_PER_TYPE(u32); group++) {
net/netlink/af_netlink.c
1014
if (!test_bit(group, &groups))
net/netlink/af_netlink.c
1016
err = nlk->netlink_bind(net, group + 1);
net/netlink/af_netlink.c
1019
netlink_undo_bind(group, groups, sk);
net/netlink/af_netlink.c
1363
int netlink_has_listeners(struct sock *sk, unsigned int group)
net/netlink/af_netlink.c
1373
if (listeners && group - 1 < nl_table[sk->sk_protocol].groups)
net/netlink/af_netlink.c
1374
res = test_bit(group - 1, listeners->masks);
net/netlink/af_netlink.c
1411
u32 group;
net/netlink/af_netlink.c
1431
if (nlk->portid == p->portid || p->group - 1 >= nlk->ngroups ||
net/netlink/af_netlink.c
1432
!test_bit(p->group - 1, nlk->groups))
net/netlink/af_netlink.c
1504
u32 group, gfp_t allocation,
net/netlink/af_netlink.c
1517
info.group = group;
net/netlink/af_netlink.c
1555
u32 group, gfp_t allocation)
net/netlink/af_netlink.c
1557
return netlink_broadcast_filtered(ssk, skb, portid, group, allocation,
net/netlink/af_netlink.c
1565
u32 group;
net/netlink/af_netlink.c
1580
if (nlk->portid == p->portid || p->group - 1 >= nlk->ngroups ||
net/netlink/af_netlink.c
1581
!test_bit(p->group - 1, nlk->groups))
net/netlink/af_netlink.c
160
static inline u32 netlink_group_mask(u32 group)
net/netlink/af_netlink.c
1605
int netlink_set_err(struct sock *ssk, u32 portid, u32 group, int code)
net/netlink/af_netlink.c
1614
info.group = group;
net/netlink/af_netlink.c
162
if (group > 32)
net/netlink/af_netlink.c
1630
unsigned int group,
net/netlink/af_netlink.c
1635
old = test_bit(group - 1, nlk->groups);
net/netlink/af_netlink.c
1637
__assign_bit(group - 1, nlk->groups, new);
net/netlink/af_netlink.c
164
return group ? 1 << (group - 1) : 0;
net/netlink/af_netlink.c
1798
info.group = NETLINK_CB(skb).dst_group;
net/netlink/af_netlink.c
2128
void __netlink_clear_multicast_users(struct sock *ksk, unsigned int group)
net/netlink/af_netlink.c
2135
netlink_update_socket_mc(nlk_sk(sk), group, 0);
net/netlink/af_netlink.c
2579
unsigned int group, int report, gfp_t flags)
net/netlink/af_netlink.c
2583
if (group) {
net/netlink/af_netlink.c
2593
err = nlmsg_multicast(sk, skb, exclude_portid, group, flags);
net/netlink/af_netlink.c
649
int (*bind)(struct net *net, int group);
net/netlink/af_netlink.c
650
void (*unbind)(struct net *net, int group);
net/netlink/af_netlink.c
955
static void netlink_undo_bind(int group, long unsigned int groups,
net/netlink/af_netlink.c
964
for (undo = 0; undo < group; undo++)
net/netlink/af_netlink.h
44
int (*netlink_bind)(struct net *net, int group);
net/netlink/af_netlink.h
45
void (*netlink_unbind)(struct net *net, int group);
net/netlink/af_netlink.h
69
int (*bind)(struct net *net, int group);
net/netlink/af_netlink.h
70
void (*unbind)(struct net *net, int group);
net/netlink/genetlink.c
1811
static int genl_bind(struct net *net, int group)
net/netlink/genetlink.c
1826
i = group - family->mcgrp_offset;
net/netlink/genetlink.c
1851
static void genl_unbind(struct net *net, int group)
net/netlink/genetlink.c
1864
i = group - family->mcgrp_offset;
net/netlink/genetlink.c
1930
static int genlmsg_mcast(struct sk_buff *skb, u32 portid, unsigned long group)
net/netlink/genetlink.c
1946
portid, group, GFP_ATOMIC);
net/netlink/genetlink.c
1955
err = nlmsg_multicast(prev->genl_sock, skb, portid, group, GFP_ATOMIC);
net/netlink/genetlink.c
1973
unsigned int group)
net/netlink/genetlink.c
1975
if (WARN_ON_ONCE(group >= family->n_mcgrps))
net/netlink/genetlink.c
1978
group = family->mcgrp_offset + group;
net/netlink/genetlink.c
1979
return genlmsg_mcast(skb, portid, group);
net/netlink/genetlink.c
1984
struct genl_info *info, u32 group, gfp_t flags)
net/netlink/genetlink.c
1989
if (WARN_ON_ONCE(group >= family->n_mcgrps))
net/netlink/genetlink.c
1992
group = family->mcgrp_offset + group;
net/netlink/genetlink.c
1993
nlmsg_notify(sk, skb, info->snd_portid, group,
net/openvswitch/datapath.c
75
unsigned int group)
net/openvswitch/datapath.c
78
genl_has_listeners(family, genl_info_net(info), group);
net/psample/psample.c
123
static void psample_group_notify(struct psample_group *group,
net/psample/psample.c
133
err = psample_group_nl_fill(msg, group, cmd, 0, 0, NLM_F_MULTI);
net/psample/psample.c
135
genlmsg_multicast_netns(&psample_nl_family, group->net, msg, 0,
net/psample/psample.c
144
struct psample_group *group;
net/psample/psample.c
146
group = kzalloc_obj(*group, GFP_ATOMIC);
net/psample/psample.c
147
if (!group)
net/psample/psample.c
150
group->net = net;
net/psample/psample.c
151
group->group_num = group_num;
net/psample/psample.c
152
list_add_tail(&group->list, &psample_groups_list);
net/psample/psample.c
154
psample_group_notify(group, PSAMPLE_CMD_NEW_GROUP);
net/psample/psample.c
155
return group;
net/psample/psample.c
158
static void psample_group_destroy(struct psample_group *group)
net/psample/psample.c
160
psample_group_notify(group, PSAMPLE_CMD_DEL_GROUP);
net/psample/psample.c
161
list_del(&group->list);
net/psample/psample.c
162
kfree_rcu(group, rcu);
net/psample/psample.c
168
struct psample_group *group;
net/psample/psample.c
170
list_for_each_entry(group, &psample_groups_list, list)
net/psample/psample.c
171
if ((group->group_num == group_num) && (group->net == net))
net/psample/psample.c
172
return group;
net/psample/psample.c
178
struct psample_group *group;
net/psample/psample.c
182
group = psample_group_lookup(net, group_num);
net/psample/psample.c
183
if (!group) {
net/psample/psample.c
184
group = psample_group_create(net, group_num);
net/psample/psample.c
185
if (!group)
net/psample/psample.c
188
group->refcount++;
net/psample/psample.c
192
return group;
net/psample/psample.c
196
void psample_group_take(struct psample_group *group)
net/psample/psample.c
199
group->refcount++;
net/psample/psample.c
204
void psample_group_put(struct psample_group *group)
net/psample/psample.c
208
if (--group->refcount == 0)
net/psample/psample.c
209
psample_group_destroy(group);
net/psample/psample.c
363
void psample_sample_packet(struct psample_group *group,
net/psample/psample.c
380
if (!genl_has_listeners(&psample_nl_family, group->net,
net/psample/psample.c
41
struct psample_group *group,
net/psample/psample.c
440
ret = nla_put_u32(nl_skb, PSAMPLE_ATTR_SAMPLE_GROUP, group->group_num);
net/psample/psample.c
444
ret = nla_put_u32(nl_skb, PSAMPLE_ATTR_GROUP_SEQ, group->seq++);
net/psample/psample.c
508
genlmsg_multicast_netns(&psample_nl_family, group->net, nl_skb, 0,
net/psample/psample.c
52
ret = nla_put_u32(msg, PSAMPLE_ATTR_SAMPLE_GROUP, group->group_num);
net/psample/psample.c
56
ret = nla_put_u32(msg, PSAMPLE_ATTR_GROUP_REFCOUNT, group->refcount);
net/psample/psample.c
60
ret = nla_put_u32(msg, PSAMPLE_ATTR_GROUP_SEQ, group->seq);
net/psample/psample.c
75
struct psample_group *group;
net/psample/psample.c
81
list_for_each_entry(group, &psample_groups_list, list) {
net/psample/psample.c
82
if (!net_eq(group->net, sock_net(msg->sk)))
net/psample/psample.c
88
err = psample_group_nl_fill(msg, group, PSAMPLE_CMD_NEW_GROUP,
net/sched/act_sample.c
266
struct psample_group *group = priv;
net/sched/act_sample.c
268
psample_group_put(group);
net/sched/act_sample.c
276
struct psample_group *group;
net/sched/act_sample.c
278
group = rcu_dereference_protected(s->psample_group,
net/sched/act_sample.c
280
if (group) {
net/sched/act_sample.c
281
psample_group_take(group);
net/sched/act_sample.c
285
return group;
net/shaper/shaper.c
1032
ret = ops->group(binding, leaves_count, leaves, node, extack);
net/shaper/shaper.c
1186
if (!net_shaper_ops(binding)->group)
net/tipc/group.c
111
*m->group->open = true;
net/tipc/group.c
315
m->group = grp;
net/tipc/group.c
374
struct tipc_group *grp = m->group;
net/tipc/group.c
68
struct tipc_group *group;
net/tipc/group.c
805
*m->group->open = true;
net/tipc/group.c
928
struct nlattr *group = nla_nest_start_noflag(skb, TIPC_NLA_SOCK_GROUP);
net/tipc/group.c
930
if (!group)
net/tipc/group.c
953
nla_nest_end(skb, group);
net/tipc/group.c
957
nla_nest_cancel(skb, group);
net/tipc/socket.c
1021
exclude = tipc_group_exclude(tsk->group);
net/tipc/socket.c
1031
cong = tipc_group_cong(tsk->group, node, port, blks,
net/tipc/socket.c
1051
tsk->group &&
net/tipc/socket.c
1052
!tipc_group_cong(tsk->group, node, port,
net/tipc/socket.c
1098
!tsk->cong_link_cnt && tsk->group &&
net/tipc/socket.c
1099
!tipc_group_bc_cong(tsk->group, blks));
net/tipc/socket.c
1103
dsts = tipc_group_dests(tsk->group);
net/tipc/socket.c
1118
msg_set_grp_bc_seqno(hdr, tipc_group_bc_snd_nxt(tsk->group));
net/tipc/socket.c
1135
tipc_group_update_bc_members(tsk->group, blks, ack);
net/tipc/socket.c
1160
struct tipc_group *grp = tsk->group;
net/tipc/socket.c
131
struct tipc_group *group;
net/tipc/socket.c
1414
struct tipc_group *grp = tsk->group;
net/tipc/socket.c
1973
if (tsk->group && msg_in_group(hdr) && !grp_evt) {
net/tipc/socket.c
1975
tipc_group_update_rcv_win(tsk->group, tsk_blocks(hlen + dlen),
net/tipc/socket.c
2151
struct tipc_group *grp = tsk->group;
net/tipc/socket.c
2170
tipc_group_member_evt(tsk->group, &wakeup, &sk->sk_rcvbuf,
net/tipc/socket.c
2339
struct tipc_group *grp = tsk->group;
net/tipc/socket.c
2586
if (tsk->group) {
net/tipc/socket.c
3070
struct tipc_group *grp = tsk->group;
net/tipc/socket.c
3086
tsk->group = grp;
net/tipc/socket.c
3096
tsk->group = NULL;
net/tipc/socket.c
3109
struct tipc_group *grp = tsk->group;
net/tipc/socket.c
3119
tsk->group = NULL;
net/tipc/socket.c
3276
if (tsk->group)
net/tipc/socket.c
3277
tipc_group_self(tsk->group, &seq, &scope);
net/tipc/socket.c
3673
if (tsk->group)
net/tipc/socket.c
3674
if (tipc_group_fill_sock_diag(tsk->group, skb))
net/tipc/socket.c
695
if (tsk->group)
net/tipc/socket.c
853
if (tsk->group)
net/tipc/socket.c
908
u16 bc_snd_nxt = tipc_group_bc_snd_nxt(tsk->group);
net/tipc/socket.c
975
tsk->group &&
net/tipc/socket.c
976
!tipc_group_cong(tsk->group, node, port, blks,
net/xfrm/xfrm_user.c
1581
u32 pid, unsigned int group)
net/xfrm/xfrm_user.c
1602
return nlmsg_multicast(nlsk, skb, pid, group, GFP_ATOMIC);
samples/configfs/configfs_sample.c
172
struct config_group group;
samples/configfs/configfs_sample.c
178
struct simple_children, group);
samples/configfs/configfs_sample.c
181
static struct config_item *simple_children_make_item(struct config_group *group,
samples/configfs/configfs_sample.c
259
struct config_group *group, const char *name)
samples/configfs/configfs_sample.c
268
config_group_init_type_name(&simple_children->group, name,
samples/configfs/configfs_sample.c
271
return &simple_children->group;
samples/connector/cn_test.c
104
NETLINK_CB(skb).dst_group = ctl->group;
samples/connector/cn_test.c
108
pr_info("request was sent: group=0x%x\n", ctl->group);
samples/connector/cn_test.c
47
u32 group = 1;
samples/connector/cn_test.c
79
ctl->group = group;
scripts/mod/devicetable-offsets.c
28
DEVID_FIELD(hid_device_id, group);
scripts/mod/file2alias.c
385
DEF_FIELD(symval, hid_device_id, group);
scripts/mod/file2alias.c
390
ADD(alias, "g", group != HID_GROUP_ANY, group);
security/commoncap_test.c
152
ns->group = KGIDT_INIT(0);
security/keys/keyctl.c
1045
if (group != (gid_t) -1)
security/keys/keyctl.c
949
long keyctl_chown_key(key_serial_t id, uid_t user, gid_t group)
security/keys/keyctl.c
960
gid = make_kgid(current_user_ns(), group);
security/keys/keyctl.c
964
if ((group != (gid_t) -1) && !gid_valid(gid))
security/keys/keyctl.c
968
if (user == (uid_t) -1 && group == (gid_t) -1)
security/keys/keyctl.c
993
if (group != (gid_t) -1 && !gid_eq(gid, key->gid) && !in_group_p(gid))
security/tomoyo/common.c
1217
set_bit(idx, domain->group);
security/tomoyo/common.c
1219
clear_bit(idx, domain->group);
security/tomoyo/common.c
1536
if (ptr->address.group) {
security/tomoyo/common.c
1538
tomoyo_set_string(head, ptr->address.group->group_name
security/tomoyo/common.c
1663
if (!test_bit(i, domain->group))
security/tomoyo/common.c
1795
unsigned int group;
security/tomoyo/common.c
1798
group = simple_strtoul(param.data, &data, 10);
security/tomoyo/common.c
1799
if (group < TOMOYO_MAX_ACL_GROUPS && *data++ == ' ')
security/tomoyo/common.c
1801
(head->w.ns, &head->w.ns->acl_group[group],
security/tomoyo/common.c
1825
list_for_each_cookie(head->r.group, list) {
security/tomoyo/common.c
1826
struct tomoyo_group *group =
security/tomoyo/common.c
1827
list_entry(head->r.group, typeof(*group), head.list);
security/tomoyo/common.c
1829
list_for_each_cookie(head->r.acl, &group->member_list) {
security/tomoyo/common.c
1839
tomoyo_set_string(head, group->group_name->name);
security/tomoyo/common.c
1864
head->r.group = NULL;
security/tomoyo/common.c
383
if (ptr->group) {
security/tomoyo/common.c
385
tomoyo_set_string(head, ptr->group->group_name->name);
security/tomoyo/common.c
402
if (ptr->group) {
security/tomoyo/common.c
404
tomoyo_set_string(head, ptr->group->group_name->name);
security/tomoyo/common.c
424
if (ptr->group) {
security/tomoyo/common.c
426
tomoyo_set_string(head, ptr->group->group_name->name);
security/tomoyo/common.h
1209
static inline void tomoyo_put_group(struct tomoyo_group *group)
security/tomoyo/common.h
1211
if (group)
security/tomoyo/common.h
1212
atomic_dec(&group->head.users);
security/tomoyo/common.h
1238
return a->filename == b->filename && a->group == b->group;
security/tomoyo/common.h
1253
a->group == b->group && a->value_type[0] == b->value_type[0] &&
security/tomoyo/common.h
1268
return !memcmp(a->ip, b->ip, sizeof(a->ip)) && a->group == b->group &&
security/tomoyo/common.h
522
struct tomoyo_group *group;
security/tomoyo/common.h
528
struct tomoyo_group *group; /* Maybe NULL. */
security/tomoyo/common.h
536
struct tomoyo_group *group; /* Pointer to address group. */
security/tomoyo/common.h
688
unsigned long group[TOMOYO_MAX_ACL_GROUPS / BITS_PER_LONG];
security/tomoyo/common.h
804
struct list_head *group;
security/tomoyo/common.h
959
const struct tomoyo_group *group);
security/tomoyo/common.h
974
const struct tomoyo_group *group);
security/tomoyo/common.h
999
(const struct tomoyo_path_info *pathname, const struct tomoyo_group *group);
security/tomoyo/condition.c
1072
if (ptr->group) {
security/tomoyo/condition.c
1075
ptr->group)
security/tomoyo/domain.c
184
if (!test_bit(i, domain->group))
security/tomoyo/domain.c
567
memcpy(e.group, domain->group, sizeof(e.group));
security/tomoyo/domain.c
597
if (test_bit(i, entry->group))
security/tomoyo/file.c
101
tomoyo_put_group(ptr->group);
security/tomoyo/file.c
115
if (ptr->group)
security/tomoyo/file.c
116
return tomoyo_number_matches_group(value, value, ptr->group);
security/tomoyo/file.c
69
tomoyo_put_group(ptr->group);
security/tomoyo/file.c
85
if (ptr->group)
security/tomoyo/file.c
86
return tomoyo_path_matches_group(name, ptr->group);
security/tomoyo/gc.c
218
tomoyo_put_group(entry->address.group);
security/tomoyo/gc.c
345
struct tomoyo_group *group =
security/tomoyo/gc.c
346
container_of(element, typeof(*group), head.list);
security/tomoyo/gc.c
348
tomoyo_put_name(group->group_name);
security/tomoyo/gc.c
49
if (head->r.domain == element || head->r.group == element ||
security/tomoyo/gc.c
566
struct tomoyo_group *group;
security/tomoyo/gc.c
580
list_for_each_entry_safe(group, tmp, list, head.list) {
security/tomoyo/gc.c
581
tomoyo_collect_member(id, &group->member_list);
security/tomoyo/gc.c
582
if (!list_empty(&group->member_list) ||
security/tomoyo/gc.c
583
atomic_read(&group->head.users) > 0)
security/tomoyo/gc.c
585
atomic_set(&group->head.users,
security/tomoyo/gc.c
588
&group->head.list);
security/tomoyo/group.c
115
tomoyo_put_group(group);
security/tomoyo/group.c
132
const struct tomoyo_group *group)
security/tomoyo/group.c
136
list_for_each_entry_rcu(member, &group->member_list, head.list,
security/tomoyo/group.c
160
const struct tomoyo_group *group)
security/tomoyo/group.c
165
list_for_each_entry_rcu(member, &group->member_list, head.list,
security/tomoyo/group.c
190
const struct tomoyo_group *group)
security/tomoyo/group.c
196
list_for_each_entry_rcu(member, &group->member_list, head.list,
security/tomoyo/group.c
76
struct tomoyo_group *group = tomoyo_get_group(param, type);
security/tomoyo/group.c
79
if (!group)
security/tomoyo/group.c
81
param->list = &group->member_list;
security/tomoyo/memory.c
112
list_for_each_entry(group, list, head.list) {
security/tomoyo/memory.c
113
if (e.group_name != group->group_name ||
security/tomoyo/memory.c
114
atomic_read(&group->head.users) == TOMOYO_GC_IN_PROGRESS)
security/tomoyo/memory.c
116
atomic_inc(&group->head.users);
security/tomoyo/memory.c
127
group = entry;
security/tomoyo/memory.c
134
return found ? group : NULL;
security/tomoyo/memory.c
99
struct tomoyo_group *group = NULL;
security/tomoyo/network.c
300
e.address.group =
security/tomoyo/network.c
302
if (!e.address.group)
security/tomoyo/network.c
315
tomoyo_put_group(e.address.group);
security/tomoyo/network.c
433
if (acl->address.group)
security/tomoyo/network.c
436
r->param.inet_network.address, acl->address.group);
security/tomoyo/util.c
254
ptr->group = tomoyo_get_group(param, TOMOYO_PATH_GROUP);
security/tomoyo/util.c
255
return ptr->group != NULL;
security/tomoyo/util.c
282
ptr->group = tomoyo_get_group(param, TOMOYO_NUMBER_GROUP);
security/tomoyo/util.c
283
return ptr->group != NULL;
sound/core/control_led.c
189
unsigned int group;
sound/core/control_led.c
191
for (group = 0; group < MAX_LED; group++) {
sound/core/control_led.c
192
controls = &snd_ctl_leds[group].controls;
sound/core/control_led.c
246
unsigned int group, bool set)
sound/core/control_led.c
263
if (access != 0 && access != group_to_access(group))
sound/core/control_led.c
267
new_access |= group_to_access(group);
sound/core/control_led.c
277
unsigned int group;
sound/core/control_led.c
279
for (group = 0; group < MAX_LED; group++)
sound/core/control_led.c
280
snd_ctl_led_set_state(NULL, group_to_access(group), NULL, 0);
sound/core/control_led.c
291
unsigned int group;
sound/core/control_led.c
295
for (group = 0; group < MAX_LED; group++) {
sound/core/control_led.c
296
led = &snd_ctl_leds[group];
sound/core/control_led.c
303
static int snd_ctl_led_reset(int card_number, unsigned int group)
sound/core/control_led.c
318
led = &snd_ctl_leds[group];
sound/core/control_led.c
322
vd->access &= ~group_to_access(group);
sound/core/control_led.c
328
snd_ctl_led_set_state(NULL, group_to_access(group), NULL, 0);
sound/core/control_led.c
40
unsigned int group;
sound/core/control_led.c
419
snd_ctl_led_set_state(NULL, group_to_access(led->group), NULL, 0);
sound/core/control_led.c
559
err = snd_ctl_led_set_id(led_card->number, &id, led_card->led->group, attach);
sound/core/control_led.c
590
err = snd_ctl_led_reset(led_card->number, led_card->led->group);
sound/core/control_led.c
60
.group = (SNDRV_CTL_ELEM_ACCESS_SPK_LED >> SNDRV_CTL_ELEM_ACCESS_LED_SHIFT) - 1,
sound/core/control_led.c
649
unsigned int group;
sound/core/control_led.c
654
for (group = 0; group < MAX_LED; group++) {
sound/core/control_led.c
655
led = &snd_ctl_leds[group];
sound/core/control_led.c
66
.group = (SNDRV_CTL_ELEM_ACCESS_MIC_LED >> SNDRV_CTL_ELEM_ACCESS_LED_SHIFT) - 1,
sound/core/control_led.c
692
unsigned int group;
sound/core/control_led.c
697
for (group = 0; group < MAX_LED; group++) {
sound/core/control_led.c
698
led = &snd_ctl_leds[group];
sound/core/control_led.c
723
unsigned int group;
sound/core/control_led.c
736
for (group = 0; group < MAX_LED; group++) {
sound/core/control_led.c
737
led = &snd_ctl_leds[group];
sound/core/control_led.c
746
for (; group > 0; group--) {
sound/core/control_led.c
747
led = &snd_ctl_leds[group - 1];
sound/core/control_led.c
762
unsigned int group, card_number;
sound/core/control_led.c
774
for (group = 0; group < MAX_LED; group++) {
sound/core/control_led.c
775
led = &snd_ctl_leds[group];
sound/core/control_led.c
88
static inline unsigned int group_to_access(unsigned int group)
sound/core/control_led.c
90
return (group + 1) << SNDRV_CTL_ELEM_ACCESS_LED_SHIFT;
sound/core/control_led.c
95
unsigned int group = access_to_group(access);
sound/core/control_led.c
96
if (group >= MAX_LED)
sound/core/control_led.c
98
return &snd_ctl_leds[group];
sound/core/init.c
837
const struct attribute_group *group)
sound/core/init.c
844
card->dev_groups[i] = group;
sound/core/pcm.c
687
substream->group = &substream->self_group;
sound/core/pcm_local.h
65
void snd_pcm_group_init(struct snd_pcm_group *group);
sound/core/pcm_native.c
1326
substream->group = new_group;
sound/core/pcm_native.c
1334
static void snd_pcm_group_unref(struct snd_pcm_group *group,
sound/core/pcm_native.c
1339
if (!group)
sound/core/pcm_native.c
1341
do_free = refcount_dec_and_test(&group->refs);
sound/core/pcm_native.c
1342
snd_pcm_group_unlock(group, substream->pcm->nonatomic);
sound/core/pcm_native.c
1344
kfree(group);
sound/core/pcm_native.c
1355
struct snd_pcm_group *group;
sound/core/pcm_native.c
1361
group = substream->group;
sound/core/pcm_native.c
1363
refcount_inc(&group->refs);
sound/core/pcm_native.c
1365
trylock = nonatomic ? mutex_trylock(&group->mutex) :
sound/core/pcm_native.c
1366
spin_trylock(&group->lock);
sound/core/pcm_native.c
1372
snd_pcm_group_lock(group, nonatomic);
sound/core/pcm_native.c
1376
if (substream->group == group)
sound/core/pcm_native.c
1379
snd_pcm_group_unref(group, substream);
sound/core/pcm_native.c
1381
return group;
sound/core/pcm_native.c
1391
struct snd_pcm_group *group;
sound/core/pcm_native.c
1394
group = snd_pcm_stream_group_ref(substream);
sound/core/pcm_native.c
1395
if (group)
sound/core/pcm_native.c
1399
snd_pcm_group_unref(group, substream);
sound/core/pcm_native.c
149
struct snd_pcm_group *group = &substream->self_group;
sound/core/pcm_native.c
152
mutex_lock_nested(&group->mutex, SINGLE_DEPTH_NESTING);
sound/core/pcm_native.c
154
spin_lock_nested(&group->lock, SINGLE_DEPTH_NESTING);
sound/core/pcm_native.c
2111
struct snd_pcm_group *group;
sound/core/pcm_native.c
2157
group = snd_pcm_stream_group_ref(substream);
sound/core/pcm_native.c
2167
snd_pcm_group_unref(group, substream);
sound/core/pcm_native.c
2196
group = snd_pcm_stream_group_ref(substream);
sound/core/pcm_native.c
2203
snd_pcm_group_unref(group, substream);
sound/core/pcm_native.c
2298
struct snd_pcm_group *group __free(kfree) =
sound/core/pcm_native.c
2299
kzalloc(sizeof(*group), GFP_KERNEL);
sound/core/pcm_native.c
2300
if (!group)
sound/core/pcm_native.c
2302
snd_pcm_group_init(group);
sound/core/pcm_native.c
2314
snd_pcm_group_assign(substream, group);
sound/core/pcm_native.c
2315
group = NULL; /* assigned, don't free this one below */
sound/core/pcm_native.c
2317
target_group = substream->group;
sound/core/pcm_native.c
2338
struct snd_pcm_group *group;
sound/core/pcm_native.c
2347
group = substream->group;
sound/core/pcm_native.c
2348
snd_pcm_group_lock_irq(group, nonatomic);
sound/core/pcm_native.c
2351
refcount_dec(&group->refs);
sound/core/pcm_native.c
2354
if (list_is_singular(&group->substreams)) {
sound/core/pcm_native.c
2355
relink_to_local(list_first_entry(&group->substreams,
sound/core/pcm_native.c
2358
do_free = refcount_dec_and_test(&group->refs);
sound/core/pcm_native.c
2361
snd_pcm_group_unlock_irq(group, nonatomic);
sound/core/pcm_native.c
2363
kfree(group);
sound/core/pcm_native.c
78
void snd_pcm_group_init(struct snd_pcm_group *group)
sound/core/pcm_native.c
80
spin_lock_init(&group->lock);
sound/core/pcm_native.c
81
mutex_init(&group->mutex);
sound/core/pcm_native.c
82
INIT_LIST_HEAD(&group->substreams);
sound/core/pcm_native.c
83
refcount_set(&group->refs, 1);
sound/core/pcm_native.c
88
static void snd_pcm_group_ ## action(struct snd_pcm_group *group, bool nonatomic) \
sound/core/pcm_native.c
91
mutex_ ## mutex_action(&group->mutex); \
sound/core/pcm_native.c
95
spin_ ## action(&group->lock); \
sound/core/seq/seq_clientmgr.c
1912
struct snd_seq_port_subs_info *group;
sound/core/seq/seq_clientmgr.c
1927
group = &port->c_src;
sound/core/seq/seq_clientmgr.c
1930
group = &port->c_dest;
sound/core/seq/seq_clientmgr.c
1936
guard(rwsem_read)(&group->list_mutex);
sound/core/seq/seq_clientmgr.c
1938
subs->num_subs = group->count;
sound/core/seq/seq_clientmgr.c
1940
list_for_each(p, &group->list_head) {
sound/core/seq/seq_clientmgr.c
2501
struct snd_seq_port_subs_info *group,
sound/core/seq/seq_clientmgr.c
2508
guard(rwsem_read)(&group->list_mutex);
sound/core/seq/seq_clientmgr.c
2509
if (list_empty(&group->list_head))
sound/core/seq/seq_clientmgr.c
2512
list_for_each(p, &group->list_head) {
sound/core/seq/seq_clientmgr.c
2524
if (group->exclusive)
sound/core/seq/seq_ump_client.c
168
struct snd_ump_group *group)
sound/core/seq/seq_ump_client.c
173
port->addr.port = ump_group_to_seq_port(group->group);
sound/core/seq/seq_ump_client.c
185
if (group->dir_bits & (1 << STR_IN))
sound/core/seq/seq_ump_client.c
187
if (group->dir_bits & (1 << STR_OUT))
sound/core/seq/seq_ump_client.c
189
port->ump_group = group->group + 1;
sound/core/seq/seq_ump_client.c
190
if (!group->active)
sound/core/seq/seq_ump_client.c
192
if (group->is_midi1)
sound/core/seq/seq_ump_client.c
199
if (*group->name)
sound/core/seq/seq_ump_client.c
201
group->group + 1, group->name);
sound/core/seq/seq_ump_client.c
203
sprintf(port->name, "Group %d", group->group + 1);
sound/core/seq/seq_ump_client.c
207
static bool skip_group(struct seq_ump_client *client, struct snd_ump_group *group)
sound/core/seq/seq_ump_client.c
209
return !group->valid &&
sound/core/seq/seq_ump_client.c
216
struct snd_ump_group *group = &client->ump->groups[group_index];
sound/core/seq/seq_ump_client.c
219
if (skip_group(client, group))
sound/core/seq/seq_ump_client.c
227
fill_port_info(port, client, group);
sound/core/seq/seq_ump_client.c
31
unsigned char group; /* parsed UMP packet group */
sound/core/seq/seq_ump_client.c
54
static unsigned char ump_group_to_seq_port(unsigned char group)
sound/core/seq/seq_ump_client.c
56
return group + 1;
sound/core/seq/seq_ump_convert.c
377
midi2->note.group = midi1->note.group;
sound/core/seq/seq_ump_convert.c
448
midi1->note.group = midi2->note.group;
sound/core/seq/seq_ump_convert.c
572
unsigned char group;
sound/core/seq/seq_ump_convert.c
574
group = ump_message_group(ev->ump[0]);
sound/core/seq/seq_ump_convert.c
578
return dest->group_filter & (1U << (group + 1));
sound/core/ump.c
1150
int group = ump->legacy_mapping[substream->number];
sound/core/ump.c
1154
if (ump->legacy_substreams[dir][group])
sound/core/ump.c
1156
if (!ump->groups[group].active)
sound/core/ump.c
1168
snd_ump_convert_reset(&ump->out_cvts[group]);
sound/core/ump.c
1171
ump->legacy_substreams[dir][group] = substream;
sound/core/ump.c
1179
int group = ump->legacy_mapping[substream->number];
sound/core/ump.c
1183
ump->legacy_substreams[dir][group] = NULL;
sound/core/ump.c
1239
int group, size = 0;
sound/core/ump.c
1245
for (group = 0; group < SNDRV_UMP_MAX_GROUPS; group++) {
sound/core/ump.c
1246
substream = ump->legacy_substreams[dir][group];
sound/core/ump.c
1249
ctx = &ump->out_cvts[group];
sound/core/ump.c
1252
ump->groups[group].is_midi1)
sound/core/ump.c
1256
snd_ump_convert_to_ump(ctx, group, protocol, c);
sound/core/ump.c
1272
unsigned char group;
sound/core/ump.c
1276
size = snd_ump_convert_from_ump(src, buf, &group);
sound/core/ump.c
1280
substream = ump->legacy_substreams[dir][group];
sound/core/ump.c
563
struct snd_ump_group *group;
sound/core/ump.c
567
group = &ump->groups[i];
sound/core/ump.c
568
*group->name = 0;
sound/core/ump.c
569
group->dir_bits = 0;
sound/core/ump.c
570
group->active = 0;
sound/core/ump.c
571
group->group = i;
sound/core/ump.c
572
group->valid = false;
sound/core/ump.c
573
group->is_midi1 = false;
sound/core/ump.c
579
group = &ump->groups[fb->info.first_group];
sound/core/ump.c
580
for (i = 0; i < fb->info.num_groups; i++, group++) {
sound/core/ump.c
581
group->valid = true;
sound/core/ump.c
583
group->active = 1;
sound/core/ump.c
585
group->is_midi1 = true;
sound/core/ump.c
588
group->dir_bits |= (1 << SNDRV_RAWMIDI_STREAM_INPUT);
sound/core/ump.c
591
group->dir_bits |= (1 << SNDRV_RAWMIDI_STREAM_OUTPUT);
sound/core/ump.c
594
group->dir_bits |= (1 << SNDRV_RAWMIDI_STREAM_INPUT) |
sound/core/ump.c
600
if (*group->name)
sound/core/ump.c
601
strlcat(group->name, ", ", sizeof(group->name));
sound/core/ump.c
602
safe_append_string(group->name, sizeof(group->name),
sound/core/ump_convert.c
246
unsigned char group, u32 *data, bool finish)
sound/core/ump_convert.c
260
*data = ump_compose(UMP_MSG_TYPE_DATA, group, status, cvt->len);
sound/core/ump_convert.c
280
unsigned char group, u32 *data)
sound/core/ump_convert.c
282
data[0] = ump_compose(UMP_MSG_TYPE_SYSTEM, group, 0, cvt->buf[0]);
sound/core/ump_convert.c
330
unsigned char group,
sound/core/ump_convert.c
346
group, 0, buf[0]);
sound/core/ump_convert.c
363
group, status, channel);
sound/core/ump_convert.c
446
static int do_convert_to_ump(struct ump_cvt_to_ump *cvt, unsigned char group,
sound/core/ump_convert.c
467
return cvt_legacy_sysex_to_ump(cvt, group, data, true);
sound/core/ump_convert.c
475
data[0] = ump_compose(UMP_MSG_TYPE_SYSTEM, group, 0, c);
sound/core/ump_convert.c
497
return cvt_legacy_sysex_to_ump(cvt, group, data, false);
sound/core/ump_convert.c
509
return cvt_legacy_system_to_ump(cvt, group, data);
sound/core/ump_convert.c
510
return cvt_legacy_cmd_to_ump(cvt, group, protocol, data, cvt->cmd_bytes);
sound/core/ump_convert.c
523
void snd_ump_convert_to_ump(struct ump_cvt_to_ump *cvt, unsigned char group,
sound/core/ump_convert.c
526
cvt->ump_bytes = do_convert_to_ump(cvt, group, protocol, c, cvt->ump);
sound/hda/codecs/ca0132.c
3650
static void ca0113_mmio_command_set(struct hda_codec *codec, unsigned int group,
sound/hda/codecs/ca0132.c
3663
writel(group, spec->mem_base + 0x804);
sound/hda/codecs/ca0132.c
3690
unsigned int group, unsigned int target, unsigned int value)
sound/hda/codecs/ca0132.c
3702
writel(group, spec->mem_base + 0x804);
sound/hda/codecs/ca0132.c
4543
ca0113_mmio_command_set(codec, out_cmds->group[i],
sound/hda/codecs/ca0132.c
732
unsigned int group[AE_CA0113_OUT_SET_COMMANDS];
sound/hda/codecs/ca0132.c
738
.group = { 0x30, 0x30, 0x48, 0x48, 0x48, 0x30 },
sound/hda/codecs/ca0132.c
747
.group = { 0x30, 0x30, 0x48, 0x48, 0x48, 0x30 },
sound/hda/core/sysfs.c
312
const struct attribute_group *group)
sound/hda/core/sysfs.c
315
sysfs_remove_group(kobj, group);
sound/hda/core/sysfs.c
339
const struct attribute_group *group,
sound/hda/core/sysfs.c
353
err = sysfs_create_group(kobj, group);
sound/pci/mixart/mixart.c
313
pipe->group_uid = buf->sgroup_resp.group; /* id of the pipe, as returned by embedded */
sound/pci/mixart/mixart_core.h
171
struct mixart_uid group;
sound/pci/rme96.c
1352
(rme96->playback_substream->group ==
sound/pci/rme96.c
1353
rme96->capture_substream->group);
sound/pci/rme96.c
1409
(rme96->playback_substream->group ==
sound/pci/rme96.c
1410
rme96->capture_substream->group);
sound/soc/codecs/cs42l52.c
656
u8 group;
sound/soc/codecs/cs42l52.c
836
(clk_map_table[index].group << CLK_32K_SR_SHIFT) |
sound/soc/sdca/sdca_asoc.c
103
if (entity->group)
sound/soc/sdca/sdca_asoc.c
448
if (!entity->group) {
sound/soc/sdca/sdca_asoc.c
453
range = sdca_selector_find_range(dev, entity->group, SDCA_CTL_GE_SELECTED_MODE,
sound/soc/sdca/sdca_asoc.c
459
(*widget)->kcontrol_news = entity->group->ge.kctl;
sound/soc/sdca/sdca_asoc.c
463
for (i = 0; i < entity->group->ge.num_modes; i++) {
sound/soc/sdca/sdca_asoc.c
464
struct sdca_ge_mode *mode = &entity->group->ge.modes[i];
sound/soc/sdca/sdca_asoc.c
773
if (entity->group)
sound/soc/sdca/sdca_asoc.c
774
add_route(&route, entity->label, NULL, entity->group->label);
sound/soc/sdca/sdca_functions.c
1284
struct sdca_entity_ge *group = &entity->ge;
sound/soc/sdca/sdca_functions.c
1312
group->num_modes = *affected_list;
sound/soc/sdca/sdca_functions.c
1315
group->modes = devm_kcalloc(dev, group->num_modes, sizeof(*group->modes),
sound/soc/sdca/sdca_functions.c
1317
if (!group->modes)
sound/soc/sdca/sdca_functions.c
1320
for (i = 0; i < group->num_modes; i++) {
sound/soc/sdca/sdca_functions.c
1322
struct sdca_ge_mode *mode = &group->modes[i];
sound/soc/sdca/sdca_functions.c
1754
if (managed->group && managed->group != entity) {
sound/soc/sdca/sdca_functions.c
1757
managed->label, managed->group->label,
sound/soc/sdca/sdca_functions.c
1762
managed->group = entity;
sound/usb/mixer_quirks.c
4246
u8 device_idx, u8 group, u16 value)
sound/usb/mixer_quirks.c
4250
if (group >= device->ncontrols || value >= device->controls[group].noptions)
sound/usb/mixer_quirks.c
4261
device->controls[group].options[value],
sound/usb/mixer_quirks.c
4262
device->controls[group].wIndex,
sound/usb/mixer_quirks.c
4280
u8 group = (private_value & SND_DJM_GROUP_MASK) >> SND_DJM_GROUP_SHIFT;
sound/usb/mixer_quirks.c
4284
(group << SND_DJM_GROUP_SHIFT) |
sound/usb/mixer_quirks.c
4287
return snd_djm_controls_update(mixer, device, group, value);
sound/usb/mixer_quirks.c
4294
u8 group = (private_value & SND_DJM_GROUP_MASK) >> SND_DJM_GROUP_SHIFT;
sound/usb/mixer_quirks.c
4297
return snd_djm_controls_update(list->mixer, device, group, value);
tools/include/nolibc/sys.h
175
int sys_chown(const char *path, uid_t owner, gid_t group)
tools/include/nolibc/sys.h
178
return my_syscall5(__NR_fchownat, AT_FDCWD, path, owner, group, 0);
tools/include/nolibc/sys.h
180
return my_syscall3(__NR_chown, path, owner, group);
tools/include/nolibc/sys.h
185
int chown(const char *path, uid_t owner, gid_t group)
tools/include/nolibc/sys.h
187
return __sysret(sys_chown(path, owner, group));
tools/include/uapi/linux/kvm.h
1181
__u32 group; /* device-defined */
tools/include/uapi/linux/netlink.h
161
__u32 group;
tools/lib/subcmd/parse-options.c
811
struct option *opt, *ordered = NULL, *group;
tools/lib/subcmd/parse-options.c
823
group = realloc(ordered, len);
tools/lib/subcmd/parse-options.c
824
if (!group)
tools/lib/subcmd/parse-options.c
826
ordered = group;
tools/lib/subcmd/parse-options.c
835
for (opt = group = ordered; opt->type != OPTION_END; opt++) {
tools/lib/subcmd/parse-options.c
837
qsort(group, nr_group, sizeof(*opt), option__cmp);
tools/lib/subcmd/parse-options.c
838
group = opt + 1;
tools/lib/subcmd/parse-options.c
844
qsort(group, nr_group, sizeof(*opt), option__cmp);
tools/lib/thermal/thermal_nl.c
101
grp->group,
tools/lib/thermal/thermal_nl.c
114
const char *family, const char *group)
tools/lib/thermal/thermal_nl.c
119
.group = group,
tools/lib/thermal/thermal_nl.c
12
const char *group;
tools/lib/thermal/thermal_nl.c
186
const char *group)
tools/lib/thermal/thermal_nl.c
191
group);
tools/lib/thermal/thermal_nl.c
202
const char *group)
tools/lib/thermal/thermal_nl.c
207
group);
tools/lib/thermal/thermal_nl.h
33
const char *group);
tools/lib/thermal/thermal_nl.h
36
const char *group);
tools/objtool/arch/x86/special.c
14
static struct special_alt *group, *prev;
tools/objtool/arch/x86/special.c
19
if (group && group->orig_sec == alt->orig_sec &&
tools/objtool/arch/x86/special.c
20
group->orig_off == alt->orig_off) {
tools/objtool/arch/x86/special.c
22
struct special_alt *iter = group;
tools/objtool/arch/x86/special.c
33
} else group = alt;
tools/objtool/check.c
1984
if (!special_alt->group || special_alt->new_len) {
tools/objtool/check.c
1994
if (special_alt->group) {
tools/objtool/include/objtool/special.h
18
bool group;
tools/objtool/special.c
21
bool group, jump_or_nop;
tools/objtool/special.c
31
.group = true,
tools/objtool/special.c
76
alt->group = entry->group;
tools/objtool/special.c
79
if (alt->group) {
tools/objtool/special.c
98
if (!entry->group || alt->new_len) {
tools/perf/bench/sched-messaging.c
317
total_children += group(worker_tab + total_children, num_fds,
tools/perf/builtin-list.c
200
const char *group,
tools/perf/builtin-list.c
214
(!print_state->metricgroups || !group || !strglobmatch(group, print_state->event_glob)))
tools/perf/builtin-list.c
227
strcmp(print_state->last_metricgroups, group ?: "")) {
tools/perf/builtin-list.c
228
if (group && print_state->metricgroups) {
tools/perf/builtin-list.c
230
fprintf(fp, "%s ", group);
tools/perf/builtin-list.c
233
? describe_metricgroup(group)
tools/perf/builtin-list.c
243
fprintf(fp, "%s%s [%s]\n", group, print_colon, gdesc);
tools/perf/builtin-list.c
245
fprintf(fp, "%s%s\n", group, print_colon);
tools/perf/builtin-list.c
249
print_state->last_metricgroups = strdup(group ?: "");
tools/perf/builtin-list.c
458
static void json_print_metric(void *ps __maybe_unused, const char *group,
tools/perf/builtin-list.c
472
(!print_state->common.metricgroups || !group ||
tools/perf/builtin-list.c
473
!strglobmatch(group, print_state->common.event_glob)))
tools/perf/builtin-list.c
479
if (group) {
tools/perf/builtin-list.c
480
fix_escape_fprintf(fp, &buf, "\t\"MetricGroup\": \"%S\"", group);
tools/perf/builtin-probe.c
347
const char *event = NULL, *group = NULL;
tools/perf/builtin-probe.c
383
show_perf_probe_event(tev->group, tev->event, pev,
tools/perf/builtin-probe.c
388
group = tev->group;
tools/perf/builtin-probe.c
397
pr_info("\techo 1 > events/%s/%s/enable\n", group, event);
tools/perf/builtin-probe.c
400
pr_info("\tBefore removing the probe, echo 0 > events/%s/%s/enable\n", group, event);
tools/perf/builtin-probe.c
404
pr_info("\tperf record -e %s:%s -aR sleep 1\n\n", group, event);
tools/perf/pmu-events/empty-pmu-events.c
3373
const char *describe_metricgroup(const char *group)
tools/perf/pmu-events/empty-pmu-events.c
3380
int cmp = strcmp(mgroup, group);
tools/perf/pmu-events/pmu-events.h
142
const char *describe_metricgroup(const char *group);
tools/perf/tests/sample-parsing.c
305
sample.read.group.nr = 4;
tools/perf/tests/sample-parsing.c
306
sample.read.group.values = values;
tools/perf/tests/sample-parsing.c
80
COMP(read.group.nr);
tools/perf/tests/sample-parsing.c
89
for (i = 0; i < s1->read.group.nr; i++) {
tools/perf/tests/sample-parsing.c
92
MCOMP(read.group.values[i]);
tools/perf/tests/sdt.c
60
const char *group, const char *event)
tools/perf/tests/sdt.c
70
if (!probe_cache__find_by_name(cache, group, event)) {
tools/perf/tests/sdt.c
71
pr_debug("Failed to find %s:%s in the cache\n", group, event);
tools/perf/util/bpf-filter.c
135
struct perf_bpf_filter_expr *group;
tools/perf/util/bpf-filter.c
137
list_for_each_entry(group, &expr->groups, list) {
tools/perf/util/bpf-filter.c
138
if (check_sample_flags(evsel, group) < 0)
tools/perf/util/bpf-filter.c
176
struct perf_bpf_filter_expr *group;
tools/perf/util/bpf-filter.c
178
list_for_each_entry(group, &expr->groups, list) {
tools/perf/util/bpf-filter.c
182
entry[i].op = group->op;
tools/perf/util/bpf-filter.c
183
entry[i].part = group->part;
tools/perf/util/bpf-filter.c
184
entry[i].term = group->term;
tools/perf/util/bpf-filter.c
185
entry[i].value = group->val;
tools/perf/util/evsel.c
3321
data->read.group.nr = *array;
tools/perf/util/evsel.c
3344
if (data->read.group.nr > max_group_nr)
tools/perf/util/evsel.c
3347
sz = data->read.group.nr * sample_read_value_size(read_format);
tools/perf/util/evsel.c
3349
data->read.group.values =
tools/perf/util/parse-events.c
1796
bool group)
tools/perf/util/parse-events.c
1800
if (!group && mod.weak) {
tools/perf/util/parse-events.c
1808
int eu = group ? evsel->core.attr.exclude_user : 0;
tools/perf/util/parse-events.c
1809
int ek = group ? evsel->core.attr.exclude_kernel : 0;
tools/perf/util/parse-events.c
1810
int eh = group ? evsel->core.attr.exclude_hv : 0;
tools/perf/util/parse-events.c
1811
int eH = group ? evsel->core.attr.exclude_host : 0;
tools/perf/util/parse-events.c
1812
int eG = group ? evsel->core.attr.exclude_guest : 0;
tools/perf/util/parse-events.y
134
groups ',' group
tools/perf/util/parse-events.y
138
struct list_head *group = $3;
tools/perf/util/parse-events.y
140
list_splice_tail(group, groups);
tools/perf/util/parse-events.y
141
free(group);
tools/perf/util/parse-events.y
157
group
tools/perf/util/parse-events.y
161
group:
tools/perf/util/parse-events.y
92
%type <list_evsel> group
tools/perf/util/print-events.c
80
ent->pev.group, ent->pev.event, bid_nd->s);
tools/perf/util/print-events.h
22
const char *group,
tools/perf/util/probe-event.c
1472
pev->group = strdup_esq(*arg);
tools/perf/util/probe-event.c
1473
if (!pev->group)
tools/perf/util/probe-event.c
1477
pev->group = NULL;
tools/perf/util/probe-event.c
1486
zfree(&pev->group);
tools/perf/util/probe-event.c
1935
tev->group = strdup(fmt2_str);
tools/perf/util/probe-event.c
1937
if (tev->group == NULL || tev->event == NULL) {
tools/perf/util/probe-event.c
1941
pr_debug("Group:%s Event:%s probe:%c\n", tev->group, tev->event, pr);
tools/perf/util/probe-event.c
2117
if (strbuf_addf(&buf, "%s:%s=", pev->group ?: PERFPROBE_GROUP,
tools/perf/util/probe-event.c
2265
tev->group, tev->event) < 0)
tools/perf/util/probe-event.c
2365
pev->group = strdup(tev->group);
tools/perf/util/probe-event.c
2366
if (pev->event == NULL || pev->group == NULL)
tools/perf/util/probe-event.c
2404
zfree(&pev->group);
tools/perf/util/probe-event.c
2476
dst->group = strdup_or_goto(src->group, out_err);
tools/perf/util/probe-event.c
2504
zfree(&tev->group);
tools/perf/util/probe-event.c
2632
static int perf_probe_event__sprintf(const char *group, const char *event,
tools/perf/util/probe-event.c
2640
if (asprintf(&buf, "%s:%s", group, event) < 0)
tools/perf/util/probe-event.c
2674
int show_perf_probe_event(const char *group, const char *event,
tools/perf/util/probe-event.c
2681
ret = perf_probe_event__sprintf(group, event, pev, module, &buf);
tools/perf/util/probe-event.c
2703
if (e_snprintf(tmp, 128, "%s:%s", tev->group, tev->event) < 0)
tools/perf/util/probe-event.c
2733
ret = show_perf_probe_event(pev.group, pev.event,
tools/perf/util/probe-event.c
2876
tev->group, tev->event);
tools/perf/util/probe-event.c
2905
const char *event, *group;
tools/perf/util/probe-event.c
2926
if (pev->group && !pev->sdt)
tools/perf/util/probe-event.c
2927
group = pev->group;
tools/perf/util/probe-event.c
2928
else if (tev->group)
tools/perf/util/probe-event.c
2929
group = tev->group;
tools/perf/util/probe-event.c
2931
group = PERFPROBE_GROUP;
tools/perf/util/probe-event.c
2933
if (strlen(group) >= MAX_EVENT_NAME_LEN) {
tools/perf/util/probe-event.c
2935
group, MAX_EVENT_NAME_LEN);
tools/perf/util/probe-event.c
2949
tev->group = strdup(group);
tools/perf/util/probe-event.c
2950
if (tev->event == NULL || tev->group == NULL)
tools/perf/util/probe-event.c
3353
if (tev->group) {
tools/perf/util/probe-event.c
3354
tev->group = strdup(pev->group);
tools/perf/util/probe-event.c
3355
if (!tev->group)
tools/perf/util/probe-event.c
3450
if (!entry->pev.event || !entry->pev.group)
tools/perf/util/probe-event.c
3452
if ((!pev->group || strglobmatch(entry->pev.group, pev->group)) &&
tools/perf/util/probe-event.c
3579
if (!pev->group && !pev->sdt) {
tools/perf/util/probe-event.c
3582
pev->group = strdup(PERFPROBE_GROUP);
tools/perf/util/probe-event.c
3583
ret = pev->group ? 0 : -ENOMEM;
tools/perf/util/probe-event.c
3585
ret = convert_exec_to_group(pev->target, &pev->group);
tools/perf/util/probe-event.c
3714
if (pev->group && strcmp(pev->group, "probe"))
tools/perf/util/probe-event.c
3715
pr_warning("WARN: Group name %s is ignored\n", pev->group);
tools/perf/util/probe-event.h
171
int show_perf_probe_event(const char *group, const char *event,
tools/perf/util/probe-event.h
58
char *group; /* Group name */
tools/perf/util/probe-event.h
97
char *group; /* Group name */
tools/perf/util/probe-file.c
1020
snprintf(buf, 128, "%s:%s", entry->pev.group, entry->pev.event);
tools/perf/util/probe-file.c
238
ret = e_snprintf(buf, 128, "%s:%s", tev.group,
tools/perf/util/probe-file.c
634
(!pev->group ||
tools/perf/util/probe-file.c
635
streql(entry->pev.group, pev->group)))
tools/perf/util/probe-file.c
642
(streql(entry->pev.group, pev->group) &&
tools/perf/util/probe-file.c
656
const char *group, const char *event)
tools/perf/util/probe-file.c
662
if (streql(entry->pev.group, group) &&
tools/perf/util/probe-file.c
922
entry->pev.group = strdup(sdtgrp);
tools/perf/util/probe-file.h
67
const char *group, const char *event);
tools/perf/util/python.c
1070
int group = 0, inherit = 0;
tools/perf/util/python.c
1074
&pcpus, &pthreads, &group, &inherit))
tools/perf/util/sample.h
43
} group;
tools/perf/util/scripting-engines/trace-event-python.c
673
values = PyList_New(sample->read.group.nr);
tools/perf/util/scripting-engines/trace-event-python.c
681
struct sample_read_value *v = sample->read.group.values;
tools/perf/util/scripting-engines/trace-event-python.c
684
sample_read_group__for_each(v, sample->read.group.nr, read_format) {
tools/perf/util/session.c
1063
struct sample_read_value *value = sample->read.group.values;
tools/perf/util/session.c
1065
printf(".... group nr %" PRIu64 "\n", sample->read.group.nr);
tools/perf/util/session.c
1067
sample_read_group__for_each(value, sample->read.group.nr, read_format) {
tools/perf/util/session.c
1298
struct sample_read_value *v = sample->read.group.values;
tools/perf/util/session.c
1304
sample_read_group__for_each(v, sample->read.group.nr, read_format) {
tools/perf/util/synthetic-events.c
1498
result += sz * sample->read.group.nr;
tools/perf/util/synthetic-events.c
1597
struct sample_read_value *v = sample->read.group.values;
tools/perf/util/synthetic-events.c
1599
sample_read_group__for_each(v, sample->read.group.nr, read_format) {
tools/perf/util/synthetic-events.c
1671
*array = sample->read.group.nr;
tools/power/x86/intel-speed-select/hfi-events.c
118
const char *group;
tools/power/x86/intel-speed-select/hfi-events.c
142
res->group,
tools/power/x86/intel-speed-select/hfi-events.c
153
const char *family, const char *group)
tools/power/x86/intel-speed-select/hfi-events.c
157
struct family_data res = { group, -ENOENT };
tools/testing/selftests/bpf/progs/bpf_iter_netlink.c
19
unsigned long group, ino;
tools/testing/selftests/bpf/progs/bpf_iter_netlink.c
36
group = 0;
tools/testing/selftests/bpf/progs/bpf_iter_netlink.c
41
bpf_probe_read_kernel(&group, sizeof(group), &nlk->groups[0]);
tools/testing/selftests/bpf/progs/bpf_iter_netlink.c
44
nlk->portid, (u32)group,
tools/testing/selftests/bpf/usdt.h
109
#define USDT(group, name, ...) \
tools/testing/selftests/bpf/usdt.h
110
__usdt_probe(group, name, __usdt_sema_none, 0 __VA_OPT__(,) __VA_ARGS__)
tools/testing/selftests/bpf/usdt.h
112
#define USDT(group, name, ...) \
tools/testing/selftests/bpf/usdt.h
113
__usdt_probe(group, name, __usdt_sema_none, 0, ##__VA_ARGS__)
tools/testing/selftests/bpf/usdt.h
156
#define USDT_WITH_SEMA(group, name, ...) \
tools/testing/selftests/bpf/usdt.h
157
__usdt_probe(group, name, \
tools/testing/selftests/bpf/usdt.h
158
__usdt_sema_implicit, __usdt_sema_name(group, name) \
tools/testing/selftests/bpf/usdt.h
161
#define USDT_WITH_SEMA(group, name, ...) \
tools/testing/selftests/bpf/usdt.h
162
__usdt_probe(group, name, \
tools/testing/selftests/bpf/usdt.h
163
__usdt_sema_implicit, __usdt_sema_name(group, name), \
tools/testing/selftests/bpf/usdt.h
185
#define USDT_IS_ACTIVE(group, name) ({ \
tools/testing/selftests/bpf/usdt.h
186
extern struct usdt_sema __usdt_sema_name(group, name) \
tools/testing/selftests/bpf/usdt.h
187
__usdt_asm_name(__usdt_sema_name(group, name)); \
tools/testing/selftests/bpf/usdt.h
188
__usdt_sema_implicit(__usdt_sema_name(group, name)); \
tools/testing/selftests/bpf/usdt.h
189
__usdt_sema_name(group, name).active > 0; \
tools/testing/selftests/bpf/usdt.h
290
#define USDT_WITH_EXPLICIT_SEMA(sema, group, name, ...) \
tools/testing/selftests/bpf/usdt.h
291
__usdt_probe(group, name, __usdt_sema_explicit, USDT_SEMA(sema), ##__VA_ARGS__)
tools/testing/selftests/bpf/usdt.h
293
#define USDT_WITH_EXPLICIT_SEMA(sema, group, name, ...) \
tools/testing/selftests/bpf/usdt.h
294
__usdt_probe(group, name, __usdt_sema_explicit, USDT_SEMA(sema) __VA_OPT__(,) __VA_ARGS__)
tools/testing/selftests/bpf/usdt.h
324
#define __usdt_sema_name(group, name) __usdt_sema_##group##__##name
tools/testing/selftests/bpf/usdt.h
407
#define __usdt_probe(group, name, sema_def, sema, ...) do { \
tools/testing/selftests/bpf/usdt.h
419
__usdt_asm_strz(group) \
tools/testing/selftests/kvm/arm64/vpmu_counter_access.c
408
.group = KVM_ARM_VCPU_PMU_V3_CTRL,
tools/testing/selftests/kvm/include/arm64/gic.h
60
void gic_irq_set_group(unsigned int intid, bool group);
tools/testing/selftests/kvm/include/kvm_util.h
883
int __kvm_has_device_attr(int dev_fd, uint32_t group, uint64_t attr);
tools/testing/selftests/kvm/include/kvm_util.h
885
static inline void kvm_has_device_attr(int dev_fd, uint32_t group, uint64_t attr)
tools/testing/selftests/kvm/include/kvm_util.h
887
int ret = __kvm_has_device_attr(dev_fd, group, attr);
tools/testing/selftests/kvm/include/kvm_util.h
892
int __kvm_device_attr_get(int dev_fd, uint32_t group, uint64_t attr, void *val);
tools/testing/selftests/kvm/include/kvm_util.h
894
static inline void kvm_device_attr_get(int dev_fd, uint32_t group,
tools/testing/selftests/kvm/include/kvm_util.h
897
int ret = __kvm_device_attr_get(dev_fd, group, attr, val);
tools/testing/selftests/kvm/include/kvm_util.h
902
int __kvm_device_attr_set(int dev_fd, uint32_t group, uint64_t attr, void *val);
tools/testing/selftests/kvm/include/kvm_util.h
904
static inline void kvm_device_attr_set(int dev_fd, uint32_t group,
tools/testing/selftests/kvm/include/kvm_util.h
907
int ret = __kvm_device_attr_set(dev_fd, group, attr, val);
tools/testing/selftests/kvm/include/kvm_util.h
912
static inline int __vcpu_has_device_attr(struct kvm_vcpu *vcpu, uint32_t group,
tools/testing/selftests/kvm/include/kvm_util.h
915
return __kvm_has_device_attr(vcpu->fd, group, attr);
tools/testing/selftests/kvm/include/kvm_util.h
918
static inline void vcpu_has_device_attr(struct kvm_vcpu *vcpu, uint32_t group,
tools/testing/selftests/kvm/include/kvm_util.h
921
kvm_has_device_attr(vcpu->fd, group, attr);
tools/testing/selftests/kvm/include/kvm_util.h
924
static inline int __vcpu_device_attr_get(struct kvm_vcpu *vcpu, uint32_t group,
tools/testing/selftests/kvm/include/kvm_util.h
927
return __kvm_device_attr_get(vcpu->fd, group, attr, val);
tools/testing/selftests/kvm/include/kvm_util.h
930
static inline void vcpu_device_attr_get(struct kvm_vcpu *vcpu, uint32_t group,
tools/testing/selftests/kvm/include/kvm_util.h
933
kvm_device_attr_get(vcpu->fd, group, attr, val);
tools/testing/selftests/kvm/include/kvm_util.h
936
static inline int __vcpu_device_attr_set(struct kvm_vcpu *vcpu, uint32_t group,
tools/testing/selftests/kvm/include/kvm_util.h
939
return __kvm_device_attr_set(vcpu->fd, group, attr, val);
tools/testing/selftests/kvm/include/kvm_util.h
942
static inline void vcpu_device_attr_set(struct kvm_vcpu *vcpu, uint32_t group,
tools/testing/selftests/kvm/include/kvm_util.h
945
kvm_device_attr_set(vcpu->fd, group, attr, val);
tools/testing/selftests/kvm/lib/arm64/gic.c
159
void gic_irq_set_group(unsigned int intid, bool group)
tools/testing/selftests/kvm/lib/arm64/gic.c
162
gic_common_ops->gic_irq_set_group(intid, group);
tools/testing/selftests/kvm/lib/arm64/gic_private.h
28
void (*gic_irq_set_group)(uint32_t intid, bool group);
tools/testing/selftests/kvm/lib/arm64/vgic.c
163
uint32_t group = intid_is_private ? KVM_DEV_ARM_VGIC_GRP_REDIST_REGS
tools/testing/selftests/kvm/lib/arm64/vgic.c
181
kvm_device_attr_get(gic_fd, group, attr, &val);
tools/testing/selftests/kvm/lib/arm64/vgic.c
183
kvm_device_attr_set(gic_fd, group, attr, &val);
tools/testing/selftests/kvm/lib/kvm_util.c
1821
int __kvm_has_device_attr(int dev_fd, uint32_t group, uint64_t attr)
tools/testing/selftests/kvm/lib/kvm_util.c
1824
.group = group,
tools/testing/selftests/kvm/lib/kvm_util.c
1856
int __kvm_device_attr_get(int dev_fd, uint32_t group, uint64_t attr, void *val)
tools/testing/selftests/kvm/lib/kvm_util.c
1859
.group = group,
tools/testing/selftests/kvm/lib/kvm_util.c
1868
int __kvm_device_attr_set(int dev_fd, uint32_t group, uint64_t attr, void *val)
tools/testing/selftests/kvm/lib/kvm_util.c
1871
.group = group,
tools/testing/selftests/kvm/lib/x86/processor.c
967
.group = 0,
tools/testing/selftests/kvm/s390/ucontrol_test.c
205
.group = KVM_S390_VM_MEM_CTRL,
tools/testing/selftests/kvm/s390/ucontrol_test.c
647
.group = KVM_DEV_FLIC_GET_ALL_IRQS,
tools/testing/selftests/kvm/s390/ucontrol_test.c
655
.a = { .group = KVM_DEV_FLIC_ENQUEUE, },
tools/testing/selftests/kvm/s390/ucontrol_test.c
660
.a = { .group = KVM_DEV_FLIC_CLEAR_IRQS, },
tools/testing/selftests/kvm/s390/ucontrol_test.c
666
.group = KVM_DEV_FLIC_ADAPTER_REGISTER,
tools/testing/selftests/kvm/s390/ucontrol_test.c
675
.group = KVM_DEV_FLIC_ADAPTER_MODIFY,
tools/testing/selftests/kvm/s390/ucontrol_test.c
685
.group = KVM_DEV_FLIC_CLEAR_IO_IRQ,
tools/testing/selftests/kvm/s390/ucontrol_test.c
694
.group = KVM_DEV_FLIC_AISM,
tools/testing/selftests/kvm/s390/ucontrol_test.c
701
.a = { .group = KVM_DEV_FLIC_AIRQ_INJECT, },
tools/testing/selftests/kvm/s390/ucontrol_test.c
708
.group = KVM_DEV_FLIC_AISM_ALL,
tools/testing/selftests/kvm/s390/ucontrol_test.c
717
.a = { .group = KVM_DEV_FLIC_APF_ENABLE, },
tools/testing/selftests/kvm/s390/ucontrol_test.c
723
.a = { .group = KVM_DEV_FLIC_APF_DISABLE_WAIT, },
tools/testing/selftests/kvm/s390/ucontrol_test.c
741
.group = uc_flic_attr_tests[i].a.group,
tools/testing/selftests/kvm/steal_time.c
162
.group = KVM_ARM_VCPU_PVTIME_CTRL,
tools/testing/selftests/kvm/steal_time.c
176
.group = KVM_ARM_VCPU_PVTIME_CTRL,
tools/testing/selftests/kvm/steal_time.c
351
.group = KVM_LOONGARCH_VCPU_CPUCFG,
tools/testing/selftests/kvm/steal_time.c
373
.group = KVM_LOONGARCH_VCPU_PVTIME_CTRL,
tools/testing/selftests/net/ovpn/ovpn-cli.c
1425
const char *group;
tools/testing/selftests/net/ovpn/ovpn-cli.c
1453
grp->group, nla_len(tb_mcgrp[CTRL_ATTR_MCAST_GRP_NAME])))
tools/testing/selftests/net/ovpn/ovpn-cli.c
1532
const char *group)
tools/testing/selftests/net/ovpn/ovpn-cli.c
1538
.group = group,
tools/testing/selftests/resctrl/cat_test.c
383
.group = "CAT",
tools/testing/selftests/resctrl/cat_test.c
392
.group = "CAT",
tools/testing/selftests/resctrl/cat_test.c
400
.group = "CAT",
tools/testing/selftests/resctrl/resctrl.h
106
const char *group;
tools/testing/selftests/resctrl/resctrl_tests.c
294
(resctrl_tests[i]->group &&
tools/testing/selftests/resctrl/resctrl_tests.c
295
!strcasecmp(token, resctrl_tests[i]->group))) {
tools/testing/selftests/resctrl/resctrl_tests.c
86
if (resctrl_tests[i]->group)
tools/testing/selftests/resctrl/resctrl_tests.c
87
printf("\t\t\t%s (%s)\n", resctrl_tests[i]->name, resctrl_tests[i]->group);
tools/testing/selftests/rlimits/rlimits-per-userns.c
28
static uid_t group = 60000;
tools/testing/selftests/rlimits/rlimits-per-userns.c
63
warnx("(pid=%d): Changing to uid=%d, gid=%d", pid, user, group);
tools/testing/selftests/rlimits/rlimits-per-userns.c
65
if (setgid(group) < 0)
tools/testing/selftests/rlimits/rlimits-per-userns.c
66
err(EXIT_FAILURE, "(pid=%d): setgid(%d)", pid, group);
tools/testing/selftests/safesetid/safesetid-test.c
138
struct group g;
tools/testing/selftests/vfio/lib/vfio_pci_device.c
209
unsigned int group;
tools/testing/selftests/vfio/lib/vfio_pci_device.c
217
ret = sscanf(basename(dev_iommu_group_path), "%u", &group);
tools/testing/selftests/vfio/lib/vfio_pci_device.c
220
return group;
tools/testing/selftests/vfio/lib/vfio_pci_device.c
229
int group;
tools/testing/selftests/vfio/lib/vfio_pci_device.c
231
group = vfio_pci_get_group_from_dev(bdf);
tools/testing/selftests/vfio/lib/vfio_pci_device.c
232
snprintf(group_path, sizeof(group_path), "/dev/vfio/%d", group);
virt/kvm/vfio.c
295
switch (attr->group) {
virt/kvm/vfio.c
307
switch (attr->group) {