Symbol: sched_group
include/linux/sched/topology.h
175
struct sched_group *__percpu *sg;
include/linux/sched/topology.h
64
struct sched_group;
include/linux/sched/topology.h
77
struct sched_group *groups; /* the balancing groups of the domain */
kernel/sched/ext_idle.c
264
struct sched_group *sg;
kernel/sched/ext_idle.c
283
struct sched_group *sg;
kernel/sched/fair.c
10056
struct sched_group *busiest; /* Busiest group in this sd */
kernel/sched/fair.c
10057
struct sched_group *local; /* Local group in this sd */
kernel/sched/fair.c
10118
struct sched_group *sdg = sd->groups;
kernel/sched/fair.c
10134
struct sched_group *group, *sdg = sd->groups;
kernel/sched/fair.c
10233
static inline int sg_imbalanced(struct sched_group *group)
kernel/sched/fair.c
10294
struct sched_group *group,
kernel/sched/fair.c
10363
sched_group_asym(struct lb_env *env, struct sg_lb_stats *sgs, struct sched_group *group)
kernel/sched/fair.c
10377
static inline bool smt_vs_nonsmt_groups(struct sched_group *sg1,
kernel/sched/fair.c
10378
struct sched_group *sg2)
kernel/sched/fair.c
10388
struct sched_group *group)
kernel/sched/fair.c
10465
struct sched_group *group,
kernel/sched/fair.c
10564
struct sched_group *sg,
kernel/sched/fair.c
10784
struct sched_group *group,
kernel/sched/fair.c
10839
static bool update_pick_idlest(struct sched_group *idlest,
kernel/sched/fair.c
10841
struct sched_group *group,
kernel/sched/fair.c
10897
static struct sched_group *
kernel/sched/fair.c
10900
struct sched_group *idlest = NULL, *local = NULL, *group = sd->groups;
kernel/sched/fair.c
11141
struct sched_group *sg = env->sd->groups;
kernel/sched/fair.c
11409
static struct sched_group *sched_balance_find_src_group(struct lb_env *env)
kernel/sched/fair.c
11548
struct sched_group *group)
kernel/sched/fair.c
11761
struct sched_group *sg = env->sd->groups;
kernel/sched/fair.c
11871
struct sched_group *group;
kernel/sched/fair.c
7439
static struct sched_group *
kernel/sched/fair.c
7446
sched_balance_find_dst_group_cpu(struct sched_group *group, struct task_struct *p, int this_cpu)
kernel/sched/fair.c
7518
struct sched_group *group;
kernel/sched/fair.c
7721
struct sched_group *sg = sd->groups;
kernel/sched/sched.h
1431
static inline struct cpumask *sched_group_span(struct sched_group *sg);
kernel/sched/sched.h
1514
struct sched_group *group)
kernel/sched/sched.h
1575
struct sched_group *group)
kernel/sched/sched.h
2185
struct sched_group *next; /* Must be a circular list */
kernel/sched/sched.h
2204
static inline struct cpumask *sched_group_span(struct sched_group *sg)
kernel/sched/sched.h
2212
static inline struct cpumask *group_balance_mask(struct sched_group *sg)
kernel/sched/sched.h
2217
extern int group_balance_cpu(struct sched_group *sg);
kernel/sched/sched.h
86
struct sched_group;
kernel/sched/topology.c
1032
struct sched_group *first = NULL, *last = NULL, *sg;
kernel/sched/topology.c
1195
static struct sched_group *get_group(int cpu, struct sd_data *sdd)
kernel/sched/topology.c
1199
struct sched_group *sg;
kernel/sched/topology.c
1243
struct sched_group *first = NULL, *last = NULL;
kernel/sched/topology.c
1255
struct sched_group *sg;
kernel/sched/topology.c
1288
struct sched_group *sg = sd->groups;
kernel/sched/topology.c
1337
struct sched_group *sg;
kernel/sched/topology.c
2379
sdd->sg = alloc_percpu(struct sched_group *);
kernel/sched/topology.c
2390
struct sched_group *sg;
kernel/sched/topology.c
2407
sg = kzalloc_node(sizeof(struct sched_group) + cpumask_size(),
kernel/sched/topology.c
46
struct sched_group *group = sd->groups;
kernel/sched/topology.c
604
static void free_sched_groups(struct sched_group *sg, int free_sgc)
kernel/sched/topology.c
606
struct sched_group *tmp, *first;
kernel/sched/topology.c
757
struct sched_group *sg = sd->groups;
kernel/sched/topology.c
804
int group_balance_cpu(struct sched_group *sg)
kernel/sched/topology.c
916
build_balance_mask(struct sched_domain *sd, struct sched_group *sg, struct cpumask *mask)
kernel/sched/topology.c
952
static struct sched_group *
kernel/sched/topology.c
955
struct sched_group *sg;
kernel/sched/topology.c
958
sg = kzalloc_node(sizeof(struct sched_group) + cpumask_size(),
kernel/sched/topology.c
977
struct sched_group *sg)