sched_group
struct sched_group *__percpu *sg;
struct sched_group;
struct sched_group *groups; /* the balancing groups of the domain */
struct sched_group *sg;
struct sched_group *sg;
struct sched_group *busiest; /* Busiest group in this sd */
struct sched_group *local; /* Local group in this sd */
struct sched_group *sdg = sd->groups;
struct sched_group *group, *sdg = sd->groups;
static inline int sg_imbalanced(struct sched_group *group)
struct sched_group *group,
sched_group_asym(struct lb_env *env, struct sg_lb_stats *sgs, struct sched_group *group)
static inline bool smt_vs_nonsmt_groups(struct sched_group *sg1,
struct sched_group *sg2)
struct sched_group *group)
struct sched_group *group,
struct sched_group *sg,
struct sched_group *group,
static bool update_pick_idlest(struct sched_group *idlest,
struct sched_group *group,
static struct sched_group *
struct sched_group *idlest = NULL, *local = NULL, *group = sd->groups;
struct sched_group *sg = env->sd->groups;
static struct sched_group *sched_balance_find_src_group(struct lb_env *env)
struct sched_group *group)
struct sched_group *sg = env->sd->groups;
struct sched_group *group;
static struct sched_group *
sched_balance_find_dst_group_cpu(struct sched_group *group, struct task_struct *p, int this_cpu)
struct sched_group *group;
struct sched_group *sg = sd->groups;
static inline struct cpumask *sched_group_span(struct sched_group *sg);
struct sched_group *group)
struct sched_group *group)
struct sched_group *next; /* Must be a circular list */
static inline struct cpumask *sched_group_span(struct sched_group *sg)
static inline struct cpumask *group_balance_mask(struct sched_group *sg)
extern int group_balance_cpu(struct sched_group *sg);
struct sched_group;
struct sched_group *first = NULL, *last = NULL, *sg;
static struct sched_group *get_group(int cpu, struct sd_data *sdd)
struct sched_group *sg;
struct sched_group *first = NULL, *last = NULL;
struct sched_group *sg;
struct sched_group *sg = sd->groups;
struct sched_group *sg;
sdd->sg = alloc_percpu(struct sched_group *);
struct sched_group *sg;
sg = kzalloc_node(sizeof(struct sched_group) + cpumask_size(),
struct sched_group *group = sd->groups;
static void free_sched_groups(struct sched_group *sg, int free_sgc)
struct sched_group *tmp, *first;
struct sched_group *sg = sd->groups;
int group_balance_cpu(struct sched_group *sg)
build_balance_mask(struct sched_domain *sd, struct sched_group *sg, struct cpumask *mask)
static struct sched_group *
struct sched_group *sg;
sg = kzalloc_node(sizeof(struct sched_group) + cpumask_size(),
struct sched_group *sg)