Symbol: blkcg
block/bfq-cgroup.c
1020
struct blkcg *blkcg = css_to_blkcg(css);
block/bfq-cgroup.c
1021
struct bfq_group_data *bfqgd = blkcg_to_bfqgd(blkcg);
block/bfq-cgroup.c
1029
spin_lock_irq(&blkcg->lock);
block/bfq-cgroup.c
1031
hlist_for_each_entry(blkg, &blkcg->blkg_list, blkcg_node) {
block/bfq-cgroup.c
1037
spin_unlock_irq(&blkcg->lock);
block/bfq-cgroup.c
1048
struct blkcg *blkcg = css_to_blkcg(of_css(of));
block/bfq-cgroup.c
1054
ret = blkg_conf_prep(blkcg, &blkcg_policy_bfq, &ctx);
block/bfq-cgroup.c
488
static struct bfq_group_data *blkcg_to_bfqgd(struct blkcg *blkcg)
block/bfq-cgroup.c
490
return cpd_to_bfqgd(blkcg_to_cpd(blkcg, &blkcg_policy_bfq));
block/bfq-cgroup.c
511
struct blkcg *blkcg, gfp_t gfp)
block/bfq-cgroup.c
535
struct bfq_group_data *d = blkcg_to_bfqgd(blkg->blkcg);
block/bfq-cgroup.c
611
bio_associate_blkg_from_css(bio, &blkg->blkcg->css);
block/bfq-cgroup.c
617
&bfqg_to_blkg(bfqd->root_group)->blkcg->css);
block/bfq-cgroup.c
780
serial_nr = bfqg_to_blkg(bfqg)->blkcg->css.serial_nr;
block/bfq-cgroup.c
950
struct blkcg *blkcg = css_to_blkcg(seq_css(sf));
block/bfq-cgroup.c
951
struct bfq_group_data *bfqgd = blkcg_to_bfqgd(blkcg);
block/bfq-cgroup.c
974
struct blkcg *blkcg = css_to_blkcg(seq_css(sf));
block/bfq-cgroup.c
975
struct bfq_group_data *bfqgd = blkcg_to_bfqgd(blkcg);
block/bfq-cgroup.c
978
blkcg_print_blkgs(sf, blkcg, bfqg_prfill_weight_device,
block/bfq-iosched.c
604
depth = bfqg_to_blkg(bfqq_group(bfqq))->blkcg->css.cgroup->level + 1;
block/bfq-iosched.h
1183
&bfqg_to_blkg(bfqq_group(bfqq))->blkcg->css, \
block/blk-cgroup-fc-appid.c
15
struct blkcg *blkcg;
block/blk-cgroup-fc-appid.c
29
blkcg = css_to_blkcg(css);
block/blk-cgroup-fc-appid.c
37
strscpy(blkcg->fc_app_id, app_id, app_id_len);
block/blk-cgroup-fc-appid.c
53
if (!bio->bi_blkg || bio->bi_blkg->blkcg->fc_app_id[0] == '\0')
block/blk-cgroup-fc-appid.c
55
return bio->bi_blkg->blkcg->fc_app_id;
block/blk-cgroup.c
1043
static void __blkcg_rstat_flush(struct blkcg *blkcg, int cpu)
block/blk-cgroup.c
1045
struct llist_head *lhead = per_cpu_ptr(blkcg->lhead, cpu);
block/blk-cgroup.c
1108
plhead = per_cpu_ptr(parent->blkcg->lhead, cpu);
block/blk-cgroup.c
1232
struct blkcg *blkcg = css_to_blkcg(seq_css(sf));
block/blk-cgroup.c
1238
css_rstat_flush(&blkcg->css);
block/blk-cgroup.c
1241
hlist_for_each_entry_rcu(blkg, &blkcg->blkg_list, blkcg_node) {
block/blk-cgroup.c
1305
static void blkcg_destroy_blkgs(struct blkcg *blkcg)
block/blk-cgroup.c
1309
spin_lock_irq(&blkcg->lock);
block/blk-cgroup.c
1311
while (!hlist_empty(&blkcg->blkg_list)) {
block/blk-cgroup.c
1312
struct blkcg_gq *blkg = hlist_entry(blkcg->blkg_list.first,
block/blk-cgroup.c
1322
spin_unlock_irq(&blkcg->lock);
block/blk-cgroup.c
1324
spin_lock_irq(&blkcg->lock);
block/blk-cgroup.c
1332
spin_unlock_irq(&blkcg->lock);
block/blk-cgroup.c
1359
struct blkcg *blkcg = css_to_blkcg(blkcg_css);
block/blk-cgroup.c
1362
struct blkcg *parent;
block/blk-cgroup.c
1364
if (!refcount_dec_and_test(&blkcg->online_pin))
block/blk-cgroup.c
1367
parent = blkcg_parent(blkcg);
block/blk-cgroup.c
1368
blkcg_destroy_blkgs(blkcg);
block/blk-cgroup.c
1369
blkcg = parent;
block/blk-cgroup.c
1370
} while (blkcg);
block/blk-cgroup.c
1392
struct blkcg *blkcg = css_to_blkcg(css);
block/blk-cgroup.c
1397
list_del(&blkcg->all_blkcgs_node);
block/blk-cgroup.c
1400
if (blkcg->cpd[i])
block/blk-cgroup.c
1401
blkcg_policy[i]->cpd_free_fn(blkcg->cpd[i]);
block/blk-cgroup.c
1405
free_percpu(blkcg->lhead);
block/blk-cgroup.c
1406
kfree(blkcg);
block/blk-cgroup.c
1412
struct blkcg *blkcg;
block/blk-cgroup.c
1418
blkcg = &blkcg_root;
block/blk-cgroup.c
1420
blkcg = kzalloc_obj(*blkcg);
block/blk-cgroup.c
1421
if (!blkcg)
block/blk-cgroup.c
1425
if (init_blkcg_llists(blkcg))
block/blk-cgroup.c
1445
blkcg->cpd[i] = cpd;
block/blk-cgroup.c
1446
cpd->blkcg = blkcg;
block/blk-cgroup.c
1450
spin_lock_init(&blkcg->lock);
block/blk-cgroup.c
1451
refcount_set(&blkcg->online_pin, 1);
block/blk-cgroup.c
1452
INIT_RADIX_TREE(&blkcg->blkg_tree, GFP_NOWAIT);
block/blk-cgroup.c
1453
INIT_HLIST_HEAD(&blkcg->blkg_list);
block/blk-cgroup.c
1455
INIT_LIST_HEAD(&blkcg->cgwb_list);
block/blk-cgroup.c
1457
list_add_tail(&blkcg->all_blkcgs_node, &all_blkcgs);
block/blk-cgroup.c
1460
return &blkcg->css;
block/blk-cgroup.c
1464
if (blkcg->cpd[i])
block/blk-cgroup.c
1465
blkcg_policy[i]->cpd_free_fn(blkcg->cpd[i]);
block/blk-cgroup.c
1466
free_percpu(blkcg->lhead);
block/blk-cgroup.c
1468
if (blkcg != &blkcg_root)
block/blk-cgroup.c
1469
kfree(blkcg);
block/blk-cgroup.c
1477
struct blkcg *parent = blkcg_parent(css_to_blkcg(css));
block/blk-cgroup.c
1614
pd = pol->pd_alloc_fn(disk, blkg->blkcg,
block/blk-cgroup.c
1632
pd_prealloc = pol->pd_alloc_fn(disk, blkg->blkcg,
block/blk-cgroup.c
1640
spin_lock(&blkg->blkcg->lock);
block/blk-cgroup.c
1653
spin_unlock(&blkg->blkcg->lock);
block/blk-cgroup.c
166
struct blkcg *blkcg = blkg->blkcg;
block/blk-cgroup.c
1673
struct blkcg *blkcg = blkg->blkcg;
block/blk-cgroup.c
1676
spin_lock(&blkcg->lock);
block/blk-cgroup.c
1685
spin_unlock(&blkcg->lock);
block/blk-cgroup.c
1720
struct blkcg *blkcg = blkg->blkcg;
block/blk-cgroup.c
1722
spin_lock(&blkcg->lock);
block/blk-cgroup.c
1729
spin_unlock(&blkcg->lock);
block/blk-cgroup.c
1742
struct blkcg *blkcg;
block/blk-cgroup.c
1744
list_for_each_entry(blkcg, &all_blkcgs, all_blkcgs_node) {
block/blk-cgroup.c
1745
if (blkcg->cpd[pol->plid]) {
block/blk-cgroup.c
1746
pol->cpd_free_fn(blkcg->cpd[pol->plid]);
block/blk-cgroup.c
1747
blkcg->cpd[pol->plid] = NULL;
block/blk-cgroup.c
1761
struct blkcg *blkcg;
block/blk-cgroup.c
179
__blkcg_rstat_flush(blkcg, cpu);
block/blk-cgroup.c
1791
list_for_each_entry(blkcg, &all_blkcgs, all_blkcgs_node) {
block/blk-cgroup.c
1800
blkcg->cpd[pol->plid] = cpd;
block/blk-cgroup.c
1801
cpd->blkcg = blkcg;
block/blk-cgroup.c
182
css_put(&blkg->blkcg->css);
block/blk-cgroup.c
1998
struct blkcg *blkcg;
block/blk-cgroup.c
2009
blkcg = css_to_blkcg(blkcg_css());
block/blk-cgroup.c
2010
if (!blkcg)
block/blk-cgroup.c
2012
blkg = blkg_lookup(blkcg, disk->queue);
block/blk-cgroup.c
2190
struct blkcg *blkcg = bio->bi_blkg->blkcg;
block/blk-cgroup.c
2199
if (!cgroup_parent(blkcg->css.cgroup))
block/blk-cgroup.c
2221
struct llist_head *lhead = this_cpu_ptr(blkcg->lhead);
block/blk-cgroup.c
2228
css_rstat_updated(&blkcg->css, cpu);
block/blk-cgroup.c
2234
struct blkcg *blkcg;
block/blk-cgroup.c
2238
for (blkcg = css_to_blkcg(blkcg_css()); blkcg;
block/blk-cgroup.c
2239
blkcg = blkcg_parent(blkcg)) {
block/blk-cgroup.c
2240
if (atomic_read(&blkcg->congestion_count)) {
block/blk-cgroup.c
275
return &bio->bi_blkg->blkcg->css;
block/blk-cgroup.c
285
static inline struct blkcg *blkcg_parent(struct blkcg *blkcg)
block/blk-cgroup.c
287
return css_to_blkcg(blkcg->css.parent);
block/blk-cgroup.c
298
static struct blkcg_gq *blkg_alloc(struct blkcg *blkcg, struct gendisk *disk,
block/blk-cgroup.c
318
blkg->blkcg = blkcg;
block/blk-cgroup.c
340
pd = pol->pd_alloc_fn(disk, blkcg, gfp_mask);
block/blk-cgroup.c
369
static struct blkcg_gq *blkg_create(struct blkcg *blkcg, struct gendisk *disk,
block/blk-cgroup.c
37
static void __blkcg_rstat_flush(struct blkcg *blkcg, int cpu);
block/blk-cgroup.c
384
if (!css_tryget_online(&blkcg->css)) {
block/blk-cgroup.c
391
new_blkg = blkg_alloc(blkcg, disk, GFP_NOWAIT);
block/blk-cgroup.c
400
if (blkcg_parent(blkcg)) {
block/blk-cgroup.c
401
blkg->parent = blkg_lookup(blkcg_parent(blkcg), disk->queue);
block/blk-cgroup.c
418
spin_lock(&blkcg->lock);
block/blk-cgroup.c
419
ret = radix_tree_insert(&blkcg->blkg_tree, disk->queue->id, blkg);
block/blk-cgroup.c
421
hlist_add_head_rcu(&blkg->blkcg_node, &blkcg->blkg_list);
block/blk-cgroup.c
435
spin_unlock(&blkcg->lock);
block/blk-cgroup.c
445
css_put(&blkcg->css);
block/blk-cgroup.c
465
static struct blkcg_gq *blkg_lookup_create(struct blkcg *blkcg,
block/blk-cgroup.c
474
blkg = blkg_lookup(blkcg, q);
block/blk-cgroup.c
479
blkg = blkg_lookup(blkcg, q);
block/blk-cgroup.c
481
if (blkcg != &blkcg_root &&
block/blk-cgroup.c
482
blkg != rcu_dereference(blkcg->blkg_hint))
block/blk-cgroup.c
483
rcu_assign_pointer(blkcg->blkg_hint, blkg);
block/blk-cgroup.c
49
struct blkcg blkcg_root;
block/blk-cgroup.c
493
struct blkcg *pos = blkcg;
block/blk-cgroup.c
494
struct blkcg *parent = blkcg_parent(blkcg);
block/blk-cgroup.c
513
if (pos == blkcg)
block/blk-cgroup.c
524
struct blkcg *blkcg = blkg->blkcg;
block/blk-cgroup.c
528
lockdep_assert_held(&blkcg->lock);
block/blk-cgroup.c
551
radix_tree_delete(&blkcg->blkg_tree, blkg->q->id);
block/blk-cgroup.c
559
if (rcu_access_pointer(blkcg->blkg_hint) == blkg)
block/blk-cgroup.c
560
rcu_assign_pointer(blkcg->blkg_hint, NULL);
block/blk-cgroup.c
579
struct blkcg *blkcg = blkg->blkcg;
block/blk-cgroup.c
584
spin_lock(&blkcg->lock);
block/blk-cgroup.c
586
spin_unlock(&blkcg->lock);
block/blk-cgroup.c
652
struct blkcg *blkcg = css_to_blkcg(css);
block/blk-cgroup.c
658
spin_lock_irq(&blkcg->lock);
block/blk-cgroup.c
665
hlist_for_each_entry(blkg, &blkcg->blkg_list, blkcg_node) {
block/blk-cgroup.c
675
spin_unlock_irq(&blkcg->lock);
block/blk-cgroup.c
705
void blkcg_print_blkgs(struct seq_file *sf, struct blkcg *blkcg,
block/blk-cgroup.c
715
hlist_for_each_entry_rcu(blkg, &blkcg->blkg_list, blkcg_node) {
block/blk-cgroup.c
83
static int init_blkcg_llists(struct blkcg *blkcg)
block/blk-cgroup.c
861
int blkg_conf_prep(struct blkcg *blkcg, const struct blkcg_policy *pol,
block/blk-cgroup.c
87
blkcg->lhead = alloc_percpu_gfp(struct llist_head, GFP_KERNEL);
block/blk-cgroup.c
88
if (!blkcg->lhead)
block/blk-cgroup.c
886
blkg = blkg_lookup(blkcg, q);
block/blk-cgroup.c
895
struct blkcg *pos = blkcg;
block/blk-cgroup.c
896
struct blkcg *parent;
block/blk-cgroup.c
899
parent = blkcg_parent(blkcg);
block/blk-cgroup.c
92
init_llist_head(per_cpu_ptr(blkcg->lhead, cpu));
block/blk-cgroup.c
941
if (pos == blkcg)
block/blk-cgroup.h
122
static inline struct blkcg *css_to_blkcg(struct cgroup_subsys_state *css)
block/blk-cgroup.h
124
return css ? container_of(css, struct blkcg, css) : NULL;
block/blk-cgroup.h
154
struct blkcg *blkcg;
block/blk-cgroup.h
191
extern struct blkcg blkcg_root;
block/blk-cgroup.h
206
void blkcg_print_blkgs(struct seq_file *sf, struct blkcg *blkcg,
block/blk-cgroup.h
223
int blkg_conf_prep(struct blkcg *blkcg, const struct blkcg_policy *pol,
block/blk-cgroup.h
255
static inline struct blkcg_gq *blkg_lookup(struct blkcg *blkcg,
block/blk-cgroup.h
260
if (blkcg == &blkcg_root)
block/blk-cgroup.h
263
blkg = rcu_dereference_check(blkcg->blkg_hint,
block/blk-cgroup.h
268
blkg = radix_tree_lookup(&blkcg->blkg_tree, q->id);
block/blk-cgroup.h
287
static inline struct blkcg_policy_data *blkcg_to_cpd(struct blkcg *blkcg,
block/blk-cgroup.h
290
return blkcg ? blkcg->cpd[pol->plid] : NULL;
block/blk-cgroup.h
304
static inline struct blkcg *cpd_to_blkcg(struct blkcg_policy_data *cpd)
block/blk-cgroup.h
306
return cpd ? cpd->blkcg : NULL;
block/blk-cgroup.h
354
css_for_each_descendant_pre((pos_css), &(p_blkg)->blkcg->css) \
block/blk-cgroup.h
369
css_for_each_descendant_post((pos_css), &(p_blkg)->blkcg->css) \
block/blk-cgroup.h
378
atomic_inc(&blkg->blkcg->congestion_count);
block/blk-cgroup.h
403
atomic_dec(&blkg->blkcg->congestion_count);
block/blk-cgroup.h
422
atomic_inc(&blkg->blkcg->congestion_count);
block/blk-cgroup.h
439
atomic_dec(&blkg->blkcg->congestion_count);
block/blk-cgroup.h
479
static inline struct blkcg_gq *blkg_lookup(struct blkcg *blkcg, void *key) { return NULL; }
block/blk-cgroup.h
61
struct blkcg *blkcg;
block/blk-iocost.c
1257
struct ioc_cgrp *iocc = blkcg_to_iocc(blkg->blkcg);
block/blk-iocost.c
1964
pr_cont_cgroup_path(iocg_to_blkg(iocg)->blkcg->css.cgroup);
block/blk-iocost.c
200
cgroup_path(iocg_to_blkg(iocg)->blkcg->css.cgroup, \
block/blk-iocost.c
2963
struct blkcg *blkcg, gfp_t gfp)
block/blk-iocost.c
2965
int levels = blkcg->css.cgroup->level + 1;
block/blk-iocost.c
3006
iocg->level = blkg->blkcg->css.cgroup->level;
block/blk-iocost.c
3084
struct blkcg *blkcg = css_to_blkcg(seq_css(sf));
block/blk-iocost.c
3085
struct ioc_cgrp *iocc = blkcg_to_iocc(blkcg);
block/blk-iocost.c
3088
blkcg_print_blkgs(sf, blkcg, ioc_weight_prfill,
block/blk-iocost.c
3096
struct blkcg *blkcg = css_to_blkcg(of_css(of));
block/blk-iocost.c
3097
struct ioc_cgrp *iocc = blkcg_to_iocc(blkcg);
block/blk-iocost.c
3113
spin_lock_irq(&blkcg->lock);
block/blk-iocost.c
3115
hlist_for_each_entry(blkg, &blkcg->blkg_list, blkcg_node) {
block/blk-iocost.c
3125
spin_unlock_irq(&blkcg->lock);
block/blk-iocost.c
3132
ret = blkg_conf_prep(blkcg, &blkcg_policy_iocost, &ctx);
block/blk-iocost.c
3191
struct blkcg *blkcg = css_to_blkcg(seq_css(sf));
block/blk-iocost.c
3193
blkcg_print_blkgs(sf, blkcg, ioc_qos_prfill,
block/blk-iocost.c
3383
struct blkcg *blkcg = css_to_blkcg(seq_css(sf));
block/blk-iocost.c
3385
blkcg_print_blkgs(sf, blkcg, ioc_cost_model_prfill,
block/blk-iocost.c
694
static struct ioc_cgrp *blkcg_to_iocc(struct blkcg *blkcg)
block/blk-iocost.c
696
return container_of(blkcg_to_cpd(blkcg, &blkcg_policy_iocost),
block/blk-iolatency.c
830
struct blkcg *blkcg = css_to_blkcg(of_css(of));
block/blk-iolatency.c
855
ret = blkg_conf_prep(blkcg, &blkcg_policy_iolatency, &ctx);
block/blk-iolatency.c
966
struct blkcg *blkcg, gfp_t gfp)
block/blk-ioprio.c
100
struct ioprio_blkcg *blkcg;
block/blk-ioprio.c
102
blkcg = kzalloc_obj(*blkcg, gfp);
block/blk-ioprio.c
103
if (!blkcg)
block/blk-ioprio.c
105
blkcg->prio_policy = POLICY_NO_CHANGE;
block/blk-ioprio.c
106
return &blkcg->cpd;
block/blk-ioprio.c
111
struct ioprio_blkcg *blkcg = container_of(cpd, typeof(*blkcg), cpd);
block/blk-ioprio.c
113
kfree(blkcg);
block/blk-ioprio.c
135
struct ioprio_blkcg *blkcg = blkcg_to_ioprio_blkcg(bio->bi_blkg->blkcg);
block/blk-ioprio.c
138
if (!blkcg || blkcg->prio_policy == POLICY_NO_CHANGE)
block/blk-ioprio.c
141
if (blkcg->prio_policy == POLICY_PROMOTE_TO_RT ||
block/blk-ioprio.c
142
blkcg->prio_policy == POLICY_NONE_TO_RT) {
block/blk-ioprio.c
163
IOPRIO_PRIO_VALUE(blkcg->prio_policy, 0));
block/blk-ioprio.c
62
static struct ioprio_blkcg *blkcg_to_ioprio_blkcg(struct blkcg *blkcg)
block/blk-ioprio.c
64
return container_of(blkcg_to_cpd(blkcg, &ioprio_policy),
block/blk-ioprio.c
76
struct ioprio_blkcg *blkcg = ioprio_blkcg_from_css(seq_css(sf));
block/blk-ioprio.c
78
seq_printf(sf, "%s\n", policy_name[blkcg->prio_policy]);
block/blk-ioprio.c
85
struct ioprio_blkcg *blkcg = ioprio_blkcg_from_css(of_css(of));
block/blk-ioprio.c
94
blkcg->prio_policy = ret;
block/blk-throttle.c
122
&tg_to_blkg(__tg)->blkcg->css, "throtl " fmt, ##args);\
block/blk-throttle.c
1346
struct blkcg *blkcg = css_to_blkcg(of_css(of));
block/blk-throttle.c
1364
ret = blkg_conf_prep(blkcg, &blkcg_policy_throtl, &ctx);
block/blk-throttle.c
1530
struct blkcg *blkcg = css_to_blkcg(of_css(of));
block/blk-throttle.c
1548
ret = blkg_conf_prep(blkcg, &blkcg_policy_throtl, &ctx);
block/blk-throttle.c
262
struct blkcg *blkcg, gfp_t gfp)