Symbol: smc_lgr_list
net/smc/smc_core.c
1669
spin_lock_bh(&smc_lgr_list.lock);
net/smc/smc_core.c
1671
list_splice_init(&smc_lgr_list.list, &lgr_free_list);
net/smc/smc_core.c
1675
list_for_each_entry_safe(lgr, lg, &smc_lgr_list.list, list) {
net/smc/smc_core.c
1682
spin_unlock_bh(&smc_lgr_list.lock);
net/smc/smc_core.c
1761
spin_lock_bh(&smc_lgr_list.lock);
net/smc/smc_core.c
1762
list_for_each_entry_safe(lgr, n, &smc_lgr_list.list, list) {
net/smc/smc_core.c
1780
spin_unlock_bh(&smc_lgr_list.lock);
net/smc/smc_core.c
1851
list_for_each_entry_safe(lgr, n, &smc_lgr_list.list, list) {
net/smc/smc_core.c
1987
&smc_lgr_list.list;
net/smc/smc_core.c
1989
&smc_lgr_list.lock;
net/smc/smc_core.c
43
struct smc_lgr_list smc_lgr_list = { /* established link groups */
net/smc/smc_core.c
44
.lock = __SPIN_LOCK_UNLOCKED(smc_lgr_list.lock),
net/smc/smc_core.c
45
.list = LIST_HEAD_INIT(smc_lgr_list.list),
net/smc/smc_core.c
512
static void smc_nl_fill_lgr_list(struct smc_lgr_list *smc_lgr,
net/smc/smc_core.c
664
smc_nl_fill_lgr_list(&smc_lgr_list, skb, cb, list_links);
net/smc/smc_core.c
67
*lgr_lock = &smc_lgr_list.lock;
net/smc/smc_core.c
672
smc_nl_fill_lgr_list(&smc_lgr_list, skb, cb, list_links);
net/smc/smc_core.c
68
return &smc_lgr_list.list;
net/smc/smc_core.c
933
smc_lgr_list.num += SMC_LGR_NUM_INCR;
net/smc/smc_core.c
934
memcpy(&lgr->id, (u8 *)&smc_lgr_list.num, SMC_LGR_ID_SIZE);
net/smc/smc_core.c
991
lgr_list = &smc_lgr_list.list;
net/smc/smc_core.c
992
lgr_lock = &smc_lgr_list.lock;
net/smc/smc_ib.c
338
spin_lock_bh(&smc_lgr_list.lock);
net/smc/smc_ib.c
339
list_for_each_entry(lgr, &smc_lgr_list.list, list) {
net/smc/smc_ib.c
355
spin_unlock_bh(&smc_lgr_list.lock);
net/smc/smc_ib.c
469
static bool smcr_diag_is_dev_critical(struct smc_lgr_list *smc_lgr,
net/smc/smc_ib.c
572
is_crit = smcr_diag_is_dev_critical(&smc_lgr_list, smcibdev);
net/smc/smc_ib.h
33
extern struct smc_lgr_list smc_lgr_list; /* list of linkgroups */