Symbol: ovs_net
net/openvswitch/conntrack.c
1372
struct ovs_net *ovs_net = net_generic(net, ovs_net_id);
net/openvswitch/conntrack.c
1374
return ovs_net->xt_label;
net/openvswitch/conntrack.c
1585
static int ovs_ct_limit_init(struct net *net, struct ovs_net *ovs_net)
net/openvswitch/conntrack.c
1589
ovs_net->ct_limit_info = kmalloc_obj(*ovs_net->ct_limit_info);
net/openvswitch/conntrack.c
1590
if (!ovs_net->ct_limit_info)
net/openvswitch/conntrack.c
1593
ovs_net->ct_limit_info->default_limit = OVS_CT_LIMIT_DEFAULT;
net/openvswitch/conntrack.c
1594
ovs_net->ct_limit_info->limits =
net/openvswitch/conntrack.c
1596
if (!ovs_net->ct_limit_info->limits) {
net/openvswitch/conntrack.c
1597
kfree(ovs_net->ct_limit_info);
net/openvswitch/conntrack.c
1602
INIT_HLIST_HEAD(&ovs_net->ct_limit_info->limits[i]);
net/openvswitch/conntrack.c
1604
ovs_net->ct_limit_info->data = nf_conncount_init(net, sizeof(u32));
net/openvswitch/conntrack.c
1606
if (IS_ERR(ovs_net->ct_limit_info->data)) {
net/openvswitch/conntrack.c
1607
err = PTR_ERR(ovs_net->ct_limit_info->data);
net/openvswitch/conntrack.c
1608
kfree(ovs_net->ct_limit_info->limits);
net/openvswitch/conntrack.c
1609
kfree(ovs_net->ct_limit_info);
net/openvswitch/conntrack.c
1616
static void ovs_ct_limit_exit(struct net *net, struct ovs_net *ovs_net)
net/openvswitch/conntrack.c
1618
const struct ovs_ct_limit_info *info = ovs_net->ct_limit_info;
net/openvswitch/conntrack.c
1851
struct ovs_net *ovs_net = net_generic(sock_net(skb->sk), ovs_net_id);
net/openvswitch/conntrack.c
1852
struct ovs_ct_limit_info *ct_limit_info = ovs_net->ct_limit_info;
net/openvswitch/conntrack.c
1885
struct ovs_net *ovs_net = net_generic(sock_net(skb->sk), ovs_net_id);
net/openvswitch/conntrack.c
1886
struct ovs_ct_limit_info *ct_limit_info = ovs_net->ct_limit_info;
net/openvswitch/conntrack.c
1919
struct ovs_net *ovs_net = net_generic(net, ovs_net_id);
net/openvswitch/conntrack.c
1920
struct ovs_ct_limit_info *ct_limit_info = ovs_net->ct_limit_info;
net/openvswitch/conntrack.c
2002
struct ovs_net *ovs_net = net_generic(net, ovs_net_id);
net/openvswitch/conntrack.c
2005
ovs_net->xt_label = false;
net/openvswitch/conntrack.c
2008
ovs_net->xt_label = true;
net/openvswitch/conntrack.c
2012
return ovs_ct_limit_init(net, ovs_net);
net/openvswitch/conntrack.c
2020
struct ovs_net *ovs_net = net_generic(net, ovs_net_id);
net/openvswitch/conntrack.c
2023
ovs_ct_limit_exit(net, ovs_net);
net/openvswitch/conntrack.c
2026
if (ovs_net->xt_label)
net/openvswitch/conntrack.c
934
struct ovs_net *ovs_net = net_generic(net, ovs_net_id);
net/openvswitch/conntrack.c
935
const struct ovs_ct_limit_info *ct_limit_info = ovs_net->ct_limit_info;
net/openvswitch/datapath.c
1817
struct ovs_net *ovs_net;
net/openvswitch/datapath.c
1890
ovs_net = net_generic(ovs_dp_get_net(dp), ovs_net_id);
net/openvswitch/datapath.c
1891
list_add_tail_rcu(&dp->list_node, &ovs_net->dps);
net/openvswitch/datapath.c
2055
struct ovs_net *ovs_net = net_generic(sock_net(skb->sk), ovs_net_id);
net/openvswitch/datapath.c
2061
list_for_each_entry(dp, &ovs_net->dps, list_node) {
net/openvswitch/datapath.c
2555
struct ovs_net *ovs_net = container_of(work, struct ovs_net,
net/openvswitch/datapath.c
2561
list_for_each_entry(dp, &ovs_net->dps, list_node)
net/openvswitch/datapath.c
2566
schedule_delayed_work(&ovs_net->masks_rebalance,
net/openvswitch/datapath.c
2662
struct ovs_net *ovs_net = net_generic(net, ovs_net_id);
net/openvswitch/datapath.c
2665
INIT_LIST_HEAD(&ovs_net->dps);
net/openvswitch/datapath.c
2666
INIT_WORK(&ovs_net->dp_notify_work, ovs_dp_notify_wq);
net/openvswitch/datapath.c
2667
INIT_DELAYED_WORK(&ovs_net->masks_rebalance, ovs_dp_masks_rebalance);
net/openvswitch/datapath.c
2673
schedule_delayed_work(&ovs_net->masks_rebalance,
net/openvswitch/datapath.c
2681
struct ovs_net *ovs_net = net_generic(net, ovs_net_id);
net/openvswitch/datapath.c
2684
list_for_each_entry(dp, &ovs_net->dps, list_node) {
net/openvswitch/datapath.c
2704
struct ovs_net *ovs_net = net_generic(dnet, ovs_net_id);
net/openvswitch/datapath.c
2713
list_for_each_entry_safe(dp, dp_next, &ovs_net->dps, list_node)
net/openvswitch/datapath.c
2729
cancel_delayed_work_sync(&ovs_net->masks_rebalance);
net/openvswitch/datapath.c
2730
cancel_work_sync(&ovs_net->dp_notify_work);
net/openvswitch/datapath.c
2737
.size = sizeof(struct ovs_net),
net/openvswitch/dp_notify.c
36
struct ovs_net *ovs_net = container_of(work, struct ovs_net, dp_notify_work);
net/openvswitch/dp_notify.c
40
list_for_each_entry(dp, &ovs_net->dps, list_node) {
net/openvswitch/dp_notify.c
62
struct ovs_net *ovs_net;
net/openvswitch/dp_notify.c
77
ovs_net = net_generic(dev_net(dev), ovs_net_id);
net/openvswitch/dp_notify.c
78
queue_work(system_percpu_wq, &ovs_net->dp_notify_work);