Symbol: flow_table
drivers/infiniband/hw/mlx5/fs.c
1142
opfc->rule[i] = mlx5_add_flow_rules(prio->flow_table, &spec[i],
drivers/infiniband/hw/mlx5/fs.c
1302
if (!prio->flow_table) {
drivers/infiniband/hw/mlx5/fs.c
1324
opfc->rule[i] = mlx5_add_flow_rules(prio->flow_table, &spec[i],
drivers/infiniband/hw/mlx5/fs.c
1423
WARN_ON(!prio->flow_table);
drivers/infiniband/hw/mlx5/fs.c
1516
struct mlx5_flow_table *ft = ft_prio->flow_table;
drivers/infiniband/hw/mlx5/fs.c
1627
ft_prio->flow_table = ft;
drivers/infiniband/hw/mlx5/fs.c
2036
if (prio->flow_table)
drivers/infiniband/hw/mlx5/fs.c
2060
struct mlx5_flow_table *ft = ft_prio->flow_table;
drivers/infiniband/hw/mlx5/fs.c
2089
ft_prio->flow_table = ft;
drivers/infiniband/hw/mlx5/fs.c
2748
dest.ft = ft_prio->flow_table;
drivers/infiniband/hw/mlx5/fs.c
638
mlx5_destroy_flow_table(prio->flow_table);
drivers/infiniband/hw/mlx5/fs.c
639
prio->flow_table = NULL;
drivers/infiniband/hw/mlx5/fs.c
704
prio->flow_table = ft;
drivers/infiniband/hw/mlx5/fs.c
791
ft = prio->flow_table;
drivers/infiniband/hw/mlx5/fs.c
999
if (prio->flow_table)
drivers/infiniband/hw/mlx5/mlx5_ib.h
242
struct mlx5_flow_table *flow_table;
drivers/net/ethernet/airoha/airoha_eth.h
593
struct rhashtable flow_table;
drivers/net/ethernet/airoha/airoha_ppe.c
1017
if (rhashtable_lookup(&eth->flow_table, &f->cookie,
drivers/net/ethernet/airoha/airoha_ppe.c
1189
err = rhashtable_insert_fast(&eth->flow_table, &e->node,
drivers/net/ethernet/airoha/airoha_ppe.c
1209
e = rhashtable_lookup(&eth->flow_table, &f->cookie,
drivers/net/ethernet/airoha/airoha_ppe.c
1215
rhashtable_remove_fast(&eth->flow_table, &e->node,
drivers/net/ethernet/airoha/airoha_ppe.c
1263
e = rhashtable_lookup(&eth->flow_table, &f->cookie,
drivers/net/ethernet/airoha/airoha_ppe.c
1533
err = rhashtable_init(&eth->flow_table, &airoha_flow_table_params);
drivers/net/ethernet/airoha/airoha_ppe.c
1550
rhashtable_destroy(&eth->flow_table);
drivers/net/ethernet/airoha/airoha_ppe.c
1571
rhashtable_destroy(&eth->flow_table);
drivers/net/ethernet/broadcom/bnxt/bnxt.h
1822
struct rhashtable flow_table;
drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c
1489
rc = rhashtable_remove_fast(&tc_info->flow_table, &flow_node->node,
drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c
1560
old_node = rhashtable_lookup_fast(&tc_info->flow_table,
drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c
1588
rc = rhashtable_insert_fast(&tc_info->flow_table, &new_node->node,
drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c
1618
flow_node = rhashtable_lookup_fast(&tc_info->flow_table,
drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c
1636
flow_node = rhashtable_lookup_fast(&tc_info->flow_table,
drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c
1828
num_flows = atomic_read(&tc_info->flow_table.nelems);
drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c
1832
rhashtable_walk_enter(&tc_info->flow_table, &tc_info->iter);
drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c
2031
rc = rhashtable_init(&tc_info->flow_table, &tc_info->flow_ht_params);
drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c
2079
rhashtable_destroy(&tc_info->flow_table);
drivers/net/ethernet/broadcom/bnxt/bnxt_tc.c
2095
rhashtable_destroy(&tc_info->flow_table);
drivers/net/ethernet/mediatek/mtk_eth_soc.h
1346
struct rhashtable flow_table;
drivers/net/ethernet/mediatek/mtk_ppe_offload.c
283
if (rhashtable_lookup(&eth->flow_table, &f->cookie, mtk_flow_ht_params))
drivers/net/ethernet/mediatek/mtk_ppe_offload.c
504
err = rhashtable_insert_fast(&eth->flow_table, &entry->node,
drivers/net/ethernet/mediatek/mtk_ppe_offload.c
525
entry = rhashtable_lookup(&eth->flow_table, &f->cookie,
drivers/net/ethernet/mediatek/mtk_ppe_offload.c
531
rhashtable_remove_fast(&eth->flow_table, &entry->node,
drivers/net/ethernet/mediatek/mtk_ppe_offload.c
547
entry = rhashtable_lookup(&eth->flow_table, &f->cookie,
drivers/net/ethernet/mediatek/mtk_ppe_offload.c
676
return rhashtable_init(&eth->flow_table, &mtk_flow_ht_params);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
1600
attr->ct_attr.nf_ft = act->ct.flow_table;
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
1500
struct mlx5_flow_table *flow_table;
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
1534
flow_table = mlx5_create_flow_table(ns, &ft_attr);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
1535
if (IS_ERR(flow_table)) {
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
1536
err = PTR_ERR(flow_table);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
1540
ft_crypto->t = flow_table;
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
1563
flow_table = macsec_fs_auto_group_table_create(ns,
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
1567
if (IS_ERR(flow_table)) {
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
1568
err = PTR_ERR(flow_table);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
1572
rx_tables->ft_check = flow_table;
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
1575
MLX5_SET(create_flow_group_in, flow_group_in, start_flow_index, flow_table->max_fte - 1);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
1576
MLX5_SET(create_flow_group_in, flow_group_in, end_flow_index, flow_table->max_fte - 1);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
395
struct mlx5_flow_table *flow_table;
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
425
flow_table = mlx5_create_flow_table(ns, &ft_attr);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
426
if (IS_ERR(flow_table)) {
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
427
err = PTR_ERR(flow_table);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
431
ft_crypto->t = flow_table;
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
469
flow_table = macsec_fs_auto_group_table_create(ns, 0, TX_CHECK_TABLE_LEVEL,
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
471
if (IS_ERR(flow_table)) {
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
472
err = PTR_ERR(flow_table);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
476
tx_tables->ft_check = flow_table;
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
480
MLX5_SET(create_flow_group_in, flow_group_in, start_flow_index, flow_table->max_fte - 1);
drivers/net/ethernet/mellanox/mlx5/core/lib/macsec_fs.c
481
MLX5_SET(create_flow_group_in, flow_group_in, end_flow_index, flow_table->max_fte - 1);
drivers/net/ethernet/netronome/nfp/flower/conntrack.c
1092
err = rhashtable_insert_fast(&priv->flow_table, &flow_pay->fl_node,
drivers/net/ethernet/netronome/nfp/flower/conntrack.c
1111
WARN_ON_ONCE(rhashtable_remove_fast(&priv->flow_table,
drivers/net/ethernet/netronome/nfp/flower/conntrack.c
1168
WARN_ON_ONCE(rhashtable_remove_fast(&priv->flow_table,
drivers/net/ethernet/netronome/nfp/flower/conntrack.c
1838
zt->nft = ct_act->ct.flow_table;
drivers/net/ethernet/netronome/nfp/flower/main.h
313
struct rhashtable flow_table;
drivers/net/ethernet/netronome/nfp/flower/metadata.c
111
return rhashtable_lookup_fast(&priv->flow_table, &flower_cmp_arg,
drivers/net/ethernet/netronome/nfp/flower/metadata.c
519
err = rhashtable_init(&priv->flow_table, &nfp_flower_table_params);
drivers/net/ethernet/netronome/nfp/flower/metadata.c
602
rhashtable_destroy(&priv->flow_table);
drivers/net/ethernet/netronome/nfp/flower/metadata.c
707
rhashtable_free_and_destroy(&priv->flow_table,
drivers/net/ethernet/netronome/nfp/flower/offload.c
1065
err = rhashtable_insert_fast(&priv->flow_table, &merge_flow->fl_node,
drivers/net/ethernet/netronome/nfp/flower/offload.c
1098
WARN_ON_ONCE(rhashtable_remove_fast(&priv->flow_table,
drivers/net/ethernet/netronome/nfp/flower/offload.c
1392
err = rhashtable_insert_fast(&priv->flow_table, &flow_pay->fl_node,
drivers/net/ethernet/netronome/nfp/flower/offload.c
1437
WARN_ON_ONCE(rhashtable_remove_fast(&priv->flow_table,
drivers/net/ethernet/netronome/nfp/flower/offload.c
1520
WARN_ON_ONCE(rhashtable_remove_fast(&priv->flow_table,
drivers/net/ethernet/netronome/nfp/flower/offload.c
1622
WARN_ON_ONCE(rhashtable_remove_fast(&priv->flow_table,
drivers/net/ethernet/sfc/tc.c
1051
fa->ct.flow_table);
include/net/flow_offload.h
297
struct nf_flowtable *flow_table;
include/net/netfilter/nf_flow_table.h
250
nf_flow_table_offload_add_cb(struct nf_flowtable *flow_table,
include/net/netfilter/nf_flow_table.h
253
struct flow_block *block = &flow_table->flow_block;
include/net/netfilter/nf_flow_table.h
257
down_write(&flow_table->flow_block_lock);
include/net/netfilter/nf_flow_table.h
271
up_write(&flow_table->flow_block_lock);
include/net/netfilter/nf_flow_table.h
273
if (flow_table->type->get)
include/net/netfilter/nf_flow_table.h
274
flow_table->type->get(flow_table);
include/net/netfilter/nf_flow_table.h
278
up_write(&flow_table->flow_block_lock);
include/net/netfilter/nf_flow_table.h
283
nf_flow_table_offload_del_cb(struct nf_flowtable *flow_table,
include/net/netfilter/nf_flow_table.h
286
struct flow_block *block = &flow_table->flow_block;
include/net/netfilter/nf_flow_table.h
289
down_write(&flow_table->flow_block_lock);
include/net/netfilter/nf_flow_table.h
297
up_write(&flow_table->flow_block_lock);
include/net/netfilter/nf_flow_table.h
299
if (flow_table->type->put)
include/net/netfilter/nf_flow_table.h
300
flow_table->type->put(flow_table);
include/net/netfilter/nf_flow_table.h
306
int flow_offload_add(struct nf_flowtable *flow_table, struct flow_offload *flow);
include/net/netfilter/nf_flow_table.h
307
void flow_offload_refresh(struct nf_flowtable *flow_table,
include/net/netfilter/nf_flow_table.h
310
struct flow_offload_tuple_rhash *flow_offload_lookup(struct nf_flowtable *flow_table,
include/net/netfilter/nf_flow_table.h
312
void nf_flow_table_gc_run(struct nf_flowtable *flow_table);
include/net/netfilter/nf_flow_table.h
317
int nf_flow_table_init(struct nf_flowtable *flow_table);
include/net/netfilter/nf_flow_table.h
318
void nf_flow_table_free(struct nf_flowtable *flow_table);
net/core/dev.c
4985
static u32 rfs_slot(u32 hash, const struct rps_dev_flow_table *flow_table)
net/core/dev.c
4987
return hash_32(hash, flow_table->log);
net/core/dev.c
5003
struct rps_dev_flow_table *flow_table,
net/core/dev.c
5016
(int)(10 << flow_table->log);
net/core/dev.c
5028
struct rps_dev_flow_table *flow_table;
net/core/dev.c
5045
flow_table = rcu_dereference(rxqueue->rps_flow_table);
net/core/dev.c
5046
if (!flow_table)
net/core/dev.c
5049
flow_id = rfs_slot(hash, flow_table);
net/core/dev.c
5050
tmp_rflow = &flow_table->flows[flow_id];
net/core/dev.c
5054
if (rps_flow_is_active(tmp_rflow, flow_table,
net/core/dev.c
5094
struct rps_dev_flow_table *flow_table;
net/core/dev.c
5115
flow_table = rcu_dereference(rxqueue->rps_flow_table);
net/core/dev.c
5117
if (!flow_table && !map)
net/core/dev.c
5126
if (flow_table && sock_flow_table) {
net/core/dev.c
5143
rflow = &flow_table->flows[rfs_slot(hash, flow_table)];
net/core/dev.c
5203
struct rps_dev_flow_table *flow_table;
net/core/dev.c
5208
flow_table = rcu_dereference(rxqueue->rps_flow_table);
net/core/dev.c
5209
if (flow_table && flow_id < (1UL << flow_table->log)) {
net/core/dev.c
5212
rflow = &flow_table->flows[flow_id];
net/core/dev.c
5215
rps_flow_is_active(rflow, flow_table, cpu))
net/core/net-sysfs.c
1063
struct rps_dev_flow_table *flow_table;
net/core/net-sysfs.c
1067
flow_table = rcu_dereference(queue->rps_flow_table);
net/core/net-sysfs.c
1068
if (flow_table)
net/core/net-sysfs.c
1069
val = 1UL << flow_table->log;
net/core/net-sysfs.c
1165
struct rps_dev_flow_table *flow_table;
net/core/net-sysfs.c
1173
flow_table = rcu_dereference_protected(queue->rps_flow_table, 1);
net/core/net-sysfs.c
1174
if (flow_table) {
net/core/net-sysfs.c
1176
call_rcu(&flow_table->rcu, rps_dev_flow_table_release);
net/netfilter/nf_flow_table_core.c
323
int flow_offload_add(struct nf_flowtable *flow_table, struct flow_offload *flow)
net/netfilter/nf_flow_table_core.c
329
err = rhashtable_insert_fast(&flow_table->rhashtable,
net/netfilter/nf_flow_table_core.c
335
err = rhashtable_insert_fast(&flow_table->rhashtable,
net/netfilter/nf_flow_table_core.c
339
rhashtable_remove_fast(&flow_table->rhashtable,
net/netfilter/nf_flow_table_core.c
347
if (nf_flowtable_hw_offload(flow_table)) {
net/netfilter/nf_flow_table_core.c
349
nf_flow_offload_add(flow_table, flow);
net/netfilter/nf_flow_table_core.c
356
void flow_offload_refresh(struct nf_flowtable *flow_table,
net/netfilter/nf_flow_table_core.c
367
if (likely(!nf_flowtable_hw_offload(flow_table)) ||
net/netfilter/nf_flow_table_core.c
371
nf_flow_offload_add(flow_table, flow);
net/netfilter/nf_flow_table_core.c
375
static void flow_offload_del(struct nf_flowtable *flow_table,
net/netfilter/nf_flow_table_core.c
378
rhashtable_remove_fast(&flow_table->rhashtable,
net/netfilter/nf_flow_table_core.c
381
rhashtable_remove_fast(&flow_table->rhashtable,
net/netfilter/nf_flow_table_core.c
396
flow_offload_lookup(struct nf_flowtable *flow_table,
net/netfilter/nf_flow_table_core.c
403
tuplehash = rhashtable_lookup(&flow_table->rhashtable, tuple,
net/netfilter/nf_flow_table_core.c
421
nf_flow_table_iterate(struct nf_flowtable *flow_table,
net/netfilter/nf_flow_table_core.c
431
rhashtable_walk_enter(&flow_table->rhashtable, &hti);
net/netfilter/nf_flow_table_core.c
447
iter(flow_table, flow, data);
net/netfilter/nf_flow_table_core.c
455
static bool nf_flow_custom_gc(struct nf_flowtable *flow_table,
net/netfilter/nf_flow_table_core.c
458
return flow_table->type->gc && flow_table->type->gc(flow);
net/netfilter/nf_flow_table_core.c
558
static void nf_flow_offload_gc_step(struct nf_flowtable *flow_table,
net/netfilter/nf_flow_table_core.c
565
nf_flow_custom_gc(flow_table, flow)) {
net/netfilter/nf_flow_table_core.c
575
nf_flow_offload_del(flow_table, flow);
net/netfilter/nf_flow_table_core.c
577
flow_offload_del(flow_table, flow);
net/netfilter/nf_flow_table_core.c
579
flow_offload_del(flow_table, flow);
net/netfilter/nf_flow_table_core.c
584
nf_flow_offload_del(flow_table, flow);
net/netfilter/nf_flow_table_core.c
586
nf_flow_offload_stats(flow_table, flow);
net/netfilter/nf_flow_table_core.c
590
void nf_flow_table_gc_run(struct nf_flowtable *flow_table)
net/netfilter/nf_flow_table_core.c
592
nf_flow_table_iterate(flow_table, nf_flow_offload_gc_step, NULL);
net/netfilter/nf_flow_table_core.c
597
struct nf_flowtable *flow_table;
net/netfilter/nf_flow_table_core.c
599
flow_table = container_of(work, struct nf_flowtable, gc_work.work);
net/netfilter/nf_flow_table_core.c
600
nf_flow_table_gc_run(flow_table);
net/netfilter/nf_flow_table_core.c
601
queue_delayed_work(system_power_efficient_wq, &flow_table->gc_work, HZ);
net/netfilter/nf_flow_table_core.c
716
static void nf_flow_table_do_cleanup(struct nf_flowtable *flow_table,
net/netfilter/nf_flow_table_core.c
751
void nf_flow_table_free(struct nf_flowtable *flow_table)
net/netfilter/nf_flow_table_core.c
754
list_del(&flow_table->list);
net/netfilter/nf_flow_table_core.c
757
cancel_delayed_work_sync(&flow_table->gc_work);
net/netfilter/nf_flow_table_core.c
758
nf_flow_table_offload_flush(flow_table);
net/netfilter/nf_flow_table_core.c
760
nf_flow_table_iterate(flow_table, nf_flow_table_do_cleanup, NULL);
net/netfilter/nf_flow_table_core.c
761
nf_flow_table_gc_run(flow_table);
net/netfilter/nf_flow_table_core.c
762
nf_flow_table_offload_flush_cleanup(flow_table);
net/netfilter/nf_flow_table_core.c
763
rhashtable_destroy(&flow_table->rhashtable);
net/netfilter/nf_flow_table_ip.c
1008
struct nf_flowtable *flow_table,
net/netfilter/nf_flow_table_ip.c
1043
flow_offload_refresh(flow_table, flow, false);
net/netfilter/nf_flow_table_ip.c
1053
if (flow_table->flags & NF_FLOWTABLE_COUNTER)
net/netfilter/nf_flow_table_ip.c
1061
struct nf_flowtable *flow_table,
net/netfilter/nf_flow_table_ip.c
1072
return flow_offload_lookup(flow_table, &tuple);
net/netfilter/nf_flow_table_ip.c
1081
struct nf_flowtable *flow_table = priv;
net/netfilter/nf_flow_table_ip.c
1094
tuplehash = nf_flow_offload_ipv6_lookup(&ctx, flow_table, skb);
net/netfilter/nf_flow_table_ip.c
1098
ret = nf_flow_offload_ipv6_forward(&ctx, flow_table, tuplehash, skb,
net/netfilter/nf_flow_table_ip.c
484
struct nf_flowtable *flow_table, struct sk_buff *skb)
net/netfilter/nf_flow_table_ip.c
494
return flow_offload_lookup(flow_table, &tuple);
net/netfilter/nf_flow_table_ip.c
498
struct nf_flowtable *flow_table,
net/netfilter/nf_flow_table_ip.c
530
flow_offload_refresh(flow_table, flow, false);
net/netfilter/nf_flow_table_ip.c
541
if (flow_table->flags & NF_FLOWTABLE_COUNTER)
net/netfilter/nf_flow_table_ip.c
761
struct nf_flowtable *flow_table = priv;
net/netfilter/nf_flow_table_ip.c
774
tuplehash = nf_flow_offload_lookup(&ctx, flow_table, skb);
net/netfilter/nf_flow_table_ip.c
778
ret = nf_flow_offload_forward(&ctx, flow_table, tuplehash, skb);
net/openvswitch/datapath.c
1920
struct flow_table *table = &dp->table;
net/openvswitch/datapath.h
93
struct flow_table table;
net/openvswitch/flow.h
226
} flow_table, ufid_table;
net/openvswitch/flow_table.c
1004
static int flow_mask_insert(struct flow_table *tbl, struct sw_flow *flow,
net/openvswitch/flow_table.c
103
int ovs_flow_tbl_count(const struct flow_table *table)
net/openvswitch/flow_table.c
1033
static void flow_key_insert(struct flow_table *table, struct sw_flow *flow)
net/openvswitch/flow_table.c
1038
flow->flow_table.hash = flow_hash(&flow->key, &flow->mask->range);
net/openvswitch/flow_table.c
1057
static void flow_ufid_insert(struct flow_table *table, struct sw_flow *flow)
net/openvswitch/flow_table.c
1079
int ovs_flow_tbl_insert(struct flow_table *table, struct sw_flow *flow,
net/openvswitch/flow_table.c
1103
void ovs_flow_masks_rebalance(struct flow_table *table)
net/openvswitch/flow_table.c
243
static int tbl_mask_array_realloc(struct flow_table *tbl, int size)
net/openvswitch/flow_table.c
268
static int tbl_mask_array_add_mask(struct flow_table *tbl,
net/openvswitch/flow_table.c
296
static void tbl_mask_array_del_mask(struct flow_table *tbl,
net/openvswitch/flow_table.c
329
static void flow_mask_remove(struct flow_table *tbl, struct sw_flow_mask *mask)
net/openvswitch/flow_table.c
387
int ovs_flow_tbl_masks_cache_resize(struct flow_table *table, u32 size)
net/openvswitch/flow_table.c
409
int ovs_flow_tbl_init(struct flow_table *table)
net/openvswitch/flow_table.c
457
static void table_instance_flow_free(struct flow_table *table,
net/openvswitch/flow_table.c
462
hlist_del_rcu(&flow->flow_table.node[ti->node_ver]);
net/openvswitch/flow_table.c
474
void table_instance_flow_flush(struct flow_table *table,
net/openvswitch/flow_table.c
486
flow_table.node[ti->node_ver]) {
net/openvswitch/flow_table.c
511
void ovs_flow_tbl_destroy(struct flow_table *table)
net/openvswitch/flow_table.c
535
hlist_for_each_entry_rcu(flow, head, flow_table.node[ver]) {
net/openvswitch/flow_table.c
561
head = find_bucket(ti, flow->flow_table.hash);
net/openvswitch/flow_table.c
562
hlist_add_head_rcu(&flow->flow_table.node[ti->node_ver], head);
net/openvswitch/flow_table.c
595
flow_table.node[old_ver],
net/openvswitch/flow_table.c
615
int ovs_flow_tbl_flush(struct flow_table *flow_table)
net/openvswitch/flow_table.c
627
old_ti = ovsl_dereference(flow_table->ti);
net/openvswitch/flow_table.c
628
old_ufid_ti = ovsl_dereference(flow_table->ufid_ti);
net/openvswitch/flow_table.c
630
rcu_assign_pointer(flow_table->ti, new_ti);
net/openvswitch/flow_table.c
631
rcu_assign_pointer(flow_table->ufid_ti, new_ufid_ti);
net/openvswitch/flow_table.c
632
flow_table->last_rehash = jiffies;
net/openvswitch/flow_table.c
634
table_instance_flow_flush(flow_table, old_ti, old_ufid_ti);
net/openvswitch/flow_table.c
711
hlist_for_each_entry_rcu(flow, head, flow_table.node[ti->node_ver],
net/openvswitch/flow_table.c
713
if (flow->mask == mask && flow->flow_table.hash == hash &&
net/openvswitch/flow_table.c
725
static struct sw_flow *flow_lookup(struct flow_table *tbl,
net/openvswitch/flow_table.c
781
struct sw_flow *ovs_flow_tbl_lookup_stats(struct flow_table *tbl,
net/openvswitch/flow_table.c
845
struct sw_flow *ovs_flow_tbl_lookup(struct flow_table *tbl,
net/openvswitch/flow_table.c
865
struct sw_flow *ovs_flow_tbl_lookup_exact(struct flow_table *tbl,
net/openvswitch/flow_table.c
915
struct sw_flow *ovs_flow_tbl_lookup_ufid(struct flow_table *tbl,
net/openvswitch/flow_table.c
934
int ovs_flow_tbl_num_masks(const struct flow_table *table)
net/openvswitch/flow_table.c
940
u32 ovs_flow_tbl_masks_cache_size(const struct flow_table *table)
net/openvswitch/flow_table.c
954
void ovs_flow_tbl_remove(struct flow_table *table, struct sw_flow *flow)
net/openvswitch/flow_table.c
985
static struct sw_flow_mask *flow_mask_find(const struct flow_table *tbl,
net/openvswitch/flow_table.h
100
struct sw_flow *ovs_flow_tbl_lookup_exact(struct flow_table *tbl,
net/openvswitch/flow_table.h
102
struct sw_flow *ovs_flow_tbl_lookup_ufid(struct flow_table *,
net/openvswitch/flow_table.h
110
void ovs_flow_masks_rebalance(struct flow_table *table);
net/openvswitch/flow_table.h
111
void table_instance_flow_flush(struct flow_table *table,
net/openvswitch/flow_table.h
80
int ovs_flow_tbl_init(struct flow_table *);
net/openvswitch/flow_table.h
81
int ovs_flow_tbl_count(const struct flow_table *table);
net/openvswitch/flow_table.h
82
void ovs_flow_tbl_destroy(struct flow_table *table);
net/openvswitch/flow_table.h
83
int ovs_flow_tbl_flush(struct flow_table *flow_table);
net/openvswitch/flow_table.h
85
int ovs_flow_tbl_insert(struct flow_table *table, struct sw_flow *flow,
net/openvswitch/flow_table.h
87
void ovs_flow_tbl_remove(struct flow_table *table, struct sw_flow *flow);
net/openvswitch/flow_table.h
88
int ovs_flow_tbl_num_masks(const struct flow_table *table);
net/openvswitch/flow_table.h
89
u32 ovs_flow_tbl_masks_cache_size(const struct flow_table *table);
net/openvswitch/flow_table.h
90
int ovs_flow_tbl_masks_cache_resize(struct flow_table *table, u32 size);
net/openvswitch/flow_table.h
93
struct sw_flow *ovs_flow_tbl_lookup_stats(struct flow_table *,
net/openvswitch/flow_table.h
98
struct sw_flow *ovs_flow_tbl_lookup(struct flow_table *,
net/sched/act_ct.c
1621
entry->ct.flow_table = tcf_ct_ft(act);