root/drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
// SPDX-License-Identifier: (GPL-2.0-only OR BSD-2-Clause)
/* Copyright (C) 2019 Netronome Systems, Inc. */

#include <linux/hash.h>
#include <linux/hashtable.h>
#include <linux/jhash.h>
#include <linux/math64.h>
#include <linux/vmalloc.h>
#include <net/pkt_cls.h>
#include <net/pkt_sched.h>

#include "cmsg.h"
#include "main.h"
#include "../nfp_port.h"

#define NFP_FL_QOS_UPDATE               msecs_to_jiffies(1000)
#define NFP_FL_QOS_PPS  BIT(15)
#define NFP_FL_QOS_METER  BIT(10)

struct nfp_police_cfg_head {
        __be32 flags_opts;
        union {
                __be32 meter_id;
                __be32 port;
        };
};

enum NFP_FL_QOS_TYPES {
        NFP_FL_QOS_TYPE_BPS,
        NFP_FL_QOS_TYPE_PPS,
        NFP_FL_QOS_TYPE_MAX,
};

/* Police cmsg for configuring a trTCM traffic conditioner (8W/32B)
 * See RFC 2698 for more details.
 * ----------------------------------------------------------------
 *    3                   2                   1
 *  1 0 9 8 7 6 5 4 3 2 1 0 9 8 7 6 5 4 3 2 1 0 9 8 7 6 5 4 3 2 1 0
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 * |             Reserved          |p|         Reserved            |
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 * |                          Port Ingress                         |
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 * |                        Token Bucket Peak                      |
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 * |                     Token Bucket Committed                    |
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 * |                         Peak Burst Size                       |
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 * |                      Committed Burst Size                     |
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 * |                      Peak Information Rate                    |
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 * |                    Committed Information Rate                 |
 * +-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+-+
 * Word[0](FLag options):
 * [15] p(pps) 1 for pps, 0 for bps
 *
 * Meter control message
 *  1 0 9 8 7 6 5 4 3 2 1 0 9 8 7 6 5 4 3 2 1 0 9 8 7 6 5 4 3 2 1 0
 * +-------------------------------+-+---+-----+-+---------+-+---+-+
 * |            Reserved           |p| Y |TYPE |E|TSHFV    |P| PC|R|
 * +-------------------------------+-+---+-----+-+---------+-+---+-+
 * |                            meter ID                           |
 * +-------------------------------+-------------------------------+
 *
 */
struct nfp_police_config {
        struct nfp_police_cfg_head head;
        __be32 bkt_tkn_p;
        __be32 bkt_tkn_c;
        __be32 pbs;
        __be32 cbs;
        __be32 pir;
        __be32 cir;
};

struct nfp_police_stats_reply {
        struct nfp_police_cfg_head head;
        __be64 pass_bytes;
        __be64 pass_pkts;
        __be64 drop_bytes;
        __be64 drop_pkts;
};

int nfp_flower_offload_one_police(struct nfp_app *app, bool ingress,
                                  bool pps, u32 id, u32 rate, u32 burst)
{
        struct nfp_police_config *config;
        struct sk_buff *skb;

        skb = nfp_flower_cmsg_alloc(app, sizeof(struct nfp_police_config),
                                    NFP_FLOWER_CMSG_TYPE_QOS_MOD, GFP_KERNEL);
        if (!skb)
                return -ENOMEM;

        config = nfp_flower_cmsg_get_data(skb);
        memset(config, 0, sizeof(struct nfp_police_config));
        if (pps)
                config->head.flags_opts |= cpu_to_be32(NFP_FL_QOS_PPS);
        if (!ingress)
                config->head.flags_opts |= cpu_to_be32(NFP_FL_QOS_METER);

        if (ingress)
                config->head.port = cpu_to_be32(id);
        else
                config->head.meter_id = cpu_to_be32(id);

        config->bkt_tkn_p = cpu_to_be32(burst);
        config->bkt_tkn_c = cpu_to_be32(burst);
        config->pbs = cpu_to_be32(burst);
        config->cbs = cpu_to_be32(burst);
        config->pir = cpu_to_be32(rate);
        config->cir = cpu_to_be32(rate);
        nfp_ctrl_tx(app->ctrl, skb);

        return 0;
}

static int nfp_policer_validate(const struct flow_action *action,
                                const struct flow_action_entry *act,
                                struct netlink_ext_ack *extack,
                                bool ingress)
{
        if (act->police.exceed.act_id != FLOW_ACTION_DROP) {
                NL_SET_ERR_MSG_MOD(extack,
                                   "Offload not supported when exceed action is not drop");
                return -EOPNOTSUPP;
        }

        if (ingress) {
                if (act->police.notexceed.act_id != FLOW_ACTION_CONTINUE &&
                    act->police.notexceed.act_id != FLOW_ACTION_ACCEPT) {
                        NL_SET_ERR_MSG_MOD(extack,
                                           "Offload not supported when conform action is not continue or ok");
                        return -EOPNOTSUPP;
                }
        } else {
                if (act->police.notexceed.act_id != FLOW_ACTION_PIPE &&
                    act->police.notexceed.act_id != FLOW_ACTION_ACCEPT) {
                        NL_SET_ERR_MSG_MOD(extack,
                                           "Offload not supported when conform action is not pipe or ok");
                        return -EOPNOTSUPP;
                }
        }

        if (act->police.notexceed.act_id == FLOW_ACTION_ACCEPT &&
            !flow_action_is_last_entry(action, act)) {
                NL_SET_ERR_MSG_MOD(extack,
                                   "Offload not supported when conform action is ok, but action is not last");
                return -EOPNOTSUPP;
        }

        if (act->police.peakrate_bytes_ps ||
            act->police.avrate || act->police.overhead) {
                NL_SET_ERR_MSG_MOD(extack,
                                   "Offload not supported when peakrate/avrate/overhead is configured");
                return -EOPNOTSUPP;
        }

        return 0;
}

static int
nfp_flower_install_rate_limiter(struct nfp_app *app, struct net_device *netdev,
                                struct tc_cls_matchall_offload *flow,
                                struct netlink_ext_ack *extack)
{
        struct flow_action_entry *paction = &flow->rule->action.entries[0];
        u32 action_num = flow->rule->action.num_entries;
        struct nfp_flower_priv *fl_priv = app->priv;
        struct flow_action_entry *action = NULL;
        struct nfp_flower_repr_priv *repr_priv;
        u32 netdev_port_id, i;
        struct nfp_repr *repr;
        bool pps_support;
        u32 bps_num = 0;
        u32 pps_num = 0;
        u32 burst;
        bool pps;
        u64 rate;
        int err;

        if (!nfp_netdev_is_nfp_repr(netdev)) {
                NL_SET_ERR_MSG_MOD(extack, "unsupported offload: qos rate limit offload not supported on higher level port");
                return -EOPNOTSUPP;
        }
        repr = netdev_priv(netdev);
        repr_priv = repr->app_priv;
        netdev_port_id = nfp_repr_get_port_id(netdev);
        pps_support = !!(fl_priv->flower_ext_feats & NFP_FL_FEATS_QOS_PPS);

        if (repr_priv->block_shared) {
                NL_SET_ERR_MSG_MOD(extack, "unsupported offload: qos rate limit offload not supported on shared blocks");
                return -EOPNOTSUPP;
        }

        if (repr->port->type != NFP_PORT_VF_PORT) {
                NL_SET_ERR_MSG_MOD(extack, "unsupported offload: qos rate limit offload not supported on non-VF ports");
                return -EOPNOTSUPP;
        }

        if (pps_support) {
                if (action_num > 2 || action_num == 0) {
                        NL_SET_ERR_MSG_MOD(extack,
                                           "unsupported offload: qos rate limit offload only support action number 1 or 2");
                        return -EOPNOTSUPP;
                }
        } else {
                if (!flow_offload_has_one_action(&flow->rule->action)) {
                        NL_SET_ERR_MSG_MOD(extack,
                                           "unsupported offload: qos rate limit offload requires a single action");
                        return -EOPNOTSUPP;
                }
        }

        if (flow->common.prio != 1) {
                NL_SET_ERR_MSG_MOD(extack, "unsupported offload: qos rate limit offload requires highest priority");
                return -EOPNOTSUPP;
        }

        for (i = 0 ; i < action_num; i++) {
                action = paction + i;
                if (action->id != FLOW_ACTION_POLICE) {
                        NL_SET_ERR_MSG_MOD(extack,
                                           "unsupported offload: qos rate limit offload requires police action");
                        return -EOPNOTSUPP;
                }

                err = nfp_policer_validate(&flow->rule->action, action, extack, true);
                if (err)
                        return err;

                if (action->police.rate_bytes_ps > 0) {
                        if (bps_num++) {
                                NL_SET_ERR_MSG_MOD(extack,
                                                   "unsupported offload: qos rate limit offload only support one BPS action");
                                return -EOPNOTSUPP;
                        }
                }
                if (action->police.rate_pkt_ps > 0) {
                        if (!pps_support) {
                                NL_SET_ERR_MSG_MOD(extack,
                                                   "unsupported offload: FW does not support PPS action");
                                return -EOPNOTSUPP;
                        }
                        if (pps_num++) {
                                NL_SET_ERR_MSG_MOD(extack,
                                                   "unsupported offload: qos rate limit offload only support one PPS action");
                                return -EOPNOTSUPP;
                        }
                }
        }

        for (i = 0 ; i < action_num; i++) {
                /* Set QoS data for this interface */
                action = paction + i;
                if (action->police.rate_bytes_ps > 0) {
                        rate = action->police.rate_bytes_ps;
                        burst = action->police.burst;
                } else if (action->police.rate_pkt_ps > 0) {
                        rate = action->police.rate_pkt_ps;
                        burst = action->police.burst_pkt;
                } else {
                        NL_SET_ERR_MSG_MOD(extack,
                                           "unsupported offload: qos rate limit is not BPS or PPS");
                        continue;
                }

                if (rate != 0) {
                        pps = false;
                        if (action->police.rate_pkt_ps > 0)
                                pps = true;
                        nfp_flower_offload_one_police(repr->app, true,
                                                      pps, netdev_port_id,
                                                      rate, burst);
                }
        }
        repr_priv->qos_table.netdev_port_id = netdev_port_id;
        fl_priv->qos_rate_limiters++;
        if (fl_priv->qos_rate_limiters == 1)
                schedule_delayed_work(&fl_priv->qos_stats_work,
                                      NFP_FL_QOS_UPDATE);

        return 0;
}

static int
nfp_flower_remove_rate_limiter(struct nfp_app *app, struct net_device *netdev,
                               struct tc_cls_matchall_offload *flow,
                               struct netlink_ext_ack *extack)
{
        struct nfp_flower_priv *fl_priv = app->priv;
        struct nfp_flower_repr_priv *repr_priv;
        struct nfp_police_config *config;
        u32 netdev_port_id, i;
        struct nfp_repr *repr;
        struct sk_buff *skb;
        bool pps_support;

        if (!nfp_netdev_is_nfp_repr(netdev)) {
                NL_SET_ERR_MSG_MOD(extack, "unsupported offload: qos rate limit offload not supported on higher level port");
                return -EOPNOTSUPP;
        }
        repr = netdev_priv(netdev);

        netdev_port_id = nfp_repr_get_port_id(netdev);
        repr_priv = repr->app_priv;
        pps_support = !!(fl_priv->flower_ext_feats & NFP_FL_FEATS_QOS_PPS);

        if (!repr_priv->qos_table.netdev_port_id) {
                NL_SET_ERR_MSG_MOD(extack, "unsupported offload: cannot remove qos entry that does not exist");
                return -EOPNOTSUPP;
        }

        memset(&repr_priv->qos_table, 0, sizeof(struct nfp_fl_qos));
        fl_priv->qos_rate_limiters--;
        if (!fl_priv->qos_rate_limiters)
                cancel_delayed_work_sync(&fl_priv->qos_stats_work);
        for (i = 0 ; i < NFP_FL_QOS_TYPE_MAX; i++) {
                if (i == NFP_FL_QOS_TYPE_PPS && !pps_support)
                        break;
                /* 0:bps 1:pps
                 * Clear QoS data for this interface.
                 * There is no need to check if a specific QOS_TYPE was
                 * configured as the firmware handles clearing a QoS entry
                 * safely, even if it wasn't explicitly added.
                 */
                skb = nfp_flower_cmsg_alloc(repr->app, sizeof(struct nfp_police_config),
                                            NFP_FLOWER_CMSG_TYPE_QOS_DEL, GFP_KERNEL);
                if (!skb)
                        return -ENOMEM;

                config = nfp_flower_cmsg_get_data(skb);
                memset(config, 0, sizeof(struct nfp_police_config));
                if (i == NFP_FL_QOS_TYPE_PPS)
                        config->head.flags_opts = cpu_to_be32(NFP_FL_QOS_PPS);
                config->head.port = cpu_to_be32(netdev_port_id);
                nfp_ctrl_tx(repr->app->ctrl, skb);
        }

        return 0;
}

void nfp_flower_stats_rlim_reply(struct nfp_app *app, struct sk_buff *skb)
{
        struct nfp_flower_priv *fl_priv = app->priv;
        struct nfp_flower_repr_priv *repr_priv;
        struct nfp_police_stats_reply *msg;
        struct nfp_stat_pair *curr_stats;
        struct nfp_stat_pair *prev_stats;
        struct net_device *netdev;
        struct nfp_repr *repr;
        u32 netdev_port_id;

        msg = nfp_flower_cmsg_get_data(skb);
        if (be32_to_cpu(msg->head.flags_opts) & NFP_FL_QOS_METER)
                return nfp_act_stats_reply(app, msg);

        netdev_port_id = be32_to_cpu(msg->head.port);
        rcu_read_lock();
        netdev = nfp_app_dev_get(app, netdev_port_id, NULL);
        if (!netdev)
                goto exit_unlock_rcu;

        repr = netdev_priv(netdev);
        repr_priv = repr->app_priv;
        curr_stats = &repr_priv->qos_table.curr_stats;
        prev_stats = &repr_priv->qos_table.prev_stats;

        spin_lock_bh(&fl_priv->qos_stats_lock);
        curr_stats->pkts = be64_to_cpu(msg->pass_pkts) +
                           be64_to_cpu(msg->drop_pkts);
        curr_stats->bytes = be64_to_cpu(msg->pass_bytes) +
                            be64_to_cpu(msg->drop_bytes);

        if (!repr_priv->qos_table.last_update) {
                prev_stats->pkts = curr_stats->pkts;
                prev_stats->bytes = curr_stats->bytes;
        }

        repr_priv->qos_table.last_update = jiffies;
        spin_unlock_bh(&fl_priv->qos_stats_lock);

exit_unlock_rcu:
        rcu_read_unlock();
}

static void
nfp_flower_stats_rlim_request(struct nfp_flower_priv *fl_priv,
                              u32 id, bool ingress)
{
        struct nfp_police_cfg_head *head;
        struct sk_buff *skb;

        skb = nfp_flower_cmsg_alloc(fl_priv->app,
                                    sizeof(struct nfp_police_cfg_head),
                                    NFP_FLOWER_CMSG_TYPE_QOS_STATS,
                                    GFP_ATOMIC);
        if (!skb)
                return;
        head = nfp_flower_cmsg_get_data(skb);

        memset(head, 0, sizeof(struct nfp_police_cfg_head));
        if (ingress) {
                head->port = cpu_to_be32(id);
        } else {
                head->flags_opts = cpu_to_be32(NFP_FL_QOS_METER);
                head->meter_id = cpu_to_be32(id);
        }

        nfp_ctrl_tx(fl_priv->app->ctrl, skb);
}

static void
nfp_flower_stats_rlim_request_all(struct nfp_flower_priv *fl_priv)
{
        struct nfp_reprs *repr_set;
        int i;

        rcu_read_lock();
        repr_set = rcu_dereference(fl_priv->app->reprs[NFP_REPR_TYPE_VF]);
        if (!repr_set)
                goto exit_unlock_rcu;

        for (i = 0; i < repr_set->num_reprs; i++) {
                struct net_device *netdev;

                netdev = rcu_dereference(repr_set->reprs[i]);
                if (netdev) {
                        struct nfp_repr *priv = netdev_priv(netdev);
                        struct nfp_flower_repr_priv *repr_priv;
                        u32 netdev_port_id;

                        repr_priv = priv->app_priv;
                        netdev_port_id = repr_priv->qos_table.netdev_port_id;
                        if (!netdev_port_id)
                                continue;

                        nfp_flower_stats_rlim_request(fl_priv,
                                                      netdev_port_id, true);
                }
        }

exit_unlock_rcu:
        rcu_read_unlock();
}

static void update_stats_cache(struct work_struct *work)
{
        struct delayed_work *delayed_work;
        struct nfp_flower_priv *fl_priv;

        delayed_work = to_delayed_work(work);
        fl_priv = container_of(delayed_work, struct nfp_flower_priv,
                               qos_stats_work);

        nfp_flower_stats_rlim_request_all(fl_priv);
        nfp_flower_stats_meter_request_all(fl_priv);

        schedule_delayed_work(&fl_priv->qos_stats_work, NFP_FL_QOS_UPDATE);
}

static int
nfp_flower_stats_rate_limiter(struct nfp_app *app, struct net_device *netdev,
                              struct tc_cls_matchall_offload *flow,
                              struct netlink_ext_ack *extack)
{
        struct nfp_flower_priv *fl_priv = app->priv;
        struct nfp_flower_repr_priv *repr_priv;
        struct nfp_stat_pair *curr_stats;
        struct nfp_stat_pair *prev_stats;
        u64 diff_bytes, diff_pkts;
        struct nfp_repr *repr;

        if (!nfp_netdev_is_nfp_repr(netdev)) {
                NL_SET_ERR_MSG_MOD(extack, "unsupported offload: qos rate limit offload not supported on higher level port");
                return -EOPNOTSUPP;
        }
        repr = netdev_priv(netdev);

        repr_priv = repr->app_priv;
        if (!repr_priv->qos_table.netdev_port_id) {
                NL_SET_ERR_MSG_MOD(extack, "unsupported offload: cannot find qos entry for stats update");
                return -EOPNOTSUPP;
        }

        spin_lock_bh(&fl_priv->qos_stats_lock);
        curr_stats = &repr_priv->qos_table.curr_stats;
        prev_stats = &repr_priv->qos_table.prev_stats;
        diff_pkts = curr_stats->pkts - prev_stats->pkts;
        diff_bytes = curr_stats->bytes - prev_stats->bytes;
        prev_stats->pkts = curr_stats->pkts;
        prev_stats->bytes = curr_stats->bytes;
        spin_unlock_bh(&fl_priv->qos_stats_lock);

        flow_stats_update(&flow->stats, diff_bytes, diff_pkts, 0,
                          repr_priv->qos_table.last_update,
                          FLOW_ACTION_HW_STATS_DELAYED);
        return 0;
}

void nfp_flower_qos_init(struct nfp_app *app)
{
        struct nfp_flower_priv *fl_priv = app->priv;

        spin_lock_init(&fl_priv->qos_stats_lock);
        mutex_init(&fl_priv->meter_stats_lock);
        nfp_init_meter_table(app);

        INIT_DELAYED_WORK(&fl_priv->qos_stats_work, &update_stats_cache);
}

void nfp_flower_qos_cleanup(struct nfp_app *app)
{
        struct nfp_flower_priv *fl_priv = app->priv;

        cancel_delayed_work_sync(&fl_priv->qos_stats_work);
}

int nfp_flower_setup_qos_offload(struct nfp_app *app, struct net_device *netdev,
                                 struct tc_cls_matchall_offload *flow)
{
        struct netlink_ext_ack *extack = flow->common.extack;
        struct nfp_flower_priv *fl_priv = app->priv;
        int ret;

        if (!(fl_priv->flower_ext_feats & NFP_FL_FEATS_VF_RLIM)) {
                NL_SET_ERR_MSG_MOD(extack, "unsupported offload: loaded firmware does not support qos rate limit offload");
                return -EOPNOTSUPP;
        }

        mutex_lock(&fl_priv->nfp_fl_lock);
        switch (flow->command) {
        case TC_CLSMATCHALL_REPLACE:
                ret = nfp_flower_install_rate_limiter(app, netdev, flow, extack);
                break;
        case TC_CLSMATCHALL_DESTROY:
                ret = nfp_flower_remove_rate_limiter(app, netdev, flow, extack);
                break;
        case TC_CLSMATCHALL_STATS:
                ret = nfp_flower_stats_rate_limiter(app, netdev, flow, extack);
                break;
        default:
                ret = -EOPNOTSUPP;
                break;
        }
        mutex_unlock(&fl_priv->nfp_fl_lock);

        return ret;
}

/* Offload tc action, currently only for tc police */

static const struct rhashtable_params stats_meter_table_params = {
        .key_offset     = offsetof(struct nfp_meter_entry, meter_id),
        .head_offset    = offsetof(struct nfp_meter_entry, ht_node),
        .key_len        = sizeof(u32),
};

struct nfp_meter_entry *
nfp_flower_search_meter_entry(struct nfp_app *app, u32 meter_id)
{
        struct nfp_flower_priv *priv = app->priv;

        return rhashtable_lookup_fast(&priv->meter_table, &meter_id,
                                      stats_meter_table_params);
}

static struct nfp_meter_entry *
nfp_flower_add_meter_entry(struct nfp_app *app, u32 meter_id)
{
        struct nfp_meter_entry *meter_entry = NULL;
        struct nfp_flower_priv *priv = app->priv;

        meter_entry = rhashtable_lookup_fast(&priv->meter_table,
                                             &meter_id,
                                             stats_meter_table_params);
        if (meter_entry)
                return meter_entry;

        meter_entry = kzalloc_obj(*meter_entry);
        if (!meter_entry)
                return NULL;

        meter_entry->meter_id = meter_id;
        meter_entry->used = jiffies;
        if (rhashtable_insert_fast(&priv->meter_table, &meter_entry->ht_node,
                                   stats_meter_table_params)) {
                kfree(meter_entry);
                return NULL;
        }

        priv->qos_rate_limiters++;
        if (priv->qos_rate_limiters == 1)
                schedule_delayed_work(&priv->qos_stats_work,
                                      NFP_FL_QOS_UPDATE);

        return meter_entry;
}

static void nfp_flower_del_meter_entry(struct nfp_app *app, u32 meter_id)
{
        struct nfp_meter_entry *meter_entry = NULL;
        struct nfp_flower_priv *priv = app->priv;

        meter_entry = rhashtable_lookup_fast(&priv->meter_table, &meter_id,
                                             stats_meter_table_params);
        if (!meter_entry)
                return;

        rhashtable_remove_fast(&priv->meter_table,
                               &meter_entry->ht_node,
                               stats_meter_table_params);
        kfree(meter_entry);
        priv->qos_rate_limiters--;
        if (!priv->qos_rate_limiters)
                cancel_delayed_work_sync(&priv->qos_stats_work);
}

int nfp_flower_setup_meter_entry(struct nfp_app *app,
                                 const struct flow_action_entry *action,
                                 enum nfp_meter_op op,
                                 u32 meter_id)
{
        struct nfp_flower_priv *fl_priv = app->priv;
        struct nfp_meter_entry *meter_entry = NULL;
        int err = 0;

        mutex_lock(&fl_priv->meter_stats_lock);

        switch (op) {
        case NFP_METER_DEL:
                nfp_flower_del_meter_entry(app, meter_id);
                goto exit_unlock;
        case NFP_METER_ADD:
                meter_entry = nfp_flower_add_meter_entry(app, meter_id);
                break;
        default:
                err = -EOPNOTSUPP;
                goto exit_unlock;
        }

        if (!meter_entry) {
                err = -ENOMEM;
                goto exit_unlock;
        }

        if (action->police.rate_bytes_ps > 0) {
                meter_entry->bps = true;
                meter_entry->rate = action->police.rate_bytes_ps;
                meter_entry->burst = action->police.burst;
        } else {
                meter_entry->bps = false;
                meter_entry->rate = action->police.rate_pkt_ps;
                meter_entry->burst = action->police.burst_pkt;
        }

exit_unlock:
        mutex_unlock(&fl_priv->meter_stats_lock);
        return err;
}

int nfp_init_meter_table(struct nfp_app *app)
{
        struct nfp_flower_priv *priv = app->priv;

        return rhashtable_init(&priv->meter_table, &stats_meter_table_params);
}

void
nfp_flower_stats_meter_request_all(struct nfp_flower_priv *fl_priv)
{
        struct nfp_meter_entry *meter_entry = NULL;
        struct rhashtable_iter iter;

        mutex_lock(&fl_priv->meter_stats_lock);
        rhashtable_walk_enter(&fl_priv->meter_table, &iter);
        rhashtable_walk_start(&iter);

        while ((meter_entry = rhashtable_walk_next(&iter)) != NULL) {
                if (IS_ERR(meter_entry))
                        continue;
                nfp_flower_stats_rlim_request(fl_priv,
                                              meter_entry->meter_id, false);
        }

        rhashtable_walk_stop(&iter);
        rhashtable_walk_exit(&iter);
        mutex_unlock(&fl_priv->meter_stats_lock);
}

static int
nfp_act_install_actions(struct nfp_app *app, struct flow_offload_action *fl_act,
                        struct netlink_ext_ack *extack)
{
        struct flow_action_entry *paction = &fl_act->action.entries[0];
        u32 action_num = fl_act->action.num_entries;
        struct nfp_flower_priv *fl_priv = app->priv;
        struct flow_action_entry *action = NULL;
        u32 burst, i, meter_id;
        bool pps_support, pps;
        bool add = false;
        u64 rate;
        int err;

        pps_support = !!(fl_priv->flower_ext_feats & NFP_FL_FEATS_QOS_PPS);

        for (i = 0 ; i < action_num; i++) {
                /* Set qos associate data for this interface */
                action = paction + i;
                if (action->id != FLOW_ACTION_POLICE) {
                        NL_SET_ERR_MSG_MOD(extack,
                                           "unsupported offload: qos rate limit offload requires police action");
                        continue;
                }

                err = nfp_policer_validate(&fl_act->action, action, extack, false);
                if (err)
                        return err;

                if (action->police.rate_bytes_ps > 0) {
                        rate = action->police.rate_bytes_ps;
                        burst = action->police.burst;
                } else if (action->police.rate_pkt_ps > 0 && pps_support) {
                        rate = action->police.rate_pkt_ps;
                        burst = action->police.burst_pkt;
                } else {
                        NL_SET_ERR_MSG_MOD(extack,
                                           "unsupported offload: unsupported qos rate limit");
                        continue;
                }

                if (rate != 0) {
                        meter_id = action->hw_index;
                        if (nfp_flower_setup_meter_entry(app, action, NFP_METER_ADD, meter_id))
                                continue;

                        pps = false;
                        if (action->police.rate_pkt_ps > 0)
                                pps = true;
                        nfp_flower_offload_one_police(app, false, pps, meter_id,
                                                      rate, burst);
                        add = true;
                }
        }

        return add ? 0 : -EOPNOTSUPP;
}

static int
nfp_act_remove_actions(struct nfp_app *app, struct flow_offload_action *fl_act,
                       struct netlink_ext_ack *extack)
{
        struct nfp_meter_entry *meter_entry = NULL;
        struct nfp_police_config *config;
        struct sk_buff *skb;
        u32 meter_id;
        bool pps;

        /* Delete qos associate data for this interface */
        if (fl_act->id != FLOW_ACTION_POLICE) {
                NL_SET_ERR_MSG_MOD(extack,
                                   "unsupported offload: qos rate limit offload requires police action");
                return -EOPNOTSUPP;
        }

        meter_id = fl_act->index;
        meter_entry = nfp_flower_search_meter_entry(app, meter_id);
        if (!meter_entry) {
                NL_SET_ERR_MSG_MOD(extack,
                                   "no meter entry when delete the action index.");
                return -ENOENT;
        }
        pps = !meter_entry->bps;

        skb = nfp_flower_cmsg_alloc(app, sizeof(struct nfp_police_config),
                                    NFP_FLOWER_CMSG_TYPE_QOS_DEL, GFP_KERNEL);
        if (!skb)
                return -ENOMEM;

        config = nfp_flower_cmsg_get_data(skb);
        memset(config, 0, sizeof(struct nfp_police_config));
        config->head.flags_opts = cpu_to_be32(NFP_FL_QOS_METER);
        config->head.meter_id = cpu_to_be32(meter_id);
        if (pps)
                config->head.flags_opts |= cpu_to_be32(NFP_FL_QOS_PPS);

        nfp_ctrl_tx(app->ctrl, skb);
        nfp_flower_setup_meter_entry(app, NULL, NFP_METER_DEL, meter_id);

        return 0;
}

void
nfp_act_stats_reply(struct nfp_app *app, void *pmsg)
{
        struct nfp_flower_priv *fl_priv = app->priv;
        struct nfp_meter_entry *meter_entry = NULL;
        struct nfp_police_stats_reply *msg = pmsg;
        u32 meter_id;

        meter_id = be32_to_cpu(msg->head.meter_id);
        mutex_lock(&fl_priv->meter_stats_lock);

        meter_entry = nfp_flower_search_meter_entry(app, meter_id);
        if (!meter_entry)
                goto exit_unlock;

        meter_entry->stats.curr.pkts = be64_to_cpu(msg->pass_pkts) +
                                       be64_to_cpu(msg->drop_pkts);
        meter_entry->stats.curr.bytes = be64_to_cpu(msg->pass_bytes) +
                                        be64_to_cpu(msg->drop_bytes);
        meter_entry->stats.curr.drops = be64_to_cpu(msg->drop_pkts);
        if (!meter_entry->stats.update) {
                meter_entry->stats.prev.pkts = meter_entry->stats.curr.pkts;
                meter_entry->stats.prev.bytes = meter_entry->stats.curr.bytes;
                meter_entry->stats.prev.drops = meter_entry->stats.curr.drops;
        }

        meter_entry->stats.update = jiffies;

exit_unlock:
        mutex_unlock(&fl_priv->meter_stats_lock);
}

static int
nfp_act_stats_actions(struct nfp_app *app, struct flow_offload_action *fl_act,
                      struct netlink_ext_ack *extack)
{
        struct nfp_flower_priv *fl_priv = app->priv;
        struct nfp_meter_entry *meter_entry = NULL;
        u64 diff_bytes, diff_pkts, diff_drops;
        int err = 0;

        if (fl_act->id != FLOW_ACTION_POLICE) {
                NL_SET_ERR_MSG_MOD(extack,
                                   "unsupported offload: qos rate limit offload requires police action");
                return -EOPNOTSUPP;
        }

        mutex_lock(&fl_priv->meter_stats_lock);
        meter_entry = nfp_flower_search_meter_entry(app, fl_act->index);
        if (!meter_entry) {
                err = -ENOENT;
                goto exit_unlock;
        }
        diff_pkts = meter_entry->stats.curr.pkts > meter_entry->stats.prev.pkts ?
                    meter_entry->stats.curr.pkts - meter_entry->stats.prev.pkts : 0;
        diff_bytes = meter_entry->stats.curr.bytes > meter_entry->stats.prev.bytes ?
                     meter_entry->stats.curr.bytes - meter_entry->stats.prev.bytes : 0;
        diff_drops = meter_entry->stats.curr.drops > meter_entry->stats.prev.drops ?
                     meter_entry->stats.curr.drops - meter_entry->stats.prev.drops : 0;

        flow_stats_update(&fl_act->stats, diff_bytes, diff_pkts, diff_drops,
                          meter_entry->stats.update,
                          FLOW_ACTION_HW_STATS_DELAYED);

        meter_entry->stats.prev.pkts = meter_entry->stats.curr.pkts;
        meter_entry->stats.prev.bytes = meter_entry->stats.curr.bytes;
        meter_entry->stats.prev.drops = meter_entry->stats.curr.drops;

exit_unlock:
        mutex_unlock(&fl_priv->meter_stats_lock);
        return err;
}

int nfp_setup_tc_act_offload(struct nfp_app *app,
                             struct flow_offload_action *fl_act)
{
        struct netlink_ext_ack *extack = fl_act->extack;
        struct nfp_flower_priv *fl_priv = app->priv;

        if (!(fl_priv->flower_ext_feats & NFP_FL_FEATS_QOS_METER))
                return -EOPNOTSUPP;

        switch (fl_act->command) {
        case FLOW_ACT_REPLACE:
                return nfp_act_install_actions(app, fl_act, extack);
        case FLOW_ACT_DESTROY:
                return nfp_act_remove_actions(app, fl_act, extack);
        case FLOW_ACT_STATS:
                return nfp_act_stats_actions(app, fl_act, extack);
        default:
                return -EOPNOTSUPP;
        }
}