root/net/core/rtnetlink.c
// SPDX-License-Identifier: GPL-2.0-or-later
/*
 * INET         An implementation of the TCP/IP protocol suite for the LINUX
 *              operating system.  INET is implemented using the  BSD Socket
 *              interface as the means of communication with the user level.
 *
 *              Routing netlink socket interface: protocol independent part.
 *
 * Authors:     Alexey Kuznetsov, <kuznet@ms2.inr.ac.ru>
 *
 *      Fixes:
 *      Vitaly E. Lavrov                RTA_OK arithmetic was wrong.
 */

#include <linux/bitops.h>
#include <linux/errno.h>
#include <linux/module.h>
#include <linux/types.h>
#include <linux/socket.h>
#include <linux/kernel.h>
#include <linux/timer.h>
#include <linux/string.h>
#include <linux/sockios.h>
#include <linux/net.h>
#include <linux/fcntl.h>
#include <linux/mm.h>
#include <linux/slab.h>
#include <linux/interrupt.h>
#include <linux/capability.h>
#include <linux/skbuff.h>
#include <linux/init.h>
#include <linux/security.h>
#include <linux/mutex.h>
#include <linux/if_addr.h>
#include <linux/if_bridge.h>
#include <linux/if_vlan.h>
#include <linux/pci.h>
#include <linux/etherdevice.h>
#include <linux/bpf.h>

#include <linux/uaccess.h>

#include <linux/inet.h>
#include <linux/netdevice.h>
#include <net/ip.h>
#include <net/protocol.h>
#include <net/arp.h>
#include <net/route.h>
#include <net/udp.h>
#include <net/tcp.h>
#include <net/sock.h>
#include <net/pkt_sched.h>
#include <net/fib_rules.h>
#include <net/rtnetlink.h>
#include <net/net_namespace.h>
#include <net/netdev_lock.h>
#include <net/devlink.h>
#if IS_ENABLED(CONFIG_IPV6)
#include <net/addrconf.h>
#endif
#include <linux/dpll.h>

#include "dev.h"

#define RTNL_MAX_TYPE           50
#define RTNL_SLAVE_MAX_TYPE     44

struct rtnl_link {
        rtnl_doit_func          doit;
        rtnl_dumpit_func        dumpit;
        struct module           *owner;
        unsigned int            flags;
        struct rcu_head         rcu;
};

static DEFINE_MUTEX(rtnl_mutex);

void rtnl_lock(void)
{
        mutex_lock(&rtnl_mutex);
}
EXPORT_SYMBOL(rtnl_lock);

int rtnl_lock_interruptible(void)
{
        return mutex_lock_interruptible(&rtnl_mutex);
}

int rtnl_lock_killable(void)
{
        return mutex_lock_killable(&rtnl_mutex);
}

static struct sk_buff *defer_kfree_skb_list;
void rtnl_kfree_skbs(struct sk_buff *head, struct sk_buff *tail)
{
        if (head && tail) {
                tail->next = defer_kfree_skb_list;
                defer_kfree_skb_list = head;
        }
}
EXPORT_SYMBOL(rtnl_kfree_skbs);

void __rtnl_unlock(void)
{
        struct sk_buff *head = defer_kfree_skb_list;

        defer_kfree_skb_list = NULL;

        /* Ensure that we didn't actually add any TODO item when __rtnl_unlock()
         * is used. In some places, e.g. in cfg80211, we have code that will do
         * something like
         *   rtnl_lock()
         *   wiphy_lock()
         *   ...
         *   rtnl_unlock()
         *
         * and because netdev_run_todo() acquires the RTNL for items on the list
         * we could cause a situation such as this:
         * Thread 1                     Thread 2
         *                                rtnl_lock()
         *                                unregister_netdevice()
         *                                __rtnl_unlock()
         * rtnl_lock()
         * wiphy_lock()
         * rtnl_unlock()
         *   netdev_run_todo()
         *     __rtnl_unlock()
         *
         *     // list not empty now
         *     // because of thread 2
         *                                rtnl_lock()
         *     while (!list_empty(...))
         *       rtnl_lock()
         *                                wiphy_lock()
         * **** DEADLOCK ****
         *
         * However, usage of __rtnl_unlock() is rare, and so we can ensure that
         * it's not used in cases where something is added to do the list.
         */
        WARN_ON(!list_empty(&net_todo_list));

        mutex_unlock(&rtnl_mutex);

        while (head) {
                struct sk_buff *next = head->next;

                kfree_skb(head);
                cond_resched();
                head = next;
        }
}

void rtnl_unlock(void)
{
        /* This fellow will unlock it for us. */
        netdev_run_todo();
}
EXPORT_SYMBOL(rtnl_unlock);

int rtnl_trylock(void)
{
        return mutex_trylock(&rtnl_mutex);
}
EXPORT_SYMBOL(rtnl_trylock);

int rtnl_is_locked(void)
{
        return mutex_is_locked(&rtnl_mutex);
}
EXPORT_SYMBOL(rtnl_is_locked);

bool refcount_dec_and_rtnl_lock(refcount_t *r)
{
        return refcount_dec_and_mutex_lock(r, &rtnl_mutex);
}
EXPORT_SYMBOL(refcount_dec_and_rtnl_lock);

#ifdef CONFIG_PROVE_LOCKING
bool lockdep_rtnl_is_held(void)
{
        return lockdep_is_held(&rtnl_mutex);
}
EXPORT_SYMBOL(lockdep_rtnl_is_held);
#endif /* #ifdef CONFIG_PROVE_LOCKING */

#ifdef CONFIG_DEBUG_NET_SMALL_RTNL
void __rtnl_net_lock(struct net *net)
{
        ASSERT_RTNL();

        mutex_lock(&net->rtnl_mutex);
}
EXPORT_SYMBOL(__rtnl_net_lock);

void __rtnl_net_unlock(struct net *net)
{
        ASSERT_RTNL();

        mutex_unlock(&net->rtnl_mutex);
}
EXPORT_SYMBOL(__rtnl_net_unlock);

void rtnl_net_lock(struct net *net)
{
        rtnl_lock();
        __rtnl_net_lock(net);
}
EXPORT_SYMBOL(rtnl_net_lock);

void rtnl_net_unlock(struct net *net)
{
        __rtnl_net_unlock(net);
        rtnl_unlock();
}
EXPORT_SYMBOL(rtnl_net_unlock);

int rtnl_net_trylock(struct net *net)
{
        int ret = rtnl_trylock();

        if (ret)
                __rtnl_net_lock(net);

        return ret;
}
EXPORT_SYMBOL(rtnl_net_trylock);

int rtnl_net_lock_killable(struct net *net)
{
        int ret = rtnl_lock_killable();

        if (!ret)
                __rtnl_net_lock(net);

        return ret;
}

static int rtnl_net_cmp_locks(const struct net *net_a, const struct net *net_b)
{
        if (net_eq(net_a, net_b))
                return 0;

        /* always init_net first */
        if (net_eq(net_a, &init_net))
                return -1;

        if (net_eq(net_b, &init_net))
                return 1;

        /* otherwise lock in ascending order */
        return net_a < net_b ? -1 : 1;
}

int rtnl_net_lock_cmp_fn(const struct lockdep_map *a, const struct lockdep_map *b)
{
        const struct net *net_a, *net_b;

        net_a = container_of(a, struct net, rtnl_mutex.dep_map);
        net_b = container_of(b, struct net, rtnl_mutex.dep_map);

        return rtnl_net_cmp_locks(net_a, net_b);
}

bool rtnl_net_is_locked(struct net *net)
{
        return rtnl_is_locked() && mutex_is_locked(&net->rtnl_mutex);
}
EXPORT_SYMBOL(rtnl_net_is_locked);

bool lockdep_rtnl_net_is_held(struct net *net)
{
        return lockdep_rtnl_is_held() && lockdep_is_held(&net->rtnl_mutex);
}
EXPORT_SYMBOL(lockdep_rtnl_net_is_held);
#else
static int rtnl_net_cmp_locks(const struct net *net_a, const struct net *net_b)
{
        /* No need to swap */
        return -1;
}
#endif

struct rtnl_nets {
        /* ->newlink() needs to freeze 3 netns at most;
         * 2 for the new device, 1 for its peer.
         */
        struct net *net[3];
        unsigned char len;
};

static void rtnl_nets_init(struct rtnl_nets *rtnl_nets)
{
        memset(rtnl_nets, 0, sizeof(*rtnl_nets));
}

static void rtnl_nets_destroy(struct rtnl_nets *rtnl_nets)
{
        int i;

        for (i = 0; i < rtnl_nets->len; i++) {
                put_net(rtnl_nets->net[i]);
                rtnl_nets->net[i] = NULL;
        }

        rtnl_nets->len = 0;
}

/**
 * rtnl_nets_add - Add netns to be locked before ->newlink().
 *
 * @rtnl_nets: rtnl_nets pointer passed to ->get_peer_net().
 * @net: netns pointer with an extra refcnt held.
 *
 * The extra refcnt is released in rtnl_nets_destroy().
 */
static void rtnl_nets_add(struct rtnl_nets *rtnl_nets, struct net *net)
{
        int i;

        DEBUG_NET_WARN_ON_ONCE(rtnl_nets->len == ARRAY_SIZE(rtnl_nets->net));

        for (i = 0; i < rtnl_nets->len; i++) {
                switch (rtnl_net_cmp_locks(rtnl_nets->net[i], net)) {
                case 0:
                        put_net(net);
                        return;
                case 1:
                        swap(rtnl_nets->net[i], net);
                }
        }

        rtnl_nets->net[i] = net;
        rtnl_nets->len++;
}

static void rtnl_nets_lock(struct rtnl_nets *rtnl_nets)
{
        int i;

        rtnl_lock();

        for (i = 0; i < rtnl_nets->len; i++)
                __rtnl_net_lock(rtnl_nets->net[i]);
}

static void rtnl_nets_unlock(struct rtnl_nets *rtnl_nets)
{
        int i;

        for (i = 0; i < rtnl_nets->len; i++)
                __rtnl_net_unlock(rtnl_nets->net[i]);

        rtnl_unlock();
}

static struct rtnl_link __rcu *__rcu *rtnl_msg_handlers[RTNL_FAMILY_MAX + 1];

static inline int rtm_msgindex(int msgtype)
{
        int msgindex = msgtype - RTM_BASE;

        /*
         * msgindex < 0 implies someone tried to register a netlink
         * control code. msgindex >= RTM_NR_MSGTYPES may indicate that
         * the message type has not been added to linux/rtnetlink.h
         */
        BUG_ON(msgindex < 0 || msgindex >= RTM_NR_MSGTYPES);

        return msgindex;
}

static struct rtnl_link *rtnl_get_link(int protocol, int msgtype)
{
        struct rtnl_link __rcu **tab;

        if (protocol >= ARRAY_SIZE(rtnl_msg_handlers))
                protocol = PF_UNSPEC;

        tab = rcu_dereference_rtnl(rtnl_msg_handlers[protocol]);
        if (!tab)
                tab = rcu_dereference_rtnl(rtnl_msg_handlers[PF_UNSPEC]);

        return rcu_dereference_rtnl(tab[msgtype]);
}

static int rtnl_register_internal(struct module *owner,
                                  int protocol, int msgtype,
                                  rtnl_doit_func doit, rtnl_dumpit_func dumpit,
                                  unsigned int flags)
{
        struct rtnl_link *link, *old;
        struct rtnl_link __rcu **tab;
        int msgindex;
        int ret = -ENOBUFS;

        BUG_ON(protocol < 0 || protocol > RTNL_FAMILY_MAX);
        msgindex = rtm_msgindex(msgtype);

        rtnl_lock();
        tab = rtnl_dereference(rtnl_msg_handlers[protocol]);
        if (tab == NULL) {
                tab = kcalloc(RTM_NR_MSGTYPES, sizeof(void *), GFP_KERNEL);
                if (!tab)
                        goto unlock;

                /* ensures we see the 0 stores */
                rcu_assign_pointer(rtnl_msg_handlers[protocol], tab);
        }

        old = rtnl_dereference(tab[msgindex]);
        if (old) {
                link = kmemdup(old, sizeof(*old), GFP_KERNEL);
                if (!link)
                        goto unlock;
        } else {
                link = kzalloc_obj(*link);
                if (!link)
                        goto unlock;
        }

        WARN_ON(link->owner && link->owner != owner);
        link->owner = owner;

        WARN_ON(doit && link->doit && link->doit != doit);
        if (doit)
                link->doit = doit;
        WARN_ON(dumpit && link->dumpit && link->dumpit != dumpit);
        if (dumpit)
                link->dumpit = dumpit;

        WARN_ON(rtnl_msgtype_kind(msgtype) != RTNL_KIND_DEL &&
                (flags & RTNL_FLAG_BULK_DEL_SUPPORTED));
        link->flags |= flags;

        /* publish protocol:msgtype */
        rcu_assign_pointer(tab[msgindex], link);
        ret = 0;
        if (old)
                kfree_rcu(old, rcu);
unlock:
        rtnl_unlock();
        return ret;
}

/**
 * rtnl_unregister - Unregister a rtnetlink message type
 * @protocol: Protocol family or PF_UNSPEC
 * @msgtype: rtnetlink message type
 *
 * Returns 0 on success or a negative error code.
 */
static int rtnl_unregister(int protocol, int msgtype)
{
        struct rtnl_link __rcu **tab;
        struct rtnl_link *link;
        int msgindex;

        BUG_ON(protocol < 0 || protocol > RTNL_FAMILY_MAX);
        msgindex = rtm_msgindex(msgtype);

        rtnl_lock();
        tab = rtnl_dereference(rtnl_msg_handlers[protocol]);
        if (!tab) {
                rtnl_unlock();
                return -ENOENT;
        }

        link = rcu_replace_pointer_rtnl(tab[msgindex], NULL);
        rtnl_unlock();

        kfree_rcu(link, rcu);

        return 0;
}

/**
 * rtnl_unregister_all - Unregister all rtnetlink message type of a protocol
 * @protocol : Protocol family or PF_UNSPEC
 *
 * Identical to calling rtnl_unregister() for all registered message types
 * of a certain protocol family.
 */
void rtnl_unregister_all(int protocol)
{
        struct rtnl_link __rcu **tab;
        struct rtnl_link *link;
        int msgindex;

        BUG_ON(protocol < 0 || protocol > RTNL_FAMILY_MAX);

        rtnl_lock();
        tab = rcu_replace_pointer_rtnl(rtnl_msg_handlers[protocol], NULL);
        if (!tab) {
                rtnl_unlock();
                return;
        }
        for (msgindex = 0; msgindex < RTM_NR_MSGTYPES; msgindex++) {
                link = rcu_replace_pointer_rtnl(tab[msgindex], NULL);
                kfree_rcu(link, rcu);
        }
        rtnl_unlock();

        synchronize_net();

        kfree(tab);
}
EXPORT_SYMBOL_GPL(rtnl_unregister_all);

/**
 * __rtnl_register_many - Register rtnetlink message types
 * @handlers: Array of struct rtnl_msg_handlers
 * @n: The length of @handlers
 *
 * Registers the specified function pointers (at least one of them has
 * to be non-NULL) to be called whenever a request message for the
 * specified protocol family and message type is received.
 *
 * The special protocol family PF_UNSPEC may be used to define fallback
 * function pointers for the case when no entry for the specific protocol
 * family exists.
 *
 * When one element of @handlers fails to register,
 * 1) built-in: panics.
 * 2) modules : the previous successful registrations are unwinded
 *              and an error is returned.
 *
 * Use rtnl_register_many().
 */
int __rtnl_register_many(const struct rtnl_msg_handler *handlers, int n)
{
        const struct rtnl_msg_handler *handler;
        int i, err;

        for (i = 0, handler = handlers; i < n; i++, handler++) {
                err = rtnl_register_internal(handler->owner, handler->protocol,
                                             handler->msgtype, handler->doit,
                                             handler->dumpit, handler->flags);
                if (err) {
                        if (!handler->owner)
                                panic("Unable to register rtnetlink message "
                                      "handlers, %pS\n", handlers);

                        __rtnl_unregister_many(handlers, i);
                        break;
                }
        }

        return err;
}
EXPORT_SYMBOL_GPL(__rtnl_register_many);

void __rtnl_unregister_many(const struct rtnl_msg_handler *handlers, int n)
{
        const struct rtnl_msg_handler *handler;
        int i;

        for (i = n - 1, handler = handlers + n - 1; i >= 0; i--, handler--)
                rtnl_unregister(handler->protocol, handler->msgtype);
}
EXPORT_SYMBOL_GPL(__rtnl_unregister_many);

static DEFINE_MUTEX(link_ops_mutex);
static LIST_HEAD(link_ops);

static struct rtnl_link_ops *rtnl_link_ops_get(const char *kind, int *srcu_index)
{
        struct rtnl_link_ops *ops;

        rcu_read_lock();

        list_for_each_entry_rcu(ops, &link_ops, list) {
                if (!strcmp(ops->kind, kind)) {
                        *srcu_index = srcu_read_lock(&ops->srcu);
                        goto unlock;
                }
        }

        ops = NULL;
unlock:
        rcu_read_unlock();

        return ops;
}

static void rtnl_link_ops_put(struct rtnl_link_ops *ops, int srcu_index)
{
        srcu_read_unlock(&ops->srcu, srcu_index);
}

/**
 * rtnl_link_register - Register rtnl_link_ops with rtnetlink.
 * @ops: struct rtnl_link_ops * to register
 *
 * Returns 0 on success or a negative error code.
 */
int rtnl_link_register(struct rtnl_link_ops *ops)
{
        struct rtnl_link_ops *tmp;
        int err;

        /* Sanity-check max sizes to avoid stack buffer overflow. */
        if (WARN_ON(ops->maxtype > RTNL_MAX_TYPE ||
                    ops->slave_maxtype > RTNL_SLAVE_MAX_TYPE))
                return -EINVAL;

        /* The check for alloc/setup is here because if ops
         * does not have that filled up, it is not possible
         * to use the ops for creating device. So do not
         * fill up dellink as well. That disables rtnl_dellink.
         */
        if ((ops->alloc || ops->setup) && !ops->dellink)
                ops->dellink = unregister_netdevice_queue;

        err = init_srcu_struct(&ops->srcu);
        if (err)
                return err;

        mutex_lock(&link_ops_mutex);

        list_for_each_entry(tmp, &link_ops, list) {
                if (!strcmp(ops->kind, tmp->kind)) {
                        err = -EEXIST;
                        goto unlock;
                }
        }

        list_add_tail_rcu(&ops->list, &link_ops);
unlock:
        mutex_unlock(&link_ops_mutex);

        if (err)
                cleanup_srcu_struct(&ops->srcu);

        return err;
}
EXPORT_SYMBOL_GPL(rtnl_link_register);

static void __rtnl_kill_links(struct net *net, struct rtnl_link_ops *ops)
{
        struct net_device *dev;
        LIST_HEAD(list_kill);

        for_each_netdev(net, dev) {
                if (dev->rtnl_link_ops == ops)
                        ops->dellink(dev, &list_kill);
        }
        unregister_netdevice_many(&list_kill);
}

/* Return with the rtnl_lock held when there are no network
 * devices unregistering in any network namespace.
 */
static void rtnl_lock_unregistering_all(void)
{
        DEFINE_WAIT_FUNC(wait, woken_wake_function);

        add_wait_queue(&netdev_unregistering_wq, &wait);
        for (;;) {
                rtnl_lock();
                /* We held write locked pernet_ops_rwsem, and parallel
                 * setup_net() and cleanup_net() are not possible.
                 */
                if (!atomic_read(&dev_unreg_count))
                        break;
                __rtnl_unlock();

                wait_woken(&wait, TASK_UNINTERRUPTIBLE, MAX_SCHEDULE_TIMEOUT);
        }
        remove_wait_queue(&netdev_unregistering_wq, &wait);
}

/**
 * rtnl_link_unregister - Unregister rtnl_link_ops from rtnetlink.
 * @ops: struct rtnl_link_ops * to unregister
 */
void rtnl_link_unregister(struct rtnl_link_ops *ops)
{
        struct net *net;

        mutex_lock(&link_ops_mutex);
        list_del_rcu(&ops->list);
        mutex_unlock(&link_ops_mutex);

        synchronize_srcu(&ops->srcu);
        cleanup_srcu_struct(&ops->srcu);

        /* Close the race with setup_net() and cleanup_net() */
        down_write(&pernet_ops_rwsem);
        rtnl_lock_unregistering_all();

        for_each_net(net)
                __rtnl_kill_links(net, ops);

        rtnl_unlock();
        up_write(&pernet_ops_rwsem);
}
EXPORT_SYMBOL_GPL(rtnl_link_unregister);

static size_t rtnl_link_get_slave_info_data_size(const struct net_device *dev)
{
        struct net_device *master_dev;
        const struct rtnl_link_ops *ops;
        size_t size = 0;

        rcu_read_lock();

        master_dev = netdev_master_upper_dev_get_rcu((struct net_device *)dev);
        if (!master_dev)
                goto out;

        ops = master_dev->rtnl_link_ops;
        if (!ops)
                goto out;
        size += nla_total_size(strlen(ops->kind) + 1);  /* IFLA_INFO_SLAVE_KIND */
        if (!ops->get_slave_size)
                goto out;
        /* IFLA_INFO_SLAVE_DATA + nested data */
        size += nla_total_size(sizeof(struct nlattr)) +
                ops->get_slave_size(master_dev, dev);

out:
        rcu_read_unlock();
        return size;
}

static size_t rtnl_link_get_size(const struct net_device *dev)
{
        const struct rtnl_link_ops *ops = dev->rtnl_link_ops;
        size_t size;

        if (!ops)
                return 0;

        size = nla_total_size(sizeof(struct nlattr)) + /* IFLA_LINKINFO */
               nla_total_size(strlen(ops->kind) + 1);  /* IFLA_INFO_KIND */

        if (ops->get_size)
                /* IFLA_INFO_DATA + nested data */
                size += nla_total_size(sizeof(struct nlattr)) +
                        ops->get_size(dev);

        if (ops->get_xstats_size)
                /* IFLA_INFO_XSTATS */
                size += nla_total_size(ops->get_xstats_size(dev));

        size += rtnl_link_get_slave_info_data_size(dev);

        return size;
}

static LIST_HEAD(rtnl_af_ops);

static struct rtnl_af_ops *rtnl_af_lookup(const int family, int *srcu_index)
{
        struct rtnl_af_ops *ops;

        ASSERT_RTNL();

        rcu_read_lock();

        list_for_each_entry_rcu(ops, &rtnl_af_ops, list) {
                if (ops->family == family) {
                        *srcu_index = srcu_read_lock(&ops->srcu);
                        goto unlock;
                }
        }

        ops = NULL;
unlock:
        rcu_read_unlock();

        return ops;
}

static void rtnl_af_put(struct rtnl_af_ops *ops, int srcu_index)
{
        srcu_read_unlock(&ops->srcu, srcu_index);
}

/**
 * rtnl_af_register - Register rtnl_af_ops with rtnetlink.
 * @ops: struct rtnl_af_ops * to register
 *
 * Return: 0 on success or a negative error code.
 */
int rtnl_af_register(struct rtnl_af_ops *ops)
{
        int err = init_srcu_struct(&ops->srcu);

        if (err)
                return err;

        rtnl_lock();
        list_add_tail_rcu(&ops->list, &rtnl_af_ops);
        rtnl_unlock();

        return 0;
}
EXPORT_SYMBOL_GPL(rtnl_af_register);

/**
 * rtnl_af_unregister - Unregister rtnl_af_ops from rtnetlink.
 * @ops: struct rtnl_af_ops * to unregister
 */
void rtnl_af_unregister(struct rtnl_af_ops *ops)
{
        rtnl_lock();
        list_del_rcu(&ops->list);
        rtnl_unlock();

        synchronize_rcu();
        synchronize_srcu(&ops->srcu);
        cleanup_srcu_struct(&ops->srcu);
}
EXPORT_SYMBOL_GPL(rtnl_af_unregister);

static size_t rtnl_link_get_af_size(const struct net_device *dev,
                                    u32 ext_filter_mask)
{
        struct rtnl_af_ops *af_ops;
        size_t size;

        /* IFLA_AF_SPEC */
        size = nla_total_size(sizeof(struct nlattr));

        rcu_read_lock();
        list_for_each_entry_rcu(af_ops, &rtnl_af_ops, list) {
                if (af_ops->get_link_af_size) {
                        /* AF_* + nested data */
                        size += nla_total_size(sizeof(struct nlattr)) +
                                af_ops->get_link_af_size(dev, ext_filter_mask);
                }
        }
        rcu_read_unlock();

        return size;
}

static bool rtnl_have_link_slave_info(const struct net_device *dev)
{
        struct net_device *master_dev;
        bool ret = false;

        rcu_read_lock();

        master_dev = netdev_master_upper_dev_get_rcu((struct net_device *)dev);
        if (master_dev && master_dev->rtnl_link_ops)
                ret = true;
        rcu_read_unlock();
        return ret;
}

static int rtnl_link_slave_info_fill(struct sk_buff *skb,
                                     const struct net_device *dev)
{
        struct net_device *master_dev;
        const struct rtnl_link_ops *ops;
        struct nlattr *slave_data;
        int err;

        master_dev = netdev_master_upper_dev_get((struct net_device *) dev);
        if (!master_dev)
                return 0;
        ops = master_dev->rtnl_link_ops;
        if (!ops)
                return 0;
        if (nla_put_string(skb, IFLA_INFO_SLAVE_KIND, ops->kind) < 0)
                return -EMSGSIZE;
        if (ops->fill_slave_info) {
                slave_data = nla_nest_start_noflag(skb, IFLA_INFO_SLAVE_DATA);
                if (!slave_data)
                        return -EMSGSIZE;
                err = ops->fill_slave_info(skb, master_dev, dev);
                if (err < 0)
                        goto err_cancel_slave_data;
                nla_nest_end(skb, slave_data);
        }
        return 0;

err_cancel_slave_data:
        nla_nest_cancel(skb, slave_data);
        return err;
}

static int rtnl_link_info_fill(struct sk_buff *skb,
                               const struct net_device *dev)
{
        const struct rtnl_link_ops *ops = dev->rtnl_link_ops;
        struct nlattr *data;
        int err;

        if (!ops)
                return 0;
        if (nla_put_string(skb, IFLA_INFO_KIND, ops->kind) < 0)
                return -EMSGSIZE;
        if (ops->fill_xstats) {
                err = ops->fill_xstats(skb, dev);
                if (err < 0)
                        return err;
        }
        if (ops->fill_info) {
                data = nla_nest_start_noflag(skb, IFLA_INFO_DATA);
                if (data == NULL)
                        return -EMSGSIZE;
                err = ops->fill_info(skb, dev);
                if (err < 0)
                        goto err_cancel_data;
                nla_nest_end(skb, data);
        }
        return 0;

err_cancel_data:
        nla_nest_cancel(skb, data);
        return err;
}

static int rtnl_link_fill(struct sk_buff *skb, const struct net_device *dev)
{
        struct nlattr *linkinfo;
        int err = -EMSGSIZE;

        linkinfo = nla_nest_start_noflag(skb, IFLA_LINKINFO);
        if (linkinfo == NULL)
                goto out;

        err = rtnl_link_info_fill(skb, dev);
        if (err < 0)
                goto err_cancel_link;

        err = rtnl_link_slave_info_fill(skb, dev);
        if (err < 0)
                goto err_cancel_link;

        nla_nest_end(skb, linkinfo);
        return 0;

err_cancel_link:
        nla_nest_cancel(skb, linkinfo);
out:
        return err;
}

int rtnetlink_send(struct sk_buff *skb, struct net *net, u32 pid, unsigned int group, int echo)
{
        struct sock *rtnl = net->rtnl;

        return nlmsg_notify(rtnl, skb, pid, group, echo, GFP_KERNEL);
}

int rtnl_unicast(struct sk_buff *skb, struct net *net, u32 pid)
{
        struct sock *rtnl = net->rtnl;

        return nlmsg_unicast(rtnl, skb, pid);
}
EXPORT_SYMBOL(rtnl_unicast);

void rtnl_notify(struct sk_buff *skb, struct net *net, u32 pid, u32 group,
                 const struct nlmsghdr *nlh, gfp_t flags)
{
        struct sock *rtnl = net->rtnl;

        nlmsg_notify(rtnl, skb, pid, group, nlmsg_report(nlh), flags);
}
EXPORT_SYMBOL(rtnl_notify);

void rtnl_set_sk_err(struct net *net, u32 group, int error)
{
        struct sock *rtnl = net->rtnl;

        netlink_set_err(rtnl, 0, group, error);
}
EXPORT_SYMBOL(rtnl_set_sk_err);

int rtnetlink_put_metrics(struct sk_buff *skb, u32 *metrics)
{
        struct nlattr *mx;
        int i, valid = 0;

        /* nothing is dumped for dst_default_metrics, so just skip the loop */
        if (metrics == dst_default_metrics.metrics)
                return 0;

        mx = nla_nest_start_noflag(skb, RTA_METRICS);
        if (mx == NULL)
                return -ENOBUFS;

        for (i = 0; i < RTAX_MAX; i++) {
                if (metrics[i]) {
                        if (i == RTAX_CC_ALGO - 1) {
                                char tmp[TCP_CA_NAME_MAX], *name;

                                name = tcp_ca_get_name_by_key(metrics[i], tmp);
                                if (!name)
                                        continue;
                                if (nla_put_string(skb, i + 1, name))
                                        goto nla_put_failure;
                        } else if (i == RTAX_FEATURES - 1) {
                                u32 user_features = metrics[i] & RTAX_FEATURE_MASK;

                                if (!user_features)
                                        continue;
                                BUILD_BUG_ON(RTAX_FEATURE_MASK & DST_FEATURE_MASK);
                                if (nla_put_u32(skb, i + 1, user_features))
                                        goto nla_put_failure;
                        } else {
                                if (nla_put_u32(skb, i + 1, metrics[i]))
                                        goto nla_put_failure;
                        }
                        valid++;
                }
        }

        if (!valid) {
                nla_nest_cancel(skb, mx);
                return 0;
        }

        return nla_nest_end(skb, mx);

nla_put_failure:
        nla_nest_cancel(skb, mx);
        return -EMSGSIZE;
}
EXPORT_SYMBOL(rtnetlink_put_metrics);

int rtnl_put_cacheinfo(struct sk_buff *skb, struct dst_entry *dst, u32 id,
                       long expires, u32 error)
{
        struct rta_cacheinfo ci = {
                .rta_error = error,
                .rta_id =  id,
        };
        unsigned long delta;

        if (dst) {
                delta = jiffies - READ_ONCE(dst->lastuse);
                ci.rta_lastuse = jiffies_delta_to_clock_t(delta);
                ci.rta_used = dst->__use;
                ci.rta_clntref = rcuref_read(&dst->__rcuref);
        }
        if (expires) {
                unsigned long clock;

                clock = jiffies_to_clock_t(abs(expires));
                clock = min_t(unsigned long, clock, INT_MAX);
                ci.rta_expires = (expires > 0) ? clock : -clock;
        }
        return nla_put(skb, RTA_CACHEINFO, sizeof(ci), &ci);
}
EXPORT_SYMBOL_GPL(rtnl_put_cacheinfo);

void netif_set_operstate(struct net_device *dev, int newstate)
{
        unsigned int old = READ_ONCE(dev->operstate);

        do {
                if (old == newstate)
                        return;
        } while (!try_cmpxchg(&dev->operstate, &old, newstate));

        netif_state_change(dev);
}
EXPORT_SYMBOL(netif_set_operstate);

static void set_operstate(struct net_device *dev, unsigned char transition)
{
        unsigned char operstate = READ_ONCE(dev->operstate);

        switch (transition) {
        case IF_OPER_UP:
                if ((operstate == IF_OPER_DORMANT ||
                     operstate == IF_OPER_TESTING ||
                     operstate == IF_OPER_UNKNOWN) &&
                    !netif_dormant(dev) && !netif_testing(dev))
                        operstate = IF_OPER_UP;
                break;

        case IF_OPER_TESTING:
                if (netif_oper_up(dev))
                        operstate = IF_OPER_TESTING;
                break;

        case IF_OPER_DORMANT:
                if (netif_oper_up(dev))
                        operstate = IF_OPER_DORMANT;
                break;
        }

        netif_set_operstate(dev, operstate);
}

static unsigned int rtnl_dev_get_flags(const struct net_device *dev)
{
        return (dev->flags & ~(IFF_PROMISC | IFF_ALLMULTI)) |
               (dev->gflags & (IFF_PROMISC | IFF_ALLMULTI));
}

static unsigned int rtnl_dev_combine_flags(const struct net_device *dev,
                                           const struct ifinfomsg *ifm)
{
        unsigned int flags = ifm->ifi_flags;

        /* bugwards compatibility: ifi_change == 0 is treated as ~0 */
        if (ifm->ifi_change)
                flags = (flags & ifm->ifi_change) |
                        (rtnl_dev_get_flags(dev) & ~ifm->ifi_change);

        return flags;
}

static void copy_rtnl_link_stats(struct rtnl_link_stats *a,
                                 const struct rtnl_link_stats64 *b)
{
        a->rx_packets = b->rx_packets;
        a->tx_packets = b->tx_packets;
        a->rx_bytes = b->rx_bytes;
        a->tx_bytes = b->tx_bytes;
        a->rx_errors = b->rx_errors;
        a->tx_errors = b->tx_errors;
        a->rx_dropped = b->rx_dropped;
        a->tx_dropped = b->tx_dropped;

        a->multicast = b->multicast;
        a->collisions = b->collisions;

        a->rx_length_errors = b->rx_length_errors;
        a->rx_over_errors = b->rx_over_errors;
        a->rx_crc_errors = b->rx_crc_errors;
        a->rx_frame_errors = b->rx_frame_errors;
        a->rx_fifo_errors = b->rx_fifo_errors;
        a->rx_missed_errors = b->rx_missed_errors;

        a->tx_aborted_errors = b->tx_aborted_errors;
        a->tx_carrier_errors = b->tx_carrier_errors;
        a->tx_fifo_errors = b->tx_fifo_errors;
        a->tx_heartbeat_errors = b->tx_heartbeat_errors;
        a->tx_window_errors = b->tx_window_errors;

        a->rx_compressed = b->rx_compressed;
        a->tx_compressed = b->tx_compressed;

        a->rx_nohandler = b->rx_nohandler;
}

/* All VF info */
static inline int rtnl_vfinfo_size(const struct net_device *dev,
                                   u32 ext_filter_mask)
{
        if (dev->dev.parent && (ext_filter_mask & RTEXT_FILTER_VF)) {
                int num_vfs = dev_num_vf(dev->dev.parent);
                size_t size = nla_total_size(0);
                size += num_vfs *
                        (nla_total_size(0) +
                         nla_total_size(sizeof(struct ifla_vf_mac)) +
                         nla_total_size(sizeof(struct ifla_vf_broadcast)) +
                         nla_total_size(sizeof(struct ifla_vf_vlan)) +
                         nla_total_size(0) + /* nest IFLA_VF_VLAN_LIST */
                         nla_total_size(MAX_VLAN_LIST_LEN *
                                        sizeof(struct ifla_vf_vlan_info)) +
                         nla_total_size(sizeof(struct ifla_vf_spoofchk)) +
                         nla_total_size(sizeof(struct ifla_vf_tx_rate)) +
                         nla_total_size(sizeof(struct ifla_vf_rate)) +
                         nla_total_size(sizeof(struct ifla_vf_link_state)) +
                         nla_total_size(sizeof(struct ifla_vf_rss_query_en)) +
                         nla_total_size(sizeof(struct ifla_vf_trust)));
                if (~ext_filter_mask & RTEXT_FILTER_SKIP_STATS) {
                        size += num_vfs *
                                (nla_total_size(0) + /* nest IFLA_VF_STATS */
                                 /* IFLA_VF_STATS_RX_PACKETS */
                                 nla_total_size_64bit(sizeof(__u64)) +
                                 /* IFLA_VF_STATS_TX_PACKETS */
                                 nla_total_size_64bit(sizeof(__u64)) +
                                 /* IFLA_VF_STATS_RX_BYTES */
                                 nla_total_size_64bit(sizeof(__u64)) +
                                 /* IFLA_VF_STATS_TX_BYTES */
                                 nla_total_size_64bit(sizeof(__u64)) +
                                 /* IFLA_VF_STATS_BROADCAST */
                                 nla_total_size_64bit(sizeof(__u64)) +
                                 /* IFLA_VF_STATS_MULTICAST */
                                 nla_total_size_64bit(sizeof(__u64)) +
                                 /* IFLA_VF_STATS_RX_DROPPED */
                                 nla_total_size_64bit(sizeof(__u64)) +
                                 /* IFLA_VF_STATS_TX_DROPPED */
                                 nla_total_size_64bit(sizeof(__u64)));
                }
                if (dev->netdev_ops->ndo_get_vf_guid)
                        size += num_vfs * 2 *
                                nla_total_size(sizeof(struct ifla_vf_guid));
                return size;
        } else
                return 0;
}

static size_t rtnl_port_size(const struct net_device *dev,
                             u32 ext_filter_mask)
{
        size_t port_size = nla_total_size(4)            /* PORT_VF */
                + nla_total_size(PORT_PROFILE_MAX)      /* PORT_PROFILE */
                + nla_total_size(PORT_UUID_MAX)         /* PORT_INSTANCE_UUID */
                + nla_total_size(PORT_UUID_MAX)         /* PORT_HOST_UUID */
                + nla_total_size(1)                     /* PROT_VDP_REQUEST */
                + nla_total_size(2);                    /* PORT_VDP_RESPONSE */
        size_t vf_ports_size = nla_total_size(sizeof(struct nlattr));
        size_t vf_port_size = nla_total_size(sizeof(struct nlattr))
                + port_size;
        size_t port_self_size = nla_total_size(sizeof(struct nlattr))
                + port_size;

        if (!dev->netdev_ops->ndo_get_vf_port || !dev->dev.parent ||
            !(ext_filter_mask & RTEXT_FILTER_VF))
                return 0;
        if (dev_num_vf(dev->dev.parent))
                return port_self_size + vf_ports_size +
                        vf_port_size * dev_num_vf(dev->dev.parent);
        else
                return port_self_size;
}

static size_t rtnl_xdp_size(void)
{
        size_t xdp_size = nla_total_size(0) +   /* nest IFLA_XDP */
                          nla_total_size(1) +   /* XDP_ATTACHED */
                          nla_total_size(4) +   /* XDP_PROG_ID (or 1st mode) */
                          nla_total_size(4);    /* XDP_<mode>_PROG_ID */

        return xdp_size;
}

static size_t rtnl_prop_list_size(const struct net_device *dev)
{
        struct netdev_name_node *name_node;
        unsigned int cnt = 0;

        rcu_read_lock();
        list_for_each_entry_rcu(name_node, &dev->name_node->list, list)
                cnt++;
        rcu_read_unlock();

        if (!cnt)
                return 0;

        return nla_total_size(0) + cnt * nla_total_size(ALTIFNAMSIZ);
}

static size_t rtnl_proto_down_size(const struct net_device *dev)
{
        size_t size = nla_total_size(1);

        /* Assume dev->proto_down_reason is not zero. */
        size += nla_total_size(0) + nla_total_size(4);

        return size;
}

static size_t rtnl_devlink_port_size(const struct net_device *dev)
{
        size_t size = nla_total_size(0); /* nest IFLA_DEVLINK_PORT */

        if (dev->devlink_port)
                size += devlink_nl_port_handle_size(dev->devlink_port);

        return size;
}

static size_t rtnl_dpll_pin_size(const struct net_device *dev)
{
        size_t size = nla_total_size(0); /* nest IFLA_DPLL_PIN */

        size += dpll_netdev_pin_handle_size(dev);

        return size;
}

static size_t rtnl_dev_parent_size(const struct net_device *dev)
{
        size_t size = 0;

        /* IFLA_PARENT_DEV_NAME */
        if (dev->dev.parent)
                size += nla_total_size(strlen(dev_name(dev->dev.parent)) + 1);

        /* IFLA_PARENT_DEV_BUS_NAME */
        if (dev->dev.parent && dev->dev.parent->bus)
                size += nla_total_size(strlen(dev->dev.parent->bus->name) + 1);

        return size;
}

static noinline size_t if_nlmsg_size(const struct net_device *dev,
                                     u32 ext_filter_mask)
{
        size_t size;

        size = NLMSG_ALIGN(sizeof(struct ifinfomsg))
               + nla_total_size(IFNAMSIZ) /* IFLA_IFNAME */
               + nla_total_size(IFALIASZ) /* IFLA_IFALIAS */
               + nla_total_size(IFNAMSIZ) /* IFLA_QDISC */
               + nla_total_size_64bit(sizeof(struct rtnl_link_ifmap))
               + nla_total_size(MAX_ADDR_LEN) /* IFLA_ADDRESS */
               + nla_total_size(MAX_ADDR_LEN) /* IFLA_BROADCAST */
               + nla_total_size(4) /* IFLA_TXQLEN */
               + nla_total_size(4) /* IFLA_WEIGHT */
               + nla_total_size(4) /* IFLA_MTU */
               + nla_total_size(4) /* IFLA_LINK */
               + nla_total_size(4) /* IFLA_MASTER */
               + nla_total_size(1) /* IFLA_CARRIER */
               + nla_total_size(4) /* IFLA_PROMISCUITY */
               + nla_total_size(4) /* IFLA_ALLMULTI */
               + nla_total_size(4) /* IFLA_NUM_TX_QUEUES */
               + nla_total_size(4) /* IFLA_NUM_RX_QUEUES */
               + nla_total_size(4) /* IFLA_GSO_MAX_SEGS */
               + nla_total_size(4) /* IFLA_GSO_MAX_SIZE */
               + nla_total_size(4) /* IFLA_GRO_MAX_SIZE */
               + nla_total_size(4) /* IFLA_GSO_IPV4_MAX_SIZE */
               + nla_total_size(4) /* IFLA_GRO_IPV4_MAX_SIZE */
               + nla_total_size(4) /* IFLA_TSO_MAX_SIZE */
               + nla_total_size(4) /* IFLA_TSO_MAX_SEGS */
               + nla_total_size(1) /* IFLA_OPERSTATE */
               + nla_total_size(1) /* IFLA_LINKMODE */
               + nla_total_size(1) /* IFLA_NETNS_IMMUTABLE */
               + nla_total_size(4) /* IFLA_CARRIER_CHANGES */
               + nla_total_size(4) /* IFLA_LINK_NETNSID */
               + nla_total_size(4) /* IFLA_GROUP */
               + nla_total_size(ext_filter_mask
                                & RTEXT_FILTER_VF ? 4 : 0) /* IFLA_NUM_VF */
               + rtnl_vfinfo_size(dev, ext_filter_mask) /* IFLA_VFINFO_LIST */
               + rtnl_port_size(dev, ext_filter_mask) /* IFLA_VF_PORTS + IFLA_PORT_SELF */
               + rtnl_link_get_size(dev) /* IFLA_LINKINFO */
               + rtnl_link_get_af_size(dev, ext_filter_mask) /* IFLA_AF_SPEC */
               + nla_total_size(MAX_PHYS_ITEM_ID_LEN) /* IFLA_PHYS_PORT_ID */
               + nla_total_size(MAX_PHYS_ITEM_ID_LEN) /* IFLA_PHYS_SWITCH_ID */
               + nla_total_size(IFNAMSIZ) /* IFLA_PHYS_PORT_NAME */
               + rtnl_xdp_size() /* IFLA_XDP */
               + nla_total_size(4)  /* IFLA_EVENT */
               + nla_total_size(4)  /* IFLA_NEW_NETNSID */
               + nla_total_size(4)  /* IFLA_NEW_IFINDEX */
               + rtnl_proto_down_size(dev)  /* proto down */
               + nla_total_size(4)  /* IFLA_TARGET_NETNSID */
               + nla_total_size(4)  /* IFLA_CARRIER_UP_COUNT */
               + nla_total_size(4)  /* IFLA_CARRIER_DOWN_COUNT */
               + nla_total_size(4)  /* IFLA_MIN_MTU */
               + nla_total_size(4)  /* IFLA_MAX_MTU */
               + rtnl_prop_list_size(dev)
               + nla_total_size(MAX_ADDR_LEN) /* IFLA_PERM_ADDRESS */
               + rtnl_devlink_port_size(dev)
               + rtnl_dpll_pin_size(dev)
               + nla_total_size(8)  /* IFLA_MAX_PACING_OFFLOAD_HORIZON */
               + nla_total_size(2)  /* IFLA_HEADROOM */
               + nla_total_size(2)  /* IFLA_TAILROOM */
               + rtnl_dev_parent_size(dev)
               + 0;

        if (!(ext_filter_mask & RTEXT_FILTER_SKIP_STATS))
                size += nla_total_size(sizeof(struct rtnl_link_stats)) +
                        nla_total_size_64bit(sizeof(struct rtnl_link_stats64));

        return size;
}

static int rtnl_vf_ports_fill(struct sk_buff *skb, struct net_device *dev)
{
        struct nlattr *vf_ports;
        struct nlattr *vf_port;
        int vf;
        int err;

        vf_ports = nla_nest_start_noflag(skb, IFLA_VF_PORTS);
        if (!vf_ports)
                return -EMSGSIZE;

        for (vf = 0; vf < dev_num_vf(dev->dev.parent); vf++) {
                vf_port = nla_nest_start_noflag(skb, IFLA_VF_PORT);
                if (!vf_port)
                        goto nla_put_failure;
                if (nla_put_u32(skb, IFLA_PORT_VF, vf))
                        goto nla_put_failure;
                err = dev->netdev_ops->ndo_get_vf_port(dev, vf, skb);
                if (err == -EMSGSIZE)
                        goto nla_put_failure;
                if (err) {
                        nla_nest_cancel(skb, vf_port);
                        continue;
                }
                nla_nest_end(skb, vf_port);
        }

        nla_nest_end(skb, vf_ports);

        return 0;

nla_put_failure:
        nla_nest_cancel(skb, vf_ports);
        return -EMSGSIZE;
}

static int rtnl_port_self_fill(struct sk_buff *skb, struct net_device *dev)
{
        struct nlattr *port_self;
        int err;

        port_self = nla_nest_start_noflag(skb, IFLA_PORT_SELF);
        if (!port_self)
                return -EMSGSIZE;

        err = dev->netdev_ops->ndo_get_vf_port(dev, PORT_SELF_VF, skb);
        if (err) {
                nla_nest_cancel(skb, port_self);
                return (err == -EMSGSIZE) ? err : 0;
        }

        nla_nest_end(skb, port_self);

        return 0;
}

static int rtnl_port_fill(struct sk_buff *skb, struct net_device *dev,
                          u32 ext_filter_mask)
{
        int err;

        if (!dev->netdev_ops->ndo_get_vf_port || !dev->dev.parent ||
            !(ext_filter_mask & RTEXT_FILTER_VF))
                return 0;

        err = rtnl_port_self_fill(skb, dev);
        if (err)
                return err;

        if (dev_num_vf(dev->dev.parent)) {
                err = rtnl_vf_ports_fill(skb, dev);
                if (err)
                        return err;
        }

        return 0;
}

static int rtnl_phys_port_id_fill(struct sk_buff *skb, struct net_device *dev)
{
        int err;
        struct netdev_phys_item_id ppid;

        err = dev_get_phys_port_id(dev, &ppid);
        if (err) {
                if (err == -EOPNOTSUPP)
                        return 0;
                return err;
        }

        if (nla_put(skb, IFLA_PHYS_PORT_ID, ppid.id_len, ppid.id))
                return -EMSGSIZE;

        return 0;
}

static int rtnl_phys_port_name_fill(struct sk_buff *skb, struct net_device *dev)
{
        char name[IFNAMSIZ];
        int err;

        err = dev_get_phys_port_name(dev, name, sizeof(name));
        if (err) {
                if (err == -EOPNOTSUPP)
                        return 0;
                return err;
        }

        if (nla_put_string(skb, IFLA_PHYS_PORT_NAME, name))
                return -EMSGSIZE;

        return 0;
}

static int rtnl_phys_switch_id_fill(struct sk_buff *skb, struct net_device *dev)
{
        struct netdev_phys_item_id ppid = { };
        int err;

        err = netif_get_port_parent_id(dev, &ppid, false);
        if (err) {
                if (err == -EOPNOTSUPP)
                        return 0;
                return err;
        }

        if (nla_put(skb, IFLA_PHYS_SWITCH_ID, ppid.id_len, ppid.id))
                return -EMSGSIZE;

        return 0;
}

static noinline_for_stack int rtnl_fill_stats(struct sk_buff *skb,
                                              struct net_device *dev)
{
        struct rtnl_link_stats64 *sp;
        struct nlattr *attr;

        attr = nla_reserve_64bit(skb, IFLA_STATS64,
                                 sizeof(struct rtnl_link_stats64), IFLA_PAD);
        if (!attr)
                return -EMSGSIZE;

        sp = nla_data(attr);
        dev_get_stats(dev, sp);

        attr = nla_reserve(skb, IFLA_STATS,
                           sizeof(struct rtnl_link_stats));
        if (!attr)
                return -EMSGSIZE;

        copy_rtnl_link_stats(nla_data(attr), sp);

        return 0;
}

static noinline_for_stack int rtnl_fill_vfinfo(struct sk_buff *skb,
                                               struct net_device *dev,
                                               int vfs_num,
                                               u32 ext_filter_mask)
{
        struct ifla_vf_rss_query_en vf_rss_query_en;
        struct nlattr *vf, *vfstats, *vfvlanlist;
        struct ifla_vf_link_state vf_linkstate;
        struct ifla_vf_vlan_info vf_vlan_info;
        struct ifla_vf_spoofchk vf_spoofchk;
        struct ifla_vf_tx_rate vf_tx_rate;
        struct ifla_vf_stats vf_stats;
        struct ifla_vf_trust vf_trust;
        struct ifla_vf_vlan vf_vlan;
        struct ifla_vf_rate vf_rate;
        struct ifla_vf_mac vf_mac;
        struct ifla_vf_broadcast vf_broadcast;
        struct ifla_vf_info ivi;
        struct ifla_vf_guid node_guid;
        struct ifla_vf_guid port_guid;

        memset(&ivi, 0, sizeof(ivi));

        /* Not all SR-IOV capable drivers support the
         * spoofcheck and "RSS query enable" query.  Preset to
         * -1 so the user space tool can detect that the driver
         * didn't report anything.
         */
        ivi.spoofchk = -1;
        ivi.rss_query_en = -1;
        ivi.trusted = -1;
        /* The default value for VF link state is "auto"
         * IFLA_VF_LINK_STATE_AUTO which equals zero
         */
        ivi.linkstate = 0;
        /* VLAN Protocol by default is 802.1Q */
        ivi.vlan_proto = htons(ETH_P_8021Q);
        if (dev->netdev_ops->ndo_get_vf_config(dev, vfs_num, &ivi))
                return 0;

        memset(&vf_vlan_info, 0, sizeof(vf_vlan_info));
        memset(&node_guid, 0, sizeof(node_guid));
        memset(&port_guid, 0, sizeof(port_guid));

        vf_mac.vf =
                vf_vlan.vf =
                vf_vlan_info.vf =
                vf_rate.vf =
                vf_tx_rate.vf =
                vf_spoofchk.vf =
                vf_linkstate.vf =
                vf_rss_query_en.vf =
                vf_trust.vf =
                node_guid.vf =
                port_guid.vf = ivi.vf;

        memcpy(vf_mac.mac, ivi.mac, sizeof(ivi.mac));
        memcpy(vf_broadcast.broadcast, dev->broadcast, dev->addr_len);
        vf_vlan.vlan = ivi.vlan;
        vf_vlan.qos = ivi.qos;
        vf_vlan_info.vlan = ivi.vlan;
        vf_vlan_info.qos = ivi.qos;
        vf_vlan_info.vlan_proto = ivi.vlan_proto;
        vf_tx_rate.rate = ivi.max_tx_rate;
        vf_rate.min_tx_rate = ivi.min_tx_rate;
        vf_rate.max_tx_rate = ivi.max_tx_rate;
        vf_spoofchk.setting = ivi.spoofchk;
        vf_linkstate.link_state = ivi.linkstate;
        vf_rss_query_en.setting = ivi.rss_query_en;
        vf_trust.setting = ivi.trusted;
        vf = nla_nest_start_noflag(skb, IFLA_VF_INFO);
        if (!vf)
                return -EMSGSIZE;
        if (nla_put(skb, IFLA_VF_MAC, sizeof(vf_mac), &vf_mac) ||
            nla_put(skb, IFLA_VF_BROADCAST, sizeof(vf_broadcast), &vf_broadcast) ||
            nla_put(skb, IFLA_VF_VLAN, sizeof(vf_vlan), &vf_vlan) ||
            nla_put(skb, IFLA_VF_RATE, sizeof(vf_rate),
                    &vf_rate) ||
            nla_put(skb, IFLA_VF_TX_RATE, sizeof(vf_tx_rate),
                    &vf_tx_rate) ||
            nla_put(skb, IFLA_VF_SPOOFCHK, sizeof(vf_spoofchk),
                    &vf_spoofchk) ||
            nla_put(skb, IFLA_VF_LINK_STATE, sizeof(vf_linkstate),
                    &vf_linkstate) ||
            nla_put(skb, IFLA_VF_RSS_QUERY_EN,
                    sizeof(vf_rss_query_en),
                    &vf_rss_query_en) ||
            nla_put(skb, IFLA_VF_TRUST,
                    sizeof(vf_trust), &vf_trust))
                goto nla_put_vf_failure;

        if (dev->netdev_ops->ndo_get_vf_guid &&
            !dev->netdev_ops->ndo_get_vf_guid(dev, vfs_num, &node_guid,
                                              &port_guid)) {
                if (nla_put(skb, IFLA_VF_IB_NODE_GUID, sizeof(node_guid),
                            &node_guid) ||
                    nla_put(skb, IFLA_VF_IB_PORT_GUID, sizeof(port_guid),
                            &port_guid))
                        goto nla_put_vf_failure;
        }
        vfvlanlist = nla_nest_start_noflag(skb, IFLA_VF_VLAN_LIST);
        if (!vfvlanlist)
                goto nla_put_vf_failure;
        if (nla_put(skb, IFLA_VF_VLAN_INFO, sizeof(vf_vlan_info),
                    &vf_vlan_info)) {
                nla_nest_cancel(skb, vfvlanlist);
                goto nla_put_vf_failure;
        }
        nla_nest_end(skb, vfvlanlist);
        if (~ext_filter_mask & RTEXT_FILTER_SKIP_STATS) {
                memset(&vf_stats, 0, sizeof(vf_stats));
                if (dev->netdev_ops->ndo_get_vf_stats)
                        dev->netdev_ops->ndo_get_vf_stats(dev, vfs_num,
                                                          &vf_stats);
                vfstats = nla_nest_start_noflag(skb, IFLA_VF_STATS);
                if (!vfstats)
                        goto nla_put_vf_failure;
                if (nla_put_u64_64bit(skb, IFLA_VF_STATS_RX_PACKETS,
                                      vf_stats.rx_packets, IFLA_VF_STATS_PAD) ||
                    nla_put_u64_64bit(skb, IFLA_VF_STATS_TX_PACKETS,
                                      vf_stats.tx_packets, IFLA_VF_STATS_PAD) ||
                    nla_put_u64_64bit(skb, IFLA_VF_STATS_RX_BYTES,
                                      vf_stats.rx_bytes, IFLA_VF_STATS_PAD) ||
                    nla_put_u64_64bit(skb, IFLA_VF_STATS_TX_BYTES,
                                      vf_stats.tx_bytes, IFLA_VF_STATS_PAD) ||
                    nla_put_u64_64bit(skb, IFLA_VF_STATS_BROADCAST,
                                      vf_stats.broadcast, IFLA_VF_STATS_PAD) ||
                    nla_put_u64_64bit(skb, IFLA_VF_STATS_MULTICAST,
                                      vf_stats.multicast, IFLA_VF_STATS_PAD) ||
                    nla_put_u64_64bit(skb, IFLA_VF_STATS_RX_DROPPED,
                                      vf_stats.rx_dropped, IFLA_VF_STATS_PAD) ||
                    nla_put_u64_64bit(skb, IFLA_VF_STATS_TX_DROPPED,
                                      vf_stats.tx_dropped, IFLA_VF_STATS_PAD)) {
                        nla_nest_cancel(skb, vfstats);
                        goto nla_put_vf_failure;
                }
                nla_nest_end(skb, vfstats);
        }
        nla_nest_end(skb, vf);
        return 0;

nla_put_vf_failure:
        nla_nest_cancel(skb, vf);
        return -EMSGSIZE;
}

static noinline_for_stack int rtnl_fill_vf(struct sk_buff *skb,
                                           struct net_device *dev,
                                           u32 ext_filter_mask)
{
        struct nlattr *vfinfo;
        int i, num_vfs;

        if (!dev->dev.parent || ((ext_filter_mask & RTEXT_FILTER_VF) == 0))
                return 0;

        num_vfs = dev_num_vf(dev->dev.parent);
        if (nla_put_u32(skb, IFLA_NUM_VF, num_vfs))
                return -EMSGSIZE;

        if (!dev->netdev_ops->ndo_get_vf_config)
                return 0;

        vfinfo = nla_nest_start_noflag(skb, IFLA_VFINFO_LIST);
        if (!vfinfo)
                return -EMSGSIZE;

        for (i = 0; i < num_vfs; i++) {
                if (rtnl_fill_vfinfo(skb, dev, i, ext_filter_mask)) {
                        nla_nest_cancel(skb, vfinfo);
                        return -EMSGSIZE;
                }
        }

        nla_nest_end(skb, vfinfo);
        return 0;
}

static int rtnl_fill_link_ifmap(struct sk_buff *skb,
                                const struct net_device *dev)
{
        struct rtnl_link_ifmap map;

        memset(&map, 0, sizeof(map));
        map.mem_start = READ_ONCE(dev->mem_start);
        map.mem_end   = READ_ONCE(dev->mem_end);
        map.base_addr = READ_ONCE(dev->base_addr);
        map.irq       = READ_ONCE(dev->irq);
        map.dma       = READ_ONCE(dev->dma);
        map.port      = READ_ONCE(dev->if_port);

        if (nla_put_64bit(skb, IFLA_MAP, sizeof(map), &map, IFLA_PAD))
                return -EMSGSIZE;

        return 0;
}

static u32 rtnl_xdp_prog_skb(struct net_device *dev)
{
        const struct bpf_prog *generic_xdp_prog;
        u32 res = 0;

        rcu_read_lock();
        generic_xdp_prog = rcu_dereference(dev->xdp_prog);
        if (generic_xdp_prog)
                res = generic_xdp_prog->aux->id;
        rcu_read_unlock();

        return res;
}

static u32 rtnl_xdp_prog_drv(struct net_device *dev)
{
        return dev_xdp_prog_id(dev, XDP_MODE_DRV);
}

static u32 rtnl_xdp_prog_hw(struct net_device *dev)
{
        return dev_xdp_prog_id(dev, XDP_MODE_HW);
}

static int rtnl_xdp_report_one(struct sk_buff *skb, struct net_device *dev,
                               u32 *prog_id, u8 *mode, u8 tgt_mode, u32 attr,
                               u32 (*get_prog_id)(struct net_device *dev))
{
        u32 curr_id;
        int err;

        curr_id = get_prog_id(dev);
        if (!curr_id)
                return 0;

        *prog_id = curr_id;
        err = nla_put_u32(skb, attr, curr_id);
        if (err)
                return err;

        if (*mode != XDP_ATTACHED_NONE)
                *mode = XDP_ATTACHED_MULTI;
        else
                *mode = tgt_mode;

        return 0;
}

static int rtnl_xdp_fill(struct sk_buff *skb, struct net_device *dev)
{
        struct nlattr *xdp;
        u32 prog_id;
        int err;
        u8 mode;

        xdp = nla_nest_start_noflag(skb, IFLA_XDP);
        if (!xdp)
                return -EMSGSIZE;

        prog_id = 0;
        mode = XDP_ATTACHED_NONE;
        err = rtnl_xdp_report_one(skb, dev, &prog_id, &mode, XDP_ATTACHED_SKB,
                                  IFLA_XDP_SKB_PROG_ID, rtnl_xdp_prog_skb);
        if (err)
                goto err_cancel;
        err = rtnl_xdp_report_one(skb, dev, &prog_id, &mode, XDP_ATTACHED_DRV,
                                  IFLA_XDP_DRV_PROG_ID, rtnl_xdp_prog_drv);
        if (err)
                goto err_cancel;
        err = rtnl_xdp_report_one(skb, dev, &prog_id, &mode, XDP_ATTACHED_HW,
                                  IFLA_XDP_HW_PROG_ID, rtnl_xdp_prog_hw);
        if (err)
                goto err_cancel;

        err = nla_put_u8(skb, IFLA_XDP_ATTACHED, mode);
        if (err)
                goto err_cancel;

        if (prog_id && mode != XDP_ATTACHED_MULTI) {
                err = nla_put_u32(skb, IFLA_XDP_PROG_ID, prog_id);
                if (err)
                        goto err_cancel;
        }

        nla_nest_end(skb, xdp);
        return 0;

err_cancel:
        nla_nest_cancel(skb, xdp);
        return err;
}

static u32 rtnl_get_event(unsigned long event)
{
        u32 rtnl_event_type = IFLA_EVENT_NONE;

        switch (event) {
        case NETDEV_REBOOT:
                rtnl_event_type = IFLA_EVENT_REBOOT;
                break;
        case NETDEV_FEAT_CHANGE:
                rtnl_event_type = IFLA_EVENT_FEATURES;
                break;
        case NETDEV_BONDING_FAILOVER:
                rtnl_event_type = IFLA_EVENT_BONDING_FAILOVER;
                break;
        case NETDEV_NOTIFY_PEERS:
                rtnl_event_type = IFLA_EVENT_NOTIFY_PEERS;
                break;
        case NETDEV_RESEND_IGMP:
                rtnl_event_type = IFLA_EVENT_IGMP_RESEND;
                break;
        case NETDEV_CHANGEINFODATA:
                rtnl_event_type = IFLA_EVENT_BONDING_OPTIONS;
                break;
        default:
                break;
        }

        return rtnl_event_type;
}

static int put_master_ifindex(struct sk_buff *skb, struct net_device *dev)
{
        const struct net_device *upper_dev;
        int ret = 0;

        rcu_read_lock();

        upper_dev = netdev_master_upper_dev_get_rcu(dev);
        if (upper_dev)
                ret = nla_put_u32(skb, IFLA_MASTER,
                                  READ_ONCE(upper_dev->ifindex));

        rcu_read_unlock();
        return ret;
}

static int nla_put_iflink(struct sk_buff *skb, const struct net_device *dev,
                          bool force)
{
        int iflink = dev_get_iflink(dev);

        if (force || READ_ONCE(dev->ifindex) != iflink)
                return nla_put_u32(skb, IFLA_LINK, iflink);

        return 0;
}

static noinline_for_stack int nla_put_ifalias(struct sk_buff *skb,
                                              struct net_device *dev)
{
        char buf[IFALIASZ];
        int ret;

        ret = dev_get_alias(dev, buf, sizeof(buf));
        return ret > 0 ? nla_put_string(skb, IFLA_IFALIAS, buf) : 0;
}

static int rtnl_fill_link_netnsid(struct sk_buff *skb,
                                  const struct net_device *dev,
                                  struct net *src_net, gfp_t gfp)
{
        bool put_iflink = false;

        if (dev->rtnl_link_ops && dev->rtnl_link_ops->get_link_net) {
                struct net *link_net = dev->rtnl_link_ops->get_link_net(dev);

                if (!net_eq(dev_net(dev), link_net)) {
                        int id = peernet2id_alloc(src_net, link_net, gfp);

                        if (nla_put_s32(skb, IFLA_LINK_NETNSID, id))
                                return -EMSGSIZE;

                        put_iflink = true;
                }
        }

        return nla_put_iflink(skb, dev, put_iflink);
}

static int rtnl_fill_link_af(struct sk_buff *skb,
                             const struct net_device *dev,
                             u32 ext_filter_mask)
{
        const struct rtnl_af_ops *af_ops;
        struct nlattr *af_spec;

        af_spec = nla_nest_start_noflag(skb, IFLA_AF_SPEC);
        if (!af_spec)
                return -EMSGSIZE;

        list_for_each_entry_rcu(af_ops, &rtnl_af_ops, list) {
                struct nlattr *af;
                int err;

                if (!af_ops->fill_link_af)
                        continue;

                af = nla_nest_start_noflag(skb, af_ops->family);
                if (!af)
                        return -EMSGSIZE;

                err = af_ops->fill_link_af(skb, dev, ext_filter_mask);
                /*
                 * Caller may return ENODATA to indicate that there
                 * was no data to be dumped. This is not an error, it
                 * means we should trim the attribute header and
                 * continue.
                 */
                if (err == -ENODATA)
                        nla_nest_cancel(skb, af);
                else if (err < 0)
                        return -EMSGSIZE;

                nla_nest_end(skb, af);
        }

        nla_nest_end(skb, af_spec);
        return 0;
}

static int rtnl_fill_alt_ifnames(struct sk_buff *skb,
                                 const struct net_device *dev)
{
        struct netdev_name_node *name_node;
        int count = 0;

        list_for_each_entry_rcu(name_node, &dev->name_node->list, list) {
                if (nla_put_string(skb, IFLA_ALT_IFNAME, name_node->name))
                        return -EMSGSIZE;
                count++;
        }
        return count;
}

/* RCU protected. */
static int rtnl_fill_prop_list(struct sk_buff *skb,
                               const struct net_device *dev)
{
        struct nlattr *prop_list;
        int ret;

        prop_list = nla_nest_start(skb, IFLA_PROP_LIST);
        if (!prop_list)
                return -EMSGSIZE;

        ret = rtnl_fill_alt_ifnames(skb, dev);
        if (ret <= 0)
                goto nest_cancel;

        nla_nest_end(skb, prop_list);
        return 0;

nest_cancel:
        nla_nest_cancel(skb, prop_list);
        return ret;
}

static int rtnl_fill_proto_down(struct sk_buff *skb,
                                const struct net_device *dev)
{
        struct nlattr *pr;
        u32 preason;

        if (nla_put_u8(skb, IFLA_PROTO_DOWN, READ_ONCE(dev->proto_down)))
                goto nla_put_failure;

        preason = READ_ONCE(dev->proto_down_reason);
        if (!preason)
                return 0;

        pr = nla_nest_start(skb, IFLA_PROTO_DOWN_REASON);
        if (!pr)
                return -EMSGSIZE;

        if (nla_put_u32(skb, IFLA_PROTO_DOWN_REASON_VALUE, preason)) {
                nla_nest_cancel(skb, pr);
                goto nla_put_failure;
        }

        nla_nest_end(skb, pr);
        return 0;

nla_put_failure:
        return -EMSGSIZE;
}

static int rtnl_fill_devlink_port(struct sk_buff *skb,
                                  const struct net_device *dev)
{
        struct nlattr *devlink_port_nest;
        int ret;

        devlink_port_nest = nla_nest_start(skb, IFLA_DEVLINK_PORT);
        if (!devlink_port_nest)
                return -EMSGSIZE;

        if (dev->devlink_port) {
                ret = devlink_nl_port_handle_fill(skb, dev->devlink_port);
                if (ret < 0)
                        goto nest_cancel;
        }

        nla_nest_end(skb, devlink_port_nest);
        return 0;

nest_cancel:
        nla_nest_cancel(skb, devlink_port_nest);
        return ret;
}

static int rtnl_fill_dpll_pin(struct sk_buff *skb,
                              const struct net_device *dev)
{
        struct nlattr *dpll_pin_nest;
        int ret;

        dpll_pin_nest = nla_nest_start(skb, IFLA_DPLL_PIN);
        if (!dpll_pin_nest)
                return -EMSGSIZE;

        ret = dpll_netdev_add_pin_handle(skb, dev);
        if (ret < 0)
                goto nest_cancel;

        nla_nest_end(skb, dpll_pin_nest);
        return 0;

nest_cancel:
        nla_nest_cancel(skb, dpll_pin_nest);
        return ret;
}

static int rtnl_fill_ifinfo(struct sk_buff *skb,
                            struct net_device *dev, struct net *src_net,
                            int type, u32 pid, u32 seq, u32 change,
                            unsigned int flags, u32 ext_filter_mask,
                            u32 event, int *new_nsid, int new_ifindex,
                            int tgt_netnsid, gfp_t gfp)
{
        char devname[IFNAMSIZ];
        struct ifinfomsg *ifm;
        struct nlmsghdr *nlh;
        struct Qdisc *qdisc;

        ASSERT_RTNL();
        nlh = nlmsg_put(skb, pid, seq, type, sizeof(*ifm), flags);
        if (nlh == NULL)
                return -EMSGSIZE;

        ifm = nlmsg_data(nlh);
        ifm->ifi_family = AF_UNSPEC;
        ifm->__ifi_pad = 0;
        ifm->ifi_type = READ_ONCE(dev->type);
        ifm->ifi_index = READ_ONCE(dev->ifindex);
        ifm->ifi_flags = netif_get_flags(dev);
        ifm->ifi_change = change;

        if (tgt_netnsid >= 0 && nla_put_s32(skb, IFLA_TARGET_NETNSID, tgt_netnsid))
                goto nla_put_failure;

        netdev_copy_name(dev, devname);
        if (nla_put_string(skb, IFLA_IFNAME, devname))
                goto nla_put_failure;

        if (nla_put_u32(skb, IFLA_TXQLEN, READ_ONCE(dev->tx_queue_len)) ||
            nla_put_u8(skb, IFLA_OPERSTATE,
                       netif_running(dev) ? READ_ONCE(dev->operstate) :
                                            IF_OPER_DOWN) ||
            nla_put_u8(skb, IFLA_LINKMODE, READ_ONCE(dev->link_mode)) ||
            nla_put_u8(skb, IFLA_NETNS_IMMUTABLE, dev->netns_immutable) ||
            nla_put_u32(skb, IFLA_MTU, READ_ONCE(dev->mtu)) ||
            nla_put_u32(skb, IFLA_MIN_MTU, READ_ONCE(dev->min_mtu)) ||
            nla_put_u32(skb, IFLA_MAX_MTU, READ_ONCE(dev->max_mtu)) ||
            nla_put_u32(skb, IFLA_GROUP, READ_ONCE(dev->group)) ||
            nla_put_u32(skb, IFLA_PROMISCUITY, READ_ONCE(dev->promiscuity)) ||
            nla_put_u32(skb, IFLA_ALLMULTI, READ_ONCE(dev->allmulti)) ||
            nla_put_u32(skb, IFLA_NUM_TX_QUEUES,
                        READ_ONCE(dev->num_tx_queues)) ||
            nla_put_u32(skb, IFLA_GSO_MAX_SEGS,
                        READ_ONCE(dev->gso_max_segs)) ||
            nla_put_u32(skb, IFLA_GSO_MAX_SIZE,
                        READ_ONCE(dev->gso_max_size)) ||
            nla_put_u32(skb, IFLA_GRO_MAX_SIZE,
                        READ_ONCE(dev->gro_max_size)) ||
            nla_put_u32(skb, IFLA_GSO_IPV4_MAX_SIZE,
                        READ_ONCE(dev->gso_ipv4_max_size)) ||
            nla_put_u32(skb, IFLA_GRO_IPV4_MAX_SIZE,
                        READ_ONCE(dev->gro_ipv4_max_size)) ||
            nla_put_u32(skb, IFLA_TSO_MAX_SIZE,
                        READ_ONCE(dev->tso_max_size)) ||
            nla_put_u32(skb, IFLA_TSO_MAX_SEGS,
                        READ_ONCE(dev->tso_max_segs)) ||
            nla_put_uint(skb, IFLA_MAX_PACING_OFFLOAD_HORIZON,
                         READ_ONCE(dev->max_pacing_offload_horizon)) ||
#ifdef CONFIG_RPS
            nla_put_u32(skb, IFLA_NUM_RX_QUEUES,
                        READ_ONCE(dev->num_rx_queues)) ||
#endif
            put_master_ifindex(skb, dev) ||
            nla_put_u8(skb, IFLA_CARRIER, netif_carrier_ok(dev)) ||
            nla_put_ifalias(skb, dev) ||
            nla_put_u32(skb, IFLA_CARRIER_CHANGES,
                        atomic_read(&dev->carrier_up_count) +
                        atomic_read(&dev->carrier_down_count)) ||
            nla_put_u32(skb, IFLA_CARRIER_UP_COUNT,
                        atomic_read(&dev->carrier_up_count)) ||
            nla_put_u32(skb, IFLA_CARRIER_DOWN_COUNT,
                        atomic_read(&dev->carrier_down_count)) ||
            nla_put_u16(skb, IFLA_HEADROOM,
                        READ_ONCE(dev->needed_headroom)) ||
            nla_put_u16(skb, IFLA_TAILROOM,
                        READ_ONCE(dev->needed_tailroom)))
                goto nla_put_failure;

        if (rtnl_fill_proto_down(skb, dev))
                goto nla_put_failure;

        if (event != IFLA_EVENT_NONE) {
                if (nla_put_u32(skb, IFLA_EVENT, event))
                        goto nla_put_failure;
        }

        if (dev->addr_len) {
                if (nla_put(skb, IFLA_ADDRESS, dev->addr_len, dev->dev_addr) ||
                    nla_put(skb, IFLA_BROADCAST, dev->addr_len, dev->broadcast))
                        goto nla_put_failure;
        }

        if (rtnl_phys_port_id_fill(skb, dev))
                goto nla_put_failure;

        if (rtnl_phys_port_name_fill(skb, dev))
                goto nla_put_failure;

        if (rtnl_phys_switch_id_fill(skb, dev))
                goto nla_put_failure;

        if (!(ext_filter_mask & RTEXT_FILTER_SKIP_STATS) &&
            rtnl_fill_stats(skb, dev))
                goto nla_put_failure;

        if (rtnl_fill_vf(skb, dev, ext_filter_mask))
                goto nla_put_failure;

        if (rtnl_port_fill(skb, dev, ext_filter_mask))
                goto nla_put_failure;

        if (rtnl_xdp_fill(skb, dev))
                goto nla_put_failure;

        if (dev->rtnl_link_ops || rtnl_have_link_slave_info(dev)) {
                if (rtnl_link_fill(skb, dev) < 0)
                        goto nla_put_failure;
        }

        if (new_nsid &&
            nla_put_s32(skb, IFLA_NEW_NETNSID, *new_nsid) < 0)
                goto nla_put_failure;
        if (new_ifindex &&
            nla_put_s32(skb, IFLA_NEW_IFINDEX, new_ifindex) < 0)
                goto nla_put_failure;

        if (memchr_inv(dev->perm_addr, '\0', dev->addr_len) &&
            nla_put(skb, IFLA_PERM_ADDRESS, dev->addr_len, dev->perm_addr))
                goto nla_put_failure;

        rcu_read_lock();
        if (rtnl_fill_link_netnsid(skb, dev, src_net, GFP_ATOMIC))
                goto nla_put_failure_rcu;
        qdisc = rcu_dereference(dev->qdisc);
        if (qdisc && nla_put_string(skb, IFLA_QDISC, qdisc->ops->id))
                goto nla_put_failure_rcu;
        if (rtnl_fill_link_af(skb, dev, ext_filter_mask))
                goto nla_put_failure_rcu;
        if (rtnl_fill_link_ifmap(skb, dev))
                goto nla_put_failure_rcu;
        if (rtnl_fill_prop_list(skb, dev))
                goto nla_put_failure_rcu;
        rcu_read_unlock();

        if (dev->dev.parent &&
            nla_put_string(skb, IFLA_PARENT_DEV_NAME,
                           dev_name(dev->dev.parent)))
                goto nla_put_failure;

        if (dev->dev.parent && dev->dev.parent->bus &&
            nla_put_string(skb, IFLA_PARENT_DEV_BUS_NAME,
                           dev->dev.parent->bus->name))
                goto nla_put_failure;

        if (rtnl_fill_devlink_port(skb, dev))
                goto nla_put_failure;

        if (rtnl_fill_dpll_pin(skb, dev))
                goto nla_put_failure;

        nlmsg_end(skb, nlh);
        return 0;

nla_put_failure_rcu:
        rcu_read_unlock();
nla_put_failure:
        nlmsg_cancel(skb, nlh);
        return -EMSGSIZE;
}

static const struct nla_policy ifla_policy[IFLA_MAX+1] = {
        [IFLA_UNSPEC]           = { .strict_start_type = IFLA_DPLL_PIN },
        [IFLA_IFNAME]           = { .type = NLA_STRING, .len = IFNAMSIZ-1 },
        [IFLA_ADDRESS]          = { .type = NLA_BINARY, .len = MAX_ADDR_LEN },
        [IFLA_BROADCAST]        = { .type = NLA_BINARY, .len = MAX_ADDR_LEN },
        [IFLA_MAP]              = { .len = sizeof(struct rtnl_link_ifmap) },
        [IFLA_MTU]              = { .type = NLA_U32 },
        [IFLA_LINK]             = { .type = NLA_U32 },
        [IFLA_MASTER]           = { .type = NLA_U32 },
        [IFLA_CARRIER]          = { .type = NLA_U8 },
        [IFLA_TXQLEN]           = { .type = NLA_U32 },
        [IFLA_WEIGHT]           = { .type = NLA_U32 },
        [IFLA_OPERSTATE]        = { .type = NLA_U8 },
        [IFLA_LINKMODE]         = { .type = NLA_U8 },
        [IFLA_LINKINFO]         = { .type = NLA_NESTED },
        [IFLA_NET_NS_PID]       = { .type = NLA_U32 },
        [IFLA_NET_NS_FD]        = { .type = NLA_U32 },
        /* IFLA_IFALIAS is a string, but policy is set to NLA_BINARY to
         * allow 0-length string (needed to remove an alias).
         */
        [IFLA_IFALIAS]          = { .type = NLA_BINARY, .len = IFALIASZ - 1 },
        [IFLA_VFINFO_LIST]      = {. type = NLA_NESTED },
        [IFLA_VF_PORTS]         = { .type = NLA_NESTED },
        [IFLA_PORT_SELF]        = { .type = NLA_NESTED },
        [IFLA_AF_SPEC]          = { .type = NLA_NESTED },
        [IFLA_EXT_MASK]         = { .type = NLA_U32 },
        [IFLA_PROMISCUITY]      = { .type = NLA_U32 },
        [IFLA_NUM_TX_QUEUES]    = { .type = NLA_U32 },
        [IFLA_NUM_RX_QUEUES]    = { .type = NLA_U32 },
        [IFLA_GSO_MAX_SEGS]     = { .type = NLA_U32 },
        [IFLA_GSO_MAX_SIZE]     = NLA_POLICY_MIN(NLA_U32, MAX_TCP_HEADER + 1),
        [IFLA_PHYS_PORT_ID]     = { .type = NLA_BINARY, .len = MAX_PHYS_ITEM_ID_LEN },
        [IFLA_CARRIER_CHANGES]  = { .type = NLA_U32 },  /* ignored */
        [IFLA_PHYS_SWITCH_ID]   = { .type = NLA_BINARY, .len = MAX_PHYS_ITEM_ID_LEN },
        [IFLA_LINK_NETNSID]     = { .type = NLA_S32 },
        [IFLA_PROTO_DOWN]       = { .type = NLA_U8 },
        [IFLA_XDP]              = { .type = NLA_NESTED },
        [IFLA_EVENT]            = { .type = NLA_U32 },
        [IFLA_GROUP]            = { .type = NLA_U32 },
        [IFLA_TARGET_NETNSID]   = { .type = NLA_S32 },
        [IFLA_CARRIER_UP_COUNT] = { .type = NLA_U32 },
        [IFLA_CARRIER_DOWN_COUNT] = { .type = NLA_U32 },
        [IFLA_MIN_MTU]          = { .type = NLA_U32 },
        [IFLA_MAX_MTU]          = { .type = NLA_U32 },
        [IFLA_PROP_LIST]        = { .type = NLA_NESTED },
        [IFLA_ALT_IFNAME]       = { .type = NLA_STRING,
                                    .len = ALTIFNAMSIZ - 1 },
        [IFLA_PERM_ADDRESS]     = { .type = NLA_REJECT },
        [IFLA_PROTO_DOWN_REASON] = { .type = NLA_NESTED },
        [IFLA_NEW_IFINDEX]      = NLA_POLICY_MIN(NLA_S32, 1),
        [IFLA_PARENT_DEV_NAME]  = { .type = NLA_NUL_STRING },
        [IFLA_GRO_MAX_SIZE]     = { .type = NLA_U32 },
        [IFLA_TSO_MAX_SIZE]     = { .type = NLA_REJECT },
        [IFLA_TSO_MAX_SEGS]     = { .type = NLA_REJECT },
        [IFLA_ALLMULTI]         = { .type = NLA_REJECT },
        [IFLA_GSO_IPV4_MAX_SIZE]        = NLA_POLICY_MIN(NLA_U32, MAX_TCP_HEADER + 1),
        [IFLA_GRO_IPV4_MAX_SIZE]        = { .type = NLA_U32 },
        [IFLA_NETNS_IMMUTABLE]  = { .type = NLA_REJECT },
        [IFLA_HEADROOM]         = { .type = NLA_REJECT },
        [IFLA_TAILROOM]         = { .type = NLA_REJECT },
};

static const struct nla_policy ifla_info_policy[IFLA_INFO_MAX+1] = {
        [IFLA_INFO_KIND]        = { .type = NLA_STRING },
        [IFLA_INFO_DATA]        = { .type = NLA_NESTED },
        [IFLA_INFO_SLAVE_KIND]  = { .type = NLA_STRING },
        [IFLA_INFO_SLAVE_DATA]  = { .type = NLA_NESTED },
};

static const struct nla_policy ifla_vf_policy[IFLA_VF_MAX+1] = {
        [IFLA_VF_MAC]           = { .len = sizeof(struct ifla_vf_mac) },
        [IFLA_VF_BROADCAST]     = { .type = NLA_REJECT },
        [IFLA_VF_VLAN]          = { .len = sizeof(struct ifla_vf_vlan) },
        [IFLA_VF_VLAN_LIST]     = { .type = NLA_NESTED },
        [IFLA_VF_TX_RATE]       = { .len = sizeof(struct ifla_vf_tx_rate) },
        [IFLA_VF_SPOOFCHK]      = { .len = sizeof(struct ifla_vf_spoofchk) },
        [IFLA_VF_RATE]          = { .len = sizeof(struct ifla_vf_rate) },
        [IFLA_VF_LINK_STATE]    = { .len = sizeof(struct ifla_vf_link_state) },
        [IFLA_VF_RSS_QUERY_EN]  = { .len = sizeof(struct ifla_vf_rss_query_en) },
        [IFLA_VF_STATS]         = { .type = NLA_NESTED },
        [IFLA_VF_TRUST]         = { .len = sizeof(struct ifla_vf_trust) },
        [IFLA_VF_IB_NODE_GUID]  = { .len = sizeof(struct ifla_vf_guid) },
        [IFLA_VF_IB_PORT_GUID]  = { .len = sizeof(struct ifla_vf_guid) },
};

static const struct nla_policy ifla_port_policy[IFLA_PORT_MAX+1] = {
        [IFLA_PORT_VF]          = { .type = NLA_U32 },
        [IFLA_PORT_PROFILE]     = { .type = NLA_STRING,
                                    .len = PORT_PROFILE_MAX },
        [IFLA_PORT_INSTANCE_UUID] = { .type = NLA_BINARY,
                                      .len = PORT_UUID_MAX },
        [IFLA_PORT_HOST_UUID]   = { .type = NLA_STRING,
                                    .len = PORT_UUID_MAX },
        [IFLA_PORT_REQUEST]     = { .type = NLA_U8, },
        [IFLA_PORT_RESPONSE]    = { .type = NLA_U16, },

        /* Unused, but we need to keep it here since user space could
         * fill it. It's also broken with regard to NLA_BINARY use in
         * combination with structs.
         */
        [IFLA_PORT_VSI_TYPE]    = { .type = NLA_BINARY,
                                    .len = sizeof(struct ifla_port_vsi) },
};

static const struct nla_policy ifla_xdp_policy[IFLA_XDP_MAX + 1] = {
        [IFLA_XDP_UNSPEC]       = { .strict_start_type = IFLA_XDP_EXPECTED_FD },
        [IFLA_XDP_FD]           = { .type = NLA_S32 },
        [IFLA_XDP_EXPECTED_FD]  = { .type = NLA_S32 },
        [IFLA_XDP_ATTACHED]     = { .type = NLA_U8 },
        [IFLA_XDP_FLAGS]        = { .type = NLA_U32 },
        [IFLA_XDP_PROG_ID]      = { .type = NLA_U32 },
};

static struct rtnl_link_ops *linkinfo_to_kind_ops(const struct nlattr *nla,
                                                  int *ops_srcu_index)
{
        struct nlattr *linfo[IFLA_INFO_MAX + 1];
        struct rtnl_link_ops *ops = NULL;

        if (nla_parse_nested_deprecated(linfo, IFLA_INFO_MAX, nla, ifla_info_policy, NULL) < 0)
                return NULL;

        if (linfo[IFLA_INFO_KIND]) {
                char kind[MODULE_NAME_LEN];

                nla_strscpy(kind, linfo[IFLA_INFO_KIND], sizeof(kind));
                ops = rtnl_link_ops_get(kind, ops_srcu_index);
        }

        return ops;
}

static bool link_master_filtered(struct net_device *dev, int master_idx)
{
        struct net_device *master;

        if (!master_idx)
                return false;

        master = netdev_master_upper_dev_get(dev);

        /* 0 is already used to denote IFLA_MASTER wasn't passed, therefore need
         * another invalid value for ifindex to denote "no master".
         */
        if (master_idx == -1)
                return !!master;

        if (!master || master->ifindex != master_idx)
                return true;

        return false;
}

static bool link_kind_filtered(const struct net_device *dev,
                               const struct rtnl_link_ops *kind_ops)
{
        if (kind_ops && dev->rtnl_link_ops != kind_ops)
                return true;

        return false;
}

static bool link_dump_filtered(struct net_device *dev,
                               int master_idx,
                               const struct rtnl_link_ops *kind_ops)
{
        if (link_master_filtered(dev, master_idx) ||
            link_kind_filtered(dev, kind_ops))
                return true;

        return false;
}

/**
 * rtnl_get_net_ns_capable - Get netns if sufficiently privileged.
 * @sk: netlink socket
 * @netnsid: network namespace identifier
 *
 * Returns the network namespace identified by netnsid on success or an error
 * pointer on failure.
 */
struct net *rtnl_get_net_ns_capable(struct sock *sk, int netnsid)
{
        struct net *net;

        net = get_net_ns_by_id(sock_net(sk), netnsid);
        if (!net)
                return ERR_PTR(-EINVAL);

        /* For now, the caller is required to have CAP_NET_ADMIN in
         * the user namespace owning the target net ns.
         */
        if (!sk_ns_capable(sk, net->user_ns, CAP_NET_ADMIN)) {
                put_net(net);
                return ERR_PTR(-EACCES);
        }
        return net;
}
EXPORT_SYMBOL_GPL(rtnl_get_net_ns_capable);

static int rtnl_valid_dump_ifinfo_req(const struct nlmsghdr *nlh,
                                      bool strict_check, struct nlattr **tb,
                                      struct netlink_ext_ack *extack)
{
        int hdrlen;

        if (strict_check) {
                struct ifinfomsg *ifm;

                ifm = nlmsg_payload(nlh, sizeof(*ifm));
                if (!ifm) {
                        NL_SET_ERR_MSG(extack, "Invalid header for link dump");
                        return -EINVAL;
                }

                if (ifm->__ifi_pad || ifm->ifi_type || ifm->ifi_flags ||
                    ifm->ifi_change) {
                        NL_SET_ERR_MSG(extack, "Invalid values in header for link dump request");
                        return -EINVAL;
                }
                if (ifm->ifi_index) {
                        NL_SET_ERR_MSG(extack, "Filter by device index not supported for link dumps");
                        return -EINVAL;
                }

                return nlmsg_parse_deprecated_strict(nlh, sizeof(*ifm), tb,
                                                     IFLA_MAX, ifla_policy,
                                                     extack);
        }

        /* A hack to preserve kernel<->userspace interface.
         * The correct header is ifinfomsg. It is consistent with rtnl_getlink.
         * However, before Linux v3.9 the code here assumed rtgenmsg and that's
         * what iproute2 < v3.9.0 used.
         * We can detect the old iproute2. Even including the IFLA_EXT_MASK
         * attribute, its netlink message is shorter than struct ifinfomsg.
         */
        hdrlen = nlmsg_len(nlh) < sizeof(struct ifinfomsg) ?
                 sizeof(struct rtgenmsg) : sizeof(struct ifinfomsg);

        return nlmsg_parse_deprecated(nlh, hdrlen, tb, IFLA_MAX, ifla_policy,
                                      extack);
}

static int rtnl_dump_ifinfo(struct sk_buff *skb, struct netlink_callback *cb)
{
        struct netlink_ext_ack *extack = cb->extack;
        struct rtnl_link_ops *kind_ops = NULL;
        const struct nlmsghdr *nlh = cb->nlh;
        struct net *net = sock_net(skb->sk);
        unsigned int flags = NLM_F_MULTI;
        struct nlattr *tb[IFLA_MAX+1];
        struct {
                unsigned long ifindex;
        } *ctx = (void *)cb->ctx;
        struct net *tgt_net = net;
        u32 ext_filter_mask = 0;
        struct net_device *dev;
        int ops_srcu_index;
        int master_idx = 0;
        int netnsid = -1;
        int err, i;

        err = rtnl_valid_dump_ifinfo_req(nlh, cb->strict_check, tb, extack);
        if (err < 0) {
                if (cb->strict_check)
                        return err;

                goto walk_entries;
        }

        for (i = 0; i <= IFLA_MAX; ++i) {
                if (!tb[i])
                        continue;

                /* new attributes should only be added with strict checking */
                switch (i) {
                case IFLA_TARGET_NETNSID:
                        netnsid = nla_get_s32(tb[i]);
                        tgt_net = rtnl_get_net_ns_capable(skb->sk, netnsid);
                        if (IS_ERR(tgt_net)) {
                                NL_SET_ERR_MSG(extack, "Invalid target network namespace id");
                                err = PTR_ERR(tgt_net);
                                netnsid = -1;
                                goto out;
                        }
                        break;
                case IFLA_EXT_MASK:
                        ext_filter_mask = nla_get_u32(tb[i]);
                        break;
                case IFLA_MASTER:
                        master_idx = nla_get_u32(tb[i]);
                        break;
                case IFLA_LINKINFO:
                        kind_ops = linkinfo_to_kind_ops(tb[i], &ops_srcu_index);
                        break;
                default:
                        if (cb->strict_check) {
                                NL_SET_ERR_MSG(extack, "Unsupported attribute in link dump request");
                                err = -EINVAL;
                                goto out;
                        }
                }
        }

        if (master_idx || kind_ops)
                flags |= NLM_F_DUMP_FILTERED;

walk_entries:
        err = 0;
        for_each_netdev_dump(tgt_net, dev, ctx->ifindex) {
                if (link_dump_filtered(dev, master_idx, kind_ops))
                        continue;
                err = rtnl_fill_ifinfo(skb, dev, net, RTM_NEWLINK,
                                       NETLINK_CB(cb->skb).portid,
                                       nlh->nlmsg_seq, 0, flags,
                                       ext_filter_mask, 0, NULL, 0,
                                       netnsid, GFP_KERNEL);
                if (err < 0)
                        break;
        }


        cb->seq = tgt_net->dev_base_seq;
        nl_dump_check_consistent(cb, nlmsg_hdr(skb));

out:

        if (kind_ops)
                rtnl_link_ops_put(kind_ops, ops_srcu_index);
        if (netnsid >= 0)
                put_net(tgt_net);

        return err;
}

int rtnl_nla_parse_ifinfomsg(struct nlattr **tb, const struct nlattr *nla_peer,
                             struct netlink_ext_ack *exterr)
{
        const struct ifinfomsg *ifmp;
        const struct nlattr *attrs;
        size_t len;

        ifmp = nla_data(nla_peer);
        attrs = nla_data(nla_peer) + sizeof(struct ifinfomsg);
        len = nla_len(nla_peer) - sizeof(struct ifinfomsg);

        if (ifmp->ifi_index < 0) {
                NL_SET_ERR_MSG_ATTR(exterr, nla_peer,
                                    "ifindex can't be negative");
                return -EINVAL;
        }

        return nla_parse_deprecated(tb, IFLA_MAX, attrs, len, ifla_policy,
                                    exterr);
}
EXPORT_SYMBOL(rtnl_nla_parse_ifinfomsg);

static struct net *rtnl_link_get_net_ifla(struct nlattr *tb[])
{
        struct net *net = NULL;

        /* Examine the link attributes and figure out which
         * network namespace we are talking about.
         */
        if (tb[IFLA_NET_NS_PID])
                net = get_net_ns_by_pid(nla_get_u32(tb[IFLA_NET_NS_PID]));
        else if (tb[IFLA_NET_NS_FD])
                net = get_net_ns_by_fd(nla_get_u32(tb[IFLA_NET_NS_FD]));

        return net;
}

struct net *rtnl_link_get_net(struct net *src_net, struct nlattr *tb[])
{
        struct net *net = rtnl_link_get_net_ifla(tb);

        if (!net)
                net = get_net(src_net);

        return net;
}
EXPORT_SYMBOL(rtnl_link_get_net);

/* Figure out which network namespace we are talking about by
 * examining the link attributes in the following order:
 *
 * 1. IFLA_NET_NS_PID
 * 2. IFLA_NET_NS_FD
 * 3. IFLA_TARGET_NETNSID
 */
static struct net *rtnl_link_get_net_by_nlattr(struct net *src_net,
                                               struct nlattr *tb[])
{
        struct net *net;

        if (tb[IFLA_NET_NS_PID] || tb[IFLA_NET_NS_FD])
                return rtnl_link_get_net(src_net, tb);

        if (!tb[IFLA_TARGET_NETNSID])
                return get_net(src_net);

        net = get_net_ns_by_id(src_net, nla_get_u32(tb[IFLA_TARGET_NETNSID]));
        if (!net)
                return ERR_PTR(-EINVAL);

        return net;
}

static struct net *rtnl_link_get_net_capable(const struct sk_buff *skb,
                                             struct net *src_net,
                                             struct nlattr *tb[], int cap)
{
        struct net *net;

        net = rtnl_link_get_net_by_nlattr(src_net, tb);
        if (IS_ERR(net))
                return net;

        if (!netlink_ns_capable(skb, net->user_ns, cap)) {
                put_net(net);
                return ERR_PTR(-EPERM);
        }

        return net;
}

/* Verify that rtnetlink requests do not pass additional properties
 * potentially referring to different network namespaces.
 */
static int rtnl_ensure_unique_netns(struct nlattr *tb[],
                                    struct netlink_ext_ack *extack,
                                    bool netns_id_only)
{

        if (netns_id_only) {
                if (!tb[IFLA_NET_NS_PID] && !tb[IFLA_NET_NS_FD])
                        return 0;

                NL_SET_ERR_MSG(extack, "specified netns attribute not supported");
                return -EOPNOTSUPP;
        }

        if (tb[IFLA_TARGET_NETNSID] && (tb[IFLA_NET_NS_PID] || tb[IFLA_NET_NS_FD]))
                goto invalid_attr;

        if (tb[IFLA_NET_NS_PID] && (tb[IFLA_TARGET_NETNSID] || tb[IFLA_NET_NS_FD]))
                goto invalid_attr;

        if (tb[IFLA_NET_NS_FD] && (tb[IFLA_TARGET_NETNSID] || tb[IFLA_NET_NS_PID]))
                goto invalid_attr;

        return 0;

invalid_attr:
        NL_SET_ERR_MSG(extack, "multiple netns identifying attributes specified");
        return -EINVAL;
}

static  int rtnl_set_vf_rate(struct net_device *dev, int vf, int min_tx_rate,
                             int max_tx_rate)
{
        const struct net_device_ops *ops = dev->netdev_ops;

        if (!ops->ndo_set_vf_rate)
                return -EOPNOTSUPP;
        if (max_tx_rate && max_tx_rate < min_tx_rate)
                return -EINVAL;

        return ops->ndo_set_vf_rate(dev, vf, min_tx_rate, max_tx_rate);
}

static int validate_linkmsg(struct net_device *dev, struct nlattr *tb[],
                            struct netlink_ext_ack *extack)
{
        if (tb[IFLA_ADDRESS] &&
            nla_len(tb[IFLA_ADDRESS]) < dev->addr_len)
                return -EINVAL;

        if (tb[IFLA_BROADCAST] &&
            nla_len(tb[IFLA_BROADCAST]) < dev->addr_len)
                return -EINVAL;

        if (tb[IFLA_GSO_MAX_SIZE] &&
            nla_get_u32(tb[IFLA_GSO_MAX_SIZE]) > dev->tso_max_size) {
                NL_SET_ERR_MSG(extack, "too big gso_max_size");
                return -EINVAL;
        }

        if (tb[IFLA_GSO_MAX_SEGS] &&
            (nla_get_u32(tb[IFLA_GSO_MAX_SEGS]) > GSO_MAX_SEGS ||
             nla_get_u32(tb[IFLA_GSO_MAX_SEGS]) > dev->tso_max_segs)) {
                NL_SET_ERR_MSG(extack, "too big gso_max_segs");
                return -EINVAL;
        }

        if (tb[IFLA_GRO_MAX_SIZE] &&
            nla_get_u32(tb[IFLA_GRO_MAX_SIZE]) > GRO_MAX_SIZE) {
                NL_SET_ERR_MSG(extack, "too big gro_max_size");
                return -EINVAL;
        }

        if (tb[IFLA_GSO_IPV4_MAX_SIZE] &&
            nla_get_u32(tb[IFLA_GSO_IPV4_MAX_SIZE]) > dev->tso_max_size) {
                NL_SET_ERR_MSG(extack, "too big gso_ipv4_max_size");
                return -EINVAL;
        }

        if (tb[IFLA_GRO_IPV4_MAX_SIZE] &&
            nla_get_u32(tb[IFLA_GRO_IPV4_MAX_SIZE]) > GRO_MAX_SIZE) {
                NL_SET_ERR_MSG(extack, "too big gro_ipv4_max_size");
                return -EINVAL;
        }

        if (tb[IFLA_AF_SPEC]) {
                struct nlattr *af;
                int rem, err;

                nla_for_each_nested(af, tb[IFLA_AF_SPEC], rem) {
                        struct rtnl_af_ops *af_ops;
                        int af_ops_srcu_index;

                        af_ops = rtnl_af_lookup(nla_type(af), &af_ops_srcu_index);
                        if (!af_ops)
                                return -EAFNOSUPPORT;

                        if (!af_ops->set_link_af)
                                err = -EOPNOTSUPP;
                        else if (af_ops->validate_link_af)
                                err = af_ops->validate_link_af(dev, af, extack);
                        else
                                err = 0;

                        rtnl_af_put(af_ops, af_ops_srcu_index);

                        if (err < 0)
                                return err;
                }
        }

        return 0;
}

static int handle_infiniband_guid(struct net_device *dev, struct ifla_vf_guid *ivt,
                                  int guid_type)
{
        const struct net_device_ops *ops = dev->netdev_ops;

        return ops->ndo_set_vf_guid(dev, ivt->vf, ivt->guid, guid_type);
}

static int handle_vf_guid(struct net_device *dev, struct ifla_vf_guid *ivt, int guid_type)
{
        if (dev->type != ARPHRD_INFINIBAND)
                return -EOPNOTSUPP;

        return handle_infiniband_guid(dev, ivt, guid_type);
}

static int do_setvfinfo(struct net_device *dev, struct nlattr **tb)
{
        const struct net_device_ops *ops = dev->netdev_ops;
        int err = -EINVAL;

        if (tb[IFLA_VF_MAC]) {
                struct ifla_vf_mac *ivm = nla_data(tb[IFLA_VF_MAC]);

                if (ivm->vf >= INT_MAX)
                        return -EINVAL;
                err = -EOPNOTSUPP;
                if (ops->ndo_set_vf_mac)
                        err = ops->ndo_set_vf_mac(dev, ivm->vf,
                                                  ivm->mac);
                if (err < 0)
                        return err;
        }

        if (tb[IFLA_VF_VLAN]) {
                struct ifla_vf_vlan *ivv = nla_data(tb[IFLA_VF_VLAN]);

                if (ivv->vf >= INT_MAX)
                        return -EINVAL;
                err = -EOPNOTSUPP;
                if (ops->ndo_set_vf_vlan)
                        err = ops->ndo_set_vf_vlan(dev, ivv->vf, ivv->vlan,
                                                   ivv->qos,
                                                   htons(ETH_P_8021Q));
                if (err < 0)
                        return err;
        }

        if (tb[IFLA_VF_VLAN_LIST]) {
                struct ifla_vf_vlan_info *ivvl[MAX_VLAN_LIST_LEN];
                struct nlattr *attr;
                int rem, len = 0;

                err = -EOPNOTSUPP;
                if (!ops->ndo_set_vf_vlan)
                        return err;

                nla_for_each_nested(attr, tb[IFLA_VF_VLAN_LIST], rem) {
                        if (nla_type(attr) != IFLA_VF_VLAN_INFO ||
                            nla_len(attr) < sizeof(struct ifla_vf_vlan_info)) {
                                return -EINVAL;
                        }
                        if (len >= MAX_VLAN_LIST_LEN)
                                return -EOPNOTSUPP;
                        ivvl[len] = nla_data(attr);

                        len++;
                }
                if (len == 0)
                        return -EINVAL;

                if (ivvl[0]->vf >= INT_MAX)
                        return -EINVAL;
                err = ops->ndo_set_vf_vlan(dev, ivvl[0]->vf, ivvl[0]->vlan,
                                           ivvl[0]->qos, ivvl[0]->vlan_proto);
                if (err < 0)
                        return err;
        }

        if (tb[IFLA_VF_TX_RATE]) {
                struct ifla_vf_tx_rate *ivt = nla_data(tb[IFLA_VF_TX_RATE]);
                struct ifla_vf_info ivf;

                if (ivt->vf >= INT_MAX)
                        return -EINVAL;
                err = -EOPNOTSUPP;
                if (ops->ndo_get_vf_config)
                        err = ops->ndo_get_vf_config(dev, ivt->vf, &ivf);
                if (err < 0)
                        return err;

                err = rtnl_set_vf_rate(dev, ivt->vf,
                                       ivf.min_tx_rate, ivt->rate);
                if (err < 0)
                        return err;
        }

        if (tb[IFLA_VF_RATE]) {
                struct ifla_vf_rate *ivt = nla_data(tb[IFLA_VF_RATE]);

                if (ivt->vf >= INT_MAX)
                        return -EINVAL;

                err = rtnl_set_vf_rate(dev, ivt->vf,
                                       ivt->min_tx_rate, ivt->max_tx_rate);
                if (err < 0)
                        return err;
        }

        if (tb[IFLA_VF_SPOOFCHK]) {
                struct ifla_vf_spoofchk *ivs = nla_data(tb[IFLA_VF_SPOOFCHK]);

                if (ivs->vf >= INT_MAX)
                        return -EINVAL;
                err = -EOPNOTSUPP;
                if (ops->ndo_set_vf_spoofchk)
                        err = ops->ndo_set_vf_spoofchk(dev, ivs->vf,
                                                       ivs->setting);
                if (err < 0)
                        return err;
        }

        if (tb[IFLA_VF_LINK_STATE]) {
                struct ifla_vf_link_state *ivl = nla_data(tb[IFLA_VF_LINK_STATE]);

                if (ivl->vf >= INT_MAX)
                        return -EINVAL;
                err = -EOPNOTSUPP;
                if (ops->ndo_set_vf_link_state)
                        err = ops->ndo_set_vf_link_state(dev, ivl->vf,
                                                         ivl->link_state);
                if (err < 0)
                        return err;
        }

        if (tb[IFLA_VF_RSS_QUERY_EN]) {
                struct ifla_vf_rss_query_en *ivrssq_en;

                err = -EOPNOTSUPP;
                ivrssq_en = nla_data(tb[IFLA_VF_RSS_QUERY_EN]);
                if (ivrssq_en->vf >= INT_MAX)
                        return -EINVAL;
                if (ops->ndo_set_vf_rss_query_en)
                        err = ops->ndo_set_vf_rss_query_en(dev, ivrssq_en->vf,
                                                           ivrssq_en->setting);
                if (err < 0)
                        return err;
        }

        if (tb[IFLA_VF_TRUST]) {
                struct ifla_vf_trust *ivt = nla_data(tb[IFLA_VF_TRUST]);

                if (ivt->vf >= INT_MAX)
                        return -EINVAL;
                err = -EOPNOTSUPP;
                if (ops->ndo_set_vf_trust)
                        err = ops->ndo_set_vf_trust(dev, ivt->vf, ivt->setting);
                if (err < 0)
                        return err;
        }

        if (tb[IFLA_VF_IB_NODE_GUID]) {
                struct ifla_vf_guid *ivt = nla_data(tb[IFLA_VF_IB_NODE_GUID]);

                if (ivt->vf >= INT_MAX)
                        return -EINVAL;
                if (!ops->ndo_set_vf_guid)
                        return -EOPNOTSUPP;
                return handle_vf_guid(dev, ivt, IFLA_VF_IB_NODE_GUID);
        }

        if (tb[IFLA_VF_IB_PORT_GUID]) {
                struct ifla_vf_guid *ivt = nla_data(tb[IFLA_VF_IB_PORT_GUID]);

                if (ivt->vf >= INT_MAX)
                        return -EINVAL;
                if (!ops->ndo_set_vf_guid)
                        return -EOPNOTSUPP;

                return handle_vf_guid(dev, ivt, IFLA_VF_IB_PORT_GUID);
        }

        return err;
}

static int do_set_master(struct net_device *dev, int ifindex,
                         struct netlink_ext_ack *extack)
{
        struct net_device *upper_dev = netdev_master_upper_dev_get(dev);
        const struct net_device_ops *ops;
        int err;

        /* Release the lower lock, the upper is responsible for locking
         * the lower if needed. None of the existing upper devices
         * use netdev instance lock, so don't grab it.
         */

        if (upper_dev) {
                if (upper_dev->ifindex == ifindex)
                        return 0;
                ops = upper_dev->netdev_ops;
                if (ops->ndo_del_slave) {
                        netdev_unlock_ops(dev);
                        err = ops->ndo_del_slave(upper_dev, dev);
                        netdev_lock_ops(dev);
                        if (err)
                                return err;
                } else {
                        return -EOPNOTSUPP;
                }
        }

        if (ifindex) {
                upper_dev = __dev_get_by_index(dev_net(dev), ifindex);
                if (!upper_dev)
                        return -EINVAL;
                ops = upper_dev->netdev_ops;
                if (ops->ndo_add_slave) {
                        netdev_unlock_ops(dev);
                        err = ops->ndo_add_slave(upper_dev, dev, extack);
                        netdev_lock_ops(dev);
                        if (err)
                                return err;
                } else {
                        return -EOPNOTSUPP;
                }
        }
        return 0;
}

static const struct nla_policy ifla_proto_down_reason_policy[IFLA_PROTO_DOWN_REASON_VALUE + 1] = {
        [IFLA_PROTO_DOWN_REASON_MASK]   = { .type = NLA_U32 },
        [IFLA_PROTO_DOWN_REASON_VALUE]  = { .type = NLA_U32 },
};

static int do_set_proto_down(struct net_device *dev,
                             struct nlattr *nl_proto_down,
                             struct nlattr *nl_proto_down_reason,
                             struct netlink_ext_ack *extack)
{
        struct nlattr *pdreason[IFLA_PROTO_DOWN_REASON_MAX + 1];
        unsigned long mask = 0;
        u32 value;
        bool proto_down;
        int err;

        if (!dev->change_proto_down) {
                NL_SET_ERR_MSG(extack,  "Protodown not supported by device");
                return -EOPNOTSUPP;
        }

        if (nl_proto_down_reason) {
                err = nla_parse_nested_deprecated(pdreason,
                                                  IFLA_PROTO_DOWN_REASON_MAX,
                                                  nl_proto_down_reason,
                                                  ifla_proto_down_reason_policy,
                                                  NULL);
                if (err < 0)
                        return err;

                if (!pdreason[IFLA_PROTO_DOWN_REASON_VALUE]) {
                        NL_SET_ERR_MSG(extack, "Invalid protodown reason value");
                        return -EINVAL;
                }

                value = nla_get_u32(pdreason[IFLA_PROTO_DOWN_REASON_VALUE]);

                if (pdreason[IFLA_PROTO_DOWN_REASON_MASK])
                        mask = nla_get_u32(pdreason[IFLA_PROTO_DOWN_REASON_MASK]);

                netdev_change_proto_down_reason_locked(dev, mask, value);
        }

        if (nl_proto_down) {
                proto_down = nla_get_u8(nl_proto_down);

                /* Don't turn off protodown if there are active reasons */
                if (!proto_down && dev->proto_down_reason) {
                        NL_SET_ERR_MSG(extack, "Cannot clear protodown, active reasons");
                        return -EBUSY;
                }
                err = netif_change_proto_down(dev, proto_down);
                if (err)
                        return err;
        }

        return 0;
}

#define DO_SETLINK_MODIFIED     0x01
/* notify flag means notify + modified. */
#define DO_SETLINK_NOTIFY       0x03
static int do_setlink(const struct sk_buff *skb, struct net_device *dev,
                      struct net *tgt_net, struct ifinfomsg *ifm,
                      struct netlink_ext_ack *extack,
                      struct nlattr **tb, int status)
{
        const struct net_device_ops *ops = dev->netdev_ops;
        char ifname[IFNAMSIZ];
        int err;

        err = validate_linkmsg(dev, tb, extack);
        if (err < 0)
                return err;

        if (tb[IFLA_IFNAME])
                nla_strscpy(ifname, tb[IFLA_IFNAME], IFNAMSIZ);
        else
                ifname[0] = '\0';

        if (!net_eq(tgt_net, dev_net(dev))) {
                const char *pat = ifname[0] ? ifname : NULL;
                int new_ifindex;

                new_ifindex = nla_get_s32_default(tb[IFLA_NEW_IFINDEX], 0);

                err = __dev_change_net_namespace(dev, tgt_net, pat,
                                                 new_ifindex, extack);
                if (err)
                        return err;

                status |= DO_SETLINK_MODIFIED;
        }

        netdev_lock_ops(dev);

        if (tb[IFLA_MAP]) {
                struct rtnl_link_ifmap *u_map;
                struct ifmap k_map;

                if (!ops->ndo_set_config) {
                        err = -EOPNOTSUPP;
                        goto errout;
                }

                if (!netif_device_present(dev)) {
                        err = -ENODEV;
                        goto errout;
                }

                u_map = nla_data(tb[IFLA_MAP]);
                k_map.mem_start = (unsigned long) u_map->mem_start;
                k_map.mem_end = (unsigned long) u_map->mem_end;
                k_map.base_addr = (unsigned short) u_map->base_addr;
                k_map.irq = (unsigned char) u_map->irq;
                k_map.dma = (unsigned char) u_map->dma;
                k_map.port = (unsigned char) u_map->port;

                err = ops->ndo_set_config(dev, &k_map);
                if (err < 0)
                        goto errout;

                status |= DO_SETLINK_NOTIFY;
        }

        if (tb[IFLA_ADDRESS]) {
                struct sockaddr_storage ss = { };

                netdev_unlock_ops(dev);

                /* dev_addr_sem is an outer lock, enforce proper ordering */
                down_write(&dev_addr_sem);
                netdev_lock_ops(dev);

                ss.ss_family = dev->type;
                memcpy(ss.__data, nla_data(tb[IFLA_ADDRESS]), dev->addr_len);
                err = netif_set_mac_address(dev, &ss, extack);
                if (err) {
                        up_write(&dev_addr_sem);
                        goto errout;
                }
                status |= DO_SETLINK_MODIFIED;

                up_write(&dev_addr_sem);
        }

        if (tb[IFLA_MTU]) {
                err = netif_set_mtu_ext(dev, nla_get_u32(tb[IFLA_MTU]), extack);
                if (err < 0)
                        goto errout;
                status |= DO_SETLINK_MODIFIED;
        }

        if (tb[IFLA_GROUP]) {
                netif_set_group(dev, nla_get_u32(tb[IFLA_GROUP]));
                status |= DO_SETLINK_NOTIFY;
        }

        /*
         * Interface selected by interface index but interface
         * name provided implies that a name change has been
         * requested.
         */
        if (ifm->ifi_index > 0 && ifname[0]) {
                err = netif_change_name(dev, ifname);
                if (err < 0)
                        goto errout;
                status |= DO_SETLINK_MODIFIED;
        }

        if (tb[IFLA_IFALIAS]) {
                err = netif_set_alias(dev, nla_data(tb[IFLA_IFALIAS]),
                                      nla_len(tb[IFLA_IFALIAS]));
                if (err < 0)
                        goto errout;
                status |= DO_SETLINK_NOTIFY;
        }

        if (tb[IFLA_BROADCAST]) {
                nla_memcpy(dev->broadcast, tb[IFLA_BROADCAST], dev->addr_len);
                call_netdevice_notifiers(NETDEV_CHANGEADDR, dev);
        }

        if (ifm->ifi_flags || ifm->ifi_change) {
                err = netif_change_flags(dev, rtnl_dev_combine_flags(dev, ifm),
                                         extack);
                if (err < 0)
                        goto errout;
        }

        if (tb[IFLA_MASTER]) {
                err = do_set_master(dev, nla_get_u32(tb[IFLA_MASTER]), extack);
                if (err)
                        goto errout;
                status |= DO_SETLINK_MODIFIED;
        }

        if (tb[IFLA_CARRIER]) {
                err = netif_change_carrier(dev, nla_get_u8(tb[IFLA_CARRIER]));
                if (err)
                        goto errout;
                status |= DO_SETLINK_MODIFIED;
        }

        if (tb[IFLA_TXQLEN]) {
                unsigned int value = nla_get_u32(tb[IFLA_TXQLEN]);

                err = netif_change_tx_queue_len(dev, value);
                if (err)
                        goto errout;
                status |= DO_SETLINK_MODIFIED;
        }

        if (tb[IFLA_GSO_MAX_SIZE]) {
                u32 max_size = nla_get_u32(tb[IFLA_GSO_MAX_SIZE]);

                if (dev->gso_max_size ^ max_size) {
                        netif_set_gso_max_size(dev, max_size);
                        status |= DO_SETLINK_MODIFIED;
                }
        }

        if (tb[IFLA_GSO_MAX_SEGS]) {
                u32 max_segs = nla_get_u32(tb[IFLA_GSO_MAX_SEGS]);

                if (dev->gso_max_segs ^ max_segs) {
                        netif_set_gso_max_segs(dev, max_segs);
                        status |= DO_SETLINK_MODIFIED;
                }
        }

        if (tb[IFLA_GRO_MAX_SIZE]) {
                u32 gro_max_size = nla_get_u32(tb[IFLA_GRO_MAX_SIZE]);

                if (dev->gro_max_size ^ gro_max_size) {
                        netif_set_gro_max_size(dev, gro_max_size);
                        status |= DO_SETLINK_MODIFIED;
                }
        }

        if (tb[IFLA_GSO_IPV4_MAX_SIZE]) {
                u32 max_size = nla_get_u32(tb[IFLA_GSO_IPV4_MAX_SIZE]);

                if (dev->gso_ipv4_max_size ^ max_size) {
                        netif_set_gso_ipv4_max_size(dev, max_size);
                        status |= DO_SETLINK_MODIFIED;
                }
        }

        if (tb[IFLA_GRO_IPV4_MAX_SIZE]) {
                u32 gro_max_size = nla_get_u32(tb[IFLA_GRO_IPV4_MAX_SIZE]);

                if (dev->gro_ipv4_max_size ^ gro_max_size) {
                        netif_set_gro_ipv4_max_size(dev, gro_max_size);
                        status |= DO_SETLINK_MODIFIED;
                }
        }

        if (tb[IFLA_OPERSTATE])
                set_operstate(dev, nla_get_u8(tb[IFLA_OPERSTATE]));

        if (tb[IFLA_LINKMODE]) {
                unsigned char value = nla_get_u8(tb[IFLA_LINKMODE]);

                if (dev->link_mode ^ value)
                        status |= DO_SETLINK_NOTIFY;
                WRITE_ONCE(dev->link_mode, value);
        }

        if (tb[IFLA_VFINFO_LIST]) {
                struct nlattr *vfinfo[IFLA_VF_MAX + 1];
                struct nlattr *attr;
                int rem;

                nla_for_each_nested(attr, tb[IFLA_VFINFO_LIST], rem) {
                        if (nla_type(attr) != IFLA_VF_INFO ||
                            nla_len(attr) < NLA_HDRLEN) {
                                err = -EINVAL;
                                goto errout;
                        }
                        err = nla_parse_nested_deprecated(vfinfo, IFLA_VF_MAX,
                                                          attr,
                                                          ifla_vf_policy,
                                                          NULL);
                        if (err < 0)
                                goto errout;
                        err = do_setvfinfo(dev, vfinfo);
                        if (err < 0)
                                goto errout;
                        status |= DO_SETLINK_NOTIFY;
                }
        }
        err = 0;

        if (tb[IFLA_VF_PORTS]) {
                struct nlattr *port[IFLA_PORT_MAX+1];
                struct nlattr *attr;
                int vf;
                int rem;

                err = -EOPNOTSUPP;
                if (!ops->ndo_set_vf_port)
                        goto errout;

                nla_for_each_nested(attr, tb[IFLA_VF_PORTS], rem) {
                        if (nla_type(attr) != IFLA_VF_PORT ||
                            nla_len(attr) < NLA_HDRLEN) {
                                err = -EINVAL;
                                goto errout;
                        }
                        err = nla_parse_nested_deprecated(port, IFLA_PORT_MAX,
                                                          attr,
                                                          ifla_port_policy,
                                                          NULL);
                        if (err < 0)
                                goto errout;
                        if (!port[IFLA_PORT_VF]) {
                                err = -EOPNOTSUPP;
                                goto errout;
                        }
                        vf = nla_get_u32(port[IFLA_PORT_VF]);
                        err = ops->ndo_set_vf_port(dev, vf, port);
                        if (err < 0)
                                goto errout;
                        status |= DO_SETLINK_NOTIFY;
                }
        }
        err = 0;

        if (tb[IFLA_PORT_SELF]) {
                struct nlattr *port[IFLA_PORT_MAX+1];

                err = nla_parse_nested_deprecated(port, IFLA_PORT_MAX,
                                                  tb[IFLA_PORT_SELF],
                                                  ifla_port_policy, NULL);
                if (err < 0)
                        goto errout;

                err = -EOPNOTSUPP;
                if (ops->ndo_set_vf_port)
                        err = ops->ndo_set_vf_port(dev, PORT_SELF_VF, port);
                if (err < 0)
                        goto errout;
                status |= DO_SETLINK_NOTIFY;
        }

        if (tb[IFLA_AF_SPEC]) {
                struct nlattr *af;
                int rem;

                nla_for_each_nested(af, tb[IFLA_AF_SPEC], rem) {
                        struct rtnl_af_ops *af_ops;
                        int af_ops_srcu_index;

                        af_ops = rtnl_af_lookup(nla_type(af), &af_ops_srcu_index);
                        if (!af_ops) {
                                err = -EAFNOSUPPORT;
                                goto errout;
                        }

                        err = af_ops->set_link_af(dev, af, extack);
                        rtnl_af_put(af_ops, af_ops_srcu_index);

                        if (err < 0)
                                goto errout;

                        status |= DO_SETLINK_NOTIFY;
                }
        }
        err = 0;

        if (tb[IFLA_PROTO_DOWN] || tb[IFLA_PROTO_DOWN_REASON]) {
                err = do_set_proto_down(dev, tb[IFLA_PROTO_DOWN],
                                        tb[IFLA_PROTO_DOWN_REASON], extack);
                if (err)
                        goto errout;
                status |= DO_SETLINK_NOTIFY;
        }

        if (tb[IFLA_XDP]) {
                struct nlattr *xdp[IFLA_XDP_MAX + 1];
                u32 xdp_flags = 0;

                err = nla_parse_nested_deprecated(xdp, IFLA_XDP_MAX,
                                                  tb[IFLA_XDP],
                                                  ifla_xdp_policy, NULL);
                if (err < 0)
                        goto errout;

                if (xdp[IFLA_XDP_ATTACHED] || xdp[IFLA_XDP_PROG_ID]) {
                        err = -EINVAL;
                        goto errout;
                }

                if (xdp[IFLA_XDP_FLAGS]) {
                        xdp_flags = nla_get_u32(xdp[IFLA_XDP_FLAGS]);
                        if (xdp_flags & ~XDP_FLAGS_MASK) {
                                err = -EINVAL;
                                goto errout;
                        }
                        if (hweight32(xdp_flags & XDP_FLAGS_MODES) > 1) {
                                err = -EINVAL;
                                goto errout;
                        }
                }

                if (xdp[IFLA_XDP_FD]) {
                        int expected_fd = -1;

                        if (xdp_flags & XDP_FLAGS_REPLACE) {
                                if (!xdp[IFLA_XDP_EXPECTED_FD]) {
                                        err = -EINVAL;
                                        goto errout;
                                }
                                expected_fd =
                                        nla_get_s32(xdp[IFLA_XDP_EXPECTED_FD]);
                        }

                        err = dev_change_xdp_fd(dev, extack,
                                                nla_get_s32(xdp[IFLA_XDP_FD]),
                                                expected_fd,
                                                xdp_flags);
                        if (err)
                                goto errout;
                        status |= DO_SETLINK_NOTIFY;
                }
        }

errout:
        if (status & DO_SETLINK_MODIFIED) {
                if ((status & DO_SETLINK_NOTIFY) == DO_SETLINK_NOTIFY)
                        netif_state_change(dev);

                if (err < 0)
                        net_warn_ratelimited("A link change request failed with some changes committed already. Interface %s may have been left with an inconsistent configuration, please check.\n",
                                             dev->name);
        }

        netdev_unlock_ops(dev);

        return err;
}

static struct net_device *rtnl_dev_get(struct net *net,
                                       struct nlattr *tb[])
{
        char ifname[ALTIFNAMSIZ];

        if (tb[IFLA_IFNAME])
                nla_strscpy(ifname, tb[IFLA_IFNAME], IFNAMSIZ);
        else if (tb[IFLA_ALT_IFNAME])
                nla_strscpy(ifname, tb[IFLA_ALT_IFNAME], ALTIFNAMSIZ);
        else
                return NULL;

        return __dev_get_by_name(net, ifname);
}

static int rtnl_setlink(struct sk_buff *skb, struct nlmsghdr *nlh,
                        struct netlink_ext_ack *extack)
{
        struct ifinfomsg *ifm = nlmsg_data(nlh);
        struct net *net = sock_net(skb->sk);
        struct nlattr *tb[IFLA_MAX+1];
        struct net_device *dev = NULL;
        struct rtnl_nets rtnl_nets;
        struct net *tgt_net;
        int err;

        err = nlmsg_parse_deprecated(nlh, sizeof(*ifm), tb, IFLA_MAX,
                                     ifla_policy, extack);
        if (err < 0)
                goto errout;

        err = rtnl_ensure_unique_netns(tb, extack, false);
        if (err < 0)
                goto errout;

        tgt_net = rtnl_link_get_net_capable(skb, net, tb, CAP_NET_ADMIN);
        if (IS_ERR(tgt_net)) {
                err = PTR_ERR(tgt_net);
                goto errout;
        }

        rtnl_nets_init(&rtnl_nets);
        rtnl_nets_add(&rtnl_nets, get_net(net));
        rtnl_nets_add(&rtnl_nets, tgt_net);

        rtnl_nets_lock(&rtnl_nets);

        if (ifm->ifi_index > 0)
                dev = __dev_get_by_index(net, ifm->ifi_index);
        else if (tb[IFLA_IFNAME] || tb[IFLA_ALT_IFNAME])
                dev = rtnl_dev_get(net, tb);
        else
                err = -EINVAL;

        if (dev)
                err = do_setlink(skb, dev, tgt_net, ifm, extack, tb, 0);
        else if (!err)
                err = -ENODEV;

        rtnl_nets_unlock(&rtnl_nets);
        rtnl_nets_destroy(&rtnl_nets);
errout:
        return err;
}

static int rtnl_group_dellink(const struct net *net, int group)
{
        struct net_device *dev, *aux;
        LIST_HEAD(list_kill);
        bool found = false;

        if (!group)
                return -EPERM;

        for_each_netdev(net, dev) {
                if (dev->group == group) {
                        const struct rtnl_link_ops *ops;

                        found = true;
                        ops = dev->rtnl_link_ops;
                        if (!ops || !ops->dellink)
                                return -EOPNOTSUPP;
                }
        }

        if (!found)
                return -ENODEV;

        for_each_netdev_safe(net, dev, aux) {
                if (dev->group == group) {
                        const struct rtnl_link_ops *ops;

                        ops = dev->rtnl_link_ops;
                        ops->dellink(dev, &list_kill);
                }
        }
        unregister_netdevice_many(&list_kill);

        return 0;
}

int rtnl_delete_link(struct net_device *dev, u32 portid, const struct nlmsghdr *nlh)
{
        const struct rtnl_link_ops *ops;
        LIST_HEAD(list_kill);

        ops = dev->rtnl_link_ops;
        if (!ops || !ops->dellink)
                return -EOPNOTSUPP;

        ops->dellink(dev, &list_kill);
        unregister_netdevice_many_notify(&list_kill, portid, nlh);

        return 0;
}
EXPORT_SYMBOL_GPL(rtnl_delete_link);

static int rtnl_dellink(struct sk_buff *skb, struct nlmsghdr *nlh,
                        struct netlink_ext_ack *extack)
{
        struct ifinfomsg *ifm = nlmsg_data(nlh);
        struct net *net = sock_net(skb->sk);
        u32 portid = NETLINK_CB(skb).portid;
        struct nlattr *tb[IFLA_MAX+1];
        struct net_device *dev = NULL;
        struct net *tgt_net = net;
        int netnsid = -1;
        int err;

        err = nlmsg_parse_deprecated(nlh, sizeof(*ifm), tb, IFLA_MAX,
                                     ifla_policy, extack);
        if (err < 0)
                return err;

        err = rtnl_ensure_unique_netns(tb, extack, true);
        if (err < 0)
                return err;

        if (tb[IFLA_TARGET_NETNSID]) {
                netnsid = nla_get_s32(tb[IFLA_TARGET_NETNSID]);
                tgt_net = rtnl_get_net_ns_capable(NETLINK_CB(skb).sk, netnsid);
                if (IS_ERR(tgt_net))
                        return PTR_ERR(tgt_net);
        }

        rtnl_net_lock(tgt_net);

        if (ifm->ifi_index > 0)
                dev = __dev_get_by_index(tgt_net, ifm->ifi_index);
        else if (tb[IFLA_IFNAME] || tb[IFLA_ALT_IFNAME])
                dev = rtnl_dev_get(tgt_net, tb);

        if (dev)
                err = rtnl_delete_link(dev, portid, nlh);
        else if (ifm->ifi_index > 0 || tb[IFLA_IFNAME] || tb[IFLA_ALT_IFNAME])
                err = -ENODEV;
        else if (tb[IFLA_GROUP])
                err = rtnl_group_dellink(tgt_net, nla_get_u32(tb[IFLA_GROUP]));
        else
                err = -EINVAL;

        rtnl_net_unlock(tgt_net);

        if (netnsid >= 0)
                put_net(tgt_net);

        return err;
}

int rtnl_configure_link(struct net_device *dev, const struct ifinfomsg *ifm,
                        u32 portid, const struct nlmsghdr *nlh)
{
        unsigned int old_flags, changed;
        int err;

        old_flags = dev->flags;
        if (ifm && (ifm->ifi_flags || ifm->ifi_change)) {
                err = __dev_change_flags(dev, rtnl_dev_combine_flags(dev, ifm),
                                         NULL);
                if (err < 0)
                        return err;
        }

        changed = old_flags ^ dev->flags;
        if (dev->rtnl_link_initializing) {
                dev->rtnl_link_initializing = false;
                changed = ~0U;
        }

        __dev_notify_flags(dev, old_flags, changed, portid, nlh);
        return 0;
}
EXPORT_SYMBOL(rtnl_configure_link);

struct net_device *rtnl_create_link(struct net *net, const char *ifname,
                                    unsigned char name_assign_type,
                                    const struct rtnl_link_ops *ops,
                                    struct nlattr *tb[],
                                    struct netlink_ext_ack *extack)
{
        struct net_device *dev;
        unsigned int num_tx_queues = 1;
        unsigned int num_rx_queues = 1;
        int err;

        if (tb[IFLA_NUM_TX_QUEUES])
                num_tx_queues = nla_get_u32(tb[IFLA_NUM_TX_QUEUES]);
        else if (ops->get_num_tx_queues)
                num_tx_queues = ops->get_num_tx_queues();

        if (tb[IFLA_NUM_RX_QUEUES])
                num_rx_queues = nla_get_u32(tb[IFLA_NUM_RX_QUEUES]);
        else if (ops->get_num_rx_queues)
                num_rx_queues = ops->get_num_rx_queues();

        if (num_tx_queues < 1 || num_tx_queues > 4096) {
                NL_SET_ERR_MSG(extack, "Invalid number of transmit queues");
                return ERR_PTR(-EINVAL);
        }

        if (num_rx_queues < 1 || num_rx_queues > 4096) {
                NL_SET_ERR_MSG(extack, "Invalid number of receive queues");
                return ERR_PTR(-EINVAL);
        }

        if (ops->alloc) {
                dev = ops->alloc(tb, ifname, name_assign_type,
                                 num_tx_queues, num_rx_queues);
                if (IS_ERR(dev))
                        return dev;
        } else {
                dev = alloc_netdev_mqs(ops->priv_size, ifname,
                                       name_assign_type, ops->setup,
                                       num_tx_queues, num_rx_queues);
        }

        if (!dev)
                return ERR_PTR(-ENOMEM);

        err = validate_linkmsg(dev, tb, extack);
        if (err < 0) {
                free_netdev(dev);
                return ERR_PTR(err);
        }

        dev_net_set(dev, net);
        dev->rtnl_link_ops = ops;
        dev->rtnl_link_initializing = true;

        if (tb[IFLA_MTU]) {
                u32 mtu = nla_get_u32(tb[IFLA_MTU]);

                err = dev_validate_mtu(dev, mtu, extack);
                if (err) {
                        free_netdev(dev);
                        return ERR_PTR(err);
                }
                dev->mtu = mtu;
        }
        if (tb[IFLA_ADDRESS]) {
                __dev_addr_set(dev, nla_data(tb[IFLA_ADDRESS]),
                               nla_len(tb[IFLA_ADDRESS]));
                dev->addr_assign_type = NET_ADDR_SET;
        }
        if (tb[IFLA_BROADCAST])
                memcpy(dev->broadcast, nla_data(tb[IFLA_BROADCAST]),
                                nla_len(tb[IFLA_BROADCAST]));
        if (tb[IFLA_TXQLEN])
                dev->tx_queue_len = nla_get_u32(tb[IFLA_TXQLEN]);
        if (tb[IFLA_OPERSTATE])
                set_operstate(dev, nla_get_u8(tb[IFLA_OPERSTATE]));
        if (tb[IFLA_LINKMODE])
                dev->link_mode = nla_get_u8(tb[IFLA_LINKMODE]);
        if (tb[IFLA_GROUP])
                netif_set_group(dev, nla_get_u32(tb[IFLA_GROUP]));
        if (tb[IFLA_GSO_MAX_SIZE])
                netif_set_gso_max_size(dev, nla_get_u32(tb[IFLA_GSO_MAX_SIZE]));
        if (tb[IFLA_GSO_MAX_SEGS])
                netif_set_gso_max_segs(dev, nla_get_u32(tb[IFLA_GSO_MAX_SEGS]));
        if (tb[IFLA_GRO_MAX_SIZE])
                netif_set_gro_max_size(dev, nla_get_u32(tb[IFLA_GRO_MAX_SIZE]));
        if (tb[IFLA_GSO_IPV4_MAX_SIZE])
                netif_set_gso_ipv4_max_size(dev, nla_get_u32(tb[IFLA_GSO_IPV4_MAX_SIZE]));
        if (tb[IFLA_GRO_IPV4_MAX_SIZE])
                netif_set_gro_ipv4_max_size(dev, nla_get_u32(tb[IFLA_GRO_IPV4_MAX_SIZE]));

        return dev;
}
EXPORT_SYMBOL(rtnl_create_link);

struct rtnl_newlink_tbs {
        struct nlattr *tb[IFLA_MAX + 1];
        struct nlattr *linkinfo[IFLA_INFO_MAX + 1];
        struct nlattr *attr[RTNL_MAX_TYPE + 1];
        struct nlattr *slave_attr[RTNL_SLAVE_MAX_TYPE + 1];
};

static int rtnl_changelink(const struct sk_buff *skb, struct nlmsghdr *nlh,
                           const struct rtnl_link_ops *ops,
                           struct net_device *dev, struct net *tgt_net,
                           struct rtnl_newlink_tbs *tbs,
                           struct nlattr **data,
                           struct netlink_ext_ack *extack)
{
        struct nlattr ** const linkinfo = tbs->linkinfo;
        struct nlattr ** const tb = tbs->tb;
        int status = 0;
        int err;

        if (nlh->nlmsg_flags & NLM_F_EXCL)
                return -EEXIST;

        if (nlh->nlmsg_flags & NLM_F_REPLACE)
                return -EOPNOTSUPP;

        if (linkinfo[IFLA_INFO_DATA]) {
                if (!ops || ops != dev->rtnl_link_ops || !ops->changelink)
                        return -EOPNOTSUPP;

                err = ops->changelink(dev, tb, data, extack);
                if (err < 0)
                        return err;

                status |= DO_SETLINK_NOTIFY;
        }

        if (linkinfo[IFLA_INFO_SLAVE_DATA]) {
                const struct rtnl_link_ops *m_ops = NULL;
                struct nlattr **slave_data = NULL;
                struct net_device *master_dev;

                master_dev = netdev_master_upper_dev_get(dev);
                if (master_dev)
                        m_ops = master_dev->rtnl_link_ops;

                if (!m_ops || !m_ops->slave_changelink)
                        return -EOPNOTSUPP;

                if (m_ops->slave_maxtype > RTNL_SLAVE_MAX_TYPE)
                        return -EINVAL;

                if (m_ops->slave_maxtype) {
                        err = nla_parse_nested_deprecated(tbs->slave_attr,
                                                          m_ops->slave_maxtype,
                                                          linkinfo[IFLA_INFO_SLAVE_DATA],
                                                          m_ops->slave_policy, extack);
                        if (err < 0)
                                return err;

                        slave_data = tbs->slave_attr;
                }

                err = m_ops->slave_changelink(master_dev, dev, tb, slave_data, extack);
                if (err < 0)
                        return err;

                status |= DO_SETLINK_NOTIFY;
        }

        return do_setlink(skb, dev, tgt_net, nlmsg_data(nlh), extack, tb, status);
}

static int rtnl_group_changelink(const struct sk_buff *skb,
                                 struct net *net, struct net *tgt_net,
                                 int group, struct ifinfomsg *ifm,
                                 struct netlink_ext_ack *extack,
                                 struct nlattr **tb)
{
        struct net_device *dev, *aux;
        int err;

        for_each_netdev_safe(net, dev, aux) {
                if (dev->group == group) {
                        err = do_setlink(skb, dev, tgt_net, ifm, extack, tb, 0);
                        if (err < 0)
                                return err;
                }
        }

        return 0;
}

static int rtnl_newlink_create(struct sk_buff *skb, struct ifinfomsg *ifm,
                               const struct rtnl_link_ops *ops,
                               struct net *tgt_net, struct net *link_net,
                               struct net *peer_net,
                               const struct nlmsghdr *nlh,
                               struct nlattr **tb, struct nlattr **data,
                               struct netlink_ext_ack *extack)
{
        unsigned char name_assign_type = NET_NAME_USER;
        struct rtnl_newlink_params params = {
                .src_net = sock_net(skb->sk),
                .link_net = link_net,
                .peer_net = peer_net,
                .tb = tb,
                .data = data,
        };
        u32 portid = NETLINK_CB(skb).portid;
        struct net_device *dev;
        char ifname[IFNAMSIZ];
        int err;

        if (!ops->alloc && !ops->setup)
                return -EOPNOTSUPP;

        if (tb[IFLA_IFNAME]) {
                nla_strscpy(ifname, tb[IFLA_IFNAME], IFNAMSIZ);
        } else {
                snprintf(ifname, IFNAMSIZ, "%s%%d", ops->kind);
                name_assign_type = NET_NAME_ENUM;
        }

        dev = rtnl_create_link(tgt_net, ifname, name_assign_type, ops, tb,
                               extack);
        if (IS_ERR(dev)) {
                err = PTR_ERR(dev);
                goto out;
        }

        dev->ifindex = ifm->ifi_index;

        if (ops->newlink)
                err = ops->newlink(dev, &params, extack);
        else
                err = register_netdevice(dev);
        if (err < 0) {
                free_netdev(dev);
                goto out;
        }

        netdev_lock_ops(dev);

        err = rtnl_configure_link(dev, ifm, portid, nlh);
        if (err < 0)
                goto out_unregister;
        if (tb[IFLA_MASTER]) {
                err = do_set_master(dev, nla_get_u32(tb[IFLA_MASTER]), extack);
                if (err)
                        goto out_unregister;
        }

        netdev_unlock_ops(dev);
out:
        return err;
out_unregister:
        netdev_unlock_ops(dev);
        if (ops->newlink) {
                LIST_HEAD(list_kill);

                ops->dellink(dev, &list_kill);
                unregister_netdevice_many(&list_kill);
        } else {
                unregister_netdevice(dev);
        }
        goto out;
}

static struct net *rtnl_get_peer_net(const struct rtnl_link_ops *ops,
                                     struct nlattr *tbp[],
                                     struct nlattr *data[],
                                     struct netlink_ext_ack *extack)
{
        struct nlattr *tb[IFLA_MAX + 1];
        int err;

        if (!data || !data[ops->peer_type])
                return rtnl_link_get_net_ifla(tbp);

        err = rtnl_nla_parse_ifinfomsg(tb, data[ops->peer_type], extack);
        if (err < 0)
                return ERR_PTR(err);

        if (ops->validate) {
                err = ops->validate(tb, NULL, extack);
                if (err < 0)
                        return ERR_PTR(err);
        }

        return rtnl_link_get_net_ifla(tb);
}

static int __rtnl_newlink(struct sk_buff *skb, struct nlmsghdr *nlh,
                          const struct rtnl_link_ops *ops,
                          struct net *tgt_net, struct net *link_net,
                          struct net *peer_net,
                          struct rtnl_newlink_tbs *tbs,
                          struct nlattr **data,
                          struct netlink_ext_ack *extack)
{
        struct nlattr ** const tb = tbs->tb;
        struct net *net = sock_net(skb->sk);
        struct net *device_net;
        struct net_device *dev;
        struct ifinfomsg *ifm;
        bool link_specified;

        /* When creating, lookup for existing device in target net namespace */
        device_net = (nlh->nlmsg_flags & NLM_F_CREATE) &&
                     (nlh->nlmsg_flags & NLM_F_EXCL) ?
                     tgt_net : net;

        ifm = nlmsg_data(nlh);
        if (ifm->ifi_index > 0) {
                link_specified = true;
                dev = __dev_get_by_index(device_net, ifm->ifi_index);
        } else if (ifm->ifi_index < 0) {
                NL_SET_ERR_MSG(extack, "ifindex can't be negative");
                return -EINVAL;
        } else if (tb[IFLA_IFNAME] || tb[IFLA_ALT_IFNAME]) {
                link_specified = true;
                dev = rtnl_dev_get(device_net, tb);
        } else {
                link_specified = false;
                dev = NULL;
        }

        if (dev)
                return rtnl_changelink(skb, nlh, ops, dev, tgt_net, tbs, data, extack);

        if (!(nlh->nlmsg_flags & NLM_F_CREATE)) {
                /* No dev found and NLM_F_CREATE not set. Requested dev does not exist,
                 * or it's for a group
                */
                if (link_specified || !tb[IFLA_GROUP])
                        return -ENODEV;

                return rtnl_group_changelink(skb, net, tgt_net,
                                             nla_get_u32(tb[IFLA_GROUP]),
                                             ifm, extack, tb);
        }

        if (tb[IFLA_MAP] || tb[IFLA_PROTINFO])
                return -EOPNOTSUPP;

        if (!ops) {
                NL_SET_ERR_MSG(extack, "Unknown device type");
                return -EOPNOTSUPP;
        }

        return rtnl_newlink_create(skb, ifm, ops, tgt_net, link_net, peer_net, nlh,
                                   tb, data, extack);
}

static int rtnl_newlink(struct sk_buff *skb, struct nlmsghdr *nlh,
                        struct netlink_ext_ack *extack)
{
        struct net *tgt_net, *link_net = NULL, *peer_net = NULL;
        struct nlattr **tb, **linkinfo, **data = NULL;
        struct rtnl_link_ops *ops = NULL;
        struct rtnl_newlink_tbs *tbs;
        struct rtnl_nets rtnl_nets;
        int ops_srcu_index;
        int ret;

        tbs = kmalloc_obj(*tbs);
        if (!tbs)
                return -ENOMEM;

        tb = tbs->tb;
        ret = nlmsg_parse_deprecated(nlh, sizeof(struct ifinfomsg), tb,
                                     IFLA_MAX, ifla_policy, extack);
        if (ret < 0)
                goto free;

        ret = rtnl_ensure_unique_netns(tb, extack, false);
        if (ret < 0)
                goto free;

        linkinfo = tbs->linkinfo;
        if (tb[IFLA_LINKINFO]) {
                ret = nla_parse_nested_deprecated(linkinfo, IFLA_INFO_MAX,
                                                  tb[IFLA_LINKINFO],
                                                  ifla_info_policy, NULL);
                if (ret < 0)
                        goto free;
        } else {
                memset(linkinfo, 0, sizeof(tbs->linkinfo));
        }

        if (linkinfo[IFLA_INFO_KIND]) {
                char kind[MODULE_NAME_LEN];

                nla_strscpy(kind, linkinfo[IFLA_INFO_KIND], sizeof(kind));
                ops = rtnl_link_ops_get(kind, &ops_srcu_index);
#ifdef CONFIG_MODULES
                if (!ops) {
                        request_module("rtnl-link-%s", kind);
                        ops = rtnl_link_ops_get(kind, &ops_srcu_index);
                }
#endif
        }

        rtnl_nets_init(&rtnl_nets);

        if (ops) {
                if (ops->maxtype > RTNL_MAX_TYPE) {
                        ret = -EINVAL;
                        goto put_ops;
                }

                if (ops->maxtype && linkinfo[IFLA_INFO_DATA]) {
                        ret = nla_parse_nested_deprecated(tbs->attr, ops->maxtype,
                                                          linkinfo[IFLA_INFO_DATA],
                                                          ops->policy, extack);
                        if (ret < 0)
                                goto put_ops;

                        data = tbs->attr;
                }

                if (ops->validate) {
                        ret = ops->validate(tb, data, extack);
                        if (ret < 0)
                                goto put_ops;
                }

                if (ops->peer_type) {
                        peer_net = rtnl_get_peer_net(ops, tb, data, extack);
                        if (IS_ERR(peer_net)) {
                                ret = PTR_ERR(peer_net);
                                goto put_ops;
                        }
                        if (peer_net)
                                rtnl_nets_add(&rtnl_nets, peer_net);
                }
        }

        tgt_net = rtnl_link_get_net_capable(skb, sock_net(skb->sk), tb, CAP_NET_ADMIN);
        if (IS_ERR(tgt_net)) {
                ret = PTR_ERR(tgt_net);
                goto put_net;
        }

        rtnl_nets_add(&rtnl_nets, tgt_net);

        if (tb[IFLA_LINK_NETNSID]) {
                int id = nla_get_s32(tb[IFLA_LINK_NETNSID]);

                link_net = get_net_ns_by_id(tgt_net, id);
                if (!link_net) {
                        NL_SET_ERR_MSG(extack, "Unknown network namespace id");
                        ret =  -EINVAL;
                        goto put_net;
                }

                rtnl_nets_add(&rtnl_nets, link_net);

                if (!netlink_ns_capable(skb, link_net->user_ns, CAP_NET_ADMIN)) {
                        ret = -EPERM;
                        goto put_net;
                }
        }

        rtnl_nets_lock(&rtnl_nets);
        ret = __rtnl_newlink(skb, nlh, ops, tgt_net, link_net, peer_net, tbs, data, extack);
        rtnl_nets_unlock(&rtnl_nets);

put_net:
        rtnl_nets_destroy(&rtnl_nets);
put_ops:
        if (ops)
                rtnl_link_ops_put(ops, ops_srcu_index);
free:
        kfree(tbs);
        return ret;
}

static int rtnl_valid_getlink_req(struct sk_buff *skb,
                                  const struct nlmsghdr *nlh,
                                  struct nlattr **tb,
                                  struct netlink_ext_ack *extack)
{
        struct ifinfomsg *ifm;
        int i, err;

        ifm = nlmsg_payload(nlh, sizeof(*ifm));
        if (!ifm) {
                NL_SET_ERR_MSG(extack, "Invalid header for get link");
                return -EINVAL;
        }

        if (!netlink_strict_get_check(skb))
                return nlmsg_parse_deprecated(nlh, sizeof(*ifm), tb, IFLA_MAX,
                                              ifla_policy, extack);

        if (ifm->__ifi_pad || ifm->ifi_type || ifm->ifi_flags ||
            ifm->ifi_change) {
                NL_SET_ERR_MSG(extack, "Invalid values in header for get link request");
                return -EINVAL;
        }

        err = nlmsg_parse_deprecated_strict(nlh, sizeof(*ifm), tb, IFLA_MAX,
                                            ifla_policy, extack);
        if (err)
                return err;

        for (i = 0; i <= IFLA_MAX; i++) {
                if (!tb[i])
                        continue;

                switch (i) {
                case IFLA_IFNAME:
                case IFLA_ALT_IFNAME:
                case IFLA_EXT_MASK:
                case IFLA_TARGET_NETNSID:
                        break;
                default:
                        NL_SET_ERR_MSG(extack, "Unsupported attribute in get link request");
                        return -EINVAL;
                }
        }

        return 0;
}

static int rtnl_getlink(struct sk_buff *skb, struct nlmsghdr *nlh,
                        struct netlink_ext_ack *extack)
{
        struct net *net = sock_net(skb->sk);
        struct net *tgt_net = net;
        struct ifinfomsg *ifm;
        struct nlattr *tb[IFLA_MAX+1];
        struct net_device *dev = NULL;
        struct sk_buff *nskb;
        int netnsid = -1;
        int err;
        u32 ext_filter_mask = 0;

        err = rtnl_valid_getlink_req(skb, nlh, tb, extack);
        if (err < 0)
                return err;

        err = rtnl_ensure_unique_netns(tb, extack, true);
        if (err < 0)
                return err;

        if (tb[IFLA_TARGET_NETNSID]) {
                netnsid = nla_get_s32(tb[IFLA_TARGET_NETNSID]);
                tgt_net = rtnl_get_net_ns_capable(NETLINK_CB(skb).sk, netnsid);
                if (IS_ERR(tgt_net))
                        return PTR_ERR(tgt_net);
        }

        if (tb[IFLA_EXT_MASK])
                ext_filter_mask = nla_get_u32(tb[IFLA_EXT_MASK]);

        err = -EINVAL;
        ifm = nlmsg_data(nlh);
        if (ifm->ifi_index > 0)
                dev = __dev_get_by_index(tgt_net, ifm->ifi_index);
        else if (tb[IFLA_IFNAME] || tb[IFLA_ALT_IFNAME])
                dev = rtnl_dev_get(tgt_net, tb);
        else
                goto out;

        err = -ENODEV;
        if (dev == NULL)
                goto out;

        err = -ENOBUFS;
        nskb = nlmsg_new_large(if_nlmsg_size(dev, ext_filter_mask));
        if (nskb == NULL)
                goto out;

        /* Synchronize the carrier state so we don't report a state
         * that we're not actually going to honour immediately; if
         * the driver just did a carrier off->on transition, we can
         * only TX if link watch work has run, but without this we'd
         * already report carrier on, even if it doesn't work yet.
         */
        linkwatch_sync_dev(dev);

        err = rtnl_fill_ifinfo(nskb, dev, net,
                               RTM_NEWLINK, NETLINK_CB(skb).portid,
                               nlh->nlmsg_seq, 0, 0, ext_filter_mask,
                               0, NULL, 0, netnsid, GFP_KERNEL);
        if (err < 0) {
                /* -EMSGSIZE implies BUG in if_nlmsg_size */
                WARN_ON(err == -EMSGSIZE);
                kfree_skb(nskb);
        } else
                err = rtnl_unicast(nskb, net, NETLINK_CB(skb).portid);
out:
        if (netnsid >= 0)
                put_net(tgt_net);

        return err;
}

static int rtnl_alt_ifname(int cmd, struct net_device *dev, struct nlattr *attr,
                           bool *changed, struct netlink_ext_ack *extack)
{
        char *alt_ifname;
        size_t size;
        int err;

        err = nla_validate(attr, attr->nla_len, IFLA_MAX, ifla_policy, extack);
        if (err)
                return err;

        if (cmd == RTM_NEWLINKPROP) {
                size = rtnl_prop_list_size(dev);
                size += nla_total_size(ALTIFNAMSIZ);
                if (size >= U16_MAX) {
                        NL_SET_ERR_MSG(extack,
                                       "effective property list too long");
                        return -EINVAL;
                }
        }

        alt_ifname = nla_strdup(attr, GFP_KERNEL_ACCOUNT);
        if (!alt_ifname)
                return -ENOMEM;

        if (cmd == RTM_NEWLINKPROP) {
                err = netdev_name_node_alt_create(dev, alt_ifname);
                if (!err)
                        alt_ifname = NULL;
        } else if (cmd == RTM_DELLINKPROP) {
                err = netdev_name_node_alt_destroy(dev, alt_ifname);
        } else {
                WARN_ON_ONCE(1);
                err = -EINVAL;
        }

        kfree(alt_ifname);
        if (!err)
                *changed = true;
        return err;
}

static int rtnl_linkprop(int cmd, struct sk_buff *skb, struct nlmsghdr *nlh,
                         struct netlink_ext_ack *extack)
{
        struct net *net = sock_net(skb->sk);
        struct nlattr *tb[IFLA_MAX + 1];
        struct net_device *dev;
        struct ifinfomsg *ifm;
        bool changed = false;
        struct nlattr *attr;
        int err, rem;

        err = nlmsg_parse(nlh, sizeof(*ifm), tb, IFLA_MAX, ifla_policy, extack);
        if (err)
                return err;

        err = rtnl_ensure_unique_netns(tb, extack, true);
        if (err)
                return err;

        ifm = nlmsg_data(nlh);
        if (ifm->ifi_index > 0)
                dev = __dev_get_by_index(net, ifm->ifi_index);
        else if (tb[IFLA_IFNAME] || tb[IFLA_ALT_IFNAME])
                dev = rtnl_dev_get(net, tb);
        else
                return -EINVAL;

        if (!dev)
                return -ENODEV;

        if (!tb[IFLA_PROP_LIST])
                return 0;

        nla_for_each_nested(attr, tb[IFLA_PROP_LIST], rem) {
                switch (nla_type(attr)) {
                case IFLA_ALT_IFNAME:
                        err = rtnl_alt_ifname(cmd, dev, attr, &changed, extack);
                        if (err)
                                return err;
                        break;
                }
        }

        if (changed)
                netdev_state_change(dev);
        return 0;
}

static int rtnl_newlinkprop(struct sk_buff *skb, struct nlmsghdr *nlh,
                            struct netlink_ext_ack *extack)
{
        return rtnl_linkprop(RTM_NEWLINKPROP, skb, nlh, extack);
}

static int rtnl_dellinkprop(struct sk_buff *skb, struct nlmsghdr *nlh,
                            struct netlink_ext_ack *extack)
{
        return rtnl_linkprop(RTM_DELLINKPROP, skb, nlh, extack);
}

static noinline_for_stack u32 rtnl_calcit(struct sk_buff *skb,
                                          struct nlmsghdr *nlh)
{
        struct net *net = sock_net(skb->sk);
        size_t min_ifinfo_dump_size = 0;
        u32 ext_filter_mask = 0;
        struct net_device *dev;
        struct nlattr *nla;
        int hdrlen, rem;

        /* Same kernel<->userspace interface hack as in rtnl_dump_ifinfo. */
        hdrlen = nlmsg_len(nlh) < sizeof(struct ifinfomsg) ?
                 sizeof(struct rtgenmsg) : sizeof(struct ifinfomsg);

        if (nlh->nlmsg_len < nlmsg_msg_size(hdrlen))
                return NLMSG_GOODSIZE;

        nla_for_each_attr_type(nla, IFLA_EXT_MASK,
                               nlmsg_attrdata(nlh, hdrlen),
                               nlmsg_attrlen(nlh, hdrlen), rem) {
                if (nla_len(nla) == sizeof(u32))
                        ext_filter_mask = nla_get_u32(nla);
        }

        if (!ext_filter_mask)
                return NLMSG_GOODSIZE;
        /*
         * traverse the list of net devices and compute the minimum
         * buffer size based upon the filter mask.
         */
        rcu_read_lock();
        for_each_netdev_rcu(net, dev) {
                min_ifinfo_dump_size = max(min_ifinfo_dump_size,
                                           if_nlmsg_size(dev, ext_filter_mask));
        }
        rcu_read_unlock();

        return nlmsg_total_size(min_ifinfo_dump_size);
}

static int rtnl_dump_all(struct sk_buff *skb, struct netlink_callback *cb)
{
        int idx;
        int s_idx = cb->family;
        int type = cb->nlh->nlmsg_type - RTM_BASE;
        int ret = 0;

        if (s_idx == 0)
                s_idx = 1;

        for (idx = 1; idx <= RTNL_FAMILY_MAX; idx++) {
                struct rtnl_link __rcu **tab;
                struct rtnl_link *link;
                rtnl_dumpit_func dumpit;

                if (idx < s_idx || idx == PF_PACKET)
                        continue;

                if (type < 0 || type >= RTM_NR_MSGTYPES)
                        continue;

                tab = rcu_dereference_rtnl(rtnl_msg_handlers[idx]);
                if (!tab)
                        continue;

                link = rcu_dereference_rtnl(tab[type]);
                if (!link)
                        continue;

                dumpit = link->dumpit;
                if (!dumpit)
                        continue;

                if (idx > s_idx) {
                        memset(&cb->args[0], 0, sizeof(cb->args));
                        cb->prev_seq = 0;
                        cb->seq = 0;
                }
                ret = dumpit(skb, cb);
                if (ret)
                        break;
        }
        cb->family = idx;

        return skb->len ? : ret;
}

struct sk_buff *rtmsg_ifinfo_build_skb(int type, struct net_device *dev,
                                       unsigned int change,
                                       u32 event, gfp_t flags, int *new_nsid,
                                       int new_ifindex, u32 portid,
                                       const struct nlmsghdr *nlh)
{
        struct net *net = dev_net(dev);
        struct sk_buff *skb;
        int err = -ENOBUFS;
        u32 seq = 0;

        skb = nlmsg_new(if_nlmsg_size(dev, 0), flags);
        if (skb == NULL)
                goto errout;

        if (nlmsg_report(nlh))
                seq = nlmsg_seq(nlh);
        else
                portid = 0;

        err = rtnl_fill_ifinfo(skb, dev, dev_net(dev),
                               type, portid, seq, change, 0, 0, event,
                               new_nsid, new_ifindex, -1, flags);
        if (err < 0) {
                /* -EMSGSIZE implies BUG in if_nlmsg_size() */
                WARN_ON(err == -EMSGSIZE);
                kfree_skb(skb);
                goto errout;
        }
        return skb;
errout:
        rtnl_set_sk_err(net, RTNLGRP_LINK, err);
        return NULL;
}

void rtmsg_ifinfo_send(struct sk_buff *skb, struct net_device *dev, gfp_t flags,
                       u32 portid, const struct nlmsghdr *nlh)
{
        struct net *net = dev_net(dev);

        rtnl_notify(skb, net, portid, RTNLGRP_LINK, nlh, flags);
}

static void rtmsg_ifinfo_event(int type, struct net_device *dev,
                               unsigned int change, u32 event,
                               gfp_t flags, int *new_nsid, int new_ifindex,
                               u32 portid, const struct nlmsghdr *nlh)
{
        struct sk_buff *skb;

        if (dev->reg_state != NETREG_REGISTERED)
                return;

        skb = rtmsg_ifinfo_build_skb(type, dev, change, event, flags, new_nsid,
                                     new_ifindex, portid, nlh);
        if (skb)
                rtmsg_ifinfo_send(skb, dev, flags, portid, nlh);
}

void rtmsg_ifinfo(int type, struct net_device *dev, unsigned int change,
                  gfp_t flags, u32 portid, const struct nlmsghdr *nlh)
{
        rtmsg_ifinfo_event(type, dev, change, rtnl_get_event(0), flags,
                           NULL, 0, portid, nlh);
}

void rtmsg_ifinfo_newnet(int type, struct net_device *dev, unsigned int change,
                         gfp_t flags, int *new_nsid, int new_ifindex)
{
        rtmsg_ifinfo_event(type, dev, change, rtnl_get_event(0), flags,
                           new_nsid, new_ifindex, 0, NULL);
}

static int nlmsg_populate_fdb_fill(struct sk_buff *skb,
                                   struct net_device *dev,
                                   u8 *addr, u16 vid, u32 pid, u32 seq,
                                   int type, unsigned int flags,
                                   int nlflags, u16 ndm_state)
{
        struct nlmsghdr *nlh;
        struct ndmsg *ndm;

        nlh = nlmsg_put(skb, pid, seq, type, sizeof(*ndm), nlflags);
        if (!nlh)
                return -EMSGSIZE;

        ndm = nlmsg_data(nlh);
        ndm->ndm_family  = AF_BRIDGE;
        ndm->ndm_pad1    = 0;
        ndm->ndm_pad2    = 0;
        ndm->ndm_flags   = flags;
        ndm->ndm_type    = 0;
        ndm->ndm_ifindex = dev->ifindex;
        ndm->ndm_state   = ndm_state;

        if (nla_put(skb, NDA_LLADDR, dev->addr_len, addr))
                goto nla_put_failure;
        if (vid)
                if (nla_put(skb, NDA_VLAN, sizeof(u16), &vid))
                        goto nla_put_failure;

        nlmsg_end(skb, nlh);
        return 0;

nla_put_failure:
        nlmsg_cancel(skb, nlh);
        return -EMSGSIZE;
}

static inline size_t rtnl_fdb_nlmsg_size(const struct net_device *dev)
{
        return NLMSG_ALIGN(sizeof(struct ndmsg)) +
               nla_total_size(dev->addr_len) +  /* NDA_LLADDR */
               nla_total_size(sizeof(u16)) +    /* NDA_VLAN */
               0;
}

static void rtnl_fdb_notify(struct net_device *dev, u8 *addr, u16 vid, int type,
                            u16 ndm_state)
{
        struct net *net = dev_net(dev);
        struct sk_buff *skb;
        int err = -ENOBUFS;

        skb = nlmsg_new(rtnl_fdb_nlmsg_size(dev), GFP_ATOMIC);
        if (!skb)
                goto errout;

        err = nlmsg_populate_fdb_fill(skb, dev, addr, vid,
                                      0, 0, type, NTF_SELF, 0, ndm_state);
        if (err < 0) {
                kfree_skb(skb);
                goto errout;
        }

        rtnl_notify(skb, net, 0, RTNLGRP_NEIGH, NULL, GFP_ATOMIC);
        return;
errout:
        rtnl_set_sk_err(net, RTNLGRP_NEIGH, err);
}

/*
 * ndo_dflt_fdb_add - default netdevice operation to add an FDB entry
 */
int ndo_dflt_fdb_add(struct ndmsg *ndm,
                     struct nlattr *tb[],
                     struct net_device *dev,
                     const unsigned char *addr, u16 vid,
                     u16 flags)
{
        int err = -EINVAL;

        /* If aging addresses are supported device will need to
         * implement its own handler for this.
         */
        if (ndm->ndm_state && !(ndm->ndm_state & NUD_PERMANENT)) {
                netdev_info(dev, "default FDB implementation only supports local addresses\n");
                return err;
        }

        if (tb[NDA_FLAGS_EXT]) {
                netdev_info(dev, "invalid flags given to default FDB implementation\n");
                return err;
        }

        if (vid) {
                netdev_info(dev, "vlans aren't supported yet for dev_uc|mc_add()\n");
                return err;
        }

        if (is_unicast_ether_addr(addr) || is_link_local_ether_addr(addr))
                err = dev_uc_add_excl(dev, addr);
        else if (is_multicast_ether_addr(addr))
                err = dev_mc_add_excl(dev, addr);

        /* Only return duplicate errors if NLM_F_EXCL is set */
        if (err == -EEXIST && !(flags & NLM_F_EXCL))
                err = 0;

        return err;
}
EXPORT_SYMBOL(ndo_dflt_fdb_add);

static int fdb_vid_parse(struct nlattr *vlan_attr, u16 *p_vid,
                         struct netlink_ext_ack *extack)
{
        u16 vid = 0;

        if (vlan_attr) {
                if (nla_len(vlan_attr) != sizeof(u16)) {
                        NL_SET_ERR_MSG(extack, "invalid vlan attribute size");
                        return -EINVAL;
                }

                vid = nla_get_u16(vlan_attr);

                if (!vid || vid >= VLAN_VID_MASK) {
                        NL_SET_ERR_MSG(extack, "invalid vlan id");
                        return -EINVAL;
                }
        }
        *p_vid = vid;
        return 0;
}

static int rtnl_fdb_add(struct sk_buff *skb, struct nlmsghdr *nlh,
                        struct netlink_ext_ack *extack)
{
        struct net *net = sock_net(skb->sk);
        struct ndmsg *ndm;
        struct nlattr *tb[NDA_MAX+1];
        struct net_device *dev;
        u8 *addr;
        u16 vid;
        int err;

        err = nlmsg_parse_deprecated(nlh, sizeof(*ndm), tb, NDA_MAX, NULL,
                                     extack);
        if (err < 0)
                return err;

        ndm = nlmsg_data(nlh);
        if (ndm->ndm_ifindex == 0) {
                NL_SET_ERR_MSG(extack, "invalid ifindex");
                return -EINVAL;
        }

        dev = __dev_get_by_index(net, ndm->ndm_ifindex);
        if (dev == NULL) {
                NL_SET_ERR_MSG(extack, "unknown ifindex");
                return -ENODEV;
        }

        if (!tb[NDA_LLADDR] || nla_len(tb[NDA_LLADDR]) != ETH_ALEN) {
                NL_SET_ERR_MSG(extack, "invalid address");
                return -EINVAL;
        }

        if (dev->type != ARPHRD_ETHER) {
                NL_SET_ERR_MSG(extack, "FDB add only supported for Ethernet devices");
                return -EINVAL;
        }

        addr = nla_data(tb[NDA_LLADDR]);

        err = fdb_vid_parse(tb[NDA_VLAN], &vid, extack);
        if (err)
                return err;

        err = -EOPNOTSUPP;

        /* Support fdb on master device the net/bridge default case */
        if ((!ndm->ndm_flags || ndm->ndm_flags & NTF_MASTER) &&
            netif_is_bridge_port(dev)) {
                struct net_device *br_dev = netdev_master_upper_dev_get(dev);
                const struct net_device_ops *ops = br_dev->netdev_ops;
                bool notified = false;

                err = ops->ndo_fdb_add(ndm, tb, dev, addr, vid,
                                       nlh->nlmsg_flags, &notified, extack);
                if (err)
                        goto out;
                else
                        ndm->ndm_flags &= ~NTF_MASTER;
        }

        /* Embedded bridge, macvlan, and any other device support */
        if ((ndm->ndm_flags & NTF_SELF)) {
                bool notified = false;

                if (dev->netdev_ops->ndo_fdb_add)
                        err = dev->netdev_ops->ndo_fdb_add(ndm, tb, dev, addr,
                                                           vid,
                                                           nlh->nlmsg_flags,
                                                           &notified, extack);
                else
                        err = ndo_dflt_fdb_add(ndm, tb, dev, addr, vid,
                                               nlh->nlmsg_flags);

                if (!err && !notified) {
                        rtnl_fdb_notify(dev, addr, vid, RTM_NEWNEIGH,
                                        ndm->ndm_state);
                        ndm->ndm_flags &= ~NTF_SELF;
                }
        }
out:
        return err;
}

/*
 * ndo_dflt_fdb_del - default netdevice operation to delete an FDB entry
 */
int ndo_dflt_fdb_del(struct ndmsg *ndm,
                     struct nlattr *tb[],
                     struct net_device *dev,
                     const unsigned char *addr, u16 vid)
{
        int err = -EINVAL;

        /* If aging addresses are supported device will need to
         * implement its own handler for this.
         */
        if (!(ndm->ndm_state & NUD_PERMANENT)) {
                netdev_info(dev, "default FDB implementation only supports local addresses\n");
                return err;
        }

        if (is_unicast_ether_addr(addr) || is_link_local_ether_addr(addr))
                err = dev_uc_del(dev, addr);
        else if (is_multicast_ether_addr(addr))
                err = dev_mc_del(dev, addr);

        return err;
}
EXPORT_SYMBOL(ndo_dflt_fdb_del);

static int rtnl_fdb_del(struct sk_buff *skb, struct nlmsghdr *nlh,
                        struct netlink_ext_ack *extack)
{
        bool del_bulk = !!(nlh->nlmsg_flags & NLM_F_BULK);
        struct net *net = sock_net(skb->sk);
        const struct net_device_ops *ops;
        struct ndmsg *ndm;
        struct nlattr *tb[NDA_MAX+1];
        struct net_device *dev;
        __u8 *addr = NULL;
        int err;
        u16 vid;

        if (!del_bulk) {
                err = nlmsg_parse_deprecated(nlh, sizeof(*ndm), tb, NDA_MAX,
                                             NULL, extack);
        } else {
                /* For bulk delete, the drivers will parse the message with
                 * policy.
                 */
                err = nlmsg_parse(nlh, sizeof(*ndm), tb, NDA_MAX, NULL, extack);
        }
        if (err < 0)
                return err;

        ndm = nlmsg_data(nlh);
        if (ndm->ndm_ifindex == 0) {
                NL_SET_ERR_MSG(extack, "invalid ifindex");
                return -EINVAL;
        }

        dev = __dev_get_by_index(net, ndm->ndm_ifindex);
        if (dev == NULL) {
                NL_SET_ERR_MSG(extack, "unknown ifindex");
                return -ENODEV;
        }

        if (!del_bulk) {
                if (!tb[NDA_LLADDR] || nla_len(tb[NDA_LLADDR]) != ETH_ALEN) {
                        NL_SET_ERR_MSG(extack, "invalid address");
                        return -EINVAL;
                }
                addr = nla_data(tb[NDA_LLADDR]);

                err = fdb_vid_parse(tb[NDA_VLAN], &vid, extack);
                if (err)
                        return err;
        }

        if (dev->type != ARPHRD_ETHER) {
                NL_SET_ERR_MSG(extack, "FDB delete only supported for Ethernet devices");
                return -EINVAL;
        }

        err = -EOPNOTSUPP;

        /* Support fdb on master device the net/bridge default case */
        if ((!ndm->ndm_flags || ndm->ndm_flags & NTF_MASTER) &&
            netif_is_bridge_port(dev)) {
                struct net_device *br_dev = netdev_master_upper_dev_get(dev);
                bool notified = false;

                ops = br_dev->netdev_ops;
                if (!del_bulk) {
                        if (ops->ndo_fdb_del)
                                err = ops->ndo_fdb_del(ndm, tb, dev, addr, vid,
                                                       &notified, extack);
                } else {
                        if (ops->ndo_fdb_del_bulk)
                                err = ops->ndo_fdb_del_bulk(nlh, dev, extack);
                }

                if (err)
                        goto out;
                else
                        ndm->ndm_flags &= ~NTF_MASTER;
        }

        /* Embedded bridge, macvlan, and any other device support */
        if (ndm->ndm_flags & NTF_SELF) {
                bool notified = false;

                ops = dev->netdev_ops;
                if (!del_bulk) {
                        if (ops->ndo_fdb_del)
                                err = ops->ndo_fdb_del(ndm, tb, dev, addr, vid,
                                                       &notified, extack);
                        else
                                err = ndo_dflt_fdb_del(ndm, tb, dev, addr, vid);
                } else {
                        /* in case err was cleared by NTF_MASTER call */
                        err = -EOPNOTSUPP;
                        if (ops->ndo_fdb_del_bulk)
                                err = ops->ndo_fdb_del_bulk(nlh, dev, extack);
                }

                if (!err) {
                        if (!del_bulk && !notified)
                                rtnl_fdb_notify(dev, addr, vid, RTM_DELNEIGH,
                                                ndm->ndm_state);
                        ndm->ndm_flags &= ~NTF_SELF;
                }
        }
out:
        return err;
}

static int nlmsg_populate_fdb(struct sk_buff *skb,
                              struct netlink_callback *cb,
                              struct net_device *dev,
                              int *idx,
                              struct netdev_hw_addr_list *list)
{
        struct ndo_fdb_dump_context *ctx = (void *)cb->ctx;
        struct netdev_hw_addr *ha;
        u32 portid, seq;
        int err;

        portid = NETLINK_CB(cb->skb).portid;
        seq = cb->nlh->nlmsg_seq;

        list_for_each_entry(ha, &list->list, list) {
                if (*idx < ctx->fdb_idx)
                        goto skip;

                err = nlmsg_populate_fdb_fill(skb, dev, ha->addr, 0,
                                              portid, seq,
                                              RTM_NEWNEIGH, NTF_SELF,
                                              NLM_F_MULTI, NUD_PERMANENT);
                if (err < 0)
                        return err;
skip:
                *idx += 1;
        }
        return 0;
}

/**
 * ndo_dflt_fdb_dump - default netdevice operation to dump an FDB table.
 * @skb: socket buffer to store message in
 * @cb: netlink callback
 * @dev: netdevice
 * @filter_dev: ignored
 * @idx: the number of FDB table entries dumped is added to *@idx
 *
 * Default netdevice operation to dump the existing unicast address list.
 * Returns number of addresses from list put in skb.
 */
int ndo_dflt_fdb_dump(struct sk_buff *skb,
                      struct netlink_callback *cb,
                      struct net_device *dev,
                      struct net_device *filter_dev,
                      int *idx)
{
        int err;

        if (dev->type != ARPHRD_ETHER)
                return -EINVAL;

        netif_addr_lock_bh(dev);
        err = nlmsg_populate_fdb(skb, cb, dev, idx, &dev->uc);
        if (err)
                goto out;
        err = nlmsg_populate_fdb(skb, cb, dev, idx, &dev->mc);
out:
        netif_addr_unlock_bh(dev);
        return err;
}
EXPORT_SYMBOL(ndo_dflt_fdb_dump);

static int valid_fdb_dump_strict(const struct nlmsghdr *nlh,
                                 int *br_idx, int *brport_idx,
                                 struct netlink_ext_ack *extack)
{
        struct nlattr *tb[NDA_MAX + 1];
        struct ndmsg *ndm;
        int err, i;

        ndm = nlmsg_payload(nlh, sizeof(*ndm));
        if (!ndm) {
                NL_SET_ERR_MSG(extack, "Invalid header for fdb dump request");
                return -EINVAL;
        }

        if (ndm->ndm_pad1  || ndm->ndm_pad2  || ndm->ndm_state ||
            ndm->ndm_flags || ndm->ndm_type) {
                NL_SET_ERR_MSG(extack, "Invalid values in header for fdb dump request");
                return -EINVAL;
        }

        err = nlmsg_parse_deprecated_strict(nlh, sizeof(struct ndmsg), tb,
                                            NDA_MAX, NULL, extack);
        if (err < 0)
                return err;

        *brport_idx = ndm->ndm_ifindex;
        for (i = 0; i <= NDA_MAX; ++i) {
                if (!tb[i])
                        continue;

                switch (i) {
                case NDA_IFINDEX:
                        if (nla_len(tb[i]) != sizeof(u32)) {
                                NL_SET_ERR_MSG(extack, "Invalid IFINDEX attribute in fdb dump request");
                                return -EINVAL;
                        }
                        *brport_idx = nla_get_u32(tb[NDA_IFINDEX]);
                        break;
                case NDA_MASTER:
                        if (nla_len(tb[i]) != sizeof(u32)) {
                                NL_SET_ERR_MSG(extack, "Invalid MASTER attribute in fdb dump request");
                                return -EINVAL;
                        }
                        *br_idx = nla_get_u32(tb[NDA_MASTER]);
                        break;
                default:
                        NL_SET_ERR_MSG(extack, "Unsupported attribute in fdb dump request");
                        return -EINVAL;
                }
        }

        return 0;
}

static int valid_fdb_dump_legacy(const struct nlmsghdr *nlh,
                                 int *br_idx, int *brport_idx,
                                 struct netlink_ext_ack *extack)
{
        struct nlattr *tb[IFLA_MAX+1];
        int err;

        /* A hack to preserve kernel<->userspace interface.
         * Before Linux v4.12 this code accepted ndmsg since iproute2 v3.3.0.
         * However, ndmsg is shorter than ifinfomsg thus nlmsg_parse() bails.
         * So, check for ndmsg with an optional u32 attribute (not used here).
         * Fortunately these sizes don't conflict with the size of ifinfomsg
         * with an optional attribute.
         */
        if (nlmsg_len(nlh) != sizeof(struct ndmsg) &&
            (nlmsg_len(nlh) != sizeof(struct ndmsg) +
             nla_attr_size(sizeof(u32)))) {
                struct ifinfomsg *ifm;

                err = nlmsg_parse_deprecated(nlh, sizeof(struct ifinfomsg),
                                             tb, IFLA_MAX, ifla_policy,
                                             extack);
                if (err < 0) {
                        return -EINVAL;
                } else if (err == 0) {
                        if (tb[IFLA_MASTER])
                                *br_idx = nla_get_u32(tb[IFLA_MASTER]);
                }

                ifm = nlmsg_data(nlh);
                *brport_idx = ifm->ifi_index;
        }
        return 0;
}

static int rtnl_fdb_dump(struct sk_buff *skb, struct netlink_callback *cb)
{
        const struct net_device_ops *ops = NULL, *cops = NULL;
        struct ndo_fdb_dump_context *ctx = (void *)cb->ctx;
        struct net_device *dev, *br_dev = NULL;
        struct net *net = sock_net(skb->sk);
        int brport_idx = 0;
        int br_idx = 0;
        int fidx = 0;
        int err;

        NL_ASSERT_CTX_FITS(struct ndo_fdb_dump_context);

        if (cb->strict_check)
                err = valid_fdb_dump_strict(cb->nlh, &br_idx, &brport_idx,
                                            cb->extack);
        else
                err = valid_fdb_dump_legacy(cb->nlh, &br_idx, &brport_idx,
                                            cb->extack);
        if (err < 0)
                return err;

        if (br_idx) {
                br_dev = __dev_get_by_index(net, br_idx);
                if (!br_dev)
                        return -ENODEV;

                ops = br_dev->netdev_ops;
        }

        for_each_netdev_dump(net, dev, ctx->ifindex) {
                if (brport_idx && (dev->ifindex != brport_idx))
                        continue;

                if (!br_idx) { /* user did not specify a specific bridge */
                        if (netif_is_bridge_port(dev)) {
                                br_dev = netdev_master_upper_dev_get(dev);
                                cops = br_dev->netdev_ops;
                        }
                } else {
                        if (dev != br_dev &&
                            !netif_is_bridge_port(dev))
                                continue;

                        if (br_dev != netdev_master_upper_dev_get(dev) &&
                            !netif_is_bridge_master(dev))
                                continue;
                        cops = ops;
                }

                if (netif_is_bridge_port(dev)) {
                        if (cops && cops->ndo_fdb_dump) {
                                err = cops->ndo_fdb_dump(skb, cb, br_dev, dev,
                                                        &fidx);
                                if (err == -EMSGSIZE)
                                        break;
                        }
                }

                if (dev->netdev_ops->ndo_fdb_dump)
                        err = dev->netdev_ops->ndo_fdb_dump(skb, cb, dev, NULL,
                                                            &fidx);
                else
                        err = ndo_dflt_fdb_dump(skb, cb, dev, NULL, &fidx);
                if (err == -EMSGSIZE)
                        break;

                cops = NULL;

                /* reset fdb offset to 0 for rest of the interfaces */
                ctx->fdb_idx = 0;
                fidx = 0;
        }

        ctx->fdb_idx = fidx;

        return skb->len;
}

static int valid_fdb_get_strict(const struct nlmsghdr *nlh,
                                struct nlattr **tb, u8 *ndm_flags,
                                int *br_idx, int *brport_idx, u8 **addr,
                                u16 *vid, struct netlink_ext_ack *extack)
{
        struct ndmsg *ndm;
        int err, i;

        ndm = nlmsg_payload(nlh, sizeof(*ndm));
        if (!ndm) {
                NL_SET_ERR_MSG(extack, "Invalid header for fdb get request");
                return -EINVAL;
        }

        if (ndm->ndm_pad1  || ndm->ndm_pad2  || ndm->ndm_state ||
            ndm->ndm_type) {
                NL_SET_ERR_MSG(extack, "Invalid values in header for fdb get request");
                return -EINVAL;
        }

        if (ndm->ndm_flags & ~(NTF_MASTER | NTF_SELF)) {
                NL_SET_ERR_MSG(extack, "Invalid flags in header for fdb get request");
                return -EINVAL;
        }

        err = nlmsg_parse_deprecated_strict(nlh, sizeof(struct ndmsg), tb,
                                            NDA_MAX, nda_policy, extack);
        if (err < 0)
                return err;

        *ndm_flags = ndm->ndm_flags;
        *brport_idx = ndm->ndm_ifindex;
        for (i = 0; i <= NDA_MAX; ++i) {
                if (!tb[i])
                        continue;

                switch (i) {
                case NDA_MASTER:
                        *br_idx = nla_get_u32(tb[i]);
                        break;
                case NDA_LLADDR:
                        if (nla_len(tb[i]) != ETH_ALEN) {
                                NL_SET_ERR_MSG(extack, "Invalid address in fdb get request");
                                return -EINVAL;
                        }
                        *addr = nla_data(tb[i]);
                        break;
                case NDA_VLAN:
                        err = fdb_vid_parse(tb[i], vid, extack);
                        if (err)
                                return err;
                        break;
                case NDA_VNI:
                        break;
                default:
                        NL_SET_ERR_MSG(extack, "Unsupported attribute in fdb get request");
                        return -EINVAL;
                }
        }

        return 0;
}

static int rtnl_fdb_get(struct sk_buff *in_skb, struct nlmsghdr *nlh,
                        struct netlink_ext_ack *extack)
{
        struct net_device *dev = NULL, *br_dev = NULL;
        const struct net_device_ops *ops = NULL;
        struct net *net = sock_net(in_skb->sk);
        struct nlattr *tb[NDA_MAX + 1];
        struct sk_buff *skb;
        int brport_idx = 0;
        u8 ndm_flags = 0;
        int br_idx = 0;
        u8 *addr = NULL;
        u16 vid = 0;
        int err;

        err = valid_fdb_get_strict(nlh, tb, &ndm_flags, &br_idx,
                                   &brport_idx, &addr, &vid, extack);
        if (err < 0)
                return err;

        if (!addr) {
                NL_SET_ERR_MSG(extack, "Missing lookup address for fdb get request");
                return -EINVAL;
        }

        if (brport_idx) {
                dev = __dev_get_by_index(net, brport_idx);
                if (!dev) {
                        NL_SET_ERR_MSG(extack, "Unknown device ifindex");
                        return -ENODEV;
                }
        }

        if (br_idx) {
                if (dev) {
                        NL_SET_ERR_MSG(extack, "Master and device are mutually exclusive");
                        return -EINVAL;
                }

                br_dev = __dev_get_by_index(net, br_idx);
                if (!br_dev) {
                        NL_SET_ERR_MSG(extack, "Invalid master ifindex");
                        return -EINVAL;
                }
                ops = br_dev->netdev_ops;
        }

        if (dev) {
                if (!ndm_flags || (ndm_flags & NTF_MASTER)) {
                        if (!netif_is_bridge_port(dev)) {
                                NL_SET_ERR_MSG(extack, "Device is not a bridge port");
                                return -EINVAL;
                        }
                        br_dev = netdev_master_upper_dev_get(dev);
                        if (!br_dev) {
                                NL_SET_ERR_MSG(extack, "Master of device not found");
                                return -EINVAL;
                        }
                        ops = br_dev->netdev_ops;
                } else {
                        if (!(ndm_flags & NTF_SELF)) {
                                NL_SET_ERR_MSG(extack, "Missing NTF_SELF");
                                return -EINVAL;
                        }
                        ops = dev->netdev_ops;
                }
        }

        if (!br_dev && !dev) {
                NL_SET_ERR_MSG(extack, "No device specified");
                return -ENODEV;
        }

        if (!ops || !ops->ndo_fdb_get) {
                NL_SET_ERR_MSG(extack, "Fdb get operation not supported by device");
                return -EOPNOTSUPP;
        }

        skb = nlmsg_new(NLMSG_GOODSIZE, GFP_KERNEL);
        if (!skb)
                return -ENOBUFS;

        if (br_dev)
                dev = br_dev;
        err = ops->ndo_fdb_get(skb, tb, dev, addr, vid,
                               NETLINK_CB(in_skb).portid,
                               nlh->nlmsg_seq, extack);
        if (err)
                goto out;

        return rtnl_unicast(skb, net, NETLINK_CB(in_skb).portid);
out:
        kfree_skb(skb);
        return err;
}

static int brport_nla_put_flag(struct sk_buff *skb, u32 flags, u32 mask,
                               unsigned int attrnum, unsigned int flag)
{
        if (mask & flag)
                return nla_put_u8(skb, attrnum, !!(flags & flag));
        return 0;
}

int ndo_dflt_bridge_getlink(struct sk_buff *skb, u32 pid, u32 seq,
                            struct net_device *dev, u16 mode,
                            u32 flags, u32 mask, int nlflags,
                            u32 filter_mask,
                            int (*vlan_fill)(struct sk_buff *skb,
                                             struct net_device *dev,
                                             u32 filter_mask))
{
        struct nlmsghdr *nlh;
        struct ifinfomsg *ifm;
        struct nlattr *br_afspec;
        struct nlattr *protinfo;
        u8 operstate = netif_running(dev) ? dev->operstate : IF_OPER_DOWN;
        struct net_device *br_dev = netdev_master_upper_dev_get(dev);
        int err = 0;

        nlh = nlmsg_put(skb, pid, seq, RTM_NEWLINK, sizeof(*ifm), nlflags);
        if (nlh == NULL)
                return -EMSGSIZE;

        ifm = nlmsg_data(nlh);
        ifm->ifi_family = AF_BRIDGE;
        ifm->__ifi_pad = 0;
        ifm->ifi_type = dev->type;
        ifm->ifi_index = dev->ifindex;
        ifm->ifi_flags = netif_get_flags(dev);
        ifm->ifi_change = 0;


        if (nla_put_string(skb, IFLA_IFNAME, dev->name) ||
            nla_put_u32(skb, IFLA_MTU, dev->mtu) ||
            nla_put_u8(skb, IFLA_OPERSTATE, operstate) ||
            (br_dev &&
             nla_put_u32(skb, IFLA_MASTER, br_dev->ifindex)) ||
            (dev->addr_len &&
             nla_put(skb, IFLA_ADDRESS, dev->addr_len, dev->dev_addr)) ||
            (dev->ifindex != dev_get_iflink(dev) &&
             nla_put_u32(skb, IFLA_LINK, dev_get_iflink(dev))))
                goto nla_put_failure;

        br_afspec = nla_nest_start_noflag(skb, IFLA_AF_SPEC);
        if (!br_afspec)
                goto nla_put_failure;

        if (nla_put_u16(skb, IFLA_BRIDGE_FLAGS, BRIDGE_FLAGS_SELF)) {
                nla_nest_cancel(skb, br_afspec);
                goto nla_put_failure;
        }

        if (mode != BRIDGE_MODE_UNDEF) {
                if (nla_put_u16(skb, IFLA_BRIDGE_MODE, mode)) {
                        nla_nest_cancel(skb, br_afspec);
                        goto nla_put_failure;
                }
        }
        if (vlan_fill) {
                err = vlan_fill(skb, dev, filter_mask);
                if (err) {
                        nla_nest_cancel(skb, br_afspec);
                        goto nla_put_failure;
                }
        }
        nla_nest_end(skb, br_afspec);

        protinfo = nla_nest_start(skb, IFLA_PROTINFO);
        if (!protinfo)
                goto nla_put_failure;

        if (brport_nla_put_flag(skb, flags, mask,
                                IFLA_BRPORT_MODE, BR_HAIRPIN_MODE) ||
            brport_nla_put_flag(skb, flags, mask,
                                IFLA_BRPORT_GUARD, BR_BPDU_GUARD) ||
            brport_nla_put_flag(skb, flags, mask,
                                IFLA_BRPORT_FAST_LEAVE,
                                BR_MULTICAST_FAST_LEAVE) ||
            brport_nla_put_flag(skb, flags, mask,
                                IFLA_BRPORT_PROTECT, BR_ROOT_BLOCK) ||
            brport_nla_put_flag(skb, flags, mask,
                                IFLA_BRPORT_LEARNING, BR_LEARNING) ||
            brport_nla_put_flag(skb, flags, mask,
                                IFLA_BRPORT_LEARNING_SYNC, BR_LEARNING_SYNC) ||
            brport_nla_put_flag(skb, flags, mask,
                                IFLA_BRPORT_UNICAST_FLOOD, BR_FLOOD) ||
            brport_nla_put_flag(skb, flags, mask,
                                IFLA_BRPORT_PROXYARP, BR_PROXYARP) ||
            brport_nla_put_flag(skb, flags, mask,
                                IFLA_BRPORT_MCAST_FLOOD, BR_MCAST_FLOOD) ||
            brport_nla_put_flag(skb, flags, mask,
                                IFLA_BRPORT_BCAST_FLOOD, BR_BCAST_FLOOD)) {
                nla_nest_cancel(skb, protinfo);
                goto nla_put_failure;
        }

        nla_nest_end(skb, protinfo);

        nlmsg_end(skb, nlh);
        return 0;
nla_put_failure:
        nlmsg_cancel(skb, nlh);
        return err ? err : -EMSGSIZE;
}
EXPORT_SYMBOL_GPL(ndo_dflt_bridge_getlink);

static int valid_bridge_getlink_req(const struct nlmsghdr *nlh,
                                    bool strict_check, u32 *filter_mask,
                                    struct netlink_ext_ack *extack)
{
        struct nlattr *tb[IFLA_MAX+1];
        int err, i;

        if (strict_check) {
                struct ifinfomsg *ifm;

                ifm = nlmsg_payload(nlh, sizeof(*ifm));
                if (!ifm) {
                        NL_SET_ERR_MSG(extack, "Invalid header for bridge link dump");
                        return -EINVAL;
                }

                if (ifm->__ifi_pad || ifm->ifi_type || ifm->ifi_flags ||
                    ifm->ifi_change || ifm->ifi_index) {
                        NL_SET_ERR_MSG(extack, "Invalid values in header for bridge link dump request");
                        return -EINVAL;
                }

                err = nlmsg_parse_deprecated_strict(nlh,
                                                    sizeof(struct ifinfomsg),
                                                    tb, IFLA_MAX, ifla_policy,
                                                    extack);
        } else {
                err = nlmsg_parse_deprecated(nlh, sizeof(struct ifinfomsg),
                                             tb, IFLA_MAX, ifla_policy,
                                             extack);
        }
        if (err < 0)
                return err;

        /* new attributes should only be added with strict checking */
        for (i = 0; i <= IFLA_MAX; ++i) {
                if (!tb[i])
                        continue;

                switch (i) {
                case IFLA_EXT_MASK:
                        *filter_mask = nla_get_u32(tb[i]);
                        break;
                default:
                        if (strict_check) {
                                NL_SET_ERR_MSG(extack, "Unsupported attribute in bridge link dump request");
                                return -EINVAL;
                        }
                }
        }

        return 0;
}

static int rtnl_bridge_getlink(struct sk_buff *skb, struct netlink_callback *cb)
{
        const struct nlmsghdr *nlh = cb->nlh;
        struct net *net = sock_net(skb->sk);
        struct net_device *dev;
        int idx = 0;
        u32 portid = NETLINK_CB(cb->skb).portid;
        u32 seq = nlh->nlmsg_seq;
        u32 filter_mask = 0;
        int err;

        err = valid_bridge_getlink_req(nlh, cb->strict_check, &filter_mask,
                                       cb->extack);
        if (err < 0 && cb->strict_check)
                return err;

        rcu_read_lock();
        for_each_netdev_rcu(net, dev) {
                const struct net_device_ops *ops = dev->netdev_ops;
                struct net_device *br_dev = netdev_master_upper_dev_get(dev);

                if (br_dev && br_dev->netdev_ops->ndo_bridge_getlink) {
                        if (idx >= cb->args[0]) {
                                err = br_dev->netdev_ops->ndo_bridge_getlink(
                                                skb, portid, seq, dev,
                                                filter_mask, NLM_F_MULTI);
                                if (err < 0 && err != -EOPNOTSUPP) {
                                        if (likely(skb->len))
                                                break;

                                        goto out_err;
                                }
                        }
                        idx++;
                }

                if (ops->ndo_bridge_getlink) {
                        if (idx >= cb->args[0]) {
                                err = ops->ndo_bridge_getlink(skb, portid,
                                                              seq, dev,
                                                              filter_mask,
                                                              NLM_F_MULTI);
                                if (err < 0 && err != -EOPNOTSUPP) {
                                        if (likely(skb->len))
                                                break;

                                        goto out_err;
                                }
                        }
                        idx++;
                }
        }
        err = skb->len;
out_err:
        rcu_read_unlock();
        cb->args[0] = idx;

        return err;
}

static inline size_t bridge_nlmsg_size(void)
{
        return NLMSG_ALIGN(sizeof(struct ifinfomsg))
                + nla_total_size(IFNAMSIZ)      /* IFLA_IFNAME */
                + nla_total_size(MAX_ADDR_LEN)  /* IFLA_ADDRESS */
                + nla_total_size(sizeof(u32))   /* IFLA_MASTER */
                + nla_total_size(sizeof(u32))   /* IFLA_MTU */
                + nla_total_size(sizeof(u32))   /* IFLA_LINK */
                + nla_total_size(sizeof(u32))   /* IFLA_OPERSTATE */
                + nla_total_size(sizeof(u8))    /* IFLA_PROTINFO */
                + nla_total_size(sizeof(struct nlattr)) /* IFLA_AF_SPEC */
                + nla_total_size(sizeof(u16))   /* IFLA_BRIDGE_FLAGS */
                + nla_total_size(sizeof(u16));  /* IFLA_BRIDGE_MODE */
}

static int rtnl_bridge_notify(struct net_device *dev)
{
        struct net *net = dev_net(dev);
        struct sk_buff *skb;
        int err = -EOPNOTSUPP;

        if (!dev->netdev_ops->ndo_bridge_getlink)
                return 0;

        skb = nlmsg_new(bridge_nlmsg_size(), GFP_ATOMIC);
        if (!skb) {
                err = -ENOMEM;
                goto errout;
        }

        err = dev->netdev_ops->ndo_bridge_getlink(skb, 0, 0, dev, 0, 0);
        if (err < 0)
                goto errout;

        /* Notification info is only filled for bridge ports, not the bridge
         * device itself. Therefore, a zero notification length is valid and
         * should not result in an error.
         */
        if (!skb->len)
                goto errout;

        rtnl_notify(skb, net, 0, RTNLGRP_LINK, NULL, GFP_ATOMIC);
        return 0;
errout:
        WARN_ON(err == -EMSGSIZE);
        kfree_skb(skb);
        if (err)
                rtnl_set_sk_err(net, RTNLGRP_LINK, err);
        return err;
}

static int rtnl_bridge_setlink(struct sk_buff *skb, struct nlmsghdr *nlh,
                               struct netlink_ext_ack *extack)
{
        struct net *net = sock_net(skb->sk);
        struct ifinfomsg *ifm;
        struct net_device *dev;
        struct nlattr *br_spec, *attr, *br_flags_attr = NULL;
        int rem, err = -EOPNOTSUPP;
        u16 flags = 0;

        if (nlmsg_len(nlh) < sizeof(*ifm))
                return -EINVAL;

        ifm = nlmsg_data(nlh);
        if (ifm->ifi_family != AF_BRIDGE)
                return -EPFNOSUPPORT;

        dev = __dev_get_by_index(net, ifm->ifi_index);
        if (!dev) {
                NL_SET_ERR_MSG(extack, "unknown ifindex");
                return -ENODEV;
        }

        br_spec = nlmsg_find_attr(nlh, sizeof(struct ifinfomsg), IFLA_AF_SPEC);
        if (br_spec) {
                nla_for_each_nested(attr, br_spec, rem) {
                        if (nla_type(attr) == IFLA_BRIDGE_FLAGS && !br_flags_attr) {
                                if (nla_len(attr) < sizeof(flags))
                                        return -EINVAL;

                                br_flags_attr = attr;
                                flags = nla_get_u16(attr);
                        }

                        if (nla_type(attr) == IFLA_BRIDGE_MODE) {
                                if (nla_len(attr) < sizeof(u16))
                                        return -EINVAL;
                        }
                }
        }

        if (!flags || (flags & BRIDGE_FLAGS_MASTER)) {
                struct net_device *br_dev = netdev_master_upper_dev_get(dev);

                if (!br_dev || !br_dev->netdev_ops->ndo_bridge_setlink) {
                        err = -EOPNOTSUPP;
                        goto out;
                }

                err = br_dev->netdev_ops->ndo_bridge_setlink(dev, nlh, flags,
                                                             extack);
                if (err)
                        goto out;

                flags &= ~BRIDGE_FLAGS_MASTER;
        }

        if ((flags & BRIDGE_FLAGS_SELF)) {
                if (!dev->netdev_ops->ndo_bridge_setlink)
                        err = -EOPNOTSUPP;
                else
                        err = dev->netdev_ops->ndo_bridge_setlink(dev, nlh,
                                                                  flags,
                                                                  extack);
                if (!err) {
                        flags &= ~BRIDGE_FLAGS_SELF;

                        /* Generate event to notify upper layer of bridge
                         * change
                         */
                        err = rtnl_bridge_notify(dev);
                }
        }

        if (br_flags_attr)
                memcpy(nla_data(br_flags_attr), &flags, sizeof(flags));
out:
        return err;
}

static int rtnl_bridge_dellink(struct sk_buff *skb, struct nlmsghdr *nlh,
                               struct netlink_ext_ack *extack)
{
        struct net *net = sock_net(skb->sk);
        struct ifinfomsg *ifm;
        struct net_device *dev;
        struct nlattr *br_spec, *attr = NULL;
        int rem, err = -EOPNOTSUPP;
        u16 flags = 0;
        bool have_flags = false;

        if (nlmsg_len(nlh) < sizeof(*ifm))
                return -EINVAL;

        ifm = nlmsg_data(nlh);
        if (ifm->ifi_family != AF_BRIDGE)
                return -EPFNOSUPPORT;

        dev = __dev_get_by_index(net, ifm->ifi_index);
        if (!dev) {
                NL_SET_ERR_MSG(extack, "unknown ifindex");
                return -ENODEV;
        }

        br_spec = nlmsg_find_attr(nlh, sizeof(struct ifinfomsg), IFLA_AF_SPEC);
        if (br_spec) {
                nla_for_each_nested_type(attr, IFLA_BRIDGE_FLAGS, br_spec,
                                         rem) {
                        if (nla_len(attr) < sizeof(flags))
                                return -EINVAL;

                        have_flags = true;
                        flags = nla_get_u16(attr);
                        break;
                }
        }

        if (!flags || (flags & BRIDGE_FLAGS_MASTER)) {
                struct net_device *br_dev = netdev_master_upper_dev_get(dev);

                if (!br_dev || !br_dev->netdev_ops->ndo_bridge_dellink) {
                        err = -EOPNOTSUPP;
                        goto out;
                }

                err = br_dev->netdev_ops->ndo_bridge_dellink(dev, nlh, flags);
                if (err)
                        goto out;

                flags &= ~BRIDGE_FLAGS_MASTER;
        }

        if ((flags & BRIDGE_FLAGS_SELF)) {
                if (!dev->netdev_ops->ndo_bridge_dellink)
                        err = -EOPNOTSUPP;
                else
                        err = dev->netdev_ops->ndo_bridge_dellink(dev, nlh,
                                                                  flags);

                if (!err) {
                        flags &= ~BRIDGE_FLAGS_SELF;

                        /* Generate event to notify upper layer of bridge
                         * change
                         */
                        err = rtnl_bridge_notify(dev);
                }
        }

        if (have_flags)
                memcpy(nla_data(attr), &flags, sizeof(flags));
out:
        return err;
}

static bool stats_attr_valid(unsigned int mask, int attrid, int idxattr)
{
        return (mask & IFLA_STATS_FILTER_BIT(attrid)) &&
               (!idxattr || idxattr == attrid);
}

static bool
rtnl_offload_xstats_have_ndo(const struct net_device *dev, int attr_id)
{
        return dev->netdev_ops &&
               dev->netdev_ops->ndo_has_offload_stats &&
               dev->netdev_ops->ndo_get_offload_stats &&
               dev->netdev_ops->ndo_has_offload_stats(dev, attr_id);
}

static unsigned int
rtnl_offload_xstats_get_size_ndo(const struct net_device *dev, int attr_id)
{
        return rtnl_offload_xstats_have_ndo(dev, attr_id) ?
               sizeof(struct rtnl_link_stats64) : 0;
}

static int
rtnl_offload_xstats_fill_ndo(struct net_device *dev, int attr_id,
                             struct sk_buff *skb)
{
        unsigned int size = rtnl_offload_xstats_get_size_ndo(dev, attr_id);
        struct nlattr *attr = NULL;
        void *attr_data;
        int err;

        if (!size)
                return -ENODATA;

        attr = nla_reserve_64bit(skb, attr_id, size,
                                 IFLA_OFFLOAD_XSTATS_UNSPEC);
        if (!attr)
                return -EMSGSIZE;

        attr_data = nla_data(attr);
        memset(attr_data, 0, size);

        err = dev->netdev_ops->ndo_get_offload_stats(attr_id, dev, attr_data);
        if (err)
                return err;

        return 0;
}

static unsigned int
rtnl_offload_xstats_get_size_stats(const struct net_device *dev,
                                   enum netdev_offload_xstats_type type)
{
        bool enabled = netdev_offload_xstats_enabled(dev, type);

        return enabled ? sizeof(struct rtnl_hw_stats64) : 0;
}

struct rtnl_offload_xstats_request_used {
        bool request;
        bool used;
};

static int
rtnl_offload_xstats_get_stats(struct net_device *dev,
                              enum netdev_offload_xstats_type type,
                              struct rtnl_offload_xstats_request_used *ru,
                              struct rtnl_hw_stats64 *stats,
                              struct netlink_ext_ack *extack)
{
        bool request;
        bool used;
        int err;

        request = netdev_offload_xstats_enabled(dev, type);
        if (!request) {
                used = false;
                goto out;
        }

        err = netdev_offload_xstats_get(dev, type, stats, &used, extack);
        if (err)
                return err;

out:
        if (ru) {
                ru->request = request;
                ru->used = used;
        }
        return 0;
}

static int
rtnl_offload_xstats_fill_hw_s_info_one(struct sk_buff *skb, int attr_id,
                                       struct rtnl_offload_xstats_request_used *ru)
{
        struct nlattr *nest;

        nest = nla_nest_start(skb, attr_id);
        if (!nest)
                return -EMSGSIZE;

        if (nla_put_u8(skb, IFLA_OFFLOAD_XSTATS_HW_S_INFO_REQUEST, ru->request))
                goto nla_put_failure;

        if (nla_put_u8(skb, IFLA_OFFLOAD_XSTATS_HW_S_INFO_USED, ru->used))
                goto nla_put_failure;

        nla_nest_end(skb, nest);
        return 0;

nla_put_failure:
        nla_nest_cancel(skb, nest);
        return -EMSGSIZE;
}

static int
rtnl_offload_xstats_fill_hw_s_info(struct sk_buff *skb, struct net_device *dev,
                                   struct netlink_ext_ack *extack)
{
        enum netdev_offload_xstats_type t_l3 = NETDEV_OFFLOAD_XSTATS_TYPE_L3;
        struct rtnl_offload_xstats_request_used ru_l3;
        struct nlattr *nest;
        int err;

        err = rtnl_offload_xstats_get_stats(dev, t_l3, &ru_l3, NULL, extack);
        if (err)
                return err;

        nest = nla_nest_start(skb, IFLA_OFFLOAD_XSTATS_HW_S_INFO);
        if (!nest)
                return -EMSGSIZE;

        if (rtnl_offload_xstats_fill_hw_s_info_one(skb,
                                                   IFLA_OFFLOAD_XSTATS_L3_STATS,
                                                   &ru_l3))
                goto nla_put_failure;

        nla_nest_end(skb, nest);
        return 0;

nla_put_failure:
        nla_nest_cancel(skb, nest);
        return -EMSGSIZE;
}

static int rtnl_offload_xstats_fill(struct sk_buff *skb, struct net_device *dev,
                                    int *prividx, u32 off_filter_mask,
                                    struct netlink_ext_ack *extack)
{
        enum netdev_offload_xstats_type t_l3 = NETDEV_OFFLOAD_XSTATS_TYPE_L3;
        int attr_id_hw_s_info = IFLA_OFFLOAD_XSTATS_HW_S_INFO;
        int attr_id_l3_stats = IFLA_OFFLOAD_XSTATS_L3_STATS;
        int attr_id_cpu_hit = IFLA_OFFLOAD_XSTATS_CPU_HIT;
        bool have_data = false;
        int err;

        if (*prividx <= attr_id_cpu_hit &&
            (off_filter_mask &
             IFLA_STATS_FILTER_BIT(attr_id_cpu_hit))) {
                err = rtnl_offload_xstats_fill_ndo(dev, attr_id_cpu_hit, skb);
                if (!err) {
                        have_data = true;
                } else if (err != -ENODATA) {
                        *prividx = attr_id_cpu_hit;
                        return err;
                }
        }

        if (*prividx <= attr_id_hw_s_info &&
            (off_filter_mask & IFLA_STATS_FILTER_BIT(attr_id_hw_s_info))) {
                *prividx = attr_id_hw_s_info;

                err = rtnl_offload_xstats_fill_hw_s_info(skb, dev, extack);
                if (err)
                        return err;

                have_data = true;
                *prividx = 0;
        }

        if (*prividx <= attr_id_l3_stats &&
            (off_filter_mask & IFLA_STATS_FILTER_BIT(attr_id_l3_stats))) {
                unsigned int size_l3;
                struct nlattr *attr;

                *prividx = attr_id_l3_stats;

                size_l3 = rtnl_offload_xstats_get_size_stats(dev, t_l3);
                if (!size_l3)
                        goto skip_l3_stats;
                attr = nla_reserve_64bit(skb, attr_id_l3_stats, size_l3,
                                         IFLA_OFFLOAD_XSTATS_UNSPEC);
                if (!attr)
                        return -EMSGSIZE;

                err = rtnl_offload_xstats_get_stats(dev, t_l3, NULL,
                                                    nla_data(attr), extack);
                if (err)
                        return err;

                have_data = true;
skip_l3_stats:
                *prividx = 0;
        }

        if (!have_data)
                return -ENODATA;

        *prividx = 0;
        return 0;
}

static unsigned int
rtnl_offload_xstats_get_size_hw_s_info_one(const struct net_device *dev,
                                           enum netdev_offload_xstats_type type)
{
        return nla_total_size(0) +
                /* IFLA_OFFLOAD_XSTATS_HW_S_INFO_REQUEST */
                nla_total_size(sizeof(u8)) +
                /* IFLA_OFFLOAD_XSTATS_HW_S_INFO_USED */
                nla_total_size(sizeof(u8)) +
                0;
}

static unsigned int
rtnl_offload_xstats_get_size_hw_s_info(const struct net_device *dev)
{
        enum netdev_offload_xstats_type t_l3 = NETDEV_OFFLOAD_XSTATS_TYPE_L3;

        return nla_total_size(0) +
                /* IFLA_OFFLOAD_XSTATS_L3_STATS */
                rtnl_offload_xstats_get_size_hw_s_info_one(dev, t_l3) +
                0;
}

static int rtnl_offload_xstats_get_size(const struct net_device *dev,
                                        u32 off_filter_mask)
{
        enum netdev_offload_xstats_type t_l3 = NETDEV_OFFLOAD_XSTATS_TYPE_L3;
        int attr_id_cpu_hit = IFLA_OFFLOAD_XSTATS_CPU_HIT;
        int nla_size = 0;
        int size;

        if (off_filter_mask &
            IFLA_STATS_FILTER_BIT(attr_id_cpu_hit)) {
                size = rtnl_offload_xstats_get_size_ndo(dev, attr_id_cpu_hit);
                nla_size += nla_total_size_64bit(size);
        }

        if (off_filter_mask &
            IFLA_STATS_FILTER_BIT(IFLA_OFFLOAD_XSTATS_HW_S_INFO))
                nla_size += rtnl_offload_xstats_get_size_hw_s_info(dev);

        if (off_filter_mask &
            IFLA_STATS_FILTER_BIT(IFLA_OFFLOAD_XSTATS_L3_STATS)) {
                size = rtnl_offload_xstats_get_size_stats(dev, t_l3);
                nla_size += nla_total_size_64bit(size);
        }

        if (nla_size != 0)
                nla_size += nla_total_size(0);

        return nla_size;
}

struct rtnl_stats_dump_filters {
        /* mask[0] filters outer attributes. Then individual nests have their
         * filtering mask at the index of the nested attribute.
         */
        u32 mask[IFLA_STATS_MAX + 1];
};

static int rtnl_fill_statsinfo(struct sk_buff *skb, struct net_device *dev,
                               int type, u32 pid, u32 seq, u32 change,
                               unsigned int flags,
                               const struct rtnl_stats_dump_filters *filters,
                               int *idxattr, int *prividx,
                               struct netlink_ext_ack *extack)
{
        unsigned int filter_mask = filters->mask[0];
        struct if_stats_msg *ifsm;
        struct nlmsghdr *nlh;
        struct nlattr *attr;
        int s_prividx = *prividx;
        int err;

        ASSERT_RTNL();

        nlh = nlmsg_put(skb, pid, seq, type, sizeof(*ifsm), flags);
        if (!nlh)
                return -EMSGSIZE;

        ifsm = nlmsg_data(nlh);
        ifsm->family = PF_UNSPEC;
        ifsm->pad1 = 0;
        ifsm->pad2 = 0;
        ifsm->ifindex = dev->ifindex;
        ifsm->filter_mask = filter_mask;

        if (stats_attr_valid(filter_mask, IFLA_STATS_LINK_64, *idxattr)) {
                struct rtnl_link_stats64 *sp;

                attr = nla_reserve_64bit(skb, IFLA_STATS_LINK_64,
                                         sizeof(struct rtnl_link_stats64),
                                         IFLA_STATS_UNSPEC);
                if (!attr) {
                        err = -EMSGSIZE;
                        goto nla_put_failure;
                }

                sp = nla_data(attr);
                dev_get_stats(dev, sp);
        }

        if (stats_attr_valid(filter_mask, IFLA_STATS_LINK_XSTATS, *idxattr)) {
                const struct rtnl_link_ops *ops = dev->rtnl_link_ops;

                if (ops && ops->fill_linkxstats) {
                        *idxattr = IFLA_STATS_LINK_XSTATS;
                        attr = nla_nest_start_noflag(skb,
                                                     IFLA_STATS_LINK_XSTATS);
                        if (!attr) {
                                err = -EMSGSIZE;
                                goto nla_put_failure;
                        }

                        err = ops->fill_linkxstats(skb, dev, prividx, *idxattr);
                        nla_nest_end(skb, attr);
                        if (err)
                                goto nla_put_failure;
                        *idxattr = 0;
                }
        }

        if (stats_attr_valid(filter_mask, IFLA_STATS_LINK_XSTATS_SLAVE,
                             *idxattr)) {
                const struct rtnl_link_ops *ops = NULL;
                const struct net_device *master;

                master = netdev_master_upper_dev_get(dev);
                if (master)
                        ops = master->rtnl_link_ops;
                if (ops && ops->fill_linkxstats) {
                        *idxattr = IFLA_STATS_LINK_XSTATS_SLAVE;
                        attr = nla_nest_start_noflag(skb,
                                                     IFLA_STATS_LINK_XSTATS_SLAVE);
                        if (!attr) {
                                err = -EMSGSIZE;
                                goto nla_put_failure;
                        }

                        err = ops->fill_linkxstats(skb, dev, prividx, *idxattr);
                        nla_nest_end(skb, attr);
                        if (err)
                                goto nla_put_failure;
                        *idxattr = 0;
                }
        }

        if (stats_attr_valid(filter_mask, IFLA_STATS_LINK_OFFLOAD_XSTATS,
                             *idxattr)) {
                u32 off_filter_mask;

                off_filter_mask = filters->mask[IFLA_STATS_LINK_OFFLOAD_XSTATS];
                *idxattr = IFLA_STATS_LINK_OFFLOAD_XSTATS;
                attr = nla_nest_start_noflag(skb,
                                             IFLA_STATS_LINK_OFFLOAD_XSTATS);
                if (!attr) {
                        err = -EMSGSIZE;
                        goto nla_put_failure;
                }

                err = rtnl_offload_xstats_fill(skb, dev, prividx,
                                               off_filter_mask, extack);
                if (err == -ENODATA)
                        nla_nest_cancel(skb, attr);
                else
                        nla_nest_end(skb, attr);

                if (err && err != -ENODATA)
                        goto nla_put_failure;
                *idxattr = 0;
        }

        if (stats_attr_valid(filter_mask, IFLA_STATS_AF_SPEC, *idxattr)) {
                struct rtnl_af_ops *af_ops;

                *idxattr = IFLA_STATS_AF_SPEC;
                attr = nla_nest_start_noflag(skb, IFLA_STATS_AF_SPEC);
                if (!attr) {
                        err = -EMSGSIZE;
                        goto nla_put_failure;
                }

                rcu_read_lock();
                list_for_each_entry_rcu(af_ops, &rtnl_af_ops, list) {
                        if (af_ops->fill_stats_af) {
                                struct nlattr *af;

                                af = nla_nest_start_noflag(skb,
                                                           af_ops->family);
                                if (!af) {
                                        rcu_read_unlock();
                                        err = -EMSGSIZE;
                                        goto nla_put_failure;
                                }
                                err = af_ops->fill_stats_af(skb, dev);

                                if (err == -ENODATA) {
                                        nla_nest_cancel(skb, af);
                                } else if (err < 0) {
                                        rcu_read_unlock();
                                        goto nla_put_failure;
                                }

                                nla_nest_end(skb, af);
                        }
                }
                rcu_read_unlock();

                nla_nest_end(skb, attr);

                *idxattr = 0;
        }

        nlmsg_end(skb, nlh);

        return 0;

nla_put_failure:
        /* not a multi message or no progress mean a real error */
        if (!(flags & NLM_F_MULTI) || s_prividx == *prividx)
                nlmsg_cancel(skb, nlh);
        else
                nlmsg_end(skb, nlh);

        return err;
}

static size_t if_nlmsg_stats_size(const struct net_device *dev,
                                  const struct rtnl_stats_dump_filters *filters)
{
        size_t size = NLMSG_ALIGN(sizeof(struct if_stats_msg));
        unsigned int filter_mask = filters->mask[0];

        if (stats_attr_valid(filter_mask, IFLA_STATS_LINK_64, 0))
                size += nla_total_size_64bit(sizeof(struct rtnl_link_stats64));

        if (stats_attr_valid(filter_mask, IFLA_STATS_LINK_XSTATS, 0)) {
                const struct rtnl_link_ops *ops = dev->rtnl_link_ops;
                int attr = IFLA_STATS_LINK_XSTATS;

                if (ops && ops->get_linkxstats_size) {
                        size += nla_total_size(ops->get_linkxstats_size(dev,
                                                                        attr));
                        /* for IFLA_STATS_LINK_XSTATS */
                        size += nla_total_size(0);
                }
        }

        if (stats_attr_valid(filter_mask, IFLA_STATS_LINK_XSTATS_SLAVE, 0)) {
                struct net_device *_dev = (struct net_device *)dev;
                const struct rtnl_link_ops *ops = NULL;
                const struct net_device *master;

                /* netdev_master_upper_dev_get can't take const */
                master = netdev_master_upper_dev_get(_dev);
                if (master)
                        ops = master->rtnl_link_ops;
                if (ops && ops->get_linkxstats_size) {
                        int attr = IFLA_STATS_LINK_XSTATS_SLAVE;

                        size += nla_total_size(ops->get_linkxstats_size(dev,
                                                                        attr));
                        /* for IFLA_STATS_LINK_XSTATS_SLAVE */
                        size += nla_total_size(0);
                }
        }

        if (stats_attr_valid(filter_mask, IFLA_STATS_LINK_OFFLOAD_XSTATS, 0)) {
                u32 off_filter_mask;

                off_filter_mask = filters->mask[IFLA_STATS_LINK_OFFLOAD_XSTATS];
                size += rtnl_offload_xstats_get_size(dev, off_filter_mask);
        }

        if (stats_attr_valid(filter_mask, IFLA_STATS_AF_SPEC, 0)) {
                struct rtnl_af_ops *af_ops;

                /* for IFLA_STATS_AF_SPEC */
                size += nla_total_size(0);

                rcu_read_lock();
                list_for_each_entry_rcu(af_ops, &rtnl_af_ops, list) {
                        if (af_ops->get_stats_af_size) {
                                size += nla_total_size(
                                        af_ops->get_stats_af_size(dev));

                                /* for AF_* */
                                size += nla_total_size(0);
                        }
                }
                rcu_read_unlock();
        }

        return size;
}

#define RTNL_STATS_OFFLOAD_XSTATS_VALID ((1 << __IFLA_OFFLOAD_XSTATS_MAX) - 1)

static const struct nla_policy
rtnl_stats_get_policy_filters[IFLA_STATS_MAX + 1] = {
        [IFLA_STATS_LINK_OFFLOAD_XSTATS] =
                    NLA_POLICY_MASK(NLA_U32, RTNL_STATS_OFFLOAD_XSTATS_VALID),
};

static const struct nla_policy
rtnl_stats_get_policy[IFLA_STATS_GETSET_MAX + 1] = {
        [IFLA_STATS_GET_FILTERS] =
                    NLA_POLICY_NESTED(rtnl_stats_get_policy_filters),
};

static const struct nla_policy
ifla_stats_set_policy[IFLA_STATS_GETSET_MAX + 1] = {
        [IFLA_STATS_SET_OFFLOAD_XSTATS_L3_STATS] = NLA_POLICY_MAX(NLA_U8, 1),
};

static int rtnl_stats_get_parse_filters(struct nlattr *ifla_filters,
                                        struct rtnl_stats_dump_filters *filters,
                                        struct netlink_ext_ack *extack)
{
        struct nlattr *tb[IFLA_STATS_MAX + 1];
        int err;
        int at;

        err = nla_parse_nested(tb, IFLA_STATS_MAX, ifla_filters,
                               rtnl_stats_get_policy_filters, extack);
        if (err < 0)
                return err;

        for (at = 1; at <= IFLA_STATS_MAX; at++) {
                if (tb[at]) {
                        if (!(filters->mask[0] & IFLA_STATS_FILTER_BIT(at))) {
                                NL_SET_ERR_MSG(extack, "Filtered attribute not enabled in filter_mask");
                                return -EINVAL;
                        }
                        filters->mask[at] = nla_get_u32(tb[at]);
                }
        }

        return 0;
}

static int rtnl_stats_get_parse(const struct nlmsghdr *nlh,
                                u32 filter_mask,
                                struct rtnl_stats_dump_filters *filters,
                                struct netlink_ext_ack *extack)
{
        struct nlattr *tb[IFLA_STATS_GETSET_MAX + 1];
        int err;
        int i;

        filters->mask[0] = filter_mask;
        for (i = 1; i < ARRAY_SIZE(filters->mask); i++)
                filters->mask[i] = -1U;

        err = nlmsg_parse(nlh, sizeof(struct if_stats_msg), tb,
                          IFLA_STATS_GETSET_MAX, rtnl_stats_get_policy, extack);
        if (err < 0)
                return err;

        if (tb[IFLA_STATS_GET_FILTERS]) {
                err = rtnl_stats_get_parse_filters(tb[IFLA_STATS_GET_FILTERS],
                                                   filters, extack);
                if (err)
                        return err;
        }

        return 0;
}

static int rtnl_valid_stats_req(const struct nlmsghdr *nlh, bool strict_check,
                                bool is_dump, struct netlink_ext_ack *extack)
{
        struct if_stats_msg *ifsm;

        ifsm = nlmsg_payload(nlh, sizeof(*ifsm));
        if (!ifsm) {
                NL_SET_ERR_MSG(extack, "Invalid header for stats dump");
                return -EINVAL;
        }

        if (!strict_check)
                return 0;

        /* only requests using strict checks can pass data to influence
         * the dump. The legacy exception is filter_mask.
         */
        if (ifsm->pad1 || ifsm->pad2 || (is_dump && ifsm->ifindex)) {
                NL_SET_ERR_MSG(extack, "Invalid values in header for stats dump request");
                return -EINVAL;
        }
        if (ifsm->filter_mask >= IFLA_STATS_FILTER_BIT(IFLA_STATS_MAX + 1)) {
                NL_SET_ERR_MSG(extack, "Invalid stats requested through filter mask");
                return -EINVAL;
        }

        return 0;
}

static int rtnl_stats_get(struct sk_buff *skb, struct nlmsghdr *nlh,
                          struct netlink_ext_ack *extack)
{
        struct rtnl_stats_dump_filters filters;
        struct net *net = sock_net(skb->sk);
        struct net_device *dev = NULL;
        int idxattr = 0, prividx = 0;
        struct if_stats_msg *ifsm;
        struct sk_buff *nskb;
        int err;

        err = rtnl_valid_stats_req(nlh, netlink_strict_get_check(skb),
                                   false, extack);
        if (err)
                return err;

        ifsm = nlmsg_data(nlh);
        if (ifsm->ifindex > 0)
                dev = __dev_get_by_index(net, ifsm->ifindex);
        else
                return -EINVAL;

        if (!dev)
                return -ENODEV;

        if (!ifsm->filter_mask) {
                NL_SET_ERR_MSG(extack, "Filter mask must be set for stats get");
                return -EINVAL;
        }

        err = rtnl_stats_get_parse(nlh, ifsm->filter_mask, &filters, extack);
        if (err)
                return err;

        nskb = nlmsg_new(if_nlmsg_stats_size(dev, &filters), GFP_KERNEL);
        if (!nskb)
                return -ENOBUFS;

        err = rtnl_fill_statsinfo(nskb, dev, RTM_NEWSTATS,
                                  NETLINK_CB(skb).portid, nlh->nlmsg_seq, 0,
                                  0, &filters, &idxattr, &prividx, extack);
        if (err < 0) {
                /* -EMSGSIZE implies BUG in if_nlmsg_stats_size */
                WARN_ON(err == -EMSGSIZE);
                kfree_skb(nskb);
        } else {
                err = rtnl_unicast(nskb, net, NETLINK_CB(skb).portid);
        }

        return err;
}

static int rtnl_stats_dump(struct sk_buff *skb, struct netlink_callback *cb)
{
        struct netlink_ext_ack *extack = cb->extack;
        struct rtnl_stats_dump_filters filters;
        struct net *net = sock_net(skb->sk);
        unsigned int flags = NLM_F_MULTI;
        struct if_stats_msg *ifsm;
        struct {
                unsigned long ifindex;
                int idxattr;
                int prividx;
        } *ctx = (void *)cb->ctx;
        struct net_device *dev;
        int err;

        cb->seq = net->dev_base_seq;

        err = rtnl_valid_stats_req(cb->nlh, cb->strict_check, true, extack);
        if (err)
                return err;

        ifsm = nlmsg_data(cb->nlh);
        if (!ifsm->filter_mask) {
                NL_SET_ERR_MSG(extack, "Filter mask must be set for stats dump");
                return -EINVAL;
        }

        err = rtnl_stats_get_parse(cb->nlh, ifsm->filter_mask, &filters,
                                   extack);
        if (err)
                return err;

        for_each_netdev_dump(net, dev, ctx->ifindex) {
                err = rtnl_fill_statsinfo(skb, dev, RTM_NEWSTATS,
                                          NETLINK_CB(cb->skb).portid,
                                          cb->nlh->nlmsg_seq, 0,
                                          flags, &filters,
                                          &ctx->idxattr, &ctx->prividx,
                                          extack);
                /* If we ran out of room on the first message,
                 * we're in trouble.
                 */
                WARN_ON((err == -EMSGSIZE) && (skb->len == 0));

                if (err < 0)
                        break;
                ctx->prividx = 0;
                ctx->idxattr = 0;
                nl_dump_check_consistent(cb, nlmsg_hdr(skb));
        }

        return err;
}

void rtnl_offload_xstats_notify(struct net_device *dev)
{
        struct rtnl_stats_dump_filters response_filters = {};
        struct net *net = dev_net(dev);
        int idxattr = 0, prividx = 0;
        struct sk_buff *skb;
        int err = -ENOBUFS;

        ASSERT_RTNL();

        response_filters.mask[0] |=
                IFLA_STATS_FILTER_BIT(IFLA_STATS_LINK_OFFLOAD_XSTATS);
        response_filters.mask[IFLA_STATS_LINK_OFFLOAD_XSTATS] |=
                IFLA_STATS_FILTER_BIT(IFLA_OFFLOAD_XSTATS_HW_S_INFO);

        skb = nlmsg_new(if_nlmsg_stats_size(dev, &response_filters),
                        GFP_KERNEL);
        if (!skb)
                goto errout;

        err = rtnl_fill_statsinfo(skb, dev, RTM_NEWSTATS, 0, 0, 0, 0,
                                  &response_filters, &idxattr, &prividx, NULL);
        if (err < 0) {
                kfree_skb(skb);
                goto errout;
        }

        rtnl_notify(skb, net, 0, RTNLGRP_STATS, NULL, GFP_KERNEL);
        return;

errout:
        rtnl_set_sk_err(net, RTNLGRP_STATS, err);
}
EXPORT_SYMBOL(rtnl_offload_xstats_notify);

static int rtnl_stats_set(struct sk_buff *skb, struct nlmsghdr *nlh,
                          struct netlink_ext_ack *extack)
{
        enum netdev_offload_xstats_type t_l3 = NETDEV_OFFLOAD_XSTATS_TYPE_L3;
        struct rtnl_stats_dump_filters response_filters = {};
        struct nlattr *tb[IFLA_STATS_GETSET_MAX + 1];
        struct net *net = sock_net(skb->sk);
        struct net_device *dev = NULL;
        struct if_stats_msg *ifsm;
        bool notify = false;
        int err;

        err = rtnl_valid_stats_req(nlh, netlink_strict_get_check(skb),
                                   false, extack);
        if (err)
                return err;

        ifsm = nlmsg_data(nlh);
        if (ifsm->family != AF_UNSPEC) {
                NL_SET_ERR_MSG(extack, "Address family should be AF_UNSPEC");
                return -EINVAL;
        }

        if (ifsm->ifindex > 0)
                dev = __dev_get_by_index(net, ifsm->ifindex);
        else
                return -EINVAL;

        if (!dev)
                return -ENODEV;

        if (ifsm->filter_mask) {
                NL_SET_ERR_MSG(extack, "Filter mask must be 0 for stats set");
                return -EINVAL;
        }

        err = nlmsg_parse(nlh, sizeof(*ifsm), tb, IFLA_STATS_GETSET_MAX,
                          ifla_stats_set_policy, extack);
        if (err < 0)
                return err;

        if (tb[IFLA_STATS_SET_OFFLOAD_XSTATS_L3_STATS]) {
                u8 req = nla_get_u8(tb[IFLA_STATS_SET_OFFLOAD_XSTATS_L3_STATS]);

                if (req)
                        err = netdev_offload_xstats_enable(dev, t_l3, extack);
                else
                        err = netdev_offload_xstats_disable(dev, t_l3);

                if (!err)
                        notify = true;
                else if (err != -EALREADY)
                        return err;

                response_filters.mask[0] |=
                        IFLA_STATS_FILTER_BIT(IFLA_STATS_LINK_OFFLOAD_XSTATS);
                response_filters.mask[IFLA_STATS_LINK_OFFLOAD_XSTATS] |=
                        IFLA_STATS_FILTER_BIT(IFLA_OFFLOAD_XSTATS_HW_S_INFO);
        }

        if (notify)
                rtnl_offload_xstats_notify(dev);

        return 0;
}

static int rtnl_mdb_valid_dump_req(const struct nlmsghdr *nlh,
                                   struct netlink_ext_ack *extack)
{
        struct br_port_msg *bpm;

        bpm = nlmsg_payload(nlh, sizeof(*bpm));
        if (!bpm) {
                NL_SET_ERR_MSG(extack, "Invalid header for mdb dump request");
                return -EINVAL;
        }

        if (bpm->ifindex) {
                NL_SET_ERR_MSG(extack, "Filtering by device index is not supported for mdb dump request");
                return -EINVAL;
        }
        if (nlmsg_attrlen(nlh, sizeof(*bpm))) {
                NL_SET_ERR_MSG(extack, "Invalid data after header in mdb dump request");
                return -EINVAL;
        }

        return 0;
}

struct rtnl_mdb_dump_ctx {
        long idx;
};

static int rtnl_mdb_dump(struct sk_buff *skb, struct netlink_callback *cb)
{
        struct rtnl_mdb_dump_ctx *ctx = (void *)cb->ctx;
        struct net *net = sock_net(skb->sk);
        struct net_device *dev;
        int idx, s_idx;
        int err;

        NL_ASSERT_CTX_FITS(struct rtnl_mdb_dump_ctx);

        if (cb->strict_check) {
                err = rtnl_mdb_valid_dump_req(cb->nlh, cb->extack);
                if (err)
                        return err;
        }

        s_idx = ctx->idx;
        idx = 0;

        for_each_netdev(net, dev) {
                if (idx < s_idx)
                        goto skip;
                if (!dev->netdev_ops->ndo_mdb_dump)
                        goto skip;

                err = dev->netdev_ops->ndo_mdb_dump(dev, skb, cb);
                if (err == -EMSGSIZE)
                        goto out;
                /* Moving on to next device, reset markers and sequence
                 * counters since they are all maintained per-device.
                 */
                memset(cb->ctx, 0, sizeof(cb->ctx));
                cb->prev_seq = 0;
                cb->seq = 0;
skip:
                idx++;
        }

out:
        ctx->idx = idx;
        return skb->len;
}

static int rtnl_validate_mdb_entry_get(const struct nlattr *attr,
                                       struct netlink_ext_ack *extack)
{
        struct br_mdb_entry *entry = nla_data(attr);

        if (nla_len(attr) != sizeof(struct br_mdb_entry)) {
                NL_SET_ERR_MSG_ATTR(extack, attr, "Invalid attribute length");
                return -EINVAL;
        }

        if (entry->ifindex) {
                NL_SET_ERR_MSG(extack, "Entry ifindex cannot be specified");
                return -EINVAL;
        }

        if (entry->state) {
                NL_SET_ERR_MSG(extack, "Entry state cannot be specified");
                return -EINVAL;
        }

        if (entry->flags) {
                NL_SET_ERR_MSG(extack, "Entry flags cannot be specified");
                return -EINVAL;
        }

        if (entry->vid >= VLAN_VID_MASK) {
                NL_SET_ERR_MSG(extack, "Invalid entry VLAN id");
                return -EINVAL;
        }

        if (entry->addr.proto != htons(ETH_P_IP) &&
            entry->addr.proto != htons(ETH_P_IPV6) &&
            entry->addr.proto != 0) {
                NL_SET_ERR_MSG(extack, "Unknown entry protocol");
                return -EINVAL;
        }

        return 0;
}

static const struct nla_policy mdba_get_policy[MDBA_GET_ENTRY_MAX + 1] = {
        [MDBA_GET_ENTRY] = NLA_POLICY_VALIDATE_FN(NLA_BINARY,
                                                  rtnl_validate_mdb_entry_get,
                                                  sizeof(struct br_mdb_entry)),
        [MDBA_GET_ENTRY_ATTRS] = { .type = NLA_NESTED },
};

static int rtnl_mdb_get(struct sk_buff *in_skb, struct nlmsghdr *nlh,
                        struct netlink_ext_ack *extack)
{
        struct nlattr *tb[MDBA_GET_ENTRY_MAX + 1];
        struct net *net = sock_net(in_skb->sk);
        struct br_port_msg *bpm;
        struct net_device *dev;
        int err;

        err = nlmsg_parse(nlh, sizeof(struct br_port_msg), tb,
                          MDBA_GET_ENTRY_MAX, mdba_get_policy, extack);
        if (err)
                return err;

        bpm = nlmsg_data(nlh);
        if (!bpm->ifindex) {
                NL_SET_ERR_MSG(extack, "Invalid ifindex");
                return -EINVAL;
        }

        dev = __dev_get_by_index(net, bpm->ifindex);
        if (!dev) {
                NL_SET_ERR_MSG(extack, "Device doesn't exist");
                return -ENODEV;
        }

        if (NL_REQ_ATTR_CHECK(extack, NULL, tb, MDBA_GET_ENTRY)) {
                NL_SET_ERR_MSG(extack, "Missing MDBA_GET_ENTRY attribute");
                return -EINVAL;
        }

        if (!dev->netdev_ops->ndo_mdb_get) {
                NL_SET_ERR_MSG(extack, "Device does not support MDB operations");
                return -EOPNOTSUPP;
        }

        return dev->netdev_ops->ndo_mdb_get(dev, tb, NETLINK_CB(in_skb).portid,
                                            nlh->nlmsg_seq, extack);
}

static int rtnl_validate_mdb_entry(const struct nlattr *attr,
                                   struct netlink_ext_ack *extack)
{
        struct br_mdb_entry *entry = nla_data(attr);

        if (nla_len(attr) != sizeof(struct br_mdb_entry)) {
                NL_SET_ERR_MSG_ATTR(extack, attr, "Invalid attribute length");
                return -EINVAL;
        }

        if (entry->ifindex == 0) {
                NL_SET_ERR_MSG(extack, "Zero entry ifindex is not allowed");
                return -EINVAL;
        }

        if (entry->addr.proto == htons(ETH_P_IP)) {
                if (!ipv4_is_multicast(entry->addr.u.ip4) &&
                    !ipv4_is_zeronet(entry->addr.u.ip4)) {
                        NL_SET_ERR_MSG(extack, "IPv4 entry group address is not multicast or 0.0.0.0");
                        return -EINVAL;
                }
                if (ipv4_is_local_multicast(entry->addr.u.ip4)) {
                        NL_SET_ERR_MSG(extack, "IPv4 entry group address is local multicast");
                        return -EINVAL;
                }
#if IS_ENABLED(CONFIG_IPV6)
        } else if (entry->addr.proto == htons(ETH_P_IPV6)) {
                if (ipv6_addr_is_ll_all_nodes(&entry->addr.u.ip6)) {
                        NL_SET_ERR_MSG(extack, "IPv6 entry group address is link-local all nodes");
                        return -EINVAL;
                }
#endif
        } else if (entry->addr.proto == 0) {
                /* L2 mdb */
                if (!is_multicast_ether_addr(entry->addr.u.mac_addr)) {
                        NL_SET_ERR_MSG(extack, "L2 entry group is not multicast");
                        return -EINVAL;
                }
        } else {
                NL_SET_ERR_MSG(extack, "Unknown entry protocol");
                return -EINVAL;
        }

        if (entry->state != MDB_PERMANENT && entry->state != MDB_TEMPORARY) {
                NL_SET_ERR_MSG(extack, "Unknown entry state");
                return -EINVAL;
        }
        if (entry->vid >= VLAN_VID_MASK) {
                NL_SET_ERR_MSG(extack, "Invalid entry VLAN id");
                return -EINVAL;
        }

        return 0;
}

static const struct nla_policy mdba_policy[MDBA_SET_ENTRY_MAX + 1] = {
        [MDBA_SET_ENTRY_UNSPEC] = { .strict_start_type = MDBA_SET_ENTRY_ATTRS + 1 },
        [MDBA_SET_ENTRY] = NLA_POLICY_VALIDATE_FN(NLA_BINARY,
                                                  rtnl_validate_mdb_entry,
                                                  sizeof(struct br_mdb_entry)),
        [MDBA_SET_ENTRY_ATTRS] = { .type = NLA_NESTED },
};

static int rtnl_mdb_add(struct sk_buff *skb, struct nlmsghdr *nlh,
                        struct netlink_ext_ack *extack)
{
        struct nlattr *tb[MDBA_SET_ENTRY_MAX + 1];
        struct net *net = sock_net(skb->sk);
        struct br_port_msg *bpm;
        struct net_device *dev;
        int err;

        err = nlmsg_parse_deprecated(nlh, sizeof(*bpm), tb,
                                     MDBA_SET_ENTRY_MAX, mdba_policy, extack);
        if (err)
                return err;

        bpm = nlmsg_data(nlh);
        if (!bpm->ifindex) {
                NL_SET_ERR_MSG(extack, "Invalid ifindex");
                return -EINVAL;
        }

        dev = __dev_get_by_index(net, bpm->ifindex);
        if (!dev) {
                NL_SET_ERR_MSG(extack, "Device doesn't exist");
                return -ENODEV;
        }

        if (NL_REQ_ATTR_CHECK(extack, NULL, tb, MDBA_SET_ENTRY)) {
                NL_SET_ERR_MSG(extack, "Missing MDBA_SET_ENTRY attribute");
                return -EINVAL;
        }

        if (!dev->netdev_ops->ndo_mdb_add) {
                NL_SET_ERR_MSG(extack, "Device does not support MDB operations");
                return -EOPNOTSUPP;
        }

        return dev->netdev_ops->ndo_mdb_add(dev, tb, nlh->nlmsg_flags, extack);
}

static int rtnl_validate_mdb_entry_del_bulk(const struct nlattr *attr,
                                            struct netlink_ext_ack *extack)
{
        struct br_mdb_entry *entry = nla_data(attr);
        struct br_mdb_entry zero_entry = {};

        if (nla_len(attr) != sizeof(struct br_mdb_entry)) {
                NL_SET_ERR_MSG_ATTR(extack, attr, "Invalid attribute length");
                return -EINVAL;
        }

        if (entry->state != MDB_PERMANENT && entry->state != MDB_TEMPORARY) {
                NL_SET_ERR_MSG(extack, "Unknown entry state");
                return -EINVAL;
        }

        if (entry->flags) {
                NL_SET_ERR_MSG(extack, "Entry flags cannot be set");
                return -EINVAL;
        }

        if (entry->vid >= VLAN_N_VID - 1) {
                NL_SET_ERR_MSG(extack, "Invalid entry VLAN id");
                return -EINVAL;
        }

        if (memcmp(&entry->addr, &zero_entry.addr, sizeof(entry->addr))) {
                NL_SET_ERR_MSG(extack, "Entry address cannot be set");
                return -EINVAL;
        }

        return 0;
}

static const struct nla_policy mdba_del_bulk_policy[MDBA_SET_ENTRY_MAX + 1] = {
        [MDBA_SET_ENTRY] = NLA_POLICY_VALIDATE_FN(NLA_BINARY,
                                                  rtnl_validate_mdb_entry_del_bulk,
                                                  sizeof(struct br_mdb_entry)),
        [MDBA_SET_ENTRY_ATTRS] = { .type = NLA_NESTED },
};

static int rtnl_mdb_del(struct sk_buff *skb, struct nlmsghdr *nlh,
                        struct netlink_ext_ack *extack)
{
        bool del_bulk = !!(nlh->nlmsg_flags & NLM_F_BULK);
        struct nlattr *tb[MDBA_SET_ENTRY_MAX + 1];
        struct net *net = sock_net(skb->sk);
        struct br_port_msg *bpm;
        struct net_device *dev;
        int err;

        if (!del_bulk)
                err = nlmsg_parse_deprecated(nlh, sizeof(*bpm), tb,
                                             MDBA_SET_ENTRY_MAX, mdba_policy,
                                             extack);
        else
                err = nlmsg_parse(nlh, sizeof(*bpm), tb, MDBA_SET_ENTRY_MAX,
                                  mdba_del_bulk_policy, extack);
        if (err)
                return err;

        bpm = nlmsg_data(nlh);
        if (!bpm->ifindex) {
                NL_SET_ERR_MSG(extack, "Invalid ifindex");
                return -EINVAL;
        }

        dev = __dev_get_by_index(net, bpm->ifindex);
        if (!dev) {
                NL_SET_ERR_MSG(extack, "Device doesn't exist");
                return -ENODEV;
        }

        if (NL_REQ_ATTR_CHECK(extack, NULL, tb, MDBA_SET_ENTRY)) {
                NL_SET_ERR_MSG(extack, "Missing MDBA_SET_ENTRY attribute");
                return -EINVAL;
        }

        if (del_bulk) {
                if (!dev->netdev_ops->ndo_mdb_del_bulk) {
                        NL_SET_ERR_MSG(extack, "Device does not support MDB bulk deletion");
                        return -EOPNOTSUPP;
                }
                return dev->netdev_ops->ndo_mdb_del_bulk(dev, tb, extack);
        }

        if (!dev->netdev_ops->ndo_mdb_del) {
                NL_SET_ERR_MSG(extack, "Device does not support MDB operations");
                return -EOPNOTSUPP;
        }

        return dev->netdev_ops->ndo_mdb_del(dev, tb, extack);
}

/* Process one rtnetlink message. */

static int rtnl_dumpit(struct sk_buff *skb, struct netlink_callback *cb)
{
        const bool needs_lock = !(cb->flags & RTNL_FLAG_DUMP_UNLOCKED);
        rtnl_dumpit_func dumpit = cb->data;
        int err;

        /* Previous iteration have already finished, avoid calling->dumpit()
         * again, it may not expect to be called after it reached the end.
         */
        if (!dumpit)
                return 0;

        if (needs_lock)
                rtnl_lock();
        err = dumpit(skb, cb);
        if (needs_lock)
                rtnl_unlock();

        /* Old dump handlers used to send NLM_DONE as in a separate recvmsg().
         * Some applications which parse netlink manually depend on this.
         */
        if (cb->flags & RTNL_FLAG_DUMP_SPLIT_NLM_DONE) {
                if (err < 0 && err != -EMSGSIZE)
                        return err;
                if (!err)
                        cb->data = NULL;

                return skb->len;
        }
        return err;
}

static int rtnetlink_dump_start(struct sock *ssk, struct sk_buff *skb,
                                const struct nlmsghdr *nlh,
                                struct netlink_dump_control *control)
{
        if (control->flags & RTNL_FLAG_DUMP_SPLIT_NLM_DONE ||
            !(control->flags & RTNL_FLAG_DUMP_UNLOCKED)) {
                WARN_ON(control->data);
                control->data = control->dump;
                control->dump = rtnl_dumpit;
        }

        return netlink_dump_start(ssk, skb, nlh, control);
}

static int rtnetlink_rcv_msg(struct sk_buff *skb, struct nlmsghdr *nlh,
                             struct netlink_ext_ack *extack)
{
        struct net *net = sock_net(skb->sk);
        struct rtnl_link *link;
        enum rtnl_kinds kind;
        struct module *owner;
        int err = -EOPNOTSUPP;
        rtnl_doit_func doit;
        unsigned int flags;
        int family;
        int type;

        type = nlh->nlmsg_type;
        if (type > RTM_MAX)
                return -EOPNOTSUPP;

        type -= RTM_BASE;

        /* All the messages must have at least 1 byte length */
        if (nlmsg_len(nlh) < sizeof(struct rtgenmsg))
                return 0;

        family = ((struct rtgenmsg *)nlmsg_data(nlh))->rtgen_family;
        kind = rtnl_msgtype_kind(type);

        if (kind != RTNL_KIND_GET && !netlink_net_capable(skb, CAP_NET_ADMIN))
                return -EPERM;

        rcu_read_lock();
        if (kind == RTNL_KIND_GET && (nlh->nlmsg_flags & NLM_F_DUMP)) {
                struct sock *rtnl;
                rtnl_dumpit_func dumpit;
                u32 min_dump_alloc = 0;

                link = rtnl_get_link(family, type);
                if (!link || !link->dumpit) {
                        family = PF_UNSPEC;
                        link = rtnl_get_link(family, type);
                        if (!link || !link->dumpit)
                                goto err_unlock;
                }
                owner = link->owner;
                dumpit = link->dumpit;
                flags = link->flags;

                if (type == RTM_GETLINK - RTM_BASE)
                        min_dump_alloc = rtnl_calcit(skb, nlh);

                err = 0;
                /* need to do this before rcu_read_unlock() */
                if (!try_module_get(owner))
                        err = -EPROTONOSUPPORT;

                rcu_read_unlock();

                rtnl = net->rtnl;
                if (err == 0) {
                        struct netlink_dump_control c = {
                                .dump           = dumpit,
                                .min_dump_alloc = min_dump_alloc,
                                .module         = owner,
                                .flags          = flags,
                        };
                        err = rtnetlink_dump_start(rtnl, skb, nlh, &c);
                        /* netlink_dump_start() will keep a reference on
                         * module if dump is still in progress.
                         */
                        module_put(owner);
                }
                return err;
        }

        link = rtnl_get_link(family, type);
        if (!link || !link->doit) {
                family = PF_UNSPEC;
                link = rtnl_get_link(PF_UNSPEC, type);
                if (!link || !link->doit)
                        goto out_unlock;
        }

        owner = link->owner;
        if (!try_module_get(owner)) {
                err = -EPROTONOSUPPORT;
                goto out_unlock;
        }

        flags = link->flags;
        if (kind == RTNL_KIND_DEL && (nlh->nlmsg_flags & NLM_F_BULK) &&
            !(flags & RTNL_FLAG_BULK_DEL_SUPPORTED)) {
                NL_SET_ERR_MSG(extack, "Bulk delete is not supported");
                module_put(owner);
                goto err_unlock;
        }

        if (flags & RTNL_FLAG_DOIT_UNLOCKED) {
                doit = link->doit;
                rcu_read_unlock();
                if (doit)
                        err = doit(skb, nlh, extack);
                module_put(owner);
                return err;
        }
        rcu_read_unlock();

        rtnl_lock();
        link = rtnl_get_link(family, type);
        if (link && link->doit)
                err = link->doit(skb, nlh, extack);
        rtnl_unlock();

        module_put(owner);

        return err;

out_unlock:
        rcu_read_unlock();
        return err;

err_unlock:
        rcu_read_unlock();
        return -EOPNOTSUPP;
}

static void rtnetlink_rcv(struct sk_buff *skb)
{
        netlink_rcv_skb(skb, &rtnetlink_rcv_msg);
}

static int rtnetlink_bind(struct net *net, int group)
{
        switch (group) {
        case RTNLGRP_IPV4_MROUTE_R:
        case RTNLGRP_IPV6_MROUTE_R:
                if (!ns_capable(net->user_ns, CAP_NET_ADMIN))
                        return -EPERM;
                break;
        }
        return 0;
}

static int rtnetlink_event(struct notifier_block *this, unsigned long event, void *ptr)
{
        struct net_device *dev = netdev_notifier_info_to_dev(ptr);

        switch (event) {
        case NETDEV_REBOOT:
        case NETDEV_CHANGEMTU:
        case NETDEV_CHANGEADDR:
        case NETDEV_CHANGENAME:
        case NETDEV_FEAT_CHANGE:
        case NETDEV_BONDING_FAILOVER:
        case NETDEV_POST_TYPE_CHANGE:
        case NETDEV_NOTIFY_PEERS:
        case NETDEV_CHANGEUPPER:
        case NETDEV_RESEND_IGMP:
        case NETDEV_CHANGEINFODATA:
        case NETDEV_CHANGELOWERSTATE:
        case NETDEV_CHANGE_TX_QUEUE_LEN:
                rtmsg_ifinfo_event(RTM_NEWLINK, dev, 0, rtnl_get_event(event),
                                   GFP_KERNEL, NULL, 0, 0, NULL);
                break;
        default:
                break;
        }
        return NOTIFY_DONE;
}

static struct notifier_block rtnetlink_dev_notifier = {
        .notifier_call  = rtnetlink_event,
};


static int __net_init rtnetlink_net_init(struct net *net)
{
        struct sock *sk;
        struct netlink_kernel_cfg cfg = {
                .groups         = RTNLGRP_MAX,
                .input          = rtnetlink_rcv,
                .flags          = NL_CFG_F_NONROOT_RECV,
                .bind           = rtnetlink_bind,
        };

        sk = netlink_kernel_create(net, NETLINK_ROUTE, &cfg);
        if (!sk)
                return -ENOMEM;
        net->rtnl = sk;
        return 0;
}

static void __net_exit rtnetlink_net_exit(struct net *net)
{
        netlink_kernel_release(net->rtnl);
        net->rtnl = NULL;
}

static struct pernet_operations rtnetlink_net_ops = {
        .init = rtnetlink_net_init,
        .exit = rtnetlink_net_exit,
};

static const struct rtnl_msg_handler rtnetlink_rtnl_msg_handlers[] __initconst = {
        {.msgtype = RTM_NEWLINK, .doit = rtnl_newlink,
         .flags = RTNL_FLAG_DOIT_PERNET},
        {.msgtype = RTM_DELLINK, .doit = rtnl_dellink,
         .flags = RTNL_FLAG_DOIT_PERNET_WIP},
        {.msgtype = RTM_GETLINK, .doit = rtnl_getlink,
         .dumpit = rtnl_dump_ifinfo, .flags = RTNL_FLAG_DUMP_SPLIT_NLM_DONE},
        {.msgtype = RTM_SETLINK, .doit = rtnl_setlink,
         .flags = RTNL_FLAG_DOIT_PERNET_WIP},
        {.msgtype = RTM_GETADDR, .dumpit = rtnl_dump_all},
        {.msgtype = RTM_GETROUTE, .dumpit = rtnl_dump_all},
        {.msgtype = RTM_GETNETCONF, .dumpit = rtnl_dump_all},
        {.msgtype = RTM_GETSTATS, .doit = rtnl_stats_get,
         .dumpit = rtnl_stats_dump},
        {.msgtype = RTM_SETSTATS, .doit = rtnl_stats_set},
        {.msgtype = RTM_NEWLINKPROP, .doit = rtnl_newlinkprop},
        {.msgtype = RTM_DELLINKPROP, .doit = rtnl_dellinkprop},
        {.protocol = PF_BRIDGE, .msgtype = RTM_GETLINK,
         .dumpit = rtnl_bridge_getlink},
        {.protocol = PF_BRIDGE, .msgtype = RTM_DELLINK,
         .doit = rtnl_bridge_dellink},
        {.protocol = PF_BRIDGE, .msgtype = RTM_SETLINK,
         .doit = rtnl_bridge_setlink},
        {.protocol = PF_BRIDGE, .msgtype = RTM_NEWNEIGH, .doit = rtnl_fdb_add},
        {.protocol = PF_BRIDGE, .msgtype = RTM_DELNEIGH, .doit = rtnl_fdb_del,
         .flags = RTNL_FLAG_BULK_DEL_SUPPORTED},
        {.protocol = PF_BRIDGE, .msgtype = RTM_GETNEIGH, .doit = rtnl_fdb_get,
         .dumpit = rtnl_fdb_dump},
        {.protocol = PF_BRIDGE, .msgtype = RTM_NEWMDB, .doit = rtnl_mdb_add},
        {.protocol = PF_BRIDGE, .msgtype = RTM_DELMDB, .doit = rtnl_mdb_del,
         .flags = RTNL_FLAG_BULK_DEL_SUPPORTED},
        {.protocol = PF_BRIDGE, .msgtype = RTM_GETMDB, .doit = rtnl_mdb_get,
         .dumpit = rtnl_mdb_dump},
};

void __init rtnetlink_init(void)
{
        if (register_pernet_subsys(&rtnetlink_net_ops))
                panic("rtnetlink_init: cannot initialize rtnetlink\n");

        register_netdevice_notifier(&rtnetlink_dev_notifier);

        rtnl_register_many(rtnetlink_rtnl_msg_handlers);
}