root/sys/netlink/netlink_domain.c
/*-
 * SPDX-License-Identifier: BSD-2-Clause
 *
 * Copyright (c) 2021 Ng Peng Nam Sean
 * Copyright (c) 2022 Alexander V. Chernikov <melifaro@FreeBSD.org>
 * Copyright (c) 2023 Gleb Smirnoff <glebius@FreeBSD.org>
 *
 * Redistribution and use in source and binary forms, with or without
 * modification, are permitted provided that the following conditions
 * are met:
 * 1. Redistributions of source code must retain the above copyright
 *    notice, this list of conditions and the following disclaimer.
 * 2. Redistributions in binary form must reproduce the above copyright
 *    notice, this list of conditions and the following disclaimer in the
 *    documentation and/or other materials provided with the distribution.
 *
 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
 * SUCH DAMAGE.
 */

/*
 * This file contains socket and protocol bindings for netlink.
 */

#include <sys/param.h>
#include <sys/kernel.h>
#include <sys/malloc.h>
#include <sys/lock.h>
#include <sys/rmlock.h>
#include <sys/domain.h>
#include <sys/jail.h>
#include <sys/mbuf.h>
#include <sys/osd.h>
#include <sys/protosw.h>
#include <sys/proc.h>
#include <sys/ck.h>
#include <sys/socket.h>
#include <sys/socketvar.h>
#include <sys/sysent.h>
#include <sys/syslog.h>
#include <sys/priv.h>
#include <sys/uio.h>

#include <netlink/netlink.h>
#include <netlink/netlink_ctl.h>
#include <netlink/netlink_var.h>

#define DEBUG_MOD_NAME  nl_domain
#define DEBUG_MAX_LEVEL LOG_DEBUG3
#include <netlink/netlink_debug.h>
_DECLARE_DEBUG(LOG_INFO);

_Static_assert((NLP_MAX_GROUPS % 64) == 0,
    "NLP_MAX_GROUPS has to be multiple of 64");
_Static_assert(NLP_MAX_GROUPS >= 64,
    "NLP_MAX_GROUPS has to be at least 64");

#define NLCTL_TRACKER           struct rm_priotracker nl_tracker
#define NLCTL_RLOCK()           rm_rlock(&V_nl_ctl.ctl_lock, &nl_tracker)
#define NLCTL_RUNLOCK()         rm_runlock(&V_nl_ctl.ctl_lock, &nl_tracker)
#define NLCTL_LOCK_ASSERT()     rm_assert(&V_nl_ctl.ctl_lock, RA_LOCKED)

#define NLCTL_WLOCK()           rm_wlock(&V_nl_ctl.ctl_lock)
#define NLCTL_WUNLOCK()         rm_wunlock(&V_nl_ctl.ctl_lock)
#define NLCTL_WLOCK_ASSERT()    rm_assert(&V_nl_ctl.ctl_lock, RA_WLOCKED)

static u_long nl_sendspace = NLSNDQ;
SYSCTL_ULONG(_net_netlink, OID_AUTO, sendspace, CTLFLAG_RW, &nl_sendspace, 0,
    "Default netlink socket send space");

static u_long nl_recvspace = NLSNDQ;
SYSCTL_ULONG(_net_netlink, OID_AUTO, recvspace, CTLFLAG_RW, &nl_recvspace, 0,
    "Default netlink socket receive space");

extern u_long sb_max_adj;
static u_long nl_maxsockbuf = 512 * 1024 * 1024; /* 512M, XXX: init based on physmem */
static int sysctl_handle_nl_maxsockbuf(SYSCTL_HANDLER_ARGS);
SYSCTL_OID(_net_netlink, OID_AUTO, nl_maxsockbuf,
    CTLTYPE_ULONG | CTLFLAG_RW | CTLFLAG_MPSAFE, &nl_maxsockbuf, 0,
    sysctl_handle_nl_maxsockbuf, "LU",
    "Maximum Netlink socket buffer size");


static unsigned int osd_slot_id = 0;

void
nl_osd_register(void)
{
        osd_slot_id = osd_register(OSD_THREAD, NULL, NULL);
}

void
nl_osd_unregister(void)
{
        osd_deregister(OSD_THREAD, osd_slot_id);
}

struct nlpcb *
_nl_get_thread_nlp(struct thread *td)
{
        return (osd_get(OSD_THREAD, &td->td_osd, osd_slot_id));
}

void
nl_set_thread_nlp(struct thread *td, struct nlpcb *nlp)
{
        NLP_LOG(LOG_DEBUG2, nlp, "Set thread %p nlp to %p (slot %u)", td, nlp, osd_slot_id);
        if (osd_set(OSD_THREAD, &td->td_osd, osd_slot_id, nlp) == 0)
                return;
        /* Failed, need to realloc */
        void **rsv = osd_reserve(osd_slot_id);
        osd_set_reserved(OSD_THREAD, &td->td_osd, osd_slot_id, rsv, nlp);
}

/*
 * Looks up a nlpcb struct based on the @portid. Need to claim nlsock_mtx.
 * Returns nlpcb pointer if present else NULL
 */
static struct nlpcb *
nl_port_lookup(uint32_t port_id)
{
        struct nlpcb *nlp;

        CK_LIST_FOREACH(nlp, &V_nl_ctl.ctl_port_head, nl_port_next) {
                if (nlp->nl_port == port_id)
                        return (nlp);
        }
        return (NULL);
}

static void
nlp_join_group(struct nlpcb *nlp, unsigned int group_id)
{
        MPASS(group_id < NLP_MAX_GROUPS);
        NLCTL_WLOCK_ASSERT();

        /* TODO: add family handler callback */
        if (!nlp_unconstrained_vnet(nlp))
                return;

        BIT_SET(NLP_MAX_GROUPS, group_id, &nlp->nl_groups);
}

static void
nlp_leave_group(struct nlpcb *nlp, unsigned int group_id)
{
        MPASS(group_id < NLP_MAX_GROUPS);
        NLCTL_WLOCK_ASSERT();

        BIT_CLR(NLP_MAX_GROUPS, group_id, &nlp->nl_groups);
}

static bool
nlp_memberof_group(struct nlpcb *nlp, unsigned int group_id)
{
        MPASS(group_id < NLP_MAX_GROUPS);
        NLCTL_LOCK_ASSERT();

        return (BIT_ISSET(NLP_MAX_GROUPS, group_id, &nlp->nl_groups));
}

static uint32_t
nlp_get_groups_compat(struct nlpcb *nlp)
{
        uint32_t groups_mask = 0;

        NLCTL_LOCK_ASSERT();

        for (int i = 0; i < 32; i++) {
                if (nlp_memberof_group(nlp, i + 1))
                        groups_mask |= (1 << i);
        }

        return (groups_mask);
}

static struct nl_buf *
nl_buf_copy(struct nl_buf *nb)
{
        struct nl_buf *copy;

        copy = nl_buf_alloc(nb->buflen, M_NOWAIT);
        if (__predict_false(copy == NULL))
                return (NULL);
        memcpy(copy, nb, sizeof(*nb) + nb->buflen);

        return (copy);
}

/*
 * Broadcasts in the writer's buffer.
 */
bool
nl_send_group(struct nl_writer *nw)
{
        struct nl_buf *nb = nw->buf;
        struct nlpcb *nlp_last = NULL;
        struct nlpcb *nlp;
        NLCTL_TRACKER;

        IF_DEBUG_LEVEL(LOG_DEBUG2) {
                struct nlmsghdr *hdr = (struct nlmsghdr *)nb->data;
                NL_LOG(LOG_DEBUG2, "MCAST len %u msg type %d len %u to group %d/%d",
                    nb->datalen, hdr->nlmsg_type, hdr->nlmsg_len,
                    nw->group.proto, nw->group.id);
        }

        nw->buf = NULL;

        NLCTL_RLOCK();
        CK_LIST_FOREACH(nlp, &V_nl_ctl.ctl_pcb_head, nl_next) {
                if ((nw->group.priv == 0 || priv_check_cred(
                    nlp->nl_socket->so_cred, nw->group.priv) == 0) &&
                    nlp->nl_proto == nw->group.proto &&
                    nlp_memberof_group(nlp, nw->group.id)) {
                        if (nlp_last != NULL) {
                                struct nl_buf *copy;

                                copy = nl_buf_copy(nb);
                                if (copy != NULL) {
                                        nw->buf = copy;
                                        (void)nl_send(nw, nlp_last);
                                } else {
                                        NLP_LOCK(nlp_last);
                                        if (nlp_last->nl_socket != NULL)
                                                sorwakeup(nlp_last->nl_socket);
                                        NLP_UNLOCK(nlp_last);
                                }
                        }
                        nlp_last = nlp;
                }
        }
        if (nlp_last != NULL) {
                nw->buf = nb;
                (void)nl_send(nw, nlp_last);
        } else
                nl_buf_free(nb);

        NLCTL_RUNLOCK();

        return (true);
}

void
nl_clear_group(u_int group)
{
        struct nlpcb *nlp;

        NLCTL_WLOCK();
        CK_LIST_FOREACH(nlp, &V_nl_ctl.ctl_pcb_head, nl_next)
                if (nlp_memberof_group(nlp, group))
                        nlp_leave_group(nlp, group);
        NLCTL_WUNLOCK();
}

static uint32_t
nl_find_port(void)
{
        /*
         * app can open multiple netlink sockets.
         * Start with current pid, if already taken,
         * try random numbers in 65k..256k+65k space,
         * avoiding clash with pids.
         */
        if (nl_port_lookup(curproc->p_pid) == NULL)
                return (curproc->p_pid);
        for (int i = 0; i < 16; i++) {
                uint32_t nl_port = (arc4random() % 65536) + 65536 * 4;
                if (nl_port_lookup(nl_port) == 0)
                        return (nl_port);
                NL_LOG(LOG_DEBUG3, "tried %u\n", nl_port);
        }
        return (curproc->p_pid);
}

static int
nl_bind_locked(struct nlpcb *nlp, struct sockaddr_nl *snl)
{
        if (nlp->nl_bound) {
                if (nlp->nl_port != snl->nl_pid) {
                        NL_LOG(LOG_DEBUG,
                            "bind() failed: program pid %d "
                            "is different from provided pid %d",
                            nlp->nl_port, snl->nl_pid);
                        return (EINVAL); // XXX: better error
                }
        } else {
                if (snl->nl_pid == 0)
                        snl->nl_pid = nl_find_port();
                if (nl_port_lookup(snl->nl_pid) != NULL)
                        return (EADDRINUSE);
                nlp->nl_port = snl->nl_pid;
                nlp->nl_bound = true;
                CK_LIST_INSERT_HEAD(&V_nl_ctl.ctl_port_head, nlp, nl_port_next);
        }
        for (int i = 0; i < 32; i++) {
                if (snl->nl_groups & ((uint32_t)1 << i))
                        nlp_join_group(nlp, i + 1);
                else
                        nlp_leave_group(nlp, i + 1);
        }

        return (0);
}

static int
nl_attach(struct socket *so, int proto, struct thread *td)
{
        struct nlpcb *nlp;
        int error;

        if (__predict_false(netlink_unloading != 0))
                return (EAFNOSUPPORT);

        error = nl_verify_proto(proto);
        if (error != 0)
                return (error);

        bool is_linux = SV_PROC_ABI(td->td_proc) == SV_ABI_LINUX;
        NL_LOG(LOG_DEBUG2, "socket %p, %sPID %d: attaching socket to %s",
            so, is_linux ? "(linux) " : "", curproc->p_pid,
            nl_get_proto_name(proto));

        mtx_init(&so->so_snd_mtx, "netlink so_snd", NULL, MTX_DEF);
        mtx_init(&so->so_rcv_mtx, "netlink so_rcv", NULL, MTX_DEF);
        error = soreserve(so, nl_sendspace, nl_recvspace);
        if (error != 0) {
                mtx_destroy(&so->so_snd_mtx);
                mtx_destroy(&so->so_rcv_mtx);
                return (error);
        }
        TAILQ_INIT(&so->so_rcv.nl_queue);
        TAILQ_INIT(&so->so_snd.nl_queue);
        nlp = malloc(sizeof(struct nlpcb), M_PCB, M_WAITOK | M_ZERO);
        so->so_pcb = nlp;
        nlp->nl_socket = so;
        nlp->nl_proto = proto;
        nlp->nl_process_id = curproc->p_pid;
        nlp->nl_linux = is_linux;
        nlp->nl_unconstrained_vnet = !jailed_without_vnet(so->so_cred);
        nlp->nl_need_thread_setup = true;
        NLP_LOCK_INIT(nlp);
        refcount_init(&nlp->nl_refcount, 1);

        nlp->nl_taskqueue = taskqueue_create("netlink_socket", M_WAITOK,
            taskqueue_thread_enqueue, &nlp->nl_taskqueue);
        TASK_INIT(&nlp->nl_task, 0, nl_taskqueue_handler, nlp);
        taskqueue_start_threads(&nlp->nl_taskqueue, 1, PWAIT,
            "netlink_socket (PID %u)", nlp->nl_process_id);

        NLCTL_WLOCK();
        CK_LIST_INSERT_HEAD(&V_nl_ctl.ctl_pcb_head, nlp, nl_next);
        NLCTL_WUNLOCK();

        soisconnected(so);

        return (0);
}

static int
nl_bind(struct socket *so, struct sockaddr *sa, struct thread *td)
{
        struct nlpcb *nlp = sotonlpcb(so);
        struct sockaddr_nl *snl = (struct sockaddr_nl *)sa;
        int error;

        NL_LOG(LOG_DEBUG3, "socket %p, PID %d", so, curproc->p_pid);
        if (snl->nl_len != sizeof(*snl)) {
                NL_LOG(LOG_DEBUG, "socket %p, wrong sizeof(), ignoring bind()", so);
                return (EINVAL);
        }


        NLCTL_WLOCK();
        NLP_LOCK(nlp);
        error = nl_bind_locked(nlp, snl);
        NLP_UNLOCK(nlp);
        NLCTL_WUNLOCK();
        NL_LOG(LOG_DEBUG2, "socket %p, bind() to %u, groups %u, error %d", so,
            snl->nl_pid, snl->nl_groups, error);

        return (error);
}


static int
nl_assign_port(struct nlpcb *nlp, uint32_t port_id)
{
        struct sockaddr_nl snl = {
                .nl_pid = port_id,
        };
        int error;

        NLCTL_WLOCK();
        NLP_LOCK(nlp);
        snl.nl_groups = nlp_get_groups_compat(nlp);
        error = nl_bind_locked(nlp, &snl);
        NLP_UNLOCK(nlp);
        NLCTL_WUNLOCK();

        NL_LOG(LOG_DEBUG3, "socket %p, port assign: %d, error: %d", nlp->nl_socket, port_id, error);
        return (error);
}

/*
 * nl_autobind_port binds a unused portid to @nlp
 * @nlp: pcb data for the netlink socket
 * @candidate_id: first id to consider
 */
static int
nl_autobind_port(struct nlpcb *nlp, uint32_t candidate_id)
{
        uint32_t port_id = candidate_id;
        NLCTL_TRACKER;
        bool exist;
        int error = EADDRINUSE;

        for (int i = 0; i < 10; i++) {
                NL_LOG(LOG_DEBUG3, "socket %p, trying to assign port %d", nlp->nl_socket, port_id);
                NLCTL_RLOCK();
                exist = nl_port_lookup(port_id) != 0;
                NLCTL_RUNLOCK();
                if (!exist) {
                        error = nl_assign_port(nlp, port_id);
                        if (error != EADDRINUSE)
                                break;
                }
                port_id++;
        }
        NL_LOG(LOG_DEBUG3, "socket %p, autobind to %d, error: %d", nlp->nl_socket, port_id, error);
        return (error);
}

static int
nl_connect(struct socket *so, struct sockaddr *sa, struct thread *td)
{
        struct sockaddr_nl *snl = (struct sockaddr_nl *)sa;
        struct nlpcb *nlp;

        NL_LOG(LOG_DEBUG3, "socket %p, PID %d", so, curproc->p_pid);
        if (snl->nl_len != sizeof(*snl)) {
                NL_LOG(LOG_DEBUG, "socket %p, wrong sizeof(), ignoring bind()", so);
                return (EINVAL);
        }

        nlp = sotonlpcb(so);
        if (!nlp->nl_bound) {
                int error = nl_autobind_port(nlp, td->td_proc->p_pid);
                if (error != 0) {
                        NL_LOG(LOG_DEBUG, "socket %p, nl_autobind() failed: %d", so, error);
                        return (error);
                }
        }
        /* XXX: Handle socket flags & multicast */
        soisconnected(so);

        NL_LOG(LOG_DEBUG2, "socket %p, connect to %u", so, snl->nl_pid);

        return (0);
}

static void
destroy_nlpcb_epoch(epoch_context_t ctx)
{
        struct nlpcb *nlp;

        nlp = __containerof(ctx, struct nlpcb, nl_epoch_ctx);

        NLP_LOCK_DESTROY(nlp);
        free(nlp, M_PCB);
}

static void
nl_close(struct socket *so)
{
        MPASS(sotonlpcb(so) != NULL);
        struct nlpcb *nlp;
        struct nl_buf *nb;

        NL_LOG(LOG_DEBUG2, "detaching socket %p, PID %d", so, curproc->p_pid);
        nlp = sotonlpcb(so);

        /* Mark as inactive so no new work can be enqueued */
        NLP_LOCK(nlp);
        bool was_bound = nlp->nl_bound;
        NLP_UNLOCK(nlp);

        /* Wait till all scheduled work has been completed  */
        taskqueue_drain_all(nlp->nl_taskqueue);
        taskqueue_free(nlp->nl_taskqueue);

        NLCTL_WLOCK();
        NLP_LOCK(nlp);
        if (was_bound) {
                CK_LIST_REMOVE(nlp, nl_port_next);
                NL_LOG(LOG_DEBUG3, "socket %p, unlinking bound pid %u", so, nlp->nl_port);
        }
        CK_LIST_REMOVE(nlp, nl_next);
        nlp->nl_socket = NULL;
        NLP_UNLOCK(nlp);
        NLCTL_WUNLOCK();

        so->so_pcb = NULL;

        while ((nb = TAILQ_FIRST(&so->so_snd.nl_queue)) != NULL) {
                TAILQ_REMOVE(&so->so_snd.nl_queue, nb, tailq);
                nl_buf_free(nb);
        }
        while ((nb = TAILQ_FIRST(&so->so_rcv.nl_queue)) != NULL) {
                TAILQ_REMOVE(&so->so_rcv.nl_queue, nb, tailq);
                nl_buf_free(nb);
        }

        mtx_destroy(&so->so_snd_mtx);
        mtx_destroy(&so->so_rcv_mtx);

        NL_LOG(LOG_DEBUG3, "socket %p, detached", so);

        /* XXX: is delayed free needed? */
        NET_EPOCH_CALL(destroy_nlpcb_epoch, &nlp->nl_epoch_ctx);
}

static int
nl_disconnect(struct socket *so)
{
        NL_LOG(LOG_DEBUG3, "socket %p, PID %d", so, curproc->p_pid);
        MPASS(sotonlpcb(so) != NULL);
        return (ENOTCONN);
}

static int
nl_sockaddr(struct socket *so, struct sockaddr *sa)
{

        *(struct sockaddr_nl *)sa = (struct sockaddr_nl ){
                /* TODO: set other fields */
                .nl_len = sizeof(struct sockaddr_nl),
                .nl_family = AF_NETLINK,
                .nl_pid = sotonlpcb(so)->nl_port,
        };

        return (0);
}

static int
nl_sosend(struct socket *so, struct sockaddr *addr, struct uio *uio,
    struct mbuf *m, struct mbuf *control, int flags, struct thread *td)
{
        struct nlpcb *nlp = sotonlpcb(so);
        struct sockbuf *sb = &so->so_snd;
        struct nl_buf *nb;
        size_t len;
        int error;

        MPASS(m == NULL && uio != NULL);

        if (__predict_false(control != NULL)) {
                m_freem(control);
                return (EINVAL);
        }

        if (__predict_false(flags & MSG_OOB))   /* XXXGL: or just ignore? */
                return (EOPNOTSUPP);

        if (__predict_false(uio->uio_resid < sizeof(struct nlmsghdr)))
                return (ENOBUFS);               /* XXXGL: any better error? */

        if (__predict_false(uio->uio_resid > sb->sb_hiwat))
                return (EMSGSIZE);

        error = SOCK_IO_SEND_LOCK(so, SBLOCKWAIT(flags));
        if (error)
                return (error);

        len = roundup2(uio->uio_resid, 8) + SCRATCH_BUFFER_SIZE;
        if (nlp->nl_linux)
                len += roundup2(uio->uio_resid, 8);
        nb = nl_buf_alloc(len, M_WAITOK);
        nb->datalen = uio->uio_resid;
        error = uiomove(&nb->data[0], uio->uio_resid, uio);
        if (__predict_false(error))
                goto out;

        NL_LOG(LOG_DEBUG2, "sending message to kernel %u bytes", nb->datalen);

        SOCK_SENDBUF_LOCK(so);
restart:
        if (sb->sb_hiwat - sb->sb_ccc >= nb->datalen) {
                TAILQ_INSERT_TAIL(&sb->nl_queue, nb, tailq);
                sb->sb_acc += nb->datalen;
                sb->sb_ccc += nb->datalen;
                nb = NULL;
        } else if ((so->so_state & SS_NBIO) ||
            (flags & (MSG_NBIO | MSG_DONTWAIT)) != 0) {
                SOCK_SENDBUF_UNLOCK(so);
                error = EWOULDBLOCK;
                goto out;
        } else {
                if ((error = sbwait(so, SO_SND)) != 0) {
                        SOCK_SENDBUF_UNLOCK(so);
                        goto out;
                } else
                        goto restart;
        }
        SOCK_SENDBUF_UNLOCK(so);

        if (nb == NULL) {
                NL_LOG(LOG_DEBUG3, "success");
                NLP_LOCK(nlp);
                nl_schedule_taskqueue(nlp);
                NLP_UNLOCK(nlp);
        }

out:
        SOCK_IO_SEND_UNLOCK(so);
        if (nb != NULL) {
                NL_LOG(LOG_DEBUG3, "failure, error %d", error);
                nl_buf_free(nb);
        }
        return (error);
}

/* Create control data for recvmsg(2) on Netlink socket. */
static struct mbuf *
nl_createcontrol(struct nlpcb *nlp)
{
        struct {
                struct nlattr nla;
                uint32_t val;
        } data[] = {
                {
                        .nla.nla_len = sizeof(struct nlattr) + sizeof(uint32_t),
                        .nla.nla_type = NLMSGINFO_ATTR_PROCESS_ID,
                        .val = nlp->nl_process_id,
                },
                {
                        .nla.nla_len = sizeof(struct nlattr) + sizeof(uint32_t),
                        .nla.nla_type = NLMSGINFO_ATTR_PORT_ID,
                        .val = nlp->nl_port,
                },
        };

        return (sbcreatecontrol(data, sizeof(data), NETLINK_MSG_INFO,
            SOL_NETLINK, M_WAITOK));
}

static int
nl_soreceive(struct socket *so, struct sockaddr **psa, struct uio *uio,
    struct mbuf **mp, struct mbuf **controlp, int *flagsp)
{
        static const struct sockaddr_nl nl_empty_src = {
                .nl_len = sizeof(struct sockaddr_nl),
                .nl_family = PF_NETLINK,
                .nl_pid = 0 /* comes from the kernel */
        };
        struct sockbuf *sb = &so->so_rcv;
        struct nlpcb *nlp = sotonlpcb(so);
        struct nl_buf *first, *last, *nb, *next;
        struct nlmsghdr *hdr;
        int flags, error;
        u_int len, overflow, partoff, partlen, msgrcv, datalen;
        bool nonblock, trunc, peek;

        MPASS(mp == NULL && uio != NULL);

        NL_LOG(LOG_DEBUG3, "socket %p, PID %d", so, curproc->p_pid);

        if (psa != NULL)
                *psa = sodupsockaddr((const struct sockaddr *)&nl_empty_src,
                    M_WAITOK);

        if (controlp != NULL && (nlp->nl_flags & NLF_MSG_INFO))
                *controlp = nl_createcontrol(nlp);

        flags = flagsp != NULL ? *flagsp & ~MSG_TRUNC : 0;
        trunc = flagsp != NULL ? *flagsp & MSG_TRUNC : false;
        nonblock = (so->so_state & SS_NBIO) ||
            (flags & (MSG_DONTWAIT | MSG_NBIO));
        peek = flags & MSG_PEEK;

        error = SOCK_IO_RECV_LOCK(so, SBLOCKWAIT(flags));
        if (__predict_false(error))
                return (error);

        len = 0;
        overflow = 0;
        msgrcv = 0;
        datalen = 0;

        SOCK_RECVBUF_LOCK(so);
        while ((first = TAILQ_FIRST(&sb->nl_queue)) == NULL) {
                if (nonblock) {
                        SOCK_RECVBUF_UNLOCK(so);
                        SOCK_IO_RECV_UNLOCK(so);
                        return (EWOULDBLOCK);
                }
                error = sbwait(so, SO_RCV);
                if (error) {
                        SOCK_RECVBUF_UNLOCK(so);
                        SOCK_IO_RECV_UNLOCK(so);
                        return (error);
                }
        }

        /*
         * Netlink socket buffer consists of a queue of nl_bufs, but for the
         * userland there should be no boundaries.  However, there are Netlink
         * messages, that shouldn't be split.  Internal invariant is that a
         * message never spans two nl_bufs.
         * If a large userland buffer is provided, we would traverse the queue
         * until either queue end is reached or the buffer is fulfilled.  If
         * an application provides a buffer that isn't able to fit a single
         * message, we would truncate it and lose its tail.  This is the only
         * condition where we would lose data.  If buffer is able to fit at
         * least one message, we would return it and won't truncate the next.
         *
         * We use same code for normal and MSG_PEEK case.  At first queue pass
         * we scan nl_bufs and count lenght.  In case we can read entire buffer
         * at one write everything is trivial.  In case we can not, we save
         * pointer to the last (or partial) nl_buf and in the !peek case we
         * split the queue into two pieces.  We can safely drop the queue lock,
         * as kernel would only append nl_bufs to the end of the queue, and
         * we are the exclusive owner of queue beginning due to sleepable lock.
         * At the second pass we copy data out and in !peek case free nl_bufs.
         */
        TAILQ_FOREACH(nb, &sb->nl_queue, tailq) {
                u_int offset;

                MPASS(nb->offset < nb->datalen);
                offset = nb->offset;
                while (offset < nb->datalen) {
                        hdr = (struct nlmsghdr *)&nb->data[offset];
                        MPASS(nb->offset + hdr->nlmsg_len <= nb->datalen);
                        if (uio->uio_resid < len + hdr->nlmsg_len) {
                                overflow = len + hdr->nlmsg_len -
                                    uio->uio_resid;
                                partoff = nb->offset;
                                if (offset > partoff) {
                                        partlen = offset - partoff;
                                        if (!peek) {
                                                nb->offset = offset;
                                                datalen += partlen;
                                        }
                                } else if (len == 0 && uio->uio_resid > 0) {
                                        flags |= MSG_TRUNC;
                                        partlen = uio->uio_resid;
                                        if (peek)
                                                goto nospace;
                                        datalen += hdr->nlmsg_len;
                                        if (nb->offset + hdr->nlmsg_len ==
                                            nb->datalen) {
                                                /*
                                                 * Avoid leaving empty nb.
                                                 * Process last nb normally.
                                                 * Trust uiomove() to care
                                                 * about negative uio_resid.
                                                 */
                                                nb = TAILQ_NEXT(nb, tailq);
                                                overflow = 0;
                                                partlen = 0;
                                        } else
                                                nb->offset += hdr->nlmsg_len;
                                        msgrcv++;
                                } else
                                        partlen = 0;
                                goto nospace;
                        }
                        len += hdr->nlmsg_len;
                        offset += hdr->nlmsg_len;
                        MPASS(offset <= nb->buflen);
                        msgrcv++;
                }
                MPASS(offset == nb->datalen);
                datalen += nb->datalen - nb->offset;
        }
nospace:
        last = nb;
        if (!peek) {
                if (last == NULL)
                        TAILQ_INIT(&sb->nl_queue);
                else {
                        /* XXXGL: create TAILQ_SPLIT */
                        TAILQ_FIRST(&sb->nl_queue) = last;
                        last->tailq.tqe_prev = &TAILQ_FIRST(&sb->nl_queue);
                }
                MPASS(sb->sb_acc >= datalen);
                sb->sb_acc -= datalen;
                sb->sb_ccc -= datalen;
        }
        SOCK_RECVBUF_UNLOCK(so);

        for (nb = first; nb != last; nb = next) {
                next = TAILQ_NEXT(nb, tailq);
                if (__predict_true(error == 0))
                        error = uiomove(&nb->data[nb->offset],
                            (int)(nb->datalen - nb->offset), uio);
                if (!peek)
                        nl_buf_free(nb);
        }
        if (last != NULL && partlen > 0 && __predict_true(error == 0))
                error = uiomove(&nb->data[partoff], (int)partlen, uio);

        if (trunc && overflow > 0) {
                uio->uio_resid -= overflow;
                MPASS(uio->uio_resid < 0);
        } else
                MPASS(uio->uio_resid >= 0);

        if (uio->uio_td)
                uio->uio_td->td_ru.ru_msgrcv += msgrcv;

        if (flagsp != NULL)
                *flagsp |= flags;

        SOCK_IO_RECV_UNLOCK(so);

        nl_on_transmit(sotonlpcb(so));

        return (error);
}

static int
nl_getoptflag(int sopt_name)
{
        switch (sopt_name) {
        case NETLINK_CAP_ACK:
                return (NLF_CAP_ACK);
        case NETLINK_EXT_ACK:
                return (NLF_EXT_ACK);
        case NETLINK_GET_STRICT_CHK:
                return (NLF_STRICT);
        case NETLINK_MSG_INFO:
                return (NLF_MSG_INFO);
        }

        return (0);
}

static int
nl_ctloutput(struct socket *so, struct sockopt *sopt)
{
        struct nlpcb *nlp = sotonlpcb(so);
        uint32_t flag;
        int optval, error = 0;
        NLCTL_TRACKER;

        NL_LOG(LOG_DEBUG2, "%ssockopt(%p, %d)", (sopt->sopt_dir) ? "set" : "get",
            so, sopt->sopt_name);

        switch (sopt->sopt_dir) {
        case SOPT_SET:
                switch (sopt->sopt_name) {
                case NETLINK_ADD_MEMBERSHIP:
                case NETLINK_DROP_MEMBERSHIP:
                        error = sooptcopyin(sopt, &optval, sizeof(optval), sizeof(optval));
                        if (error != 0)
                                break;
                        if (optval <= 0 || optval >= NLP_MAX_GROUPS) {
                                error = ERANGE;
                                break;
                        }
                        NL_LOG(LOG_DEBUG2, "ADD/DEL group %d", (uint32_t)optval);

                        NLCTL_WLOCK();
                        if (sopt->sopt_name == NETLINK_ADD_MEMBERSHIP)
                                nlp_join_group(nlp, optval);
                        else
                                nlp_leave_group(nlp, optval);
                        NLCTL_WUNLOCK();
                        break;
                case NETLINK_CAP_ACK:
                case NETLINK_EXT_ACK:
                case NETLINK_GET_STRICT_CHK:
                case NETLINK_MSG_INFO:
                        error = sooptcopyin(sopt, &optval, sizeof(optval), sizeof(optval));
                        if (error != 0)
                                break;

                        flag = nl_getoptflag(sopt->sopt_name);

                        if ((flag == NLF_MSG_INFO) && nlp->nl_linux) {
                                error = EINVAL;
                                break;
                        }

                        NLCTL_WLOCK();
                        if (optval != 0)
                                nlp->nl_flags |= flag;
                        else
                                nlp->nl_flags &= ~flag;
                        NLCTL_WUNLOCK();
                        break;
                default:
                        error = ENOPROTOOPT;
                }
                break;
        case SOPT_GET:
                switch (sopt->sopt_name) {
                case NETLINK_LIST_MEMBERSHIPS:
                        NLCTL_RLOCK();
                        optval = nlp_get_groups_compat(nlp);
                        NLCTL_RUNLOCK();
                        error = sooptcopyout(sopt, &optval, sizeof(optval));
                        break;
                case NETLINK_CAP_ACK:
                case NETLINK_EXT_ACK:
                case NETLINK_GET_STRICT_CHK:
                case NETLINK_MSG_INFO:
                        NLCTL_RLOCK();
                        optval = (nlp->nl_flags & nl_getoptflag(sopt->sopt_name)) != 0;
                        NLCTL_RUNLOCK();
                        error = sooptcopyout(sopt, &optval, sizeof(optval));
                        break;
                default:
                        error = ENOPROTOOPT;
                }
                break;
        default:
                error = ENOPROTOOPT;
        }

        return (error);
}

static int
sysctl_handle_nl_maxsockbuf(SYSCTL_HANDLER_ARGS)
{
        int error = 0;
        u_long tmp_maxsockbuf = nl_maxsockbuf;

        error = sysctl_handle_long(oidp, &tmp_maxsockbuf, arg2, req);
        if (error || !req->newptr)
                return (error);
        if (tmp_maxsockbuf < MSIZE + MCLBYTES)
                return (EINVAL);
        nl_maxsockbuf = tmp_maxsockbuf;

        return (0);
}

static int
nl_setsbopt(struct socket *so, struct sockopt *sopt)
{
        int error, optval;
        bool result;

        if (sopt->sopt_name != SO_RCVBUF)
                return (sbsetopt(so, sopt));

        /* Allow to override max buffer size in certain conditions */

        error = sooptcopyin(sopt, &optval, sizeof optval, sizeof optval);
        if (error != 0)
                return (error);
        NL_LOG(LOG_DEBUG2, "socket %p, PID %d, SO_RCVBUF=%d", so, curproc->p_pid, optval);
        if (optval > sb_max_adj) {
                if (priv_check(curthread, PRIV_NET_ROUTE) != 0)
                        return (EPERM);
        }

        SOCK_RECVBUF_LOCK(so);
        result = sbreserve_locked_limit(so, SO_RCV, optval, nl_maxsockbuf, curthread);
        SOCK_RECVBUF_UNLOCK(so);

        return (result ? 0 : ENOBUFS);
}

#define NETLINK_PROTOSW                                         \
        .pr_flags = PR_ATOMIC | PR_ADDR | PR_SOCKBUF,           \
        .pr_ctloutput = nl_ctloutput,                           \
        .pr_setsbopt = nl_setsbopt,                             \
        .pr_attach = nl_attach,                                 \
        .pr_bind = nl_bind,                                     \
        .pr_connect = nl_connect,                               \
        .pr_disconnect = nl_disconnect,                         \
        .pr_sosend = nl_sosend,                                 \
        .pr_soreceive = nl_soreceive,                           \
        .pr_sockaddr = nl_sockaddr,                             \
        .pr_close = nl_close

static struct protosw netlink_raw_sw = {
        .pr_type = SOCK_RAW,
        NETLINK_PROTOSW
};

static struct protosw netlink_dgram_sw = {
        .pr_type = SOCK_DGRAM,
        NETLINK_PROTOSW
};

static struct domain netlinkdomain = {
        .dom_family = PF_NETLINK,
        .dom_name = "netlink",
        .dom_flags = DOMF_UNLOADABLE,
        .dom_nprotosw =         2,
        .dom_protosw =          { &netlink_raw_sw, &netlink_dgram_sw },
};

DOMAIN_SET(netlink);