root/sys/kern/kern_racct.c
/*-
 * SPDX-License-Identifier: BSD-2-Clause
 *
 * Copyright (c) 2010 The FreeBSD Foundation
 *
 * This software was developed by Edward Tomasz Napierala under sponsorship
 * from the FreeBSD Foundation.
 *
 * Redistribution and use in source and binary forms, with or without
 * modification, are permitted provided that the following conditions
 * are met:
 * 1. Redistributions of source code must retain the above copyright
 *    notice, this list of conditions and the following disclaimer.
 * 2. Redistributions in binary form must reproduce the above copyright
 *    notice, this list of conditions and the following disclaimer in the
 *    documentation and/or other materials provided with the distribution.
 *
 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
 * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
 * SUCH DAMAGE.
 */

#include <sys/cdefs.h>
#include "opt_sched.h"

#include <sys/param.h>
#include <sys/buf.h>
#include <sys/systm.h>
#include <sys/eventhandler.h>
#include <sys/jail.h>
#include <sys/kernel.h>
#include <sys/kthread.h>
#include <sys/lock.h>
#include <sys/loginclass.h>
#include <sys/malloc.h>
#include <sys/mutex.h>
#include <sys/proc.h>
#include <sys/racct.h>
#include <sys/resourcevar.h>
#include <sys/sbuf.h>
#include <sys/sched.h>
#include <sys/sdt.h>
#include <sys/smp.h>
#include <sys/sx.h>
#include <sys/sysctl.h>
#include <sys/sysproto.h>
#include <sys/umtxvar.h>
#include <machine/smp.h>

#ifdef RCTL
#include <sys/rctl.h>
#endif

FEATURE(racct, "Resource Accounting");

/*
 * Do not block processes that have their %cpu usage <= pcpu_threshold.
 */
static int pcpu_threshold = 1;
#ifdef RACCT_DEFAULT_TO_DISABLED
bool __read_frequently racct_enable = false;
#else
bool __read_frequently racct_enable = true;
#endif

SYSCTL_NODE(_kern, OID_AUTO, racct, CTLFLAG_RW | CTLFLAG_MPSAFE, 0,
    "Resource Accounting");
SYSCTL_BOOL(_kern_racct, OID_AUTO, enable, CTLFLAG_RDTUN, &racct_enable,
    0, "Enable RACCT/RCTL");
SYSCTL_UINT(_kern_racct, OID_AUTO, pcpu_threshold, CTLFLAG_RW, &pcpu_threshold,
    0, "Processes with higher %cpu usage than this value can be throttled.");

struct mtx racct_lock;
MTX_SYSINIT(racct_lock, &racct_lock, "racct lock", MTX_DEF);

static uma_zone_t racct_zone;

static void racct_sub_racct(struct racct *dest, const struct racct *src);
static void racct_sub_cred_locked(struct ucred *cred, int resource,
                uint64_t amount);
static void racct_add_cred_locked(struct ucred *cred, int resource,
                uint64_t amount);
static int racct_set_locked(struct proc *p, int resource, uint64_t amount,
                int force);
static void racct_updatepcpu_locked(struct proc *p);
static void racct_updatepcpu_racct_locked(struct racct *racct);
static void racct_updatepcpu_containers(void);
static void racct_settime_locked(struct proc *p, bool exit);
static void racct_zeropcpu_locked(struct proc *p);

SDT_PROVIDER_DEFINE(racct);
SDT_PROBE_DEFINE3(racct, , rusage, add,
    "struct proc *", "int", "uint64_t");
SDT_PROBE_DEFINE3(racct, , rusage, add__failure,
    "struct proc *", "int", "uint64_t");
SDT_PROBE_DEFINE3(racct, , rusage, add__buf,
    "struct proc *", "const struct buf *", "int");
SDT_PROBE_DEFINE3(racct, , rusage, add__cred,
    "struct ucred *", "int", "uint64_t");
SDT_PROBE_DEFINE3(racct, , rusage, add__force,
    "struct proc *", "int", "uint64_t");
SDT_PROBE_DEFINE3(racct, , rusage, set,
    "struct proc *", "int", "uint64_t");
SDT_PROBE_DEFINE3(racct, , rusage, set__failure,
    "struct proc *", "int", "uint64_t");
SDT_PROBE_DEFINE3(racct, , rusage, set__force,
    "struct proc *", "int", "uint64_t");
SDT_PROBE_DEFINE3(racct, , rusage, sub,
    "struct proc *", "int", "uint64_t");
SDT_PROBE_DEFINE3(racct, , rusage, sub__cred,
    "struct ucred *", "int", "uint64_t");
SDT_PROBE_DEFINE1(racct, , racct, create,
    "struct racct *");
SDT_PROBE_DEFINE1(racct, , racct, destroy,
    "struct racct *");
SDT_PROBE_DEFINE2(racct, , racct, join,
    "struct racct *", "struct racct *");
SDT_PROBE_DEFINE2(racct, , racct, join__failure,
    "struct racct *", "struct racct *");
SDT_PROBE_DEFINE2(racct, , racct, leave,
    "struct racct *", "struct racct *");

int racct_types[] = {
        [RACCT_CPU] =
                RACCT_IN_MILLIONS,
        [RACCT_DATA] =
                RACCT_RECLAIMABLE | RACCT_INHERITABLE | RACCT_DENIABLE,
        [RACCT_STACK] =
                RACCT_RECLAIMABLE | RACCT_INHERITABLE | RACCT_DENIABLE,
        [RACCT_CORE] =
                RACCT_DENIABLE,
        [RACCT_RSS] =
                RACCT_RECLAIMABLE,
        [RACCT_MEMLOCK] =
                RACCT_RECLAIMABLE | RACCT_DENIABLE,
        [RACCT_NPROC] =
                RACCT_RECLAIMABLE | RACCT_DENIABLE,
        [RACCT_NOFILE] =
                RACCT_RECLAIMABLE | RACCT_INHERITABLE | RACCT_DENIABLE,
        [RACCT_VMEM] =
                RACCT_RECLAIMABLE | RACCT_INHERITABLE | RACCT_DENIABLE,
        [RACCT_NPTS] =
                RACCT_RECLAIMABLE | RACCT_DENIABLE | RACCT_SLOPPY,
        [RACCT_SWAP] =
                RACCT_RECLAIMABLE | RACCT_DENIABLE | RACCT_SLOPPY,
        [RACCT_NTHR] =
                RACCT_RECLAIMABLE | RACCT_DENIABLE,
        [RACCT_MSGQQUEUED] =
                RACCT_RECLAIMABLE | RACCT_DENIABLE | RACCT_SLOPPY,
        [RACCT_MSGQSIZE] =
                RACCT_RECLAIMABLE | RACCT_DENIABLE | RACCT_SLOPPY,
        [RACCT_NMSGQ] =
                RACCT_RECLAIMABLE | RACCT_DENIABLE | RACCT_SLOPPY,
        [RACCT_NSEM] =
                RACCT_RECLAIMABLE | RACCT_DENIABLE | RACCT_SLOPPY,
        [RACCT_NSEMOP] =
                RACCT_RECLAIMABLE | RACCT_INHERITABLE | RACCT_DENIABLE,
        [RACCT_NSHM] =
                RACCT_RECLAIMABLE | RACCT_DENIABLE | RACCT_SLOPPY,
        [RACCT_SHMSIZE] =
                RACCT_RECLAIMABLE | RACCT_DENIABLE | RACCT_SLOPPY,
        [RACCT_WALLCLOCK] =
                RACCT_IN_MILLIONS,
        [RACCT_PCTCPU] =
                RACCT_DECAYING | RACCT_DENIABLE | RACCT_IN_MILLIONS,
        [RACCT_READBPS] =
                RACCT_DECAYING,
        [RACCT_WRITEBPS] =
                RACCT_DECAYING,
        [RACCT_READIOPS] =
                RACCT_DECAYING,
        [RACCT_WRITEIOPS] =
                RACCT_DECAYING };

static const fixpt_t RACCT_DECAY_FACTOR = 0.3 * FSCALE;

static void
racct_add_racct(struct racct *dest, const struct racct *src)
{
        int i;

        ASSERT_RACCT_ENABLED();
        RACCT_LOCK_ASSERT();

        /*
         * Update resource usage in dest.
         */
        for (i = 0; i <= RACCT_MAX; i++) {
                KASSERT(dest->r_resources[i] >= 0,
                    ("%s: resource %d propagation meltdown: dest < 0",
                    __func__, i));
                KASSERT(src->r_resources[i] >= 0,
                    ("%s: resource %d propagation meltdown: src < 0",
                    __func__, i));
                dest->r_resources[i] += src->r_resources[i];
        }
}

static void
racct_sub_racct(struct racct *dest, const struct racct *src)
{
        int i;

        ASSERT_RACCT_ENABLED();
        RACCT_LOCK_ASSERT();

        /*
         * Update resource usage in dest.
         */
        for (i = 0; i <= RACCT_MAX; i++) {
                if (!RACCT_IS_SLOPPY(i) && !RACCT_IS_DECAYING(i)) {
                        KASSERT(dest->r_resources[i] >= 0,
                            ("%s: resource %d propagation meltdown: dest < 0",
                            __func__, i));
                        KASSERT(src->r_resources[i] >= 0,
                            ("%s: resource %d propagation meltdown: src < 0",
                            __func__, i));
                        KASSERT(src->r_resources[i] <= dest->r_resources[i],
                            ("%s: resource %d propagation meltdown: src > dest",
                            __func__, i));
                }
                if (RACCT_CAN_DROP(i)) {
                        dest->r_resources[i] -= src->r_resources[i];
                        if (dest->r_resources[i] < 0)
                                dest->r_resources[i] = 0;
                }
        }
}

void
racct_create(struct racct **racctp)
{

        if (!racct_enable)
                return;

        SDT_PROBE1(racct, , racct, create, racctp);

        KASSERT(*racctp == NULL, ("racct already allocated"));

        *racctp = uma_zalloc(racct_zone, M_WAITOK | M_ZERO);
}

static void
racct_destroy_locked(struct racct **racctp)
{
        struct racct *racct;
        int i;

        ASSERT_RACCT_ENABLED();

        SDT_PROBE1(racct, , racct, destroy, racctp);

        RACCT_LOCK_ASSERT();
        KASSERT(racctp != NULL, ("NULL racctp"));
        KASSERT(*racctp != NULL, ("NULL racct"));

        racct = *racctp;

        for (i = 0; i <= RACCT_MAX; i++) {
                if (RACCT_IS_SLOPPY(i))
                        continue;
                if (!RACCT_IS_RECLAIMABLE(i))
                        continue;
                KASSERT(racct->r_resources[i] == 0,
                    ("destroying non-empty racct: "
                    "%ju allocated for resource %d\n",
                    racct->r_resources[i], i));
        }
        uma_zfree(racct_zone, racct);
        *racctp = NULL;
}

void
racct_destroy(struct racct **racct)
{

        if (!racct_enable)
                return;

        RACCT_LOCK();
        racct_destroy_locked(racct);
        RACCT_UNLOCK();
}

/*
 * Increase consumption of 'resource' by 'amount' for 'racct',
 * but not its parents.  Differently from other cases, 'amount' here
 * may be less than zero.
 */
static void
racct_adjust_resource(struct racct *racct, int resource,
    int64_t amount)
{

        ASSERT_RACCT_ENABLED();
        RACCT_LOCK_ASSERT();
        KASSERT(racct != NULL, ("NULL racct"));

        racct->r_resources[resource] += amount;
        if (racct->r_resources[resource] < 0) {
                KASSERT(RACCT_IS_SLOPPY(resource) || RACCT_IS_DECAYING(resource),
                    ("%s: resource %d usage < 0", __func__, resource));
                racct->r_resources[resource] = 0;
        }
}

static int
racct_add_locked(struct proc *p, int resource, uint64_t amount, int force)
{
#ifdef RCTL
        int error;
#endif

        ASSERT_RACCT_ENABLED();

        /*
         * We need proc lock to dereference p->p_ucred.
         */
        PROC_LOCK_ASSERT(p, MA_OWNED);

#ifdef RCTL
        error = rctl_enforce(p, resource, amount);
        if (error && !force && RACCT_IS_DENIABLE(resource)) {
                SDT_PROBE3(racct, , rusage, add__failure, p, resource, amount);
                return (error);
        }
#endif
        racct_adjust_resource(p->p_racct, resource, amount);
        racct_add_cred_locked(p->p_ucred, resource, amount);

        return (0);
}

/*
 * Increase allocation of 'resource' by 'amount' for process 'p'.
 * Return 0 if it's below limits, or errno, if it's not.
 */
int
racct_add(struct proc *p, int resource, uint64_t amount)
{
        int error;

        if (!racct_enable)
                return (0);

        SDT_PROBE3(racct, , rusage, add, p, resource, amount);

        RACCT_LOCK();
        error = racct_add_locked(p, resource, amount, 0);
        RACCT_UNLOCK();
        return (error);
}

/*
 * Increase allocation of 'resource' by 'amount' for process 'p'.
 * Doesn't check for limits and never fails.
 */
void
racct_add_force(struct proc *p, int resource, uint64_t amount)
{

        if (!racct_enable)
                return;

        SDT_PROBE3(racct, , rusage, add__force, p, resource, amount);

        RACCT_LOCK();
        racct_add_locked(p, resource, amount, 1);
        RACCT_UNLOCK();
}

static void
racct_add_cred_locked(struct ucred *cred, int resource, uint64_t amount)
{
        struct prison *pr;

        ASSERT_RACCT_ENABLED();

        racct_adjust_resource(cred->cr_ruidinfo->ui_racct, resource, amount);
        for (pr = cred->cr_prison; pr != NULL; pr = pr->pr_parent)
                racct_adjust_resource(pr->pr_prison_racct->prr_racct, resource,
                    amount);
        racct_adjust_resource(cred->cr_loginclass->lc_racct, resource, amount);
}

/*
 * Increase allocation of 'resource' by 'amount' for credential 'cred'.
 * Doesn't check for limits and never fails.
 */
void
racct_add_cred(struct ucred *cred, int resource, uint64_t amount)
{

        if (!racct_enable)
                return;

        SDT_PROBE3(racct, , rusage, add__cred, cred, resource, amount);

        RACCT_LOCK();
        racct_add_cred_locked(cred, resource, amount);
        RACCT_UNLOCK();
}

/*
 * Account for disk IO resource consumption.  Checks for limits,
 * but never fails, due to disk limits being undeniable.
 */
void
racct_add_buf(struct proc *p, const struct buf *bp, int is_write)
{

        ASSERT_RACCT_ENABLED();
        PROC_LOCK_ASSERT(p, MA_OWNED);

        SDT_PROBE3(racct, , rusage, add__buf, p, bp, is_write);

        RACCT_LOCK();
        if (is_write) {
                racct_add_locked(curproc, RACCT_WRITEBPS, bp->b_bcount, 1);
                racct_add_locked(curproc, RACCT_WRITEIOPS, 1, 1);
        } else {
                racct_add_locked(curproc, RACCT_READBPS, bp->b_bcount, 1);
                racct_add_locked(curproc, RACCT_READIOPS, 1, 1);
        }
        RACCT_UNLOCK();
}

static void
racct_settime_locked(struct proc *p, bool exit)
{
        struct thread *td;
        struct timeval wallclock;
        uint64_t runtime;

        ASSERT_RACCT_ENABLED();
        RACCT_LOCK_ASSERT();
        PROC_LOCK_ASSERT(p, MA_OWNED);

        if (exit) {
                /*
                 * proc_reap() has already calculated rux
                 * and added crux to rux.
                 */
                runtime = cputick2usec(p->p_rux.rux_runtime -
                    p->p_crux.rux_runtime);
        } else {
                PROC_STATLOCK(p);
                FOREACH_THREAD_IN_PROC(p, td)
                        ruxagg(p, td);
                PROC_STATUNLOCK(p);
                runtime = cputick2usec(p->p_rux.rux_runtime);
        }
        microuptime(&wallclock);
        timevalsub(&wallclock, &p->p_stats->p_start);

        racct_set_locked(p, RACCT_CPU, runtime, 0);
        racct_set_locked(p, RACCT_WALLCLOCK,
            (uint64_t)wallclock.tv_sec * 1000000 +
            wallclock.tv_usec, 0);
}

static int
racct_set_locked(struct proc *p, int resource, uint64_t amount, int force)
{
        int64_t old_amount, diff_proc, diff_cred;
#ifdef RCTL
        int error;
#endif

        ASSERT_RACCT_ENABLED();

        /*
         * We need proc lock to dereference p->p_ucred.
         */
        PROC_LOCK_ASSERT(p, MA_OWNED);

        old_amount = p->p_racct->r_resources[resource];
        /*
         * The diffs may be negative.
         */
        diff_proc = amount - old_amount;
        diff_cred = diff_proc;
#ifdef notyet
        KASSERT(diff_proc >= 0 || RACCT_CAN_DROP(resource),
            ("%s: usage of non-droppable resource %d dropping", __func__,
             resource));
#endif
#ifdef RCTL
        if (diff_proc > 0) {
                error = rctl_enforce(p, resource, diff_proc);
                if (error && !force && RACCT_IS_DENIABLE(resource)) {
                        SDT_PROBE3(racct, , rusage, set__failure, p, resource,
                            amount);
                        return (error);
                }
        }
#endif
        racct_adjust_resource(p->p_racct, resource, diff_proc);
        if (diff_cred > 0)
                racct_add_cred_locked(p->p_ucred, resource, diff_cred);
        else if (diff_cred < 0)
                racct_sub_cred_locked(p->p_ucred, resource, -diff_cred);

        return (0);
}

/*
 * Set allocation of 'resource' to 'amount' for process 'p'.
 * Return 0 if it's below limits, or errno, if it's not.
 *
 * Note that decreasing the allocation always returns 0,
 * even if it's above the limit.
 */
int
racct_set_unlocked(struct proc *p, int resource, uint64_t amount)
{
        int error;

        ASSERT_RACCT_ENABLED();
        PROC_LOCK(p);
        error = racct_set(p, resource, amount);
        PROC_UNLOCK(p);
        return (error);
}

int
racct_set(struct proc *p, int resource, uint64_t amount)
{
        int error;

        if (!racct_enable)
                return (0);

        SDT_PROBE3(racct, , rusage, set__force, p, resource, amount);

        RACCT_LOCK();
        error = racct_set_locked(p, resource, amount, 0);
        RACCT_UNLOCK();
        return (error);
}

void
racct_set_force(struct proc *p, int resource, uint64_t amount)
{

        if (!racct_enable)
                return;

        SDT_PROBE3(racct, , rusage, set, p, resource, amount);

        RACCT_LOCK();
        racct_set_locked(p, resource, amount, 1);
        RACCT_UNLOCK();
}

/*
 * Returns amount of 'resource' the process 'p' can keep allocated.
 * Allocating more than that would be denied, unless the resource
 * is marked undeniable.  Amount of already allocated resource does
 * not matter.
 */
uint64_t
racct_get_limit(struct proc *p, int resource)
{
#ifdef RCTL
        uint64_t available;

        if (!racct_enable)
                return (UINT64_MAX);

        RACCT_LOCK();
        available = rctl_get_limit(p, resource);
        RACCT_UNLOCK();

        return (available);
#else

        return (UINT64_MAX);
#endif
}

/*
 * Returns amount of 'resource' the process 'p' can keep allocated.
 * Allocating more than that would be denied, unless the resource
 * is marked undeniable.  Amount of already allocated resource does
 * matter.
 */
uint64_t
racct_get_available(struct proc *p, int resource)
{
#ifdef RCTL
        uint64_t available;

        if (!racct_enable)
                return (UINT64_MAX);

        RACCT_LOCK();
        available = rctl_get_available(p, resource);
        RACCT_UNLOCK();

        return (available);
#else

        return (UINT64_MAX);
#endif
}

/*
 * Returns amount of the %cpu resource that process 'p' can add to its %cpu
 * utilization.  Adding more than that would lead to the process being
 * throttled.
 */
static int64_t
racct_pcpu_available(struct proc *p)
{
#ifdef RCTL
        uint64_t available;

        ASSERT_RACCT_ENABLED();

        RACCT_LOCK();
        available = rctl_pcpu_available(p);
        RACCT_UNLOCK();

        return (available);
#else

        return (INT64_MAX);
#endif
}

/*
 * Decrease allocation of 'resource' by 'amount' for process 'p'.
 */
void
racct_sub(struct proc *p, int resource, uint64_t amount)
{

        if (!racct_enable)
                return;

        SDT_PROBE3(racct, , rusage, sub, p, resource, amount);

        /*
         * We need proc lock to dereference p->p_ucred.
         */
        PROC_LOCK_ASSERT(p, MA_OWNED);
        KASSERT(RACCT_CAN_DROP(resource),
            ("%s: called for non-droppable resource %d", __func__, resource));

        RACCT_LOCK();
        KASSERT(amount <= p->p_racct->r_resources[resource],
            ("%s: freeing %ju of resource %d, which is more "
             "than allocated %jd for %s (pid %d)", __func__, amount, resource,
            (intmax_t)p->p_racct->r_resources[resource], p->p_comm, p->p_pid));

        racct_adjust_resource(p->p_racct, resource, -amount);
        racct_sub_cred_locked(p->p_ucred, resource, amount);
        RACCT_UNLOCK();
}

static void
racct_sub_cred_locked(struct ucred *cred, int resource, uint64_t amount)
{
        struct prison *pr;

        ASSERT_RACCT_ENABLED();

        racct_adjust_resource(cred->cr_ruidinfo->ui_racct, resource, -amount);
        for (pr = cred->cr_prison; pr != NULL; pr = pr->pr_parent)
                racct_adjust_resource(pr->pr_prison_racct->prr_racct, resource,
                    -amount);
        racct_adjust_resource(cred->cr_loginclass->lc_racct, resource, -amount);
}

/*
 * Decrease allocation of 'resource' by 'amount' for credential 'cred'.
 */
void
racct_sub_cred(struct ucred *cred, int resource, uint64_t amount)
{

        if (!racct_enable)
                return;

        SDT_PROBE3(racct, , rusage, sub__cred, cred, resource, amount);

#ifdef notyet
        KASSERT(RACCT_CAN_DROP(resource),
            ("%s: called for resource %d which can not drop", __func__,
             resource));
#endif

        RACCT_LOCK();
        racct_sub_cred_locked(cred, resource, amount);
        RACCT_UNLOCK();
}

/*
 * Inherit resource usage information from the parent process.
 */
int
racct_proc_fork(struct proc *parent, struct proc *child)
{
        int i, error = 0;

        if (!racct_enable)
                return (0);

        /*
         * Create racct for the child process.
         */
        racct_create(&child->p_racct);

        PROC_LOCK(parent);
        PROC_LOCK(child);
        RACCT_LOCK();

#ifdef RCTL
        error = rctl_proc_fork(parent, child);
        if (error != 0)
                goto out;
#endif

        child->p_throttled = 0;

        /*
         * Inherit resource usage.
         */
        for (i = 0; i <= RACCT_MAX; i++) {
                if (parent->p_racct->r_resources[i] == 0 ||
                    !RACCT_IS_INHERITABLE(i))
                        continue;

                error = racct_set_locked(child, i,
                    parent->p_racct->r_resources[i], 0);
                if (error != 0)
                        goto out;
        }

        error = racct_add_locked(child, RACCT_NPROC, 1, 0);
        error += racct_add_locked(child, RACCT_NTHR, 1, 0);

out:
        RACCT_UNLOCK();
        PROC_UNLOCK(child);
        PROC_UNLOCK(parent);

        if (error != 0)
                racct_proc_exit(child);

        return (error);
}

/*
 * Called at the end of fork1(), to handle rules that require the process
 * to be fully initialized.
 */
void
racct_proc_fork_done(struct proc *child)
{

        if (!racct_enable)
                return;

#ifdef RCTL
        PROC_LOCK(child);
        RACCT_LOCK();
        rctl_enforce(child, RACCT_NPROC, 0);
        rctl_enforce(child, RACCT_NTHR, 0);
        RACCT_UNLOCK();
        PROC_UNLOCK(child);
#endif
}

void
racct_proc_exit(struct proc *p)
{
        int i;

        if (!racct_enable)
                return;

        PROC_LOCK(p);
        RACCT_LOCK();

        racct_settime_locked(p, true);
        racct_zeropcpu_locked(p);

        KASSERT(p->p_racct->r_resources[RACCT_RSS] == 0,
            ("process reaped with %ju allocated for RSS\n",
            p->p_racct->r_resources[RACCT_RSS]));
        for (i = 0; i <= RACCT_MAX; i++) {
                if (p->p_racct->r_resources[i] == 0)
                        continue;
                if (!RACCT_IS_RECLAIMABLE(i))
                        continue;
                racct_set_locked(p, i, 0, 0);
        }

#ifdef RCTL
        rctl_racct_release(p->p_racct);
#endif
        racct_destroy_locked(&p->p_racct);
        RACCT_UNLOCK();
        PROC_UNLOCK(p);
}

/*
 * Called to signal credentials change, to move resource utilisation
 * between raccts.  Must be called with the proc lock held, in the same span as
 * the credentials change itself (i.e., without the proc lock being unlocked
 * between the two), but the order does not matter.
 */
void
racct_proc_ucred_changed(struct proc *p, struct ucred *oldcred,
    struct ucred *newcred)
{
        struct uidinfo *olduip, *newuip;
        struct loginclass *oldlc, *newlc;
        struct prison *oldpr, *newpr, *pr;

        if (!racct_enable)
                return;

        PROC_LOCK_ASSERT(p, MA_OWNED);

        newuip = newcred->cr_ruidinfo;
        olduip = oldcred->cr_ruidinfo;
        newlc = newcred->cr_loginclass;
        oldlc = oldcred->cr_loginclass;
        newpr = newcred->cr_prison;
        oldpr = oldcred->cr_prison;

        RACCT_LOCK();
        if (newuip != olduip) {
                racct_sub_racct(olduip->ui_racct, p->p_racct);
                racct_add_racct(newuip->ui_racct, p->p_racct);
        }
        if (newlc != oldlc) {
                racct_sub_racct(oldlc->lc_racct, p->p_racct);
                racct_add_racct(newlc->lc_racct, p->p_racct);
        }
        if (newpr != oldpr) {
                for (pr = oldpr; pr != NULL; pr = pr->pr_parent)
                        racct_sub_racct(pr->pr_prison_racct->prr_racct,
                            p->p_racct);
                for (pr = newpr; pr != NULL; pr = pr->pr_parent)
                        racct_add_racct(pr->pr_prison_racct->prr_racct,
                            p->p_racct);
        }
        RACCT_UNLOCK();
}

void
racct_move(struct racct *dest, struct racct *src)
{

        ASSERT_RACCT_ENABLED();

        RACCT_LOCK();
        racct_add_racct(dest, src);
        racct_sub_racct(src, src);
        dest->r_runtime = src->r_runtime;
        dest->r_time = src->r_time;
        src->r_runtime = 0;
        timevalsub(&src->r_time, &src->r_time);
        RACCT_UNLOCK();
}

static void
ast_racct(struct thread *td, int tda __unused)
{
        struct proc *p;

        ASSERT_RACCT_ENABLED();

        p = td->td_proc;
        if (p->p_throttled == 0)
                return;

        PROC_LOCK(p);
        while (p->p_throttled != 0) {
                msleep(p->p_racct, &p->p_mtx, 0, "racct",
                    p->p_throttled < 0 ? 0 : p->p_throttled);
                if (p->p_throttled > 0)
                        p->p_throttled = 0;
        }
        PROC_UNLOCK(p);
}

/*
 * Make the process sleep in userret() for 'timeout' ticks.  Setting
 * timeout to -1 makes it sleep until woken up by racct_proc_wakeup().
 */
void
racct_proc_throttle(struct proc *p, int timeout)
{
        struct thread *td;
#ifdef SMP
        int cpuid;
#endif

        KASSERT(timeout != 0, ("timeout %d", timeout));
        ASSERT_RACCT_ENABLED();
        PROC_LOCK_ASSERT(p, MA_OWNED);

        /*
         * Do not block kernel processes.  Also do not block processes with
         * low %cpu utilization to improve interactivity.
         */
        if ((p->p_flag & (P_SYSTEM | P_KPROC)) != 0)
                return;

        if (p->p_throttled < 0 || (timeout > 0 && p->p_throttled > timeout))
                return;

        p->p_throttled = timeout;

        FOREACH_THREAD_IN_PROC(p, td) {
                thread_lock(td);
                ast_sched_locked(td, TDA_RACCT);

                switch (TD_GET_STATE(td)) {
                case TDS_RUNQ:
                        /*
                         * If the thread is on the scheduler run-queue, we can
                         * not just remove it from there.  So we set the flag
                         * TDA_SCHED for the thread, so that once it is
                         * running, it is taken off the cpu as soon as possible.
                         */
                        ast_sched_locked(td, TDA_SCHED);
                        break;
                case TDS_RUNNING:
                        /*
                         * If the thread is running, we request a context
                         * switch for it by setting the TDA_SCHED flag.
                         */
                        ast_sched_locked(td, TDA_SCHED);
#ifdef SMP
                        cpuid = td->td_oncpu;
                        if ((cpuid != NOCPU) && (td != curthread))
                                ipi_cpu(cpuid, IPI_AST);
#endif
                        break;
                default:
                        break;
                }
                thread_unlock(td);
        }
}

static void
racct_proc_wakeup(struct proc *p)
{

        ASSERT_RACCT_ENABLED();

        PROC_LOCK_ASSERT(p, MA_OWNED);

        if (p->p_throttled != 0) {
                p->p_throttled = 0;
                wakeup(p->p_racct);
        }
}

static void
racct_decay_callback(struct racct *racct, void *dummy1, void *dummy2)
{
        ASSERT_RACCT_ENABLED();
        RACCT_LOCK_ASSERT();

#ifdef RCTL
        rctl_throttle_decay(racct, RACCT_READBPS);
        rctl_throttle_decay(racct, RACCT_WRITEBPS);
        rctl_throttle_decay(racct, RACCT_READIOPS);
        rctl_throttle_decay(racct, RACCT_WRITEIOPS);
#endif
}

static void
racct_decay_pre(void)
{

        RACCT_LOCK();
}

static void
racct_decay_post(void)
{

        RACCT_UNLOCK();
}

static void
racct_decay(void)
{

        ASSERT_RACCT_ENABLED();

        ui_racct_foreach(racct_decay_callback, racct_decay_pre,
            racct_decay_post, NULL, NULL);
        loginclass_racct_foreach(racct_decay_callback, racct_decay_pre,
            racct_decay_post, NULL, NULL);
        prison_racct_foreach(racct_decay_callback, racct_decay_pre,
            racct_decay_post, NULL, NULL);
}

static void
racct_updatepcpu_racct_locked(struct racct *racct)
{
        struct timeval diff;
        uint64_t elapsed;
        uint64_t runtime;
        uint64_t newpcpu;
        uint64_t oldpcpu;

        ASSERT_RACCT_ENABLED();
        RACCT_LOCK_ASSERT();

        /* Difference between now and previously-recorded time. */
        microuptime(&diff);
        timevalsub(&diff, &racct->r_time);
        elapsed = (uint64_t)diff.tv_sec * 1000000 + diff.tv_usec;

        /* Difference between current and previously-recorded runtime. */
        runtime = racct->r_resources[RACCT_CPU] - racct->r_runtime;

        newpcpu = runtime * 100 * 1000000 / elapsed;
        oldpcpu = racct->r_resources[RACCT_PCTCPU];
        /*
         * This calculation is equivalent to
         *    (1 - 0.3) * newpcpu + 0.3 * oldpcpu
         * where RACCT_DECAY_FACTOR = 0.3 * FSCALE.
         */
        racct->r_resources[RACCT_PCTCPU] = ((FSCALE - RACCT_DECAY_FACTOR) *
            newpcpu + RACCT_DECAY_FACTOR * oldpcpu) / FSCALE;
        if (racct->r_resources[RACCT_PCTCPU] >
            100 * 1000000 * (uint64_t)mp_ncpus)
                racct->r_resources[RACCT_PCTCPU] = 100 * 1000000 *
                    (uint64_t)mp_ncpus;

        /* Record current times. */
        racct->r_runtime = racct->r_resources[RACCT_CPU];
        timevaladd(&racct->r_time, &diff);
}

static void
racct_zeropcpu_locked(struct proc *p)
{
        ASSERT_RACCT_ENABLED();
        PROC_LOCK_ASSERT(p, MA_OWNED);

        p->p_racct->r_resources[RACCT_PCTCPU] = 0;
}

static void
racct_updatepcpu_locked(struct proc *p)
{
        ASSERT_RACCT_ENABLED();
        PROC_LOCK_ASSERT(p, MA_OWNED);

        racct_updatepcpu_racct_locked(p->p_racct);
}

static void
racct_updatepcpu_pre(void)
{

        RACCT_LOCK();
}

static void
racct_updatepcpu_post(void)
{

        RACCT_UNLOCK();
}

static void
racct_updatepcpu_racct_callback(struct racct *racct, void *dummy1, void *dummy2)
{
        racct_updatepcpu_racct_locked(racct);
}

static void
racct_updatepcpu_containers(void)
{
        ASSERT_RACCT_ENABLED();

        ui_racct_foreach(racct_updatepcpu_racct_callback, racct_updatepcpu_pre,
            racct_updatepcpu_post, NULL, NULL);
        loginclass_racct_foreach(racct_updatepcpu_racct_callback, racct_updatepcpu_pre,
            racct_updatepcpu_post, NULL, NULL);
        prison_racct_foreach(racct_updatepcpu_racct_callback, racct_updatepcpu_pre,
            racct_updatepcpu_post, NULL, NULL);
}

static bool
racct_proc_to_skip(const struct proc *p)
{
        PROC_LOCK_ASSERT(p, MA_OWNED);
        return (p->p_state != PRS_NORMAL || (p->p_flag & P_IDLEPROC) != 0);
}

static void
racctd(void)
{
        struct proc *p;

        ASSERT_RACCT_ENABLED();

        for (;;) {
                racct_decay();

                sx_slock(&allproc_lock);

                FOREACH_PROC_IN_SYSTEM(p) {
                        PROC_LOCK(p);
                        if (racct_proc_to_skip(p)) {
                                PROC_UNLOCK(p);
                                continue;
                        }

                        RACCT_LOCK();
#ifdef RCTL
                        rctl_throttle_decay(p->p_racct, RACCT_READBPS);
                        rctl_throttle_decay(p->p_racct, RACCT_WRITEBPS);
                        rctl_throttle_decay(p->p_racct, RACCT_READIOPS);
                        rctl_throttle_decay(p->p_racct, RACCT_WRITEIOPS);
#endif
                        racct_settime_locked(p, false);
                        racct_updatepcpu_locked(p);
                        RACCT_UNLOCK();
                        PROC_UNLOCK(p);
                }

                /*
                 * To ensure that processes are throttled in a fair way, we need
                 * to iterate over all processes again and check the limits
                 * for %cpu resource only after ucred racct containers have been
                 * properly filled.
                 */
                FOREACH_PROC_IN_SYSTEM(p) {
                        PROC_LOCK(p);
                        if (racct_proc_to_skip(p)) {
                                PROC_UNLOCK(p);
                                continue;
                        }

                        if (racct_pcpu_available(p) <= 0) {
                                if (p->p_racct->r_resources[RACCT_PCTCPU] >
                                    pcpu_threshold)
                                        racct_proc_throttle(p, -1);
                        } else if (p->p_throttled == -1) {
                                racct_proc_wakeup(p);
                        }
                        PROC_UNLOCK(p);
                }
                sx_sunlock(&allproc_lock);

                racct_updatepcpu_containers();
                pause("-", hz);
        }
}

static struct kproc_desc racctd_kp = {
        "racctd",
        racctd,
        NULL
};

static void
racctd_init(void *dummy __unused)
{
        if (!racct_enable)
                return;

        kproc_start(&racctd_kp);
}
SYSINIT(racctd, SI_SUB_RACCTD, SI_ORDER_FIRST, racctd_init, NULL);

static void
racct_init(void *dummy __unused)
{
        if (!racct_enable)
                return;

        racct_zone = uma_zcreate("racct", sizeof(struct racct),
            NULL, NULL, NULL, NULL, UMA_ALIGN_PTR, 0);
        ast_register(TDA_RACCT, ASTR_ASTF_REQUIRED, 0, ast_racct);

        /*
         * XXX: Move this somewhere.
         */
        prison0.pr_prison_racct = prison_racct_find("0");
}
SYSINIT(racct, SI_SUB_RACCT, SI_ORDER_FIRST, racct_init, NULL);