root/kernel/locking/ww_rt_mutex.c
// SPDX-License-Identifier: GPL-2.0-only
/*
 * rtmutex API
 */
#include <linux/spinlock.h>
#include <linux/export.h>

#define RT_MUTEX_BUILD_MUTEX
#define WW_RT
#include "rtmutex.c"

int ww_mutex_trylock(struct ww_mutex *lock, struct ww_acquire_ctx *ww_ctx)
{
        struct rt_mutex *rtm = &lock->base;

        if (!ww_ctx)
                return rt_mutex_trylock(rtm);

        /*
         * Reset the wounded flag after a kill. No other process can
         * race and wound us here, since they can't have a valid owner
         * pointer if we don't have any locks held.
         */
        if (ww_ctx->acquired == 0)
                ww_ctx->wounded = 0;

        if (__rt_mutex_trylock(&rtm->rtmutex)) {
                ww_mutex_set_context_fastpath(lock, ww_ctx);
                mutex_acquire_nest(&rtm->dep_map, 0, 1, &ww_ctx->dep_map, _RET_IP_);
                return 1;
        }

        return 0;
}
EXPORT_SYMBOL(ww_mutex_trylock);

static int __sched
__ww_rt_mutex_lock(struct ww_mutex *lock, struct ww_acquire_ctx *ww_ctx,
                   unsigned int state, unsigned long ip)
{
        struct lockdep_map __maybe_unused *nest_lock = NULL;
        struct rt_mutex *rtm = &lock->base;
        int ret;

        might_sleep();

        if (ww_ctx) {
                if (unlikely(ww_ctx == READ_ONCE(lock->ctx)))
                        return -EALREADY;

                /*
                 * Reset the wounded flag after a kill. No other process can
                 * race and wound us here, since they can't have a valid owner
                 * pointer if we don't have any locks held.
                 */
                if (ww_ctx->acquired == 0)
                        ww_ctx->wounded = 0;

#ifdef CONFIG_DEBUG_LOCK_ALLOC
                nest_lock = &ww_ctx->dep_map;
#endif
        }
        mutex_acquire_nest(&rtm->dep_map, 0, 0, nest_lock, ip);

        if (likely(rt_mutex_try_acquire(&rtm->rtmutex))) {
                if (ww_ctx)
                        ww_mutex_set_context_fastpath(lock, ww_ctx);
                return 0;
        }

        ret = rt_mutex_slowlock(&rtm->rtmutex, ww_ctx, state);

        if (ret)
                mutex_release(&rtm->dep_map, ip);
        return ret;
}

int __sched
ww_mutex_lock(struct ww_mutex *lock, struct ww_acquire_ctx *ctx)
{
        return __ww_rt_mutex_lock(lock, ctx, TASK_UNINTERRUPTIBLE, _RET_IP_);
}
EXPORT_SYMBOL(ww_mutex_lock);

int __sched
ww_mutex_lock_interruptible(struct ww_mutex *lock, struct ww_acquire_ctx *ctx)
{
        return __ww_rt_mutex_lock(lock, ctx, TASK_INTERRUPTIBLE, _RET_IP_);
}
EXPORT_SYMBOL(ww_mutex_lock_interruptible);

void __sched ww_mutex_unlock(struct ww_mutex *lock)
{
        struct rt_mutex *rtm = &lock->base;

        __ww_mutex_unlock(lock);

        mutex_release(&rtm->dep_map, _RET_IP_);
        __rt_mutex_unlock(&rtm->rtmutex);
}
EXPORT_SYMBOL(ww_mutex_unlock);