mlock
volatile u32 mlock; /* Global GIO memory access lock */
mutex_init(&ce->mlock);
mutex_lock(&ce->mlock);
mutex_unlock(&ce->mlock);
struct mutex mlock;
mutex_lock(&ss->mlock);
mutex_unlock(&ss->mlock);
mutex_init(&ss->mlock);
mutex_lock(&ss->mlock);
mutex_unlock(&ss->mlock);
mutex_lock(&ss->mlock);
mutex_unlock(&ss->mlock);
struct mutex mlock;
struct mutex mlock; /* protects tmp[] if can_sleep */
mutex_lock(&fwd->mlock);
mutex_unlock(&fwd->mlock);
mutex_lock(&fwd->mlock);
mutex_unlock(&fwd->mlock);
mutex_init(&fwd->mlock);
struct mutex mlock; /* for host transfer operations */
guard(mutex)(&priv->mlock);
mutex_init(&priv->mlock);
static inline u32 host1x_opcode_acquire_mlock(unsigned mlock)
return (14 << 28) | (0 << 24) | mlock;
static inline u32 host1x_opcode_release_mlock(unsigned mlock)
return (14 << 28) | (1 << 24) | mlock;
guard(mutex)(&iio_dev_opaque->mlock);
guard(mutex)(&iio_dev_opaque->mlock);
guard(mutex)(&iio_dev_opaque->mlock);
guard(mutex)(&iio_dev_opaque->mlock);
guard(mutex)(&iio_dev_opaque->mlock);
guard(mutex)(&iio_dev_opaque->mlock);
mutex_destroy(&iio_dev_opaque->mlock);
mutex_init_with_key(&iio_dev_opaque->mlock, &iio_dev_opaque->mlock_key);
mutex_lock(&to_iio_dev_opaque(indio_dev)->mlock);
mutex_unlock(&to_iio_dev_opaque(indio_dev)->mlock);
ret = mutex_lock_interruptible(&iio_dev_opaque->mlock);
mutex_unlock(&iio_dev_opaque->mlock);
mutex_unlock(&iio_dev_opaque->mlock);
fd = mutex_lock_interruptible(&iio_dev_opaque->mlock);
mutex_unlock(&iio_dev_opaque->mlock);
guard(mutex)(&iio_dev_opaque->mlock);
scoped_guard(mutex, &iio_dev_opaque->mlock) {
guard(mutex)(&iio_dev_opaque->mlock);
guard(mutex)(&iio_dev_opaque->mlock);
struct mutex mlock;
mutex_lock(&ts->mlock);
mutex_unlock(&ts->mlock);
mutex_init(&ts->mlock);
mutex_lock(&tsc->mlock);
mutex_unlock(&tsc->mlock);
vq->lock = &pcdev->mlock;
mutex_lock(&pcdev->mlock);
mutex_unlock(&pcdev->mlock);
mutex_lock(&pcdev->mlock);
mutex_unlock(&pcdev->mlock);
mutex_lock(&pcdev->mlock);
vdev->lock = &pcdev->mlock;
mutex_unlock(&pcdev->mlock);
mutex_lock(&pcdev->mlock);
mutex_unlock(&pcdev->mlock);
mutex_init(&pcdev->mlock);
struct mutex mlock;
mutex_lock(&ceudev->mlock);
mutex_unlock(&ceudev->mlock);
mutex_lock(&ceudev->mlock);
mutex_unlock(&ceudev->mlock);
q->lock = &ceudev->mlock;
vdev->lock = &ceudev->mlock;
mutex_init(&ceudev->mlock);
struct mutex mlock;
mutex_init(&genpd->mlock);
mutex_lock(&genpd->mlock);
mutex_lock_nested(&genpd->mlock, depth);
return mutex_lock_interruptible(&genpd->mlock);
return mutex_unlock(&genpd->mlock);
lockdep_set_subclass(&domain->genpd.mlock, 1);
lockdep_set_class(&domain->genpd.mlock,
lockdep_set_class(&domain->genpd.mlock,
lockdep_set_class(&domain->genpd.mlock,
struct mutex mlock;
mutex_lock(&ospi->mlock);
mutex_unlock(&ospi->mlock);
mutex_lock(&ospi->mlock);
mutex_unlock(&ospi->mlock);
ret = devm_mutex_init(dev, &ospi->mlock);
struct mutex mlock;
struct mutex mlock;
COND_SYSCALL(mlock);
{ mlock|dirty, mlock|dirty, MF_MSG_DIRTY_MLOCKED_LRU, me_pagecache_dirty },
{ mlock|dirty, mlock, MF_MSG_CLEAN_MLOCKED_LRU, me_pagecache_clean },
unsigned long mlock;
mlock = (unsigned long)folio & (LRU_FOLIO | NEW_FOLIO);
folio = (struct folio *)((unsigned long)folio - mlock);
if (mlock & LRU_FOLIO)
else if (mlock & NEW_FOLIO)
SYSCALL_DEFINE2(mlock, unsigned long, start, size_t, len)
mutex_init(&dev->mlock);
struct mutex mlock;
mutex_lock(&dev->mlock);
mutex_unlock(&dev->mlock);
mutex_lock(&dev->mlock);
mutex_unlock(&dev->mlock);
mutex_lock(&dev->mlock);
mutex_unlock(&dev->mlock);
mutex_lock(&dev->mlock);
mutex_unlock(&dev->mlock);
mutex_lock(&dev->mlock);
mutex_unlock(&dev->mlock);
mutex_lock(&dev->mlock);
mutex_unlock(&dev->mlock);
mutex_lock(&dev->mlock);
mutex_unlock(&dev->mlock);
mutex_lock(&dev->mlock);
mutex_unlock(&dev->mlock);
mutex_lock(&dev->mlock);
mutex_unlock(&dev->mlock);
mlock(buf, size);
mlock(addr, page_size);
mlock(addr, page_size);
TEST_F(guard_regions, mlock)
ASSERT_EQ(mlock(ptr, 10 * page_size), 0);
ASSERT_EQ(mlock(&ptr[5 * page_size], 5 * page_size), 0);
ret = mlock(p + start_offset, lock_size);
ret = mlock(p + start_offset, lock_size);
if (mlock(t->commit_buf, total))
if (mlock(t->fetch[i].fetch_buf, t->fetch[i].fetch_buf_size))