down_write_killable
if (down_write_killable(&mm->context.ldt_usr_sem))
rc = down_write_killable(&dax_region_rwsem);
rc = down_write_killable(&dax_dev_rwsem);
rc = down_write_killable(&dax_region_rwsem);
rc = down_write_killable(&dax_dev_rwsem);
rc = down_write_killable(&dax_region_rwsem);
rc = down_write_killable(&dax_dev_rwsem);
rc = down_write_killable(&dax_dev_rwsem);
rc = down_write_killable(&dax_region_rwsem);
rc = down_write_killable(&dax_dev_rwsem);
rc = down_write_killable(&dax_region_rwsem);
r = down_write_killable(&adev->reset_domain->sem);
err = down_write_killable(&vm->lock);
err = down_write_killable(&group->mode_sem);
err = down_write_killable(&group->mode_sem);
lock_err = down_write_killable(&driver_migrate_lock);
ret = down_write_killable(&val->lock);
err = down_write_killable(&vm->lock);
err = down_write_killable(&vm->lock);
rc = down_write_killable(&ctx->rwsem);
rc = down_write_killable(&ctx->rwsem);
ret = down_write_killable(&vnode->rmdir_lock);
if (down_write_killable(&dvnode->validate_lock) < 0)
ret = down_write_killable(&vnode->validate_lock);
int err = down_write_killable(&inode->i_rwsem);
err = down_write_killable(&inode->i_rwsem);
err = down_write_killable(&inode->i_rwsem);
if (down_write_killable(&frag->frag_sem)) {
ret = down_write_killable(&tsk->signal->exec_update_lock);
if (down_write_killable(&inode->i_rwsem) < 0)
if (down_write_killable(&inode->i_rwsem) < 0)
if (down_write_killable(&inode->i_rwsem) < 0)
err = down_write_killable(&inode->i_rwsem);
err = down_write_killable(&inode->i_rwsem);
err = down_write_killable(&inode->i_rwsem);
return down_write_killable(&inode->i_rwsem);
ret = down_write_killable(&mm->mmap_lock);
extern int __must_check down_write_killable(struct rw_semaphore *sem) __cond_acquires(0, sem);
DEFINE_LOCK_GUARD_1_COND(rwsem_write, _kill, down_write_killable(_T->lock), _RET == 0)
# define down_write_killable_nested(sem, subclass) down_write_killable(sem)
EXPORT_SYMBOL(down_write_killable);