Searched refs:down_write_killable (Results 1 – 19 of 19) sorted by relevance
48 err = down_write_killable(&inode->i_rwsem); in nfs_start_io_read()82 err = down_write_killable(&inode->i_rwsem); in nfs_start_io_write()143 err = down_write_killable(&inode->i_rwsem); in nfs_start_io_direct()
80 err = down_write_killable(&inode->i_rwsem); in ceph_start_io_read()112 int err = down_write_killable(&inode->i_rwsem); in ceph_start_io_write()192 err = down_write_killable(&inode->i_rwsem); in ceph_start_io_direct()
69 if (down_write_killable(&inode->i_rwsem) < 0) in netfs_start_io_read()106 if (down_write_killable(&inode->i_rwsem) < 0) in netfs_start_io_write()181 if (down_write_killable(&inode->i_rwsem) < 0) in netfs_start_io_direct()
242 extern int __must_check down_write_killable(struct rw_semaphore *sem) __cond_acquires(0, sem);272 DEFINE_LOCK_GUARD_1_COND(rwsem_write, _kill, down_write_killable(_T->lock), _RET == 0)329 # define down_write_killable_nested(sem, subclass) down_write_killable(sem)
554 ret = down_write_killable(&mm->mmap_lock); in mmap_write_lock_killable()
1033 return down_write_killable(&inode->i_rwsem); in inode_lock_killable()
386 rc = down_write_killable(&dax_region_rwsem); in create_store()513 rc = down_write_killable(&dax_dev_rwsem); in free_dev_dax_id()712 rc = down_write_killable(&dax_region_rwsem); in get_dax_range()1111 rc = down_write_killable(&dax_region_rwsem); in size_store()1118 rc = down_write_killable(&dax_dev_rwsem); in size_store()1178 rc = down_write_killable(&dax_region_rwsem); in mapping_store()1185 rc = down_write_killable(&dax_dev_rwsem); in mapping_store()1242 rc = down_write_killable(&dax_region_rwsem); in align_store()1250 rc = down_write_killable(&dax_dev_rwsem); in align_store()1347 rc = down_write_killable(&dax_dev_rwsem); in memmap_on_memory_store()
71 rc = down_write_killable(&ctx->rwsem); in tm_digest_read()106 rc = down_write_killable(&ctx->rwsem); in tm_digest_write()
217 err = down_write_killable(&vm->lock); in xe_exec_ioctl()
429 err = down_write_killable(&vm->lock); in xe_vm_madvise_ioctl()
1636 lock_err = down_write_killable(&driver_migrate_lock); in xe_svm_alloc_vram()
3658 err = down_write_killable(&vm->lock); in xe_vm_bind_ioctl()
406 ret = down_write_killable(&vnode->validate_lock); in afs_validate()
333 if (down_write_killable(&dvnode->validate_lock) < 0) in afs_read_dir()1441 ret = down_write_killable(&vnode->rmdir_lock); in afs_rmdir()
1598 int __sched down_write_killable(struct rw_semaphore *sem) in down_write_killable() function1611 EXPORT_SYMBOL(down_write_killable);
619 if (down_write_killable(&mm->context.ldt_usr_sem)) in write_ldt()
848 ret = down_write_killable(&tsk->signal->exec_update_lock); in exec_mmap()
1537 if (down_write_killable(&frag->frag_sem)) { in configfs_rmdir()
1653 r = down_write_killable(&adev->reset_domain->sem); in amdgpu_debugfs_test_ib_show()