/kernel/linux/linux-5.10/kernel/sched/ |
D | completion.c | 243 long t = wait_for_common(x, MAX_SCHEDULE_TIMEOUT, TASK_KILLABLE); in wait_for_completion_killable() 266 return wait_for_common(x, timeout, TASK_KILLABLE); in wait_for_completion_killable_timeout()
|
/kernel/linux/linux-5.10/include/linux/ |
D | wait.h | 274 state == TASK_INTERRUPTIBLE || state == TASK_KILLABLE) \ 620 ___wait_event(wq, condition, TASK_KILLABLE, 1, 0, \ 898 ___wait_event(wq, condition, TASK_KILLABLE, 0, 0, schedule()) 926 TASK_KILLABLE, 0, timeout, \
|
D | wait_bit.h | 282 ___wait_var_event(var, condition, TASK_KILLABLE, 0, 0, \
|
/kernel/linux/linux-5.10/fs/xfs/ |
D | xfs_extent_busy.c | 609 prepare_to_wait(&pag->pagb_wait, &wait, TASK_KILLABLE); in xfs_extent_busy_flush() 629 prepare_to_wait(&pag->pagb_wait, &wait, TASK_KILLABLE); in xfs_extent_busy_wait_all()
|
D | xfs_trans_ail.c | 588 set_current_state(TASK_KILLABLE); in xfsaild()
|
/kernel/linux/linux-5.10/kernel/locking/ |
D | semaphore.c | 247 return __down_common(sem, TASK_KILLABLE, MAX_SCHEDULE_TIMEOUT); in __down_killable()
|
D | mutex.c | 1139 return __mutex_lock(lock, TASK_KILLABLE, subclass, NULL, _RET_IP_); in mutex_lock_killable_nested() 1376 return __mutex_lock(lock, TASK_KILLABLE, 0, NULL, _RET_IP_); in __mutex_lock_killable_slowpath()
|
D | rwsem.c | 1363 if (IS_ERR(rwsem_down_read_slowpath(sem, TASK_KILLABLE))) in __down_read_killable() 1412 if (IS_ERR(rwsem_down_write_slowpath(sem, TASK_KILLABLE))) in __down_write_killable()
|
/kernel/linux/linux-5.10/lib/ |
D | test_lockup.c | 463 wait_state = TASK_KILLABLE; in test_lockup_init()
|
/kernel/linux/linux-5.10/net/sunrpc/ |
D | sched.c | 332 action, TASK_KILLABLE); in __rpc_wait_for_completion_task() 941 TASK_KILLABLE); in __rpc_execute()
|
/kernel/linux/linux-5.10/fs/erofs/ |
D | zmap.c | 48 if (wait_on_bit_lock(&vi->flags, EROFS_I_BL_Z_BIT, TASK_KILLABLE)) in z_erofs_fill_inode_lazy()
|
D | xattr.c | 60 if (wait_on_bit_lock(&vi->flags, EROFS_I_BL_XATTR_BIT, TASK_KILLABLE)) in init_inode_xattrs()
|
/kernel/linux/linux-5.10/fs/orangefs/ |
D | file.c | 312 ret = wait_on_bit(bitlock, 1, TASK_KILLABLE); in orangefs_revalidate_mapping()
|
D | inode.c | 685 if (wait_on_bit(bitlock, 1, TASK_KILLABLE)) { in orangefs_page_mkwrite()
|
/kernel/linux/linux-5.10/fs/hmdfs/ |
D | client_writeback.c | 382 __set_current_state(TASK_KILLABLE); in hmdfs_balance_dirty_pages()
|
/kernel/linux/linux-5.10/fs/nfs/ |
D | file.c | 561 nfs_wait_bit_killable, TASK_KILLABLE); in nfs_vm_page_mkwrite()
|
D | pnfs.c | 1905 TASK_KILLABLE); in pnfs_prepare_to_retry_layoutget() 2078 TASK_KILLABLE)); in pnfs_update_layout() 3181 TASK_KILLABLE); in pnfs_layoutcommit_inode()
|
/kernel/linux/linux-5.10/fs/ |
D | exec.c | 1065 __set_current_state(TASK_KILLABLE); in de_thread() 1092 __set_current_state(TASK_KILLABLE); in de_thread()
|
D | userfaultfd.c | 350 return TASK_KILLABLE; in userfaultfd_get_blocking_state() 567 set_current_state(TASK_KILLABLE); in userfaultfd_event_wait_completion()
|
/kernel/linux/linux-5.10/fs/overlayfs/ |
D | copy_up.c | 177 if (signal_pending_state(TASK_KILLABLE, current)) { in ovl_copy_up_data()
|
/kernel/linux/linux-5.10/drivers/staging/hungtask/ |
D | hungtask_base.c | 778 (t->state == TASK_KILLABLE)) in htbase_check_tasks()
|
/kernel/linux/linux-5.10/fs/cifs/ |
D | misc.c | 570 TASK_KILLABLE); in cifs_get_writer()
|
/kernel/linux/linux-5.10/Documentation/scheduler/ |
D | completion.rst | 228 Further variants include _killable which uses TASK_KILLABLE as the
|
/kernel/linux/linux-5.10/kernel/ |
D | ptrace.c | 498 wait_on_bit(&task->jobctl, JOBCTL_TRAPPING_BIT, TASK_KILLABLE); in ptrace_attach()
|
/kernel/linux/linux-5.10/fs/btrfs/ |
D | space-info.c | 1193 ret = prepare_to_wait_event(&ticket->wait, &wait, TASK_KILLABLE); in wait_reserve_ticket()
|