/linux/kernel/locking/ |
H A D | rtmutex_api.c | 68 void __sched rt_mutex_lock_nested(struct rt_mutex *lock, unsigned int subclass) in rt_mutex_lock_nested() 74 void __sched _rt_mutex_lock_nest_lock(struct rt_mutex *lock, struct lockdep_map *nest_lock) in _rt_mutex_lock_nest_lock() 87 void __sched rt_mutex_lock(struct rt_mutex *lock) in rt_mutex_lock() 103 int __sched rt_mutex_lock_interruptible(struct rt_mutex *lock) in rt_mutex_lock_interruptible() 118 int __sched rt_mutex_lock_killable(struct rt_mutex *lock) in rt_mutex_lock_killable() 136 int __sched rt_mutex_trylock(struct rt_mutex *lock) in rt_mutex_trylock() 156 void __sched rt_mutex_unlock(struct rt_mutex *lock) in rt_mutex_unlock() 166 int __sched rt_mutex_futex_trylock(struct rt_mutex_base *lock) in rt_mutex_futex_trylock() 171 int __sched __rt_mutex_futex_trylock(struct rt_mutex_base *lock) in __rt_mutex_futex_trylock() 183 bool __sched __rt_mutex_futex_unloc [all...] |
H A D | mutex.c | 246 static void __sched __mutex_lock_slowpath(struct mutex *lock); 269 void __sched mutex_lock(struct mutex *lock) in mutex_lock() 512 static noinline void __sched __mutex_unlock_slowpath(struct mutex *lock, unsigned long ip); 530 void __sched mutex_unlock(struct mutex *lock) in mutex_unlock() 551 void __sched ww_mutex_unlock(struct ww_mutex *lock) in ww_mutex_unlock() 561 static __always_inline int __sched 756 static int __sched 763 static int __sched 809 void __sched 817 void __sched [all...] |
H A D | spinlock_rt.c | 54 void __sched rt_spin_lock(spinlock_t *lock) __acquires(RCU) in rt_spin_lock() 62 void __sched rt_spin_lock_nested(spinlock_t *lock, int subclass) in rt_spin_lock_nested() 69 void __sched rt_spin_lock_nest_lock(spinlock_t *lock, in rt_spin_lock_nest_lock() 78 void __sched rt_spin_unlock(spinlock_t *lock) __releases(RCU) in rt_spin_unlock() 94 void __sched rt_spin_lock_unlock(spinlock_t *lock) in rt_spin_lock_unlock() 116 int __sched rt_spin_trylock(spinlock_t *lock) in rt_spin_trylock() 122 int __sched rt_spin_trylock_bh(spinlock_t *lock) in rt_spin_trylock_bh() 201 int __sched rt_read_trylock(rwlock_t *rwlock) in rt_read_trylock() 215 int __sched rt_write_trylock(rwlock_t *rwlock) in rt_write_trylock() 229 void __sched rt_read_loc [all...] |
H A D | semaphore.c | 91 void __sched down(struct semaphore *sem) in down() 114 int __sched down_interruptible(struct semaphore *sem) in down_interruptible() 141 int __sched down_killable(struct semaphore *sem) in down_killable() 171 int __sched down_trylock(struct semaphore *sem) in down_trylock() 196 int __sched down_timeout(struct semaphore *sem, long timeout) in down_timeout() 220 void __sched up(struct semaphore *sem) in up() 252 static inline int __sched ___down_common(struct semaphore *sem, long state, in ___down_common() 285 static inline int __sched __down_common(struct semaphore *sem, long state, in __down_common() 301 static noinline void __sched __down(struct semaphore *sem) in __down() 306 static noinline int __sched __down_interruptibl [all...] |
H A D | ww_rt_mutex.c | 37 static int __sched 78 int __sched 85 int __sched 92 void __sched ww_mutex_unlock(struct ww_mutex *lock) in ww_mutex_unlock()
|
H A D | rtmutex.c | 309 static int __sched rt_mutex_slowtrylock(struct rt_mutex_base *lock); 678 static int __sched rt_mutex_adjust_prio_chain(struct task_struct *task, in rt_mutex_adjust_prio_chain() 1086 static int __sched 1203 static int __sched task_blocks_on_rt_mutex(struct rt_mutex_base *lock, in task_blocks_on_rt_mutex() 1312 static void __sched mark_wakeup_next_waiter(struct rt_wake_q_head *wqh, in mark_wakeup_next_waiter() 1358 static int __sched __rt_mutex_slowtrylock(struct rt_mutex_base *lock) in __rt_mutex_slowtrylock() 1374 static int __sched rt_mutex_slowtrylock(struct rt_mutex_base *lock) in rt_mutex_slowtrylock() 1411 static void __sched rt_mutex_slowunlock(struct rt_mutex_base *lock) in rt_mutex_slowunlock() 1539 static void __sched remove_waiter(struct rt_mutex_base *lock, in remove_waiter() 1604 static int __sched rt_mutex_slowlock_bloc [all...] |
H A D | rwbase_rt.c | 68 static int __sched __rwbase_read_lock(struct rwbase_rt *rwb, in __rwbase_read_lock() 150 static void __sched __rwbase_read_unlock(struct rwbase_rt *rwb, in __rwbase_read_unlock() 237 static int __sched rwbase_write_lock(struct rwbase_rt *rwb, in rwbase_write_lock()
|
H A D | rwsem.c | 992 static struct rw_semaphore __sched * 1110 static struct rw_semaphore __sched * 1489 static inline void __sched __down_write(struct rw_semaphore *sem) in __down_write() 1494 static inline int __sched __down_write_killable(struct rw_semaphore *sem) in __down_write_killable() 1534 void __sched down_read(struct rw_semaphore *sem) in down_read() 1543 int __sched down_read_interruptible(struct rw_semaphore *sem) in down_read_interruptible() 1557 int __sched down_read_killable(struct rw_semaphore *sem) in down_read_killable() 1587 void __sched down_write(struct rw_semaphore *sem) in down_write() 1598 int __sched down_write_killable(struct rw_semaphore *sem) in down_write_killable() 1711 int __sched down_write_killable_neste [all...] |
H A D | percpu-rwsem.c | 169 bool __sched __percpu_down_read(struct percpu_rw_semaphore *sem, bool try, in __percpu_down_read() 227 void __sched percpu_down_write(struct percpu_rw_semaphore *sem) in percpu_down_write()
|
/linux/kernel/time/ |
H A D | sleep_timeout.c | 61 signed long __sched schedule_timeout(signed long timeout) in schedule_timeout() 125 signed long __sched schedule_timeout_interruptible(signed long timeout) in schedule_timeout_interruptible() 140 signed long __sched schedule_timeout_killable(signed long timeout) in schedule_timeout_killable() 155 signed long __sched schedule_timeout_uninterruptible(signed long timeout) in schedule_timeout_uninterruptible() 172 signed long __sched schedule_timeout_idle(signed long timeout) in schedule_timeout_idle() 189 int __sched schedule_hrtimeout_range_clock(ktime_t *expires, u64 delta, in schedule_hrtimeout_range_clock() 260 int __sched schedule_hrtimeout_range(ktime_t *expires, u64 delta, in schedule_hrtimeout_range() 276 int __sched schedule_hrtimeout(ktime_t *expires, const enum hrtimer_mode mode) in schedule_hrtimeout() 362 void __sched usleep_range_state(unsigned long min, unsigned long max, unsigned int state) in usleep_range_state()
|
H A D | alarmtimer.c | 752 static long __sched alarm_timer_nsleep_restart(struct restart_block *restart) in alarm_timer_nsleep_restart()
|
H A D | hrtimer.c | 2091 static int __sched do_nanosleep(struct hrtimer_sleeper *t, enum hrtimer_mode mode) in do_nanosleep() 2126 static long __sched hrtimer_nanosleep_restart(struct restart_block *restart) in hrtimer_nanosleep_restart()
|
/linux/drivers/tty/ |
H A D | tty_ldsem.c | 155 static struct ld_semaphore __sched * 229 static struct ld_semaphore __sched * 337 int __sched ldsem_down_read(struct ld_semaphore *sem, long timeout) in ldsem_down_read() 363 int __sched ldsem_down_write(struct ld_semaphore *sem, long timeout) in ldsem_down_write()
|
/linux/lib/ |
H A D | closure.c | 135 void __sched __closure_sync(struct closure *cl) in __closure_sync() 160 void __sched closure_return_sync(struct closure *cl) in closure_return_sync() 188 int __sched __closure_sync_timeout(struct closure *cl, unsigned long timeout) in __closure_sync_timeout()
|
/linux/kernel/ |
H A D | latencytop.c | 110 static void __sched 175 void __sched
|
/linux/include/linux/sched/ |
H A D | debug.h | 44 #define __sched __section(".sched.text") macro 46 /* Linker adds these: start and end of __sched functions */ 49 /* Is this address in the __sched functions? */
|
/linux/kernel/sched/ |
H A D | syscalls.c | 1399 void __sched yield(void) in yield() 1421 int __sched yield_to(struct task_struct *p, bool preempt) in yield_to()
|
/linux/arch/arm64/kernel/ |
H A D | process.c | 704 __notrace_funcgraph __sched
|
H A D | entry-common.c | 274 static void __sched arm64_preempt_schedule_irq(void) in arm64_preempt_schedule_irq()
|
/linux/kernel/printk/ |
H A D | printk.c | 3340 void __sched console_conditional_schedule(void) in console_conditional_schedule()
|