/linux/include/linux/sched/ |
H A D | clock.h | 10 * sched_clock() has no promise of monotonicity or bounded drift between 15 extern u64 sched_clock(void); 22 return sched_clock(); in sched_clock_noinstr() 54 return sched_clock(); in cpu_clock() 64 return sched_clock(); in local_clock() 72 * between local_clock() and sched_clock(). 103 * An i/f to runtime opt-in for irq time accounting based off of sched_clock.
|
/linux/Documentation/timers/ |
H A D | timekeeping.rst | 2 Clock sources, Clock events, sched_clock() and delay timers 12 architecture-specific overrides of the sched_clock() function and some 18 sched_clock() is used for scheduling and timestamping, and delay timers 108 sched_clock() 112 function in the kernel called sched_clock(). This function shall return the 114 may not provide an implementation of sched_clock() on its own. If a local 116 sched_clock(). 118 As the name suggests, sched_clock() is used for scheduling the system, 123 Compared to clock sources, sched_clock() has to be very fast: it is called 126 for speed in sched_clock() [all...] |
/linux/kernel/locking/ |
H A D | qspinlock_stat.h | 110 u64 start = sched_clock(); in __pv_kick() 114 this_cpu_add(EVENT_COUNT(pv_latency_kick), sched_clock() - start); in __pv_kick() 128 sched_clock() - *pkick_time); in __pv_wait()
|
H A D | rwsem.c | 813 return sched_clock() + delta; in rwsem_rspin_threshold() 868 * avoid calling sched_clock() too frequently so in rwsem_optimistic_spin() 873 else if (!(++loop & 0xf) && (sched_clock() > rspin_threshold)) { in rwsem_optimistic_spin()
|
/linux/drivers/clocksource/ |
H A D | hyperv_timer.c | 18 #include <linux/sched_clock.h> 478 * used to calculate time for hv tsc page based sched_clock, to account for 528 static __always_inline void hv_setup_sched_clock(void *sched_clock) in hv_setup_sched_clock() argument 535 sched_clock_register(sched_clock, 64, NSEC_PER_SEC); in hv_setup_sched_clock() 538 static __always_inline void hv_setup_sched_clock(void *sched_clock) in hv_setup_sched_clock() argument 541 paravirt_set_sched_clock(sched_clock); in hv_setup_sched_clock() 544 static __always_inline void hv_setup_sched_clock(void *sched_clock) {} in hv_setup_sched_clock() argument
|
/linux/arch/arm/mach-omap2/ |
H A D | pm-debug.c | 60 t = sched_clock(); in pm_dbg_update_time() 180 t = sched_clock(); in pwrdms_setup()
|
/linux/arch/m68k/coldfire/ |
H A D | dma_timer.c | 77 unsigned long long sched_clock(void) in sched_clock() function
|
/linux/arch/powerpc/lib/ |
H A D | qspinlock.c | 66 u64 delta = sched_clock() - seen; in recently_sleepy() 239 this_cpu_write(sleepy_lock_seen_clock, sched_clock()); in seen_sleepy_owner() 248 this_cpu_write(sleepy_lock_seen_clock, sched_clock()); in seen_sleepy_lock() 255 this_cpu_write(sleepy_lock_seen_clock, sched_clock()); in seen_sleepy_node()
|
/linux/kernel/trace/ |
H A D | trace_clock.c | 37 * sched_clock() is an architecture implemented, fast, scalable, in trace_clock_local() 42 clock = sched_clock(); in trace_clock_local()
|
/linux/kernel/time/ |
H A D | Makefile | 23 obj-$(CONFIG_GENERIC_SCHED_CLOCK) += sched_clock.o
|
H A D | sched_clock.c | 3 * Generic sched_clock() support, to extend low level hardware time 17 #include <linux/sched_clock.h> 24 * struct clock_data - all data needed for sched_clock() (including 29 * @read_data: Data required to read from sched_clock. 106 unsigned long long notrace sched_clock(void) in sched_clock() function 125 * sched_clock() will never observe mis-matched data even if called from 127 * steering sched_clock() to one or the other using a sequence counter. 128 * In order to preserve the data cache profile of sched_clock() as much 150 * Atomically update the sched_clock() epoch. 189 /* Cannot register a sched_clock wit in sched_clock_register() [all...] |
/linux/arch/mips/cavium-octeon/ |
H A D | csrc-octeon.c | 114 unsigned long long notrace sched_clock(void) in sched_clock() function
|
/linux/kernel/sched/ |
H A D | psi.c | 208 group->avg_last_update = sched_clock(); in group_init() 590 now = sched_clock(); in psi_avgs_work() 661 now = sched_clock(); in psi_rtpoll_work() 1261 now = sched_clock(); in psi_show() 1351 window_reset(&t->win, sched_clock(), in psi_trigger_create()
|
H A D | cputime.c | 650 clock = sched_clock(); in vtime_delta() 799 vtime->starttime = sched_clock(); in vtime_task_switch_generic() 812 vtime->starttime = sched_clock(); in vtime_init_idle()
|
/linux/arch/microblaze/kernel/ |
H A D | timer.c | 16 #include <linux/sched_clock.h> 212 timecounter_init(&xilinx_tc, &xilinx_cc, sched_clock()); in init_xilinx_timecounter()
|
/linux/sound/soc/mediatek/common/ |
H A D | mtk-btcvsd.c | 649 t1 = sched_clock(); in wait_for_bt_irq() 660 t2 = sched_clock(); in wait_for_bt_irq() 764 bt->rx->time_stamp = sched_clock(); in mtk_btcvsd_snd_read() 788 bt->tx->time_stamp = sched_clock(); in mtk_btcvsd_snd_write()
|
/linux/fs/bcachefs/ |
H A D | six.c | 364 end_time = sched_clock() + 10 * NSEC_PER_USEC; in six_optimistic_spin() 377 if (!(++loop & 0xf) && (time_after64(sched_clock(), end_time))) { in six_optimistic_spin()
|
H A D | tests.c | 787 j->start = sched_clock(); in btree_perf_test_thread() 799 j->finish = sched_clock(); in btree_perf_test_thread()
|
/linux/drivers/perf/ |
H A D | arm_pmu.c | 444 start_clock = sched_clock(); in armpmu_dispatch_irq() 446 finish_clock = sched_clock(); in armpmu_dispatch_irq()
|
/linux/kernel/rcu/ |
H A D | refscale.c | 542 x += sched_clock(); in ref_sched_clock_section() 554 x += sched_clock(); in ref_sched_clock_delay_section()
|
/linux/drivers/acpi/apei/ |
H A D | ghes.c | 1008 now = sched_clock(); in ghes_estatus_cached() 1043 cache->time_in = sched_clock(); in ghes_estatus_cache_alloc() 1073 now = sched_clock(); in ghes_estatus_cache_add()
|
/linux/arch/arm/probes/kprobes/ |
H A D | test-core.c | 508 t0 = sched_clock(); in benchmark() 511 t = sched_clock() - t0; in benchmark()
|
/linux/arch/sparc/kernel/ |
H A D | time_64.c | 882 unsigned long long sched_clock(void) in sched_clock() function
|
/linux/arch/um/kernel/ |
H A D | time.c | 54 notrace unsigned long long sched_clock(void) in sched_clock() function 458 /* clock tick; decrease extra jiffies by keeping sched_clock constant */ in time_travel_periodic_timer() 613 /* clock tick; decrease extra jiffies by keeping sched_clock constant */ in time_travel_oneshot_timer()
|
/linux/arch/um/ |
H A D | Kconfig | 237 for the same userspace thread before the sched_clock is advanced by a
|