Searched refs:vcpu_is_preempted (Results 1 – 15 of 15) sorted by relevance
20 struct paravirt_callee_save vcpu_is_preempted; member47 return PVOP_ALT_CALLEE1(bool, pv_ops_lock, vcpu_is_preempted, cpu, in pv_vcpu_is_preempted()75 #define vcpu_is_preempted vcpu_is_preempted macro76 static inline bool vcpu_is_preempted(long cpu) in vcpu_is_preempted() function
21 #define vcpu_is_preempted vcpu_is_preempted macro22 static inline bool vcpu_is_preempted(int cpu) in vcpu_is_preempted() function
37 #define vcpu_is_preempted vcpu_is_preempted macro39 bool vcpu_is_preempted(int cpu);
126 #define vcpu_is_preempted vcpu_is_preempted macro127 static inline bool vcpu_is_preempted(int cpu) in vcpu_is_preempted() function
42 return pv_ops_lock.vcpu_is_preempted.func == in pv_is_native_vcpu_is_preempted()60 .vcpu_is_preempted = PV_CALLEE_SAVE(__native_vcpu_is_preempted),
658 if (!idle_cpu(cpu) && vcpu_is_preempted(cpu)) { in kvm_smp_send_call_func_ipi()846 pv_ops_lock.vcpu_is_preempted = in kvm_guest_init()
372 if (vcpu_is_preempted(owner)) in propagate_sleepy()392 if (node->sleepy || vcpu_is_preempted(prev_cpu)) { in yield_to_prev()404 if (vcpu_is_preempted(get_owner_cpu(val))) in yield_to_prev()692 if (vcpu_is_preempted(next_cpu)) in queued_spin_lock_mcs_queue()
264 bool vcpu_is_preempted(int cpu) in vcpu_is_preempted() function274 EXPORT_SYMBOL(vcpu_is_preempted);
85 pv_ops_lock.vcpu_is_preempted = PV_CALLEE_SAVE(hv_vcpu_is_preempted); in hv_init_spinlocks()
142 pv_ops_lock.vcpu_is_preempted = PV_CALLEE_SAVE(xen_vcpu_stolen); in xen_init_spinlocks()
41 #define vcpu_is_preempted arch_vcpu_is_preempted macro
147 vcpu_is_preempted(node_cpu(node->prev)))) in osq_lock()
2294 #ifndef vcpu_is_preempted2295 static inline bool vcpu_is_preempted(int cpu) in vcpu_is_preempted() function2314 return READ_ONCE(owner->on_cpu) && !vcpu_is_preempted(task_cpu(owner)); in owner_on_cpu()
211 if (!vcpu_is_preempted(tcpu_cpu)) in __diag_time_slice_end_directed()
1412 if (vcpu_is_preempted(cpu)) in available_idle_cpu()