/linux/Documentation/i2c/ |
H A D | i2c-topology.rst | 40 mux-locked or parent-locked muxes. 43 Mux-locked muxes 46 Mux-locked muxes does not lock the entire parent adapter during the 48 adapter are locked. Mux-locked muxes are mostly interesting if the 50 their tasks. Since the parent adapter is not fully locked during the 55 Mux-locked Example 62 | root |--+--| locked | '--------' 87 Mux-locked caveat [all...] |
/linux/kernel/locking/ |
H A D | qspinlock_paravirt.h | 92 try_cmpxchg_acquire(&lock->locked, &old, _Q_LOCKED_VAL)) { in pv_hybrid_queued_unfair_trylock() 124 return !READ_ONCE(lock->locked) && in trylock_clear_pending() 142 * Try to clear pending bit & set locked bit in trylock_clear_pending() 286 * Wait for node->locked to become true, halt the vcpu after a short spin. 299 if (READ_ONCE(node->locked)) in pv_wait_node() 309 * Order pn->state vs pn->locked thusly: in pv_wait_node() 311 * [S] pn->state = VCPU_HALTED [S] next->locked = 1 in pv_wait_node() 313 * [L] pn->locked [RmW] pn->state = VCPU_HASHED in pv_wait_node() 319 if (!READ_ONCE(node->locked)) { in pv_wait_node() 333 * If the locked fla in pv_wait_node() 499 __pv_queued_spin_unlock_slowpath(struct qspinlock * lock,u8 locked) __pv_queued_spin_unlock_slowpath() argument 545 u8 locked = _Q_LOCKED_VAL; __pv_queued_spin_unlock() local [all...] |
H A D | osq_lock.c | 17 int locked; /* 1 if lock acquired */ member 100 node->locked = 0; in osq_lock() 146 if (smp_cond_load_relaxed(&node->locked, VAL || need_resched() || in osq_lock() 170 * in which case we should observe @node->locked becoming in osq_lock() 173 if (smp_load_acquire(&node->locked)) in osq_lock() 227 WRITE_ONCE(next->locked, 1); in osq_unlock() 233 WRITE_ONCE(next->locked, 1); in osq_unlock()
|
H A D | qspinlock.c | 48 * unlock the next pending (next->locked), we compress both these: {tail, 49 * next->locked} into a single u32 value. 145 * Wait for in-progress pending->locked hand-overs with a bounded in queued_spin_lock_slowpath() 191 * store-release that clears the locked bit and create lock in queued_spin_lock_slowpath() 197 smp_cond_load_acquire(&lock->locked, !VAL); in queued_spin_lock_slowpath() 251 node->locked = 0; in queued_spin_lock_slowpath() 291 arch_mcs_spin_lock_contended(&node->locked); in queued_spin_lock_slowpath() 311 * store-release that clears the locked bit and create lock in queued_spin_lock_slowpath() 318 * been designated yet, there is no way for the locked value to become in queued_spin_lock_slowpath() 326 goto locked; in queued_spin_lock_slowpath() [all...] |
H A D | mcs_spinlock.h | 53 * on this node->locked until the previous lock holder sets the node->locked 62 node->locked = 0; in mcs_spin_lock() 74 * Lock acquired, don't need to set node->locked to 1. Threads in mcs_spin_lock() 75 * only spin on its own node->locked value for lock acquisition. in mcs_spin_lock() 77 * and does not proceed to spin on its own node->locked, this in mcs_spin_lock() 79 * audit lock status, then set node->locked value here. in mcs_spin_lock() 86 arch_mcs_spin_lock_contended(&node->locked); in mcs_spin_lock() 110 arch_mcs_spin_unlock_contended(&next->locked); in mcs_spin_unlock()
|
/linux/tools/testing/selftests/net/forwarding/ |
H A D | bridge_locked_port.sh | 102 bridge link set dev $swp1 locked on 112 bridge link set dev $swp1 locked off 133 bridge link set dev $swp1 locked on 142 bridge link set dev $swp1 locked off 161 bridge link set dev $swp1 locked on 170 bridge link set dev $swp1 locked off 187 bridge link set dev $swp1 learning on locked on 190 check_fail $? "Ping worked on a locked port without an FDB entry" 195 bridge link set dev $swp1 learning on locked on mab on 200 bridge fdb get `mac_get $h1` br br0 vlan 1 | grep "dev $swp1" | grep -q "locked" [all...] |
/linux/scripts/coccinelle/locks/ |
H A D | double_lock.cocci | 17 @locked@ 42 position p1 != locked.p1; 43 position locked.p; 45 expression x <= locked.E1; 46 expression E,locked.E1; 65 expression x <= locked.E1; 66 expression locked.E1; 69 position locked.p,p1,p2;
|
H A D | call_kern.cocci | 39 @locked exists@ 74 @depends on locked && patch@ 81 @depends on locked && !patch@ 90 p1 << locked.p1; 91 p2 << locked.p2; 101 p1 << locked.p1; 102 p2 << locked.p2;
|
/linux/drivers/net/ethernet/intel/ice/ |
H A D | ice_dcb_lib.h | 18 int ice_dcb_sw_dflt_cfg(struct ice_pf *pf, bool ets_willing, bool locked); 24 ice_pf_dcb_cfg(struct ice_pf *pf, struct ice_dcbx_cfg *new_cfg, bool locked); 26 void ice_pf_dcb_recfg(struct ice_pf *pf, bool locked); 28 int ice_init_pf_dcb(struct ice_pf *pf, bool locked); 96 ice_init_pf_dcb(struct ice_pf *pf, bool __always_unused locked) in ice_init_pf_dcb() argument 105 bool __always_unused locked) in ice_pf_dcb_cfg() argument 134 static inline void ice_pf_dcb_recfg(struct ice_pf *pf, bool locked) { } in ice_pf_dcb_recfg() argument
|
/linux/mm/ |
H A D | gup.c | 1102 * FOLL_NOWAIT, the mmap_lock may be released. If it is, *@locked will be set 1107 int *locked) in faultin_page() argument 1152 *locked = 0; in faultin_page() 1174 *locked = 0; in faultin_page() 1327 * @locked: whether we're still with the mmap_lock held 1362 * be released. If this happens *@locked will be set to 0 on return. 1375 int *locked) in __get_user_pages() argument 1447 PTR_ERR(page) == -EMLINK, locked); in __get_user_pages() 1657 * Locking: (*locked == 1) means that the mmap_lock has already been acquired by 1659 * set (*locked 1673 __get_user_pages_locked(struct mm_struct * mm,unsigned long start,unsigned long nr_pages,struct page ** pages,int * locked,unsigned int flags) __get_user_pages_locked() argument 1834 populate_vma_page_range(struct vm_area_struct * vma,unsigned long start,unsigned long end,int * locked) populate_vma_page_range() argument 1908 faultin_page_range(struct mm_struct * mm,unsigned long start,unsigned long end,bool write,int * locked) faultin_page_range() argument 1950 int locked = 0; __mm_populate() local 2001 __get_user_pages_locked(struct mm_struct * mm,unsigned long start,unsigned long nr_pages,struct page ** pages,int * locked,unsigned int foll_flags) __get_user_pages_locked() argument 2207 get_dump_page(unsigned long addr,int * locked) get_dump_page() argument 2481 __gup_longterm_locked(struct mm_struct * mm,unsigned long start,unsigned long nr_pages,struct page ** pages,int * locked,unsigned int gup_flags) __gup_longterm_locked() argument 2512 is_valid_gup_args(struct page ** pages,int * locked,unsigned int * gup_flags_p,unsigned int to_set) is_valid_gup_args() argument 2618 get_user_pages_remote(struct mm_struct * mm,unsigned long start,unsigned long nr_pages,unsigned int gup_flags,struct page ** pages,int * locked) get_user_pages_remote() argument 2636 get_user_pages_remote(struct mm_struct * mm,unsigned long start,unsigned long nr_pages,unsigned int gup_flags,struct page ** pages,int * locked) get_user_pages_remote() argument 2659 int locked = 1; get_user_pages() local 2687 int locked = 0; get_user_pages_unlocked() local 3203 int locked = 0; gup_fast_fallback() local 3368 pin_user_pages_remote(struct mm_struct * mm,unsigned long start,unsigned long nr_pages,unsigned int gup_flags,struct page ** pages,int * locked) pin_user_pages_remote() argument 3402 int locked = 1; pin_user_pages() local 3422 int locked = 0; pin_user_pages_unlocked() local [all...] |
H A D | memcontrol-v1.h | 70 bool memcg1_oom_prepare(struct mem_cgroup *memcg, bool *locked); 71 void memcg1_oom_finish(struct mem_cgroup *memcg, bool locked); 106 static inline bool memcg1_oom_prepare(struct mem_cgroup *memcg, bool *locked) { return true; } in memcg1_oom_prepare() argument 107 static inline void memcg1_oom_finish(struct mem_cgroup *memcg, bool locked) {} in memcg1_oom_finish() argument
|
H A D | mlock.c | 488 * Keep track of amount of locked VM. in mlock_fixup() 614 unsigned long locked; in do_mlock() local 628 locked = len >> PAGE_SHIFT; in do_mlock() 633 locked += current->mm->locked_vm; in do_mlock() 634 if ((locked > lock_limit) && (!capable(CAP_IPC_LOCK))) { in do_mlock() 639 * and adjust locked count if necessary. in do_mlock() 641 locked -= count_mm_mlocked_page_nr(current->mm, in do_mlock() 646 if ((locked <= lock_limit) || capable(CAP_IPC_LOCK)) in do_mlock() 793 unsigned long lock_limit, locked; in user_shm_lock() local 797 locked in user_shm_lock() [all...] |
/linux/drivers/media/dvb-frontends/ |
H A D | lgs8gxx.c | 295 static int lgs8gxx_is_locked(struct lgs8gxx_state *priv, u8 *locked) in lgs8gxx_is_locked() argument 308 *locked = ((t & 0x80) == 0x80) ? 1 : 0; in lgs8gxx_is_locked() 310 *locked = ((t & 0xC0) == 0xC0) ? 1 : 0; in lgs8gxx_is_locked() 315 static int lgs8gxx_wait_ca_lock(struct lgs8gxx_state *priv, u8 *locked) in lgs8gxx_wait_ca_lock() argument 331 *locked = (ret == 0) ? 1 : 0; in lgs8gxx_wait_ca_lock() 359 u8 *locked) in lgs8gxx_autolock_gi() argument 384 err = lgs8gxx_wait_ca_lock(priv, locked); in lgs8gxx_autolock_gi() 385 if (err || !(*locked)) in lgs8gxx_autolock_gi() 393 *locked = 0; in lgs8gxx_autolock_gi() 403 u8 locked in lgs8gxx_auto_detect() local 716 u8 t, locked = 0; lgs8gxx_read_status() local [all...] |
H A D | atbm8830.c | 147 static int is_locked(struct atbm_state *priv, u8 *locked) in is_locked() argument 153 if (locked != NULL) in is_locked() 154 *locked = (status == 1); in is_locked() 261 u8 locked = 0; in atbm8830_set_fe() local 277 is_locked(priv, &locked); in atbm8830_set_fe() 278 if (locked != 0) { in atbm8830_set_fe() 279 dprintk("ATBM8830 locked!\n"); in atbm8830_set_fe() 329 u8 locked = 0; in atbm8830_read_status() local 335 is_locked(priv, &locked); in atbm8830_read_status() 336 if (locked) { in atbm8830_read_status() [all...] |
/linux/tools/testing/selftests/bpf/progs/ |
H A D | bpf_arena_spin_lock.h | 37 u8 locked; member 52 u8 locked; member 64 int locked; member 78 * 0- 7: locked byte 204 WRITE_ONCE(lock->locked, _Q_LOCKED_VAL); in set_locked() 252 * Wait for in-progress pending->locked hand-overs with a bounded in arena_spin_lock_slowpath() 299 * store-release that clears the locked bit and create lock in arena_spin_lock_slowpath() 305 smp_cond_load_acquire_label(&lock->locked, !VAL, release_err); in arena_spin_lock_slowpath() 345 node->locked = 0; in arena_spin_lock_slowpath() 383 arch_mcs_spin_lock_contended_label(&node->locked, release_node_er in arena_spin_lock_slowpath() [all...] |
/linux/drivers/mtd/spi-nor/ |
H A D | otp.c | 189 /* no need to write the register if region is already locked */ in spi_nor_otp_lock_sr2() 254 int ret, locked; in spi_nor_mtd_otp_info() local 267 locked = ops->is_locked(nor, i); in spi_nor_mtd_otp_info() 268 if (locked < 0) { in spi_nor_mtd_otp_info() 269 ret = locked; in spi_nor_mtd_otp_info() 273 buf->locked = !!locked; in spi_nor_mtd_otp_info() 290 int locked; in spi_nor_mtd_otp_range_is_locked() local 293 * If any of the affected OTP regions are locked the entire range is in spi_nor_mtd_otp_range_is_locked() 294 * considered locked in spi_nor_mtd_otp_range_is_locked() [all...] |
/linux/drivers/thermal/intel/ |
H A D | intel_quark_dts_thermal.c | 100 bool locked; member 129 if (!aux_entry->locked) { in soc_dts_enable() 136 pr_info("DTS is locked. Cannot enable DTS\n"); in soc_dts_enable() 157 if (!aux_entry->locked) { in soc_dts_disable() 165 pr_info("DTS is locked. Cannot disable DTS\n"); in soc_dts_disable() 206 if (aux_entry->locked) { in update_trip_temp() 316 if (!aux_entry->locked) { in free_soc_dts() 345 /* Check if DTS register is locked */ in alloc_soc_dts() 351 aux_entry->locked = !!(out & QRK_DTS_LOCK_BIT); in alloc_soc_dts() 353 /* Store DTS default state if DTS registers are not locked */ in alloc_soc_dts() [all...] |
/linux/rust/pin-init/examples/ |
H A D | mutex.rs | 71 locked: Cell<bool>, field 82 locked: Cell::new(false), in new() 94 if self.locked.get() { in lock() 97 while self.locked.get() { in lock() 107 self.locked.set(true); in lock() 135 self.mtx.locked.set(false); in drop()
|
/linux/sound/core/seq/ |
H A D | seq_queue.c | 19 * - The queue is locked when struct snd_seq_queue pointer is returned via 88 static struct snd_seq_queue *queue_new(int owner, int locked) in queue_new() argument 114 q->locked = locked; in queue_new() 160 struct snd_seq_queue *snd_seq_queue_alloc(int client, int locked, unsigned int info_flags) in snd_seq_queue_alloc() argument 164 q = queue_new(client, locked); in snd_seq_queue_alloc() 342 return (q->owner == client) || (!q->locked && !q->klocked); in check_access() 385 int snd_seq_queue_set_owner(int queueid, int client, int locked) in snd_seq_queue_set_owner() argument 398 q->locked = locked in snd_seq_queue_set_owner() 702 bool locked; snd_seq_info_queues_read() local [all...] |
H A D | seq_queue.h | 29 bool locked; /* timer is only accesibble by owner if set */ member 57 struct snd_seq_queue *snd_seq_queue_alloc(int client, int locked, unsigned int flags); 85 int snd_seq_queue_set_owner(int queueid, int client, int locked);
|
/linux/drivers/watchdog/ |
H A D | intel_oc_wdt.c | 44 bool locked; member 61 if (oc_wdt->locked) in intel_oc_wdt_start() 126 oc_wdt->locked = !!(val & INTEL_OC_WDT_CTL_LCK); in intel_oc_wdt_setup() 136 if (oc_wdt->locked) { in intel_oc_wdt_setup() 144 * If we are locked read the current timeout value in intel_oc_wdt_setup() 151 "Register access locked, heartbeat fixed at: %u s\n", in intel_oc_wdt_setup() 154 } else if (oc_wdt->locked) { in intel_oc_wdt_setup() 156 * In case the watchdog is disabled and locked there in intel_oc_wdt_setup()
|
/linux/Documentation/ABI/testing/ |
H A D | sysfs-platform-intel-pmc | 15 * bit 31 - global reset is locked 20 in case the register is not locked. 21 The "global reset bit" should be locked on a production
|
/linux/drivers/tty/ |
H A D | tty_ldsem.c | 234 int locked = 0; in down_write_failed() local 265 locked = writer_trylock(sem); in down_write_failed() 266 if (locked) in down_write_failed() 270 if (!locked) in down_write_failed() 280 if (!locked && list_empty(&sem->write_wait)) in down_write_failed() 288 if (!locked) in down_write_failed()
|
/linux/drivers/gpu/drm/xe/ |
H A D | xe_pcode.c | 136 u32 *status, bool atomic, int timeout_us, bool locked) in pcode_try_request() argument 143 if (locked) in pcode_try_request() 272 * @locked: true if lock held, false otherwise 282 int xe_pcode_ready(struct xe_device *xe, bool locked) in xe_pcode_ready() argument 295 if (locked) in xe_pcode_ready() 301 &status, false, timeout_us, locked); in xe_pcode_ready() 303 if (locked) in xe_pcode_ready()
|
/linux/drivers/staging/vme_user/ |
H A D | vme_bridge.h | 19 int locked; member 32 int locked; member 65 int locked; member 76 int locked; member
|