Home
last modified time | relevance | path

Searched refs:WARN_ON_ONCE (Results 1 – 25 of 1929) sorted by relevance

12345678910>>...78

/linux/io_uring/
H A Drefs.h17 WARN_ON_ONCE(!(req->flags & REQ_F_REFCOUNT)); in req_ref_inc_not_zero()
23 WARN_ON_ONCE(!(data_race(req->flags) & REQ_F_REFCOUNT)); in req_ref_put_and_test_atomic()
24 WARN_ON_ONCE(req_ref_zero_or_close_to_overflow(req)); in req_ref_put_and_test_atomic()
33 WARN_ON_ONCE(req_ref_zero_or_close_to_overflow(req)); in req_ref_put_and_test()
39 WARN_ON_ONCE(!(req->flags & REQ_F_REFCOUNT)); in req_ref_get()
40 WARN_ON_ONCE(req_ref_zero_or_close_to_overflow(req)); in req_ref_get()
46 WARN_ON_ONCE(!(req->flags & REQ_F_REFCOUNT)); in req_ref_put()
47 WARN_ON_ONCE(req_ref_zero_or_close_to_overflow(req)); in req_ref_put()
/linux/kernel/
H A Dcontext_tracking.c71 WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !rcu_is_watching_curr_cpu()); in ct_kernel_exit_state()
92 WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !(seq & CT_RCU_WATCHING)); in ct_kernel_enter_state()
107 WARN_ON_ONCE(ct_nmi_nesting() != CT_NESTING_IRQ_NONIDLE); in ct_kernel_exit()
109 WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && in ct_kernel_exit()
120 WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !user && !is_idle_task(current)); in ct_kernel_exit()
147 WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !raw_irqs_disabled()); in ct_kernel_enter()
149 WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && oldval < 0); in ct_kernel_enter()
165 WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !user && !is_idle_task(current)); in ct_kernel_enter()
167 WARN_ON_ONCE(ct_nmi_nesting()); in ct_kernel_enter()
193 WARN_ON_ONCE(ct_nmi_nesting() <= 0); in ct_nmi_exit()
[all …]
H A Dkthread.c104 if (WARN_ON_ONCE(to_kthread(p))) in set_kthread_struct()
134 WARN_ON_ONCE(kthread->blkcg_css); in free_kthread_struct()
355 if (WARN_ON_ONCE(kthread_is_per_cpu(current))) in kthread_affine_node()
359 WARN_ON_ONCE(1); in kthread_affine_node()
364 WARN_ON_ONCE(!list_empty(&kthread->affinity_node)); in kthread_affine_node()
589 WARN_ON_ONCE(kthread->started); in kthread_bind_mask()
605 WARN_ON_ONCE(kthread->started); in kthread_bind()
642 WARN_ON_ONCE(!(k->flags & PF_NO_SETAFFINITY)); in kthread_set_per_cpu()
710 if (WARN_ON_ONCE(test_bit(KTHREAD_SHOULD_PARK, &kthread->flags))) in kthread_park()
725 WARN_ON_ONCE(!wait_task_inactive(k, TASK_PARKED)); in kthread_park()
[all …]
H A Djump_label.c180 if (WARN_ON_ONCE(!static_key_fast_inc_not_disabled(key))) in static_key_slow_inc_cpuslocked()
203 WARN_ON_ONCE(atomic_read(&key->enabled) != 1); in static_key_enable_cpuslocked()
234 WARN_ON_ONCE(atomic_read(&key->enabled) != 0); in static_key_disable_cpuslocked()
276 WARN_ON_ONCE(v < 0); in static_key_dec_not_one()
282 if (WARN_ON_ONCE(v == 0)) in static_key_dec_not_one()
306 if (WARN_ON_ONCE(val == -1)) in __static_key_slow_dec_cpuslocked()
311 if (WARN_ON_ONCE(val == 0)) in __static_key_slow_dec_cpuslocked()
411 WARN_ON_ONCE(key->type & JUMP_TYPE_LINKED); in static_key_entries()
449 WARN_ON_ONCE((unsigned long)entries & JUMP_TYPE_MASK); in static_key_set_entries()
588 if (WARN_ON_ONCE(!static_key_initialized)) in jump_label_init_ro()
[all …]
/linux/security/landlock/
H A Daudit.c57 WARN_ON_ONCE(access_bit != -1); in get_blocker()
61 WARN_ON_ONCE(access_bit != -1); in get_blocker()
65 if (WARN_ON_ONCE(access_bit >= ARRAY_SIZE(fs_access_strings))) in get_blocker()
70 if (WARN_ON_ONCE(access_bit >= ARRAY_SIZE(net_access_strings))) in get_blocker()
75 WARN_ON_ONCE(access_bit != -1); in get_blocker()
79 WARN_ON_ONCE(access_bit != -1); in get_blocker()
83 WARN_ON_ONCE(1); in get_blocker()
118 WARN_ON_ONCE(hierarchy->id == 0); in log_domain()
142 if (WARN_ON_ONCE(layer >= domain->num_layers)) in get_hierarchy()
146 if (WARN_ON_ONCE(!hierarchy->parent)) in get_hierarchy()
[all …]
H A Druleset.c103 WARN_ON_ONCE(1); in is_object_pointer()
119 if (WARN_ON_ONCE(num_layers >= LANDLOCK_MAX_NUM_LAYERS)) in create_rule()
132 WARN_ON_ONCE(!id.key.object); in create_rule()
160 WARN_ON_ONCE(1); in get_root()
217 if (WARN_ON_ONCE(!layers)) in insert_rule()
220 if (is_object_pointer(id.type) && WARN_ON_ONCE(!id.key.object)) in insert_rule()
242 if (WARN_ON_ONCE(num_layers != 1)) in insert_rule()
251 if (WARN_ON_ONCE(this->num_layers != 1)) in insert_rule()
253 if (WARN_ON_ONCE(this->layers[0].level != 0)) in insert_rule()
259 if (WARN_ON_ONCE(this->layers[0].level == 0)) in insert_rule()
[all …]
H A Ddomain.c62 if (WARN_ON_ONCE(IS_ERR(exe))) in get_current_exe()
67 if (WARN_ON_ONCE(size <= 0)) in get_current_exe()
142 WARN_ON_ONCE(all_existing_optional_access != in get_layer_deny_mask()
145 if (WARN_ON_ONCE(layer >= LANDLOCK_MAX_NUM_LAYERS)) in get_layer_deny_mask()
150 if (WARN_ON_ONCE(access_weight < 1)) in get_layer_deny_mask()
192 WARN_ON_ONCE(!access_mask_subset(optional_access, in landlock_get_deny_masks()
195 if (WARN_ON_ONCE(!masks)) in landlock_get_deny_masks()
198 if (WARN_ON_ONCE(!access_opt)) in landlock_get_deny_masks()
/linux/drivers/gpu/drm/vc4/
H A Dvc4_perfmon.c28 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4)) in vc4_perfmon_get()
42 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4)) in vc4_perfmon_put()
54 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4)) in vc4_perfmon_start()
57 if (WARN_ON_ONCE(!perfmon || vc4->active_perfmon)) in vc4_perfmon_start()
74 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4)) in vc4_perfmon_stop()
77 if (WARN_ON_ONCE(!vc4->active_perfmon || in vc4_perfmon_stop()
95 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4)) in vc4_perfmon_find()
110 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4)) in vc4_perfmon_open_file()
136 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4)) in vc4_perfmon_close_file()
156 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4)) in vc4_perfmon_create_ioctl()
[all …]
/linux/kernel/rcu/
H A Dsync.c62 WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_IDLE); in rcu_sync_func()
63 WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_PASSED); in rcu_sync_func()
113 WARN_ON_ONCE(rsp->gp_count); in rcu_sync_enter()
154 WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_IDLE); in rcu_sync_exit()
157 WARN_ON_ONCE(rsp->gp_count == 0); in rcu_sync_exit()
177 WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_PASSED); in rcu_sync_dtor()
180 WARN_ON_ONCE(rsp->gp_count); in rcu_sync_dtor()
188 WARN_ON_ONCE(rsp->gp_state != GP_IDLE); in rcu_sync_dtor()
H A Dtree_plugin.h172 WARN_ON_ONCE(rdp->mynode != rnp); in rcu_preempt_ctxt_queue()
173 WARN_ON_ONCE(!rcu_is_leaf_node(rnp)); in rcu_preempt_ctxt_queue()
175 WARN_ON_ONCE(rnp->qsmaskinitnext & ~rnp->qsmaskinit & rnp->qsmask & in rcu_preempt_ctxt_queue()
244 WARN_ON_ONCE(1); in rcu_preempt_ctxt_queue()
256 WARN_ON_ONCE(rnp->completedqs == rnp->gp_seq); in rcu_preempt_ctxt_queue()
260 WARN_ON_ONCE(!(blkd_state & RCU_GP_BLKD) != in rcu_preempt_ctxt_queue()
262 WARN_ON_ONCE(!(blkd_state & RCU_EXP_BLKD) != in rcu_preempt_ctxt_queue()
277 WARN_ON_ONCE(rdp->cpu_no_qs.b.exp); in rcu_preempt_ctxt_queue()
347 WARN_ON_ONCE(!rcu_rdp_cpu_online(rdp)); in rcu_note_context_switch()
348 WARN_ON_ONCE(!list_empty(&t->rcu_node_entry)); in rcu_note_context_switch()
[all …]
H A Dtree_nocb.h102 WARN_ON_ONCE(smp_processor_id() != rdp->cpu); in rcu_nocb_bypass_lock()
320 WARN_ON_ONCE(!rcu_rdp_is_offloaded(rdp)); in rcu_nocb_do_flush_bypass()
378 WARN_ON_ONCE(!rcu_nocb_do_flush_bypass(rdp, NULL, j, false)); in rcu_nocb_try_flush_bypass()
421 WARN_ON_ONCE(rcu_cblist_n_cbs(&rdp->nocb_bypass)); in rcu_nocb_try_bypass()
452 WARN_ON_ONCE(!rcu_nocb_flush_bypass(rdp, NULL, j, false)); in rcu_nocb_try_bypass()
453 WARN_ON_ONCE(rcu_cblist_n_cbs(&rdp->nocb_bypass)); in rcu_nocb_try_bypass()
470 WARN_ON_ONCE(rcu_cblist_n_cbs(&rdp->nocb_bypass)); in rcu_nocb_try_bypass()
650 WARN_ON_ONCE(my_rdp->nocb_gp_rdp != my_rdp); in nocb_gp_wait()
719 WARN_ON_ONCE(wasempty && in nocb_gp_wait()
873 WARN_ON_ONCE(rcu_segcblist_n_cbs(&rdp->cblist)); in nocb_cb_wait()
[all …]
/linux/fs/crypto/
H A Dcrypto.c54 if (WARN_ON_ONCE(!fscrypt_bounce_page_pool)) { in fscrypt_alloc_bounce_page()
97 WARN_ON_ONCE(index > U32_MAX); in fscrypt_generate_iv()
98 WARN_ON_ONCE(ci->ci_inode->i_ino > U32_MAX); in fscrypt_generate_iv()
101 WARN_ON_ONCE(index > U32_MAX); in fscrypt_generate_iv()
121 if (WARN_ON_ONCE(len <= 0)) in fscrypt_crypt_data_unit()
123 if (WARN_ON_ONCE(len % FSCRYPT_CONTENTS_ALIGNMENT != 0)) in fscrypt_crypt_data_unit()
186 if (WARN_ON_ONCE(!folio_test_locked(folio))) in fscrypt_encrypt_pagecache_blocks()
189 if (WARN_ON_ONCE(len <= 0 || !IS_ALIGNED(len | offs, du_size))) in fscrypt_encrypt_pagecache_blocks()
233 if (WARN_ON_ONCE(inode->i_sb->s_cop->supports_subblock_data_units)) in fscrypt_encrypt_block_inplace()
267 if (WARN_ON_ONCE(!folio_test_locked(folio))) in fscrypt_decrypt_pagecache_blocks()
[all …]
/linux/arch/powerpc/kernel/
H A Dirq_64.c107 WARN_ON_ONCE(mfmsr() & MSR_EE); in __replay_soft_interrupts()
217 WARN_ON_ONCE(in_nmi()); in arch_local_irq_restore()
218 WARN_ON_ONCE(in_hardirq()); in arch_local_irq_restore()
219 WARN_ON_ONCE(local_paca->irq_happened & PACA_IRQ_REPLAYING); in arch_local_irq_restore()
247 WARN_ON_ONCE(!(mfmsr() & MSR_EE)); in arch_local_irq_restore()
260 WARN_ON_ONCE(!irq_happened); in arch_local_irq_restore()
264 WARN_ON_ONCE(mfmsr() & MSR_EE); in arch_local_irq_restore()
283 WARN_ON_ONCE(!(irq_happened & PACA_IRQ_HARD_DIS)); in arch_local_irq_restore()
290 if (WARN_ON_ONCE(mfmsr() & MSR_EE)) in arch_local_irq_restore()
/linux/fs/notify/fanotify/
H A Dfanotify.h110 WARN_ON_ONCE(info->dir_fh_totlen < FANOTIFY_FH_HDR_LEN)) in fanotify_info_dir_fh_len()
126 WARN_ON_ONCE(info->dir2_fh_totlen < FANOTIFY_FH_HDR_LEN)) in fanotify_info_dir2_fh_len()
140 WARN_ON_ONCE(info->file_fh_totlen < FANOTIFY_FH_HDR_LEN)) in fanotify_info_file_fh_len()
183 if (WARN_ON_ONCE(info->dir2_fh_totlen > 0) || in fanotify_info_set_dir_fh()
184 WARN_ON_ONCE(info->file_fh_totlen > 0) || in fanotify_info_set_dir_fh()
185 WARN_ON_ONCE(info->name_len > 0) || in fanotify_info_set_dir_fh()
186 WARN_ON_ONCE(info->name2_len > 0)) in fanotify_info_set_dir_fh()
195 if (WARN_ON_ONCE(info->file_fh_totlen > 0) || in fanotify_info_set_dir2_fh()
196 WARN_ON_ONCE(info->name_len > 0) || in fanotify_info_set_dir2_fh()
197 WARN_ON_ONCE(info->name2_len > 0)) in fanotify_info_set_dir2_fh()
[all …]
/linux/drivers/irqchip/
H A Dirq-msi-lib.c35 if (WARN_ON_ONCE(!pops)) in msi_lib_init_dev_msi_info()
44 if (WARN_ON_ONCE(domain != real_parent)) in msi_lib_init_dev_msi_info()
47 WARN_ON_ONCE(1); in msi_lib_init_dev_msi_info()
57 if (WARN_ON_ONCE(!IS_ENABLED(CONFIG_PCI_MSI))) in msi_lib_init_dev_msi_info()
68 if (WARN_ON_ONCE(info->flags)) in msi_lib_init_dev_msi_info()
83 WARN_ON_ONCE(1); in msi_lib_init_dev_msi_info()
/linux/fs/iomap/
H A Dbuffered-io.c261 WARN_ON_ONCE(ifs->read_bytes_pending != 0); in ifs_free()
262 WARN_ON_ONCE(atomic_read(&ifs->write_bytes_pending)); in ifs_free()
263 WARN_ON_ONCE(ifs_is_fully_uptodate(folio, ifs) != in ifs_free()
377 if (WARN_ON_ONCE(!iomap->inline_data)) in iomap_read_inline_data()
383 if (WARN_ON_ONCE(size > iomap->length)) { in iomap_read_inline_data()
447 WARN_ON_ONCE(ifs->read_bytes_pending != 0); in iomap_read_init()
471 WARN_ON_ONCE(bytes_submitted); in iomap_read_end()
504 WARN_ON_ONCE(bytes_submitted); in iomap_read_end()
537 if (WARN_ON_ONCE(pos_diff + plen > length)) in iomap_read_folio_iter()
625 if (WARN_ON_ONCE(!ctx->cur_folio)) in iomap_readahead_iter()
[all …]
H A Diter.c25 if (WARN_ON_ONCE(count > iomap_length(iter))) in iomap_iter_advance()
34 WARN_ON_ONCE(iter->iomap.offset > iter->pos); in iomap_iter_done()
35 WARN_ON_ONCE(iter->iomap.length == 0); in iomap_iter_done()
36 WARN_ON_ONCE(iter->iomap.offset + iter->iomap.length <= iter->pos); in iomap_iter_done()
37 WARN_ON_ONCE(iter->iomap.flags & IOMAP_F_STALE); in iomap_iter_done()
90 if (WARN_ON_ONCE(iter->status > 0)) in iomap_iter()
/linux/lib/
H A Dinterval_tree_test.c142 WARN_ON_ONCE("Failed to allocate intxn1\n"); in intersection_range_check()
148 WARN_ON_ONCE("Failed to allocate intxn2\n"); in intersection_range_check()
185 WARN_ON_ONCE(!bitmap_equal(intxn1, intxn2, nnodes)); in intersection_range_check()
263 WARN_ON_ONCE(mtree_store_range(&tree, nodes[j].start, in span_iteration_check()
285 WARN_ON_ONCE(span.is_hole != mas_span.is_hole); in span_iteration_check()
288 WARN_ON_ONCE(span.start_hole != mas_span.start_hole); in span_iteration_check()
289 WARN_ON_ONCE(span.last_hole != mas_span.last_hole); in span_iteration_check()
291 WARN_ON_ONCE(span.start_used != mas_span.start_used); in span_iteration_check()
292 WARN_ON_ONCE(span.last_used != mas_span.last_used); in span_iteration_check()
298 WARN_ON_ONCE(mas.status != ma_overflow); in span_iteration_check()
/linux/kernel/sched/
H A Didle.c35 WARN_ON_ONCE(cpu_idle_force_poll < 0); in cpu_idle_poll_ctrl()
269 if (WARN_ON_ONCE(irqs_disabled())) in cpuidle_idle_call()
416 WARN_ON_ONCE(current->policy != SCHED_FIFO); in play_idle_precise()
417 WARN_ON_ONCE(current->nr_cpus_allowed != 1); in play_idle_precise()
418 WARN_ON_ONCE(!(current->flags & PF_KTHREAD)); in play_idle_precise()
419 WARN_ON_ONCE(!(current->flags & PF_NO_SETAFFINITY)); in play_idle_precise()
420 WARN_ON_ONCE(!duration_ns); in play_idle_precise()
421 WARN_ON_ONCE(current->mm); in play_idle_precise()
467 return WARN_ON_ONCE(1); in balance_idle()
/linux/include/linux/
H A Dhung_task.h45 WARN_ON_ONCE(!lock_ptr); in hung_task_set_blocker()
46 WARN_ON_ONCE(READ_ONCE(current->blocker)); in hung_task_set_blocker()
73 WARN_ON_ONCE(!blocker); in hung_task_get_blocker_type()
80 WARN_ON_ONCE(!blocker); in hung_task_blocker_to_lock()
/linux/kernel/time/
H A Dclocksource-wdtest.c119 WARN_ON_ONCE(clocksource_wdtest_jiffies.uncertainty_margin != TICK_NSEC); in wdtest_func()
124 WARN_ON_ONCE(j1 == j2); in wdtest_func()
134 WARN_ON_ONCE(clocksource_wdtest_ktime.uncertainty_margin < NSEC_PER_USEC); in wdtest_func()
155 WARN_ON_ONCE(READ_ONCE(wdtest_ktime_read_ndelays)); in wdtest_func()
156 WARN_ON_ONCE((i <= max_retries) != in wdtest_func()
165 WARN_ON_ONCE(!(clocksource_wdtest_ktime.flags & CLOCK_SOURCE_UNSTABLE)); in wdtest_func()
/linux/virt/lib/
H A Dirqbypass.c100 if (WARN_ON_ONCE(producer->eventfd)) in irq_bypass_register_producer()
115 WARN_ON_ONCE(xa_erase(&producers, index) != producer); in irq_bypass_register_producer()
145 WARN_ON_ONCE(xa_erase(&producers, index) != producer); in irq_bypass_unregister_producer()
165 if (WARN_ON_ONCE(consumer->eventfd)) in irq_bypass_register_consumer()
181 WARN_ON_ONCE(xa_erase(&consumers, index) != consumer); in irq_bypass_register_consumer()
211 WARN_ON_ONCE(xa_erase(&consumers, index) != consumer); in irq_bypass_unregister_consumer()
/linux/arch/x86/include/asm/
H A Dentry-common.h32 WARN_ON_ONCE(flags & mask); in arch_enter_from_user_mode()
35 WARN_ON_ONCE(!user_mode(regs)); in arch_enter_from_user_mode()
42 WARN_ON_ONCE(!on_thread_stack()); in arch_enter_from_user_mode()
43 WARN_ON_ONCE(regs != task_pt_regs(current)); in arch_enter_from_user_mode()
/linux/block/
H A Dblk-settings.c91 if (WARN_ON_ONCE(lim->max_open_zones) || in blk_validate_zoned_limits()
92 WARN_ON_ONCE(lim->max_active_zones) || in blk_validate_zoned_limits()
93 WARN_ON_ONCE(lim->zone_write_granularity) || in blk_validate_zoned_limits()
94 WARN_ON_ONCE(lim->max_zone_append_sectors)) in blk_validate_zoned_limits()
99 if (WARN_ON_ONCE(!IS_ENABLED(CONFIG_BLK_DEV_ZONED))) in blk_validate_zoned_limits()
299 if (WARN_ON_ONCE(!is_power_of_2(lim->atomic_write_hw_unit_min))) in blk_validate_atomic_write_limits()
302 if (WARN_ON_ONCE(!is_power_of_2(lim->atomic_write_hw_unit_max))) in blk_validate_atomic_write_limits()
305 if (WARN_ON_ONCE(lim->atomic_write_hw_unit_min > in blk_validate_atomic_write_limits()
309 if (WARN_ON_ONCE(lim->atomic_write_hw_unit_max > in blk_validate_atomic_write_limits()
313 if (WARN_ON_ONCE(lim->chunk_sectors && in blk_validate_atomic_write_limits()
[all …]
/linux/scripts/coccinelle/misc/
H A Dwarn.cocci71 WARN_ON_ONCE(1);
78 *WARN_ON_ONCE(1);
84 cocci.print_main("printk + WARN_ON_ONCE can be just WARN_ONCE",p)
90 msg = "SUGGESTION: printk + WARN_ON_ONCE can be just WARN_ONCE"
101 -WARN_ON_ONCE(1);

12345678910>>...78