| /linux/io_uring/ |
| H A D | refs.h | 17 WARN_ON_ONCE(!(req->flags & REQ_F_REFCOUNT)); in req_ref_inc_not_zero() 23 WARN_ON_ONCE(!(data_race(req->flags) & REQ_F_REFCOUNT)); in req_ref_put_and_test_atomic() 24 WARN_ON_ONCE(req_ref_zero_or_close_to_overflow(req)); in req_ref_put_and_test_atomic() 33 WARN_ON_ONCE(req_ref_zero_or_close_to_overflow(req)); in req_ref_put_and_test() 39 WARN_ON_ONCE(!(req->flags & REQ_F_REFCOUNT)); in req_ref_get() 40 WARN_ON_ONCE(req_ref_zero_or_close_to_overflow(req)); in req_ref_get() 46 WARN_ON_ONCE(!(req->flags & REQ_F_REFCOUNT)); in req_ref_put() 47 WARN_ON_ONCE(req_ref_zero_or_close_to_overflow(req)); in req_ref_put()
|
| /linux/kernel/ |
| H A D | context_tracking.c | 71 WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !rcu_is_watching_curr_cpu()); in ct_kernel_exit_state() 92 WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !(seq & CT_RCU_WATCHING)); in ct_kernel_enter_state() 107 WARN_ON_ONCE(ct_nmi_nesting() != CT_NESTING_IRQ_NONIDLE); in ct_kernel_exit() 109 WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && in ct_kernel_exit() 120 WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !user && !is_idle_task(current)); in ct_kernel_exit() 147 WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !raw_irqs_disabled()); in ct_kernel_enter() 149 WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && oldval < 0); in ct_kernel_enter() 165 WARN_ON_ONCE(IS_ENABLED(CONFIG_RCU_EQS_DEBUG) && !user && !is_idle_task(current)); in ct_kernel_enter() 167 WARN_ON_ONCE(ct_nmi_nesting()); in ct_kernel_enter() 193 WARN_ON_ONCE(ct_nmi_nesting() <= 0); in ct_nmi_exit() [all …]
|
| H A D | kthread.c | 104 if (WARN_ON_ONCE(to_kthread(p))) in set_kthread_struct() 134 WARN_ON_ONCE(kthread->blkcg_css); in free_kthread_struct() 355 if (WARN_ON_ONCE(kthread_is_per_cpu(current))) in kthread_affine_node() 359 WARN_ON_ONCE(1); in kthread_affine_node() 364 WARN_ON_ONCE(!list_empty(&kthread->affinity_node)); in kthread_affine_node() 589 WARN_ON_ONCE(kthread->started); in kthread_bind_mask() 605 WARN_ON_ONCE(kthread->started); in kthread_bind() 642 WARN_ON_ONCE(!(k->flags & PF_NO_SETAFFINITY)); in kthread_set_per_cpu() 710 if (WARN_ON_ONCE(test_bit(KTHREAD_SHOULD_PARK, &kthread->flags))) in kthread_park() 725 WARN_ON_ONCE(!wait_task_inactive(k, TASK_PARKED)); in kthread_park() [all …]
|
| H A D | jump_label.c | 180 if (WARN_ON_ONCE(!static_key_fast_inc_not_disabled(key))) in static_key_slow_inc_cpuslocked() 203 WARN_ON_ONCE(atomic_read(&key->enabled) != 1); in static_key_enable_cpuslocked() 234 WARN_ON_ONCE(atomic_read(&key->enabled) != 0); in static_key_disable_cpuslocked() 276 WARN_ON_ONCE(v < 0); in static_key_dec_not_one() 282 if (WARN_ON_ONCE(v == 0)) in static_key_dec_not_one() 306 if (WARN_ON_ONCE(val == -1)) in __static_key_slow_dec_cpuslocked() 311 if (WARN_ON_ONCE(val == 0)) in __static_key_slow_dec_cpuslocked() 411 WARN_ON_ONCE(key->type & JUMP_TYPE_LINKED); in static_key_entries() 449 WARN_ON_ONCE((unsigned long)entries & JUMP_TYPE_MASK); in static_key_set_entries() 588 if (WARN_ON_ONCE(!static_key_initialized)) in jump_label_init_ro() [all …]
|
| /linux/security/landlock/ |
| H A D | audit.c | 57 WARN_ON_ONCE(access_bit != -1); in get_blocker() 61 WARN_ON_ONCE(access_bit != -1); in get_blocker() 65 if (WARN_ON_ONCE(access_bit >= ARRAY_SIZE(fs_access_strings))) in get_blocker() 70 if (WARN_ON_ONCE(access_bit >= ARRAY_SIZE(net_access_strings))) in get_blocker() 75 WARN_ON_ONCE(access_bit != -1); in get_blocker() 79 WARN_ON_ONCE(access_bit != -1); in get_blocker() 83 WARN_ON_ONCE(1); in get_blocker() 118 WARN_ON_ONCE(hierarchy->id == 0); in log_domain() 142 if (WARN_ON_ONCE(layer >= domain->num_layers)) in get_hierarchy() 146 if (WARN_ON_ONCE(!hierarchy->parent)) in get_hierarchy() [all …]
|
| H A D | ruleset.c | 103 WARN_ON_ONCE(1); in is_object_pointer() 119 if (WARN_ON_ONCE(num_layers >= LANDLOCK_MAX_NUM_LAYERS)) in create_rule() 132 WARN_ON_ONCE(!id.key.object); in create_rule() 160 WARN_ON_ONCE(1); in get_root() 217 if (WARN_ON_ONCE(!layers)) in insert_rule() 220 if (is_object_pointer(id.type) && WARN_ON_ONCE(!id.key.object)) in insert_rule() 242 if (WARN_ON_ONCE(num_layers != 1)) in insert_rule() 251 if (WARN_ON_ONCE(this->num_layers != 1)) in insert_rule() 253 if (WARN_ON_ONCE(this->layers[0].level != 0)) in insert_rule() 259 if (WARN_ON_ONCE(this->layers[0].level == 0)) in insert_rule() [all …]
|
| H A D | domain.c | 62 if (WARN_ON_ONCE(IS_ERR(exe))) in get_current_exe() 67 if (WARN_ON_ONCE(size <= 0)) in get_current_exe() 142 WARN_ON_ONCE(all_existing_optional_access != in get_layer_deny_mask() 145 if (WARN_ON_ONCE(layer >= LANDLOCK_MAX_NUM_LAYERS)) in get_layer_deny_mask() 150 if (WARN_ON_ONCE(access_weight < 1)) in get_layer_deny_mask() 192 WARN_ON_ONCE(!access_mask_subset(optional_access, in landlock_get_deny_masks() 195 if (WARN_ON_ONCE(!masks)) in landlock_get_deny_masks() 198 if (WARN_ON_ONCE(!access_opt)) in landlock_get_deny_masks()
|
| /linux/drivers/gpu/drm/vc4/ |
| H A D | vc4_perfmon.c | 28 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4)) in vc4_perfmon_get() 42 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4)) in vc4_perfmon_put() 54 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4)) in vc4_perfmon_start() 57 if (WARN_ON_ONCE(!perfmon || vc4->active_perfmon)) in vc4_perfmon_start() 74 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4)) in vc4_perfmon_stop() 77 if (WARN_ON_ONCE(!vc4->active_perfmon || in vc4_perfmon_stop() 95 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4)) in vc4_perfmon_find() 110 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4)) in vc4_perfmon_open_file() 136 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4)) in vc4_perfmon_close_file() 156 if (WARN_ON_ONCE(vc4->gen > VC4_GEN_4)) in vc4_perfmon_create_ioctl() [all …]
|
| /linux/kernel/rcu/ |
| H A D | sync.c | 62 WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_IDLE); in rcu_sync_func() 63 WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_PASSED); in rcu_sync_func() 113 WARN_ON_ONCE(rsp->gp_count); in rcu_sync_enter() 154 WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_IDLE); in rcu_sync_exit() 157 WARN_ON_ONCE(rsp->gp_count == 0); in rcu_sync_exit() 177 WARN_ON_ONCE(READ_ONCE(rsp->gp_state) == GP_PASSED); in rcu_sync_dtor() 180 WARN_ON_ONCE(rsp->gp_count); in rcu_sync_dtor() 188 WARN_ON_ONCE(rsp->gp_state != GP_IDLE); in rcu_sync_dtor()
|
| H A D | tree_plugin.h | 172 WARN_ON_ONCE(rdp->mynode != rnp); in rcu_preempt_ctxt_queue() 173 WARN_ON_ONCE(!rcu_is_leaf_node(rnp)); in rcu_preempt_ctxt_queue() 175 WARN_ON_ONCE(rnp->qsmaskinitnext & ~rnp->qsmaskinit & rnp->qsmask & in rcu_preempt_ctxt_queue() 244 WARN_ON_ONCE(1); in rcu_preempt_ctxt_queue() 256 WARN_ON_ONCE(rnp->completedqs == rnp->gp_seq); in rcu_preempt_ctxt_queue() 260 WARN_ON_ONCE(!(blkd_state & RCU_GP_BLKD) != in rcu_preempt_ctxt_queue() 262 WARN_ON_ONCE(!(blkd_state & RCU_EXP_BLKD) != in rcu_preempt_ctxt_queue() 277 WARN_ON_ONCE(rdp->cpu_no_qs.b.exp); in rcu_preempt_ctxt_queue() 347 WARN_ON_ONCE(!rcu_rdp_cpu_online(rdp)); in rcu_note_context_switch() 348 WARN_ON_ONCE(!list_empty(&t->rcu_node_entry)); in rcu_note_context_switch() [all …]
|
| H A D | tree_nocb.h | 102 WARN_ON_ONCE(smp_processor_id() != rdp->cpu); in rcu_nocb_bypass_lock() 320 WARN_ON_ONCE(!rcu_rdp_is_offloaded(rdp)); in rcu_nocb_do_flush_bypass() 378 WARN_ON_ONCE(!rcu_nocb_do_flush_bypass(rdp, NULL, j, false)); in rcu_nocb_try_flush_bypass() 421 WARN_ON_ONCE(rcu_cblist_n_cbs(&rdp->nocb_bypass)); in rcu_nocb_try_bypass() 452 WARN_ON_ONCE(!rcu_nocb_flush_bypass(rdp, NULL, j, false)); in rcu_nocb_try_bypass() 453 WARN_ON_ONCE(rcu_cblist_n_cbs(&rdp->nocb_bypass)); in rcu_nocb_try_bypass() 470 WARN_ON_ONCE(rcu_cblist_n_cbs(&rdp->nocb_bypass)); in rcu_nocb_try_bypass() 650 WARN_ON_ONCE(my_rdp->nocb_gp_rdp != my_rdp); in nocb_gp_wait() 719 WARN_ON_ONCE(wasempty && in nocb_gp_wait() 873 WARN_ON_ONCE(rcu_segcblist_n_cbs(&rdp->cblist)); in nocb_cb_wait() [all …]
|
| /linux/fs/crypto/ |
| H A D | crypto.c | 54 if (WARN_ON_ONCE(!fscrypt_bounce_page_pool)) { in fscrypt_alloc_bounce_page() 97 WARN_ON_ONCE(index > U32_MAX); in fscrypt_generate_iv() 98 WARN_ON_ONCE(ci->ci_inode->i_ino > U32_MAX); in fscrypt_generate_iv() 101 WARN_ON_ONCE(index > U32_MAX); in fscrypt_generate_iv() 121 if (WARN_ON_ONCE(len <= 0)) in fscrypt_crypt_data_unit() 123 if (WARN_ON_ONCE(len % FSCRYPT_CONTENTS_ALIGNMENT != 0)) in fscrypt_crypt_data_unit() 186 if (WARN_ON_ONCE(!folio_test_locked(folio))) in fscrypt_encrypt_pagecache_blocks() 189 if (WARN_ON_ONCE(len <= 0 || !IS_ALIGNED(len | offs, du_size))) in fscrypt_encrypt_pagecache_blocks() 233 if (WARN_ON_ONCE(inode->i_sb->s_cop->supports_subblock_data_units)) in fscrypt_encrypt_block_inplace() 267 if (WARN_ON_ONCE(!folio_test_locked(folio))) in fscrypt_decrypt_pagecache_blocks() [all …]
|
| /linux/arch/powerpc/kernel/ |
| H A D | irq_64.c | 107 WARN_ON_ONCE(mfmsr() & MSR_EE); in __replay_soft_interrupts() 217 WARN_ON_ONCE(in_nmi()); in arch_local_irq_restore() 218 WARN_ON_ONCE(in_hardirq()); in arch_local_irq_restore() 219 WARN_ON_ONCE(local_paca->irq_happened & PACA_IRQ_REPLAYING); in arch_local_irq_restore() 247 WARN_ON_ONCE(!(mfmsr() & MSR_EE)); in arch_local_irq_restore() 260 WARN_ON_ONCE(!irq_happened); in arch_local_irq_restore() 264 WARN_ON_ONCE(mfmsr() & MSR_EE); in arch_local_irq_restore() 283 WARN_ON_ONCE(!(irq_happened & PACA_IRQ_HARD_DIS)); in arch_local_irq_restore() 290 if (WARN_ON_ONCE(mfmsr() & MSR_EE)) in arch_local_irq_restore()
|
| /linux/fs/notify/fanotify/ |
| H A D | fanotify.h | 110 WARN_ON_ONCE(info->dir_fh_totlen < FANOTIFY_FH_HDR_LEN)) in fanotify_info_dir_fh_len() 126 WARN_ON_ONCE(info->dir2_fh_totlen < FANOTIFY_FH_HDR_LEN)) in fanotify_info_dir2_fh_len() 140 WARN_ON_ONCE(info->file_fh_totlen < FANOTIFY_FH_HDR_LEN)) in fanotify_info_file_fh_len() 183 if (WARN_ON_ONCE(info->dir2_fh_totlen > 0) || in fanotify_info_set_dir_fh() 184 WARN_ON_ONCE(info->file_fh_totlen > 0) || in fanotify_info_set_dir_fh() 185 WARN_ON_ONCE(info->name_len > 0) || in fanotify_info_set_dir_fh() 186 WARN_ON_ONCE(info->name2_len > 0)) in fanotify_info_set_dir_fh() 195 if (WARN_ON_ONCE(info->file_fh_totlen > 0) || in fanotify_info_set_dir2_fh() 196 WARN_ON_ONCE(info->name_len > 0) || in fanotify_info_set_dir2_fh() 197 WARN_ON_ONCE(info->name2_len > 0)) in fanotify_info_set_dir2_fh() [all …]
|
| /linux/drivers/irqchip/ |
| H A D | irq-msi-lib.c | 35 if (WARN_ON_ONCE(!pops)) in msi_lib_init_dev_msi_info() 44 if (WARN_ON_ONCE(domain != real_parent)) in msi_lib_init_dev_msi_info() 47 WARN_ON_ONCE(1); in msi_lib_init_dev_msi_info() 57 if (WARN_ON_ONCE(!IS_ENABLED(CONFIG_PCI_MSI))) in msi_lib_init_dev_msi_info() 68 if (WARN_ON_ONCE(info->flags)) in msi_lib_init_dev_msi_info() 83 WARN_ON_ONCE(1); in msi_lib_init_dev_msi_info()
|
| /linux/fs/iomap/ |
| H A D | buffered-io.c | 261 WARN_ON_ONCE(ifs->read_bytes_pending != 0); in ifs_free() 262 WARN_ON_ONCE(atomic_read(&ifs->write_bytes_pending)); in ifs_free() 263 WARN_ON_ONCE(ifs_is_fully_uptodate(folio, ifs) != in ifs_free() 377 if (WARN_ON_ONCE(!iomap->inline_data)) in iomap_read_inline_data() 383 if (WARN_ON_ONCE(size > iomap->length)) { in iomap_read_inline_data() 447 WARN_ON_ONCE(ifs->read_bytes_pending != 0); in iomap_read_init() 471 WARN_ON_ONCE(bytes_submitted); in iomap_read_end() 504 WARN_ON_ONCE(bytes_submitted); in iomap_read_end() 537 if (WARN_ON_ONCE(pos_diff + plen > length)) in iomap_read_folio_iter() 625 if (WARN_ON_ONCE(!ctx->cur_folio)) in iomap_readahead_iter() [all …]
|
| H A D | iter.c | 25 if (WARN_ON_ONCE(count > iomap_length(iter))) in iomap_iter_advance() 34 WARN_ON_ONCE(iter->iomap.offset > iter->pos); in iomap_iter_done() 35 WARN_ON_ONCE(iter->iomap.length == 0); in iomap_iter_done() 36 WARN_ON_ONCE(iter->iomap.offset + iter->iomap.length <= iter->pos); in iomap_iter_done() 37 WARN_ON_ONCE(iter->iomap.flags & IOMAP_F_STALE); in iomap_iter_done() 90 if (WARN_ON_ONCE(iter->status > 0)) in iomap_iter()
|
| /linux/lib/ |
| H A D | interval_tree_test.c | 142 WARN_ON_ONCE("Failed to allocate intxn1\n"); in intersection_range_check() 148 WARN_ON_ONCE("Failed to allocate intxn2\n"); in intersection_range_check() 185 WARN_ON_ONCE(!bitmap_equal(intxn1, intxn2, nnodes)); in intersection_range_check() 263 WARN_ON_ONCE(mtree_store_range(&tree, nodes[j].start, in span_iteration_check() 285 WARN_ON_ONCE(span.is_hole != mas_span.is_hole); in span_iteration_check() 288 WARN_ON_ONCE(span.start_hole != mas_span.start_hole); in span_iteration_check() 289 WARN_ON_ONCE(span.last_hole != mas_span.last_hole); in span_iteration_check() 291 WARN_ON_ONCE(span.start_used != mas_span.start_used); in span_iteration_check() 292 WARN_ON_ONCE(span.last_used != mas_span.last_used); in span_iteration_check() 298 WARN_ON_ONCE(mas.status != ma_overflow); in span_iteration_check()
|
| /linux/kernel/sched/ |
| H A D | idle.c | 35 WARN_ON_ONCE(cpu_idle_force_poll < 0); in cpu_idle_poll_ctrl() 269 if (WARN_ON_ONCE(irqs_disabled())) in cpuidle_idle_call() 416 WARN_ON_ONCE(current->policy != SCHED_FIFO); in play_idle_precise() 417 WARN_ON_ONCE(current->nr_cpus_allowed != 1); in play_idle_precise() 418 WARN_ON_ONCE(!(current->flags & PF_KTHREAD)); in play_idle_precise() 419 WARN_ON_ONCE(!(current->flags & PF_NO_SETAFFINITY)); in play_idle_precise() 420 WARN_ON_ONCE(!duration_ns); in play_idle_precise() 421 WARN_ON_ONCE(current->mm); in play_idle_precise() 467 return WARN_ON_ONCE(1); in balance_idle()
|
| /linux/include/linux/ |
| H A D | hung_task.h | 45 WARN_ON_ONCE(!lock_ptr); in hung_task_set_blocker() 46 WARN_ON_ONCE(READ_ONCE(current->blocker)); in hung_task_set_blocker() 73 WARN_ON_ONCE(!blocker); in hung_task_get_blocker_type() 80 WARN_ON_ONCE(!blocker); in hung_task_blocker_to_lock()
|
| /linux/kernel/time/ |
| H A D | clocksource-wdtest.c | 119 WARN_ON_ONCE(clocksource_wdtest_jiffies.uncertainty_margin != TICK_NSEC); in wdtest_func() 124 WARN_ON_ONCE(j1 == j2); in wdtest_func() 134 WARN_ON_ONCE(clocksource_wdtest_ktime.uncertainty_margin < NSEC_PER_USEC); in wdtest_func() 155 WARN_ON_ONCE(READ_ONCE(wdtest_ktime_read_ndelays)); in wdtest_func() 156 WARN_ON_ONCE((i <= max_retries) != in wdtest_func() 165 WARN_ON_ONCE(!(clocksource_wdtest_ktime.flags & CLOCK_SOURCE_UNSTABLE)); in wdtest_func()
|
| /linux/virt/lib/ |
| H A D | irqbypass.c | 100 if (WARN_ON_ONCE(producer->eventfd)) in irq_bypass_register_producer() 115 WARN_ON_ONCE(xa_erase(&producers, index) != producer); in irq_bypass_register_producer() 145 WARN_ON_ONCE(xa_erase(&producers, index) != producer); in irq_bypass_unregister_producer() 165 if (WARN_ON_ONCE(consumer->eventfd)) in irq_bypass_register_consumer() 181 WARN_ON_ONCE(xa_erase(&consumers, index) != consumer); in irq_bypass_register_consumer() 211 WARN_ON_ONCE(xa_erase(&consumers, index) != consumer); in irq_bypass_unregister_consumer()
|
| /linux/arch/x86/include/asm/ |
| H A D | entry-common.h | 32 WARN_ON_ONCE(flags & mask); in arch_enter_from_user_mode() 35 WARN_ON_ONCE(!user_mode(regs)); in arch_enter_from_user_mode() 42 WARN_ON_ONCE(!on_thread_stack()); in arch_enter_from_user_mode() 43 WARN_ON_ONCE(regs != task_pt_regs(current)); in arch_enter_from_user_mode()
|
| /linux/block/ |
| H A D | blk-settings.c | 91 if (WARN_ON_ONCE(lim->max_open_zones) || in blk_validate_zoned_limits() 92 WARN_ON_ONCE(lim->max_active_zones) || in blk_validate_zoned_limits() 93 WARN_ON_ONCE(lim->zone_write_granularity) || in blk_validate_zoned_limits() 94 WARN_ON_ONCE(lim->max_zone_append_sectors)) in blk_validate_zoned_limits() 99 if (WARN_ON_ONCE(!IS_ENABLED(CONFIG_BLK_DEV_ZONED))) in blk_validate_zoned_limits() 299 if (WARN_ON_ONCE(!is_power_of_2(lim->atomic_write_hw_unit_min))) in blk_validate_atomic_write_limits() 302 if (WARN_ON_ONCE(!is_power_of_2(lim->atomic_write_hw_unit_max))) in blk_validate_atomic_write_limits() 305 if (WARN_ON_ONCE(lim->atomic_write_hw_unit_min > in blk_validate_atomic_write_limits() 309 if (WARN_ON_ONCE(lim->atomic_write_hw_unit_max > in blk_validate_atomic_write_limits() 313 if (WARN_ON_ONCE(lim->chunk_sectors && in blk_validate_atomic_write_limits() [all …]
|
| /linux/scripts/coccinelle/misc/ |
| H A D | warn.cocci | 71 WARN_ON_ONCE(1); 78 *WARN_ON_ONCE(1); 84 cocci.print_main("printk + WARN_ON_ONCE can be just WARN_ONCE",p) 90 msg = "SUGGESTION: printk + WARN_ON_ONCE can be just WARN_ONCE" 101 -WARN_ON_ONCE(1);
|