| /linux/fs/xfs/ |
| H A D | xfs_trans_buf.c | 43 ASSERT(blip->bli_buf->b_map_count == nmaps); in xfs_trans_buf_item_match() 68 ASSERT(bp->b_transp == NULL); in _xfs_trans_bjoin() 77 ASSERT(!(bip->bli_flags & XFS_BLI_STALE)); in _xfs_trans_bjoin() 78 ASSERT(!(bip->__bli_format.blf_flags & XFS_BLF_CANCEL)); in _xfs_trans_bjoin() 79 ASSERT(!(bip->bli_flags & XFS_BLI_LOGGED)); in _xfs_trans_bjoin() 140 ASSERT(xfs_buf_islocked(bp)); in xfs_trans_get_buf_map() 146 ASSERT(bp->b_transp == tp); in xfs_trans_get_buf_map() 148 ASSERT(bip != NULL); in xfs_trans_get_buf_map() 149 ASSERT(atomic_read(&bip->bli_refcount) > 0); in xfs_trans_get_buf_map() 160 ASSERT(!bp->b_error); in xfs_trans_get_buf_map() [all …]
|
| H A D | xfs_buf_item.c | 40 ASSERT(bip->bli_formats == NULL); in xfs_buf_item_get_format() 82 ASSERT(!test_bit(XFS_LI_IN_AIL, &bip->bli_item.li_flags)); in xfs_buf_item_relse() 83 ASSERT(atomic_read(&bip->bli_refcount) == 0); in xfs_buf_item_relse() 144 ASSERT(nbits > 0); in xfs_buf_item_size_segment() 210 ASSERT(atomic_read(&bip->bli_refcount) > 0); in xfs_buf_item_size() 218 ASSERT(bip->__bli_format.blf_flags & XFS_BLF_CANCEL); in xfs_buf_item_size() 226 ASSERT(bip->bli_flags & XFS_BLI_LOGGED); in xfs_buf_item_size() 318 ASSERT(blfp->blf_flags & XFS_BLF_CANCEL); in xfs_buf_item_format_segment() 327 ASSERT(first_bit >= 0); in xfs_buf_item_format_segment() 330 ASSERT(nbits > 0); in xfs_buf_item_format_segment() [all …]
|
| H A D | xfs_dquot_item_recover.c | 44 ASSERT(type); in xlog_recover_dquot_ra_pass2() 49 ASSERT(dq_f); in xlog_recover_dquot_ra_pass2() 50 ASSERT(dq_f->qlf_len == 1); in xlog_recover_dquot_ra_pass2() 97 ASSERT(type); in xlog_recover_dquot_commit_pass2() 112 ASSERT(dq_f); in xlog_recover_dquot_commit_pass2() 119 ASSERT(dq_f->qlf_len == 1); in xlog_recover_dquot_commit_pass2() 134 ASSERT(bp); in xlog_recover_dquot_commit_pass2() 169 ASSERT(dq_f->qlf_size == 2); in xlog_recover_dquot_commit_pass2() 170 ASSERT(bp->b_mount == mp); in xlog_recover_dquot_commit_pass2() 196 ASSERT(qoff_f); in xlog_recover_quotaoff_commit_pass1()
|
| H A D | xfs_qm.c | 178 ASSERT(atomic_read(&dqp->q_pincount) == 0); in xfs_qm_dqpurge() 179 ASSERT(xlog_is_shutdown(dqp->q_logitem.qli_item.li_log) || in xfs_qm_dqpurge() 192 ASSERT(!list_empty(&dqp->q_lru)); in xfs_qm_dqpurge() 273 ASSERT(mp->m_rootip); in xfs_qm_unmount_quotas() 326 ASSERT(!xfs_is_metadir_inode(ip)); in xfs_qm_dqattach_locked() 333 ASSERT(ip->i_udquot); in xfs_qm_dqattach_locked() 341 ASSERT(ip->i_gdquot); in xfs_qm_dqattach_locked() 349 ASSERT(ip->i_pdquot); in xfs_qm_dqattach_locked() 393 ASSERT(!xfs_is_quota_inode(&ip->i_mount->m_sb, ip->i_ino)); in xfs_qm_dqdetach() 467 ASSERT(!XFS_DQ_IS_DIRTY(dqp)); in xfs_qm_dquot_isolate() [all …]
|
| H A D | xfs_inode.c | 99 ASSERT((lock_flags & (XFS_IOLOCK_SHARED | XFS_IOLOCK_EXCL)) != in xfs_lock_flags_assert() 101 ASSERT((lock_flags & (XFS_MMAPLOCK_SHARED | XFS_MMAPLOCK_EXCL)) != in xfs_lock_flags_assert() 103 ASSERT((lock_flags & (XFS_ILOCK_SHARED | XFS_ILOCK_EXCL)) != in xfs_lock_flags_assert() 105 ASSERT((lock_flags & ~(XFS_LOCK_MASK | XFS_LOCK_SUBCLASS_MASK)) == 0); in xfs_lock_flags_assert() 106 ASSERT(lock_flags != 0); in xfs_lock_flags_assert() 276 ASSERT(lock_flags & (XFS_IOLOCK_EXCL|XFS_MMAPLOCK_EXCL|XFS_ILOCK_EXCL)); in xfs_ilock_demote() 277 ASSERT((lock_flags & in xfs_ilock_demote() 345 ASSERT(!(lock_mode & XFS_ILOCK_PARENT)); in xfs_lock_inumorder() 346 ASSERT(xfs_lockdep_subclass_ok(subclass)); in xfs_lock_inumorder() 349 ASSERT(subclass <= XFS_IOLOCK_MAX_SUBCLASS); in xfs_lock_inumorder() [all …]
|
| /linux/tools/testing/selftests/sync/ |
| H A D | sync_fence.c | 38 ASSERT(valid, "Failure allocating timeline\n"); in test_fence_one_timeline_wait() 42 ASSERT(valid, "Failure allocating fence\n"); in test_fence_one_timeline_wait() 46 ASSERT(ret == 0, "Failure waiting on fence until timeout\n"); in test_fence_one_timeline_wait() 50 ASSERT(ret == 0, "Failure advancing timeline\n"); in test_fence_one_timeline_wait() 54 ASSERT(ret == 0, "Failure waiting on fence until timeout\n"); in test_fence_one_timeline_wait() 58 ASSERT(ret == 0, "Failure signaling the fence\n"); in test_fence_one_timeline_wait() 62 ASSERT(ret > 0, "Failure waiting on fence\n"); in test_fence_one_timeline_wait() 66 ASSERT(ret == 0, "Failure going further\n"); in test_fence_one_timeline_wait() 68 ASSERT(ret > 0, "Failure waiting ahead\n"); in test_fence_one_timeline_wait() 89 ASSERT(valid, "Failure allocating fences\n"); in test_fence_one_timeline_merge() [all …]
|
| H A D | sync_stress_consumer.c | 46 ASSERT(error == 0, "Error occurred on fence\n"); in busy_wait_on_fence() 74 ASSERT(valid, "Failure creating fence\n"); in mpsc_producer_thread() 82 ASSERT(sync_wait(fence, -1) > 0, in mpsc_producer_thread() 85 ASSERT(busy_wait_on_fence(fence) == 0, in mpsc_producer_thread() 97 ASSERT(sw_sync_timeline_inc(producer_timelines[id], 1) == 0, in mpsc_producer_thread() 126 ASSERT(valid, "Failure merging fences\n"); in mpcs_consumer_thread() 133 ASSERT(sync_wait(fence, -1) > 0, in mpcs_consumer_thread() 136 ASSERT(busy_wait_on_fence(fence) == 0, in mpcs_consumer_thread() 140 ASSERT(test_data_mpsc.counter == n * it, in mpcs_consumer_thread() 144 ASSERT(sw_sync_timeline_inc(consumer_timeline, 1) == 0, in mpcs_consumer_thread()
|
| H A D | sync_wait.c | 50 ASSERT(valid, "Failure merging fence from various timelines\n"); in test_fence_multi_timeline_wait() 54 ASSERT(active == 3, "Fence signaled too early!\n"); in test_fence_multi_timeline_wait() 57 ASSERT(ret == 0, in test_fence_multi_timeline_wait() 63 ASSERT(active == 2 && signaled == 1, in test_fence_multi_timeline_wait() 69 ASSERT(active == 1 && signaled == 2, in test_fence_multi_timeline_wait() 75 ASSERT(active == 0 && signaled == 3, in test_fence_multi_timeline_wait() 80 ASSERT(ret > 0, "Failure waiting on signaled fence\n"); in test_fence_multi_timeline_wait()
|
| H A D | sync_stress_parallelism.c | 51 ASSERT(valid, "Failure allocating fence\n"); in test_stress_two_threads_shared_timeline_thread() 55 ASSERT(ret > 0, "Problem occurred on prior thread\n"); in test_stress_two_threads_shared_timeline_thread() 61 ASSERT(test_data_two_threads.counter == i * 2 + thread_id, in test_stress_two_threads_shared_timeline_thread() 67 ASSERT(ret == 0, "Advancing timeline failed\n"); in test_stress_two_threads_shared_timeline_thread() 82 ASSERT(valid, "Failure allocating timeline\n"); in test_stress_two_threads_shared_timeline() 104 ASSERT(test_data_two_threads.counter == in test_stress_two_threads_shared_timeline()
|
| H A D | sync_merge.c | 38 ASSERT(valid, "Failure allocating timeline\n"); in test_fence_merge_same_fence() 42 ASSERT(valid, "Failure allocating fence\n"); in test_fence_merge_same_fence() 46 ASSERT(valid, "Failure merging fence\n"); in test_fence_merge_same_fence() 48 ASSERT(sync_fence_count_with_status(merged, FENCE_STATUS_SIGNALED) == 0, in test_fence_merge_same_fence() 52 ASSERT(sync_fence_count_with_status(merged, FENCE_STATUS_SIGNALED) == 1, in test_fence_merge_same_fence()
|
| H A D | sync_alloc.c | 38 ASSERT(valid, "Failure allocating timeline\n"); in test_alloc_timeline() 50 ASSERT(valid, "Failure allocating timeline\n"); in test_alloc_fence() 54 ASSERT(valid, "Failure allocating fence\n"); in test_alloc_fence() 66 ASSERT(timeline > 0, "Failure allocating timeline\n"); in test_alloc_fence_negative() 69 ASSERT(fence < 0, "Success allocating negative fence\n"); in test_alloc_fence_negative()
|
| /linux/arch/arm64/kernel/ |
| H A D | vmlinux.lds.S | 211 ASSERT(SIZEOF(.got.plt) == 0 || SIZEOF(.got.plt) == 0x18, 302 ASSERT(SIZEOF(.data.rel.ro) == 0, "Unexpected RELRO detected!") 364 ASSERT(SIZEOF(.plt) == 0, "Unexpected run-time procedure linkages detected!") 374 ASSERT(__hyp_idmap_text_end - __hyp_idmap_text_start <= PAGE_SIZE, 376 ASSERT(__idmap_text_end - (__idmap_text_start & ~(SZ_4K - 1)) <= SZ_4K, 379 ASSERT(__hibernate_exit_text_end - __hibernate_exit_text_start <= SZ_4K, 381 ASSERT(__hibernate_exit_text_start == swsusp_arch_suspend_exit, 385 ASSERT((__entry_tramp_text_end - __entry_tramp_text_start) <= 3*PAGE_SIZE, 389 ASSERT(__hyp_bss_start == __bss_start, "HYP and Host BSS are misaligned") 394 ASSERT(_text == KIMAGE_VADDR, "HEAD is misaligned") [all …]
|
| /linux/fs/xfs/libxfs/ |
| H A D | xfs_dir2_sf.c | 96 ASSERT(ino <= XFS_MAXINUMBER); in xfs_dir2_sf_put_ino() 121 ASSERT(ino <= XFS_MAXINUMBER); in xfs_dir2_sf_put_parent_ino() 154 ASSERT(ftype < XFS_DIR3_FT_MAX); in xfs_dir2_sf_put_ftype() 304 ASSERT(be64_to_cpu(dep->inumber) == dp->i_ino); in xfs_dir2_block_to_sf() 310 ASSERT(be64_to_cpu(dep->inumber) == in xfs_dir2_block_to_sf() 328 ASSERT((char *)sfep - (char *)sfp == size); in xfs_dir2_block_to_sf() 334 ASSERT(error != -ENOSPC); in xfs_dir2_block_to_sf() 344 ASSERT(dp->i_df.if_bytes == 0); in xfs_dir2_block_to_sf() 379 ASSERT(xfs_dir2_sf_lookup(args) == -ENOENT); in xfs_dir2_sf_addname() 380 ASSERT(dp->i_df.if_format == XFS_DINODE_FMT_LOCAL); in xfs_dir2_sf_addname() [all …]
|
| H A D | xfs_bmap_btree.c | 65 ASSERT(be16_to_cpu(rblock->bb_level) > 0); in xfs_bmdr_to_bmbt() 125 ASSERT(s->br_state == XFS_EXT_NORM || s->br_state == XFS_EXT_UNWRITTEN); in xfs_bmbt_disk_set_all() 126 ASSERT(!(s->br_startoff & xfs_mask64hi(64-BMBT_STARTOFF_BITLEN))); in xfs_bmbt_disk_set_all() 127 ASSERT(!(s->br_blockcount & xfs_mask64hi(64-BMBT_BLOCKCOUNT_BITLEN))); in xfs_bmbt_disk_set_all() 128 ASSERT(!(s->br_startblock & xfs_mask64hi(64-BMBT_STARTBLOCK_BITLEN))); in xfs_bmbt_disk_set_all() 158 ASSERT(rblock->bb_magic == cpu_to_be32(XFS_BMAP_CRC_MAGIC)); in xfs_bmbt_to_bmdr() 159 ASSERT(uuid_equal(&rblock->bb_u.l.bb_uuid, in xfs_bmbt_to_bmdr() 161 ASSERT(rblock->bb_u.l.bb_blkno == in xfs_bmbt_to_bmdr() 164 ASSERT(rblock->bb_magic == cpu_to_be32(XFS_BMAP_MAGIC)); in xfs_bmbt_to_bmdr() 165 ASSERT(rblock->bb_u.l.bb_leftsib == cpu_to_be64(NULLFSBLOCK)); in xfs_bmbt_to_bmdr() [all …]
|
| H A D | xfs_attr_leaf.c | 168 ASSERT(!to->count && !to->usedbytes); in xfs_attr3_leaf_firstused_from_disk() 169 ASSERT(geo->blksize > USHRT_MAX); in xfs_attr3_leaf_firstused_from_disk() 184 ASSERT(from->firstused != XFS_ATTR3_LEAF_NULLOFF); in xfs_attr3_leaf_firstused_to_disk() 193 ASSERT(from->firstused == geo->blksize); in xfs_attr3_leaf_firstused_to_disk() 213 ASSERT(from->hdr.info.magic == cpu_to_be16(XFS_ATTR_LEAF_MAGIC) || in xfs_attr3_leaf_hdr_from_disk() 255 ASSERT(from->magic == XFS_ATTR_LEAF_MAGIC || in xfs_attr3_leaf_hdr_to_disk() 275 ASSERT(xfs_attr_leaf_ichdr_freemaps_verify(from, to) == NULL); in xfs_attr3_leaf_hdr_to_disk() 292 ASSERT(xfs_attr_leaf_ichdr_freemaps_verify(from, to) == NULL); in xfs_attr3_leaf_hdr_to_disk() 584 ASSERT(args->value != NULL); in xfs_attr_parent_match() 813 ASSERT(ifp->if_bytes == 0); in xfs_attr_shortform_create() [all …]
|
| H A D | xfs_btree_staging.c | 63 ASSERT(!(cur->bc_flags & XFS_BTREE_STAGING)); in xfs_btree_stage_afakeroot() 64 ASSERT(cur->bc_ops->type != XFS_BTREE_TYPE_INODE); in xfs_btree_stage_afakeroot() 65 ASSERT(cur->bc_tp == NULL); in xfs_btree_stage_afakeroot() 84 ASSERT(cur->bc_flags & XFS_BTREE_STAGING); in xfs_btree_commit_afakeroot() 85 ASSERT(cur->bc_tp == NULL); in xfs_btree_commit_afakeroot() 130 ASSERT(!(cur->bc_flags & XFS_BTREE_STAGING)); in xfs_btree_stage_ifakeroot() 131 ASSERT(cur->bc_ops->type == XFS_BTREE_TYPE_INODE); in xfs_btree_stage_ifakeroot() 132 ASSERT(cur->bc_tp == NULL); in xfs_btree_stage_ifakeroot() 153 ASSERT(cur->bc_flags & XFS_BTREE_STAGING); in xfs_btree_commit_ifakeroot() 154 ASSERT(cur->bc_tp == NULL); in xfs_btree_commit_ifakeroot() [all …]
|
| H A D | xfs_bmap.c | 98 ASSERT(mp->m_bm_maxlevels[whichfork] <= xfs_bmbt_maxlevels_ondisk()); in xfs_bmap_compute_maxlevels() 295 ASSERT(be16_to_cpu(block->bb_level) > 0); in xfs_check_block() 303 ASSERT(be64_to_cpu(prevp->br_startoff) < in xfs_check_block() 373 ASSERT(level > 0); in xfs_bmap_check_leaf_extents() 378 ASSERT(bno != NULLFSBLOCK); in xfs_bmap_check_leaf_extents() 379 ASSERT(XFS_FSB_TO_AGNO(mp, bno) < mp->m_sb.sb_agcount); in xfs_bmap_check_leaf_extents() 380 ASSERT(XFS_FSB_TO_AGBNO(mp, bno) < mp->m_sb.sb_agblocks); in xfs_bmap_check_leaf_extents() 450 ASSERT(xfs_bmbt_disk_get_startoff(&last) + in xfs_bmap_check_leaf_extents() 456 ASSERT(xfs_bmbt_disk_get_startoff(ep) + in xfs_bmap_check_leaf_extents() 519 ASSERT(ret_nmap <= nmap); in xfs_bmap_validate_ret() [all …]
|
| H A D | xfs_iext_tree.c | 57 ASSERT((irec->br_startoff & ~XFS_IEXT_STARTOFF_MASK) == 0); in xfs_iext_set() 58 ASSERT((irec->br_blockcount & ~XFS_IEXT_LENGTH_MASK) == 0); in xfs_iext_set() 59 ASSERT((irec->br_startblock & ~XFS_IEXT_STARTBLOCK_MASK) == 0); in xfs_iext_set() 169 ASSERT(node); in xfs_iext_find_first_leaf() 190 ASSERT(node); in xfs_iext_find_last_leaf() 231 ASSERT(cur->pos <= 0 || cur->pos >= RECS_PER_LEAF); in xfs_iext_next() 236 ASSERT(cur->pos >= 0); in xfs_iext_next() 237 ASSERT(cur->pos < xfs_iext_max_recs(ifp)); in xfs_iext_next() 253 ASSERT(cur->pos <= 0 || cur->pos >= RECS_PER_LEAF); in xfs_iext_prev() 258 ASSERT(cur->pos >= 0); in xfs_iext_prev() [all …]
|
| /linux/arch/x86/kernel/ |
| H A D | vmlinux.lds.S | 113 ASSERT(__relocate_kernel_end - __relocate_kernel_start <= KEXEC_CONTROL_CODE_MAX_SIZE, 337 ASSERT(__per_cpu_hot_end - __per_cpu_hot_start <= 64, "percpu cache hot data too large") 440 ASSERT(SIZEOF(.got.plt) == 0 || 455 ASSERT(SIZEOF(.got) == 0, "Unexpected GOT entries detected!") 460 ASSERT(SIZEOF(.plt) == 0, "Unexpected run-time procedure linkages detected!") 465 ASSERT(SIZEOF(.rel.dyn) == 0, "Unexpected run-time relocations (.rel) detected!") 470 ASSERT(SIZEOF(.rela.dyn) == 0, "Unexpected run-time relocations (.rela) detected!") 483 . = ASSERT((_end - LOAD_OFFSET <= KERNEL_IMAGE_SIZE), 493 . = ASSERT((retbleed_return_thunk & 0x3f) == 0, "retbleed_return_thunk not cacheline-aligned"); 497 . = ASSERT((srso_safe_ret & 0x3f) == 0, "srso_safe_ret not cacheline-aligned"); [all …]
|
| /linux/drivers/gpu/drm/amd/display/dc/dml2_0/ |
| H A D | dml_display_rq_dlg_calc.c | 133 ASSERT(pte_row_height_linear >= 8); in dml_rq_dlg_get_rq_reg() 140 ASSERT(p1_pte_row_height_linear >= 8); in dml_rq_dlg_get_rq_reg() 282 ASSERT(refclk_freq_in_mhz != 0); in dml_rq_dlg_get_dlg_reg() 283 ASSERT(pclk_freq_in_mhz != 0); in dml_rq_dlg_get_dlg_reg() 284 ASSERT(ref_freq_to_pix_freq < 4.0); in dml_rq_dlg_get_dlg_reg() 315 ASSERT(disp_dlg_regs->refcyc_h_blank_end < (dml_uint_t)dml_pow(2, 13)); in dml_rq_dlg_get_dlg_reg() 358 ASSERT(dst_y_per_vm_vblank < max_dst_y_per_vm_vblank); in dml_rq_dlg_get_dlg_reg() 359 ASSERT(dst_y_per_row_vblank < max_dst_y_per_row_vblank); in dml_rq_dlg_get_dlg_reg() 360 ASSERT(dst_y_prefetch > (dst_y_per_vm_vblank + dst_y_per_row_vblank)); in dml_rq_dlg_get_dlg_reg() 401 ASSERT(num_cursors <= 1); in dml_rq_dlg_get_dlg_reg() [all …]
|
| /linux/drivers/gpu/drm/amd/display/dc/dml/dcn32/ |
| H A D | display_rq_dlg_calc_32.c | 135 ASSERT(pte_row_height_linear >= 8); in dml32_rq_dlg_get_rq_reg() 143 ASSERT(p1_pte_row_height_linear >= 8); in dml32_rq_dlg_get_rq_reg() 272 ASSERT(ref_freq_to_pix_freq < 4.0); in dml32_rq_dlg_get_dlg_reg() 344 ASSERT(dlg_regs->refcyc_h_blank_end < (unsigned int)dml_pow(2, 13)); in dml32_rq_dlg_get_dlg_reg() 370 ASSERT(dst_y_per_vm_vblank < max_dst_y_per_vm_vblank); in dml32_rq_dlg_get_dlg_reg() 371 ASSERT(dst_y_per_row_vblank < max_dst_y_per_row_vblank); in dml32_rq_dlg_get_dlg_reg() 372 ASSERT(dst_y_prefetch > (dst_y_per_vm_vblank + dst_y_per_row_vblank)); in dml32_rq_dlg_get_dlg_reg() 426 ASSERT(src->num_cursors <= 1); in dml32_rq_dlg_get_dlg_reg() 441 ASSERT(dlg_regs->min_dst_y_next_start < (unsigned int)dml_pow(2, 18)); in dml32_rq_dlg_get_dlg_reg() 554 ASSERT(refcyc_per_req_delivery_pre_l < dml_pow(2, 13)); in dml32_rq_dlg_get_dlg_reg() [all …]
|
| /linux/fs/btrfs/ |
| H A D | lzo.c | 140 ASSERT(out_folio && *out_folio); in write_and_queue_folio() 142 ASSERT(foffset + write_len <= fsize); in write_and_queue_folio() 156 ASSERT(IS_ALIGNED(*total_out, fsize)); in write_and_queue_folio() 208 ASSERT(out_folio); in copy_compressed_data_to_bio() 211 ASSERT(old_size); in copy_compressed_data_to_bio() 212 ASSERT(old_size == *total_out); in copy_compressed_data_to_bio() 218 ASSERT((old_size >> sectorsize_bits) == (old_size + LZO_LEN - 1) >> sectorsize_bits); in copy_compressed_data_to_bio() 269 ASSERT(*out_folio); in copy_compressed_data_to_bio() 296 ASSERT(bio->bi_iter.bi_size == 0); in lzo_compress_bio() 297 ASSERT(len); in lzo_compress_bio() [all …]
|
| H A D | subpage.c | 73 ASSERT(!folio_test_large(folio)); in btrfs_attach_folio_state() 80 ASSERT(folio_test_locked(folio)); in btrfs_attach_folio_state() 112 ASSERT(bfs); in btrfs_detach_folio_state() 122 ASSERT(fs_info->sectorsize < fsize); in btrfs_alloc_folio_state() 155 ASSERT(folio_test_private(folio) && folio->mapping); in btrfs_folio_inc_eb_refs() 169 ASSERT(folio_test_private(folio) && folio->mapping); in btrfs_folio_dec_eb_refs() 173 ASSERT(atomic_read(&bfs->eb_refs)); in btrfs_folio_dec_eb_refs() 181 ASSERT(folio_test_private(folio) && folio_get_private(folio)); in btrfs_subpage_assert() 182 ASSERT(IS_ALIGNED(start, fs_info->sectorsize) && in btrfs_subpage_assert() 189 ASSERT(folio_pos(folio) <= start && in btrfs_subpage_assert() [all …]
|
| /linux/fs/xfs/scrub/ |
| H A D | tempfile.c | 66 ASSERT(sc->tp == NULL); in xrep_tempfile_create() 67 ASSERT(sc->tempip == NULL); in xrep_tempfile_create() 209 ASSERT(sc->tp == NULL); in xrep_tempfile_adjust_directory_tree() 210 ASSERT(!xfs_is_metadir_inode(sc->tempip)); in xrep_tempfile_adjust_directory_tree() 258 ASSERT(sc->tp == NULL); in xrep_tempfile_remove_metadir() 422 ASSERT(sc->tempip != NULL); in xrep_tempfile_prealloc() 423 ASSERT(!XFS_NOT_DQATTACHED(sc->mp, sc->tempip)); in xrep_tempfile_prealloc() 437 ASSERT(nmaps != 0); in xrep_tempfile_prealloc() 495 ASSERT(S_ISREG(VFS_I(sc->tempip)->i_mode)); in xrep_tempfile_copyin() 550 ASSERT(list_empty(&buffers_list)); in xrep_tempfile_copyin() [all …]
|
| /linux/drivers/gpu/drm/amd/display/dc/basics/ |
| H A D | bw_fixed.c | 53 ASSERT(value < BW_FIXED_MAX_I32 && value > BW_FIXED_MIN_I32); in bw_int_to_fixed_nonconst() 70 ASSERT(denominator != 0); in bw_frc_to_fixed() 76 ASSERT(res_value <= BW_FIXED_MAX_I32); in bw_frc_to_fixed() 98 ASSERT(res_value <= MAX_I64 - summand); in bw_frc_to_fixed() 118 ASSERT(abs_i64(result.value) <= abs_i64(arg.value)); in bw_floor2() 159 ASSERT(res.value <= BW_FIXED_MAX_I32); in bw_mul() 165 ASSERT(tmp <= (uint64_t)(MAX_I64 - res.value)); in bw_mul() 171 ASSERT(tmp <= (uint64_t)(MAX_I64 - res.value)); in bw_mul() 180 ASSERT(tmp <= (uint64_t)(MAX_I64 - res.value)); in bw_mul()
|