/linux/block/ |
H A D | blk-settings.c | 79 io_opt = (u64)lim->max_sectors << SECTOR_SHIFT; in blk_apply_bdi_limits() 213 unsigned int unit_limit = min(lim->max_hw_sectors << SECTOR_SHIFT, in blk_atomic_writes_update_limits() 219 min(lim->atomic_write_hw_max >> SECTOR_SHIFT, in blk_atomic_writes_update_limits() 226 lim->atomic_write_hw_boundary >> SECTOR_SHIFT; in blk_atomic_writes_update_limits() 233 lim->atomic_write_hw_max >> SECTOR_SHIFT; in blk_validate_atomic_write_limits() 259 boundary_sectors = lim->atomic_write_hw_boundary >> SECTOR_SHIFT; in blk_validate_atomic_write_limits() 357 logical_block_sectors = lim->logical_block_size >> SECTOR_SHIFT; in blk_validate_limits() 375 } else if (lim->io_opt > (BLK_DEF_MAX_SECTORS_CAP << SECTOR_SHIFT)) { in blk_validate_limits() 377 min(max_hw_sectors, lim->io_opt >> SECTOR_SHIFT); in blk_validate_limits() 378 } else if (lim->io_min > (BLK_DEF_MAX_SECTORS_CAP << SECTOR_SHIFT)) { in blk_validate_limits() [all...] |
H A D | blk-merge.c | 104 return round_down(UINT_MAX, lim->logical_block_size) >> SECTOR_SHIFT; in bio_allowed_max_sectors() 197 unsigned pbs = lim->physical_block_size >> SECTOR_SHIFT; in get_max_io_size() 198 unsigned lbs = lim->logical_block_size >> SECTOR_SHIFT; in get_max_io_size() 351 return bytes >> SECTOR_SHIFT; in bio_split_rw_at() 360 get_max_io_size(bio, lim) << SECTOR_SHIFT)); in bio_split_rw() 376 lim->max_zone_append_sectors << SECTOR_SHIFT); in bio_split_zone_append()
|
H A D | ioctl.c | 47 start = p.start >> SECTOR_SHIFT; in blkpg_do_ioctl() 48 length = p.length >> SECTOR_SHIFT; in blkpg_do_ioctl() 152 sector = start >> SECTOR_SHIFT; in blk_ioctl_discard() 153 nr_sects = len >> SECTOR_SHIFT; in blk_ioctl_discard() 800 sector_t sector = start >> SECTOR_SHIFT; in blkdev_cmd_discard() 801 sector_t nr_sects = len >> SECTOR_SHIFT; in blkdev_cmd_discard()
|
/linux/fs/nfs/blocklayout/ |
H A D | blocklayout.c | 141 disk_addr = (u64)isect << SECTOR_SHIFT; in do_add_page_to_bio() 158 bio->bi_iter.bi_sector = disk_addr >> SECTOR_SHIFT; in do_add_page_to_bio() 176 isect = header->args.offset >> SECTOR_SHIFT; in bl_mark_devices_unavailable() 177 bytes_left += header->args.offset - (isect << SECTOR_SHIFT); in bl_mark_devices_unavailable() 185 if (bytes_left > extent_length << SECTOR_SHIFT) in bl_mark_devices_unavailable() 186 bytes_left -= extent_length << SECTOR_SHIFT; in bl_mark_devices_unavailable() 258 isect = (sector_t) (f_offset >> SECTOR_SHIFT); in bl_read_pagelist() 304 isect += (pg_len >> SECTOR_SHIFT); in bl_read_pagelist() 305 extent_length -= (pg_len >> SECTOR_SHIFT); in bl_read_pagelist() 310 if ((isect << SECTOR_SHIFT) > in bl_read_pagelist() [all...] |
/linux/fs/btrfs/ |
H A D | zoned.c | 64 #define SUPER_INFO_SECTORS ((u64)BTRFS_SUPER_INFO_SIZE >> SECTOR_SHIFT) 116 *wp_ret = zones[0].start << SECTOR_SHIFT; in sb_write_pointer() 125 u64 zone_end = (zones[i].start + zones[i].capacity) << SECTOR_SHIFT; in sb_write_pointer() 154 *wp_ret = sector << SECTOR_SHIFT; in sb_write_pointer() 197 const sector_t zone_sectors = device->fs_info->zone_size >> SECTOR_SHIFT; in emulate_report_zones() 201 pos >>= SECTOR_SHIFT; in emulate_report_zones() 263 ret = blkdev_report_zones(device->bdev, pos >> SECTOR_SHIFT, *nr_zones, in btrfs_get_dev_zones() 388 zone_sectors = fs_info->zone_size >> SECTOR_SHIFT; in btrfs_get_dev_zone_info() 394 zone_info->zone_size = zone_sectors << SECTOR_SHIFT; in btrfs_get_dev_zone_info() 475 ret = btrfs_get_dev_zones(device, sector << SECTOR_SHIFT, zone in btrfs_get_dev_zone_info() [all...] |
H A D | bio.c | 82 bio = bio_split(&orig_bbio->bio, map_length >> SECTOR_SHIFT, GFP_NOFS, in btrfs_split_bio() 188 repair_bbio->saved_iter.bi_sector << SECTOR_SHIFT, in btrfs_end_repair_bio() 212 const u64 logical = (failed_bbio->saved_iter.bi_sector << SECTOR_SHIFT); in repair_one_sector() 391 stripe->physical = bio->bi_iter.bi_sector << SECTOR_SHIFT; in btrfs_orig_write_end_io() 405 stripe->physical = bio->bi_iter.bi_sector << SECTOR_SHIFT; in btrfs_clone_write_end_io() 430 u64 physical = bio->bi_iter.bi_sector << SECTOR_SHIFT; in btrfs_submit_dev_bio() 434 bio->bi_iter.bi_sector = zone_start >> SECTOR_SHIFT; in btrfs_submit_dev_bio() 473 bio->bi_iter.bi_sector = bioc->stripes[dev_nr].physical >> SECTOR_SHIFT; in btrfs_submit_mirrored_bio() 485 bio->bi_iter.bi_sector = smap->physical >> SECTOR_SHIFT; in btrfs_submit_bio() 655 return ALIGN_DOWN(sector_offset << SECTOR_SHIFT, bbi in btrfs_append_map_length() [all...] |
/linux/drivers/md/ |
H A D | dm-integrity.c | 125 #define JOURNAL_SECTOR_DATA ((1 << SECTOR_SHIFT) - sizeof(commit_id_t)) 439 ms += offset >> (SECTOR_SHIFT + ic->log2_buffer_sectors - ic->log2_tag_size); in get_metadata_sector_and_offset() 440 mo = (offset << ic->log2_tag_size) & ((1U << SECTOR_SHIFT << ic->log2_buffer_sectors) - 1); in get_metadata_sector_and_offset() 442 ms += (__u64)offset * ic->tag_size >> (SECTOR_SHIFT + ic->log2_buffer_sectors); in get_metadata_sector_and_offset() 443 mo = (offset * ic->tag_size) & ((1U << SECTOR_SHIFT << ic->log2_buffer_sectors) - 1); in get_metadata_sector_and_offset() 496 __u8 *mac = sb + (1 << SECTOR_SHIFT) - mac_size; in sb_mac() 498 if (sizeof(struct superblock) + mac_size > 1 << SECTOR_SHIFT || in sb_mac() 727 *pl_index = sector >> (PAGE_SHIFT - SECTOR_SHIFT); in page_list_location() 728 *pl_offset = (sector << SECTOR_SHIFT) & (PAGE_SIZE - 1); in page_list_location() 740 *n_sectors = (PAGE_SIZE - pl_offset) >> SECTOR_SHIFT; in access_page_list() [all...] |
H A D | dm-writecache.c | 338 wc->memory_map += (size_t)wc->start_sector << SECTOR_SHIFT; in persistent_memory_claim() 339 wc->memory_map_size -= (size_t)wc->start_sector << SECTOR_SHIFT; in persistent_memory_claim() 359 vunmap(wc->memory_map - ((size_t)wc->start_sector << SECTOR_SHIFT)); in persistent_memory_release() 405 ((sector_t)e->index << (wc->block_size_bits - SECTOR_SHIFT)); in cache_sector() 515 region.sector = (sector_t)i * (BITMAP_GRANULARITY >> SECTOR_SHIFT); in ssd_commit_flushed() 516 region.count = (sector_t)(j - i) * (BITMAP_GRANULARITY >> SECTOR_SHIFT); in ssd_commit_flushed() 556 region.count = max(4096U, wc->block_size) >> SECTOR_SHIFT; in ssd_commit_superblock() 1017 (wc->metadata_sectors << SECTOR_SHIFT) - sb_entries_offset); in writecache_resume() 1344 if (next_boundary < bio->bi_iter.bi_size >> SECTOR_SHIFT) in writecache_map_remap_origin() 1365 dm_accept_partial_bio(bio, wc->block_size >> SECTOR_SHIFT); in writecache_map_read() [all...] |
H A D | dm-snap-persistent.c | 175 len = ps->store->chunk_size << SECTOR_SHIFT; in alloc_area() 297 memset(ps->area, 0, ps->store->chunk_size << SECTOR_SHIFT); in zero_memory_area() 387 memset(ps->header_area, 0, ps->store->chunk_size << SECTOR_SHIFT); in write_header() 498 ps->store->chunk_size << SECTOR_SHIFT, in read_exceptions() 546 memcpy(ps->area, area, ps->store->chunk_size << SECTOR_SHIFT); in read_exceptions() 627 ps->exceptions_per_area = (ps->store->chunk_size << SECTOR_SHIFT) / in persistent_read_metadata()
|
H A D | dm-verity-fec.c | 608 ((sector_t)(num_ll << (v->data_dev_block_bits - SECTOR_SHIFT)) in verity_fec_parse_opt_args() 609 >> (v->data_dev_block_bits - SECTOR_SHIFT) != num_ll)) { in verity_fec_parse_opt_args() 617 ((sector_t)(num_ll << (v->data_dev_block_bits - SECTOR_SHIFT)) >> in verity_fec_parse_opt_args() 618 (v->data_dev_block_bits - SECTOR_SHIFT) != num_ll)) { in verity_fec_parse_opt_args() 745 dm_bufio_set_sector_offset(f->bufio, f->start << (v->data_dev_block_bits - SECTOR_SHIFT)); in verity_fec_ctr() 747 fec_blocks = div64_u64(f->rounds * f->roots, v->fec->roots << SECTOR_SHIFT); in verity_fec_ctr()
|
/linux/drivers/mtd/ |
H A D | ssfdc.c | 35 #define SECTOR_SHIFT 9 macro 132 cis_sector = (int)(offset >> SECTOR_SHIFT); in get_valid_cis_sector() 153 loff_t offset = (loff_t)sect_no << SECTOR_SHIFT; in read_physical_sector() 307 ssfdc->cis_block = cis_sector / (mtd->erasesize >> SECTOR_SHIFT); in ssfdcr_add_mtd() 319 ssfdc->cylinders = (unsigned short)(((u32)mtd->size >> SECTOR_SHIFT) / in ssfdcr_add_mtd() 373 sectors_per_block = ssfdc->erase_size >> SECTOR_SHIFT; in ssfdcr_readsect()
|
/linux/drivers/vdpa/vdpa_sim/ |
H A D | vdpa_sim_blk.c | 157 offset = sector << SECTOR_SHIFT; in vdpasim_blk_handle_req() 172 to_push >> SECTOR_SHIFT, in vdpasim_blk_handle_req() 195 to_pull >> SECTOR_SHIFT, in vdpasim_blk_handle_req() 256 offset = sector << SECTOR_SHIFT; in vdpasim_blk_handle_req() 286 num_sectors << SECTOR_SHIFT); in vdpasim_blk_handle_req() 440 blk->buffer = kvzalloc(VDPASIM_BLK_CAPACITY << SECTOR_SHIFT, in vdpasim_blk_dev_add() 498 shared_buffer = kvzalloc(VDPASIM_BLK_CAPACITY << SECTOR_SHIFT, in vdpasim_blk_init()
|
/linux/fs/iomap/ |
H A D | ioend.c | 323 if (ioend->io_sector + (ioend->io_size >> SECTOR_SHIFT) != in iomap_ioend_can_merge() 390 lim->max_zone_append_sectors << SECTOR_SHIFT); in iomap_split_ioend() 400 sector_offset = max_len >> SECTOR_SHIFT; in iomap_split_ioend() 404 sector_offset = ALIGN_DOWN(sector_offset << SECTOR_SHIFT, in iomap_split_ioend() 405 i_blocksize(ioend->io_inode)) >> SECTOR_SHIFT; in iomap_split_ioend() 423 ioend->io_sector += (split_ioend->io_size >> SECTOR_SHIFT); in iomap_split_ioend()
|
/linux/fs/gfs2/ |
H A D | inode.h | 47 inode->i_blocks = blocks << (inode->i_blkbits - SECTOR_SHIFT); in gfs2_set_inode_blocks() 52 return inode->i_blocks >> (inode->i_blkbits - SECTOR_SHIFT); in gfs2_get_inode_blocks() 57 change <<= inode->i_blkbits - SECTOR_SHIFT; in gfs2_add_inode_blocks()
|
/linux/include/linux/ |
H A D | t10-pi.h | 47 return blk_rq_pos(rq) >> (shift - SECTOR_SHIFT) & 0xffffffff; in t10_pi_ref_tag() 72 return lower_48_bits(blk_rq_pos(rq) >> (shift - SECTOR_SHIFT)); in ext_pi_ref_tag()
|
H A D | blkdev.h | 830 return (loff_t)bdev_nr_sectors(bdev) << SECTOR_SHIFT; in bdev_nr_bytes() 841 (sb->s_blocksize_bits - SECTOR_SHIFT); in sb_bdev_nr_blocks() 1260 SECTOR_SHIFT), in sb_issue_discard() 1262 SECTOR_SHIFT), in sb_issue_discard() 1270 SECTOR_SHIFT), in sb_issue_zeroout() 1272 SECTOR_SHIFT), in sb_issue_zeroout() 1578 return q->limits.atomic_write_boundary_sectors << SECTOR_SHIFT; in queue_atomic_write_boundary_bytes() 1584 return q->limits.atomic_write_max_sectors << SECTOR_SHIFT; in queue_atomic_write_max_bytes() 1616 return order_base_2(size >> SECTOR_SHIFT) + SECTOR_SHIFT; in blksize_bits() [all...] |
/linux/fs/crypto/ |
H A D | bio.c | 71 pblk << (blockbits - SECTOR_SHIFT); in fscrypt_zeroout_range_inline_crypt() 123 sector_t sector = pblk << (inode->i_blkbits - SECTOR_SHIFT); in fscrypt_zeroout_range() 174 sector += 1U << (du_bits - SECTOR_SHIFT); in fscrypt_zeroout_range()
|
/linux/drivers/block/ |
H A D | zloop.c | 52 #define ZLOOP_DEF_ZONE_SIZE ((256ULL * SZ_1M) >> SECTOR_SHIFT) 162 file_sectors = stat.size >> SECTOR_SHIFT; in zloop_update_seq_zone() 169 if (file_sectors & ((zlo->block_size >> SECTOR_SHIFT) - 1)) { in zloop_update_seq_zone() 322 if (vfs_truncate(&zone->file->f_path, zlo->zone_size << SECTOR_SHIFT)) { in zloop_finish_zone() 476 cmd->iocb.ki_pos = (sector - zone->start) << SECTOR_SHIFT; in zloop_rw() 767 if (zlo->zone_capacity & ((zlo->block_size >> SECTOR_SHIFT) - 1)) { in zloop_get_block_size() 820 file_sectors = stat.size >> SECTOR_SHIFT; in zloop_init_zone() 829 zlo->zone_size << SECTOR_SHIFT); in zloop_init_zone() 886 .max_hw_sectors = SZ_1M >> SECTOR_SHIFT, in zloop_ctl_add() 887 .max_hw_zone_append_sectors = SZ_1M >> SECTOR_SHIFT, in zloop_ctl_add() [all...] |
H A D | n64cart.c | 92 u32 pos = bio->bi_iter.bi_sector << SECTOR_SHIFT; in n64cart_submit_bio() 150 set_capacity(disk, size >> SECTOR_SHIFT); in n64cart_probe()
|
/linux/drivers/block/zram/ |
H A D | zram_drv.h | 23 #define SECTORS_PER_PAGE_SHIFT (PAGE_SHIFT - SECTOR_SHIFT) 28 (1 << (ZRAM_LOGICAL_BLOCK_SHIFT - SECTOR_SHIFT))
|
/linux/drivers/nvdimm/ |
H A D | pmem.c | 55 return (offset - pmem->data_offset) >> SECTOR_SHIFT; in to_sect() 60 return (sector << SECTOR_SHIFT) + pmem->data_offset; in to_offset() 118 pmem_clear_bb(pmem, to_sect(pmem, offset), cleared >> SECTOR_SHIFT); in pmem_clear_poison() 247 sector_t sector = PFN_PHYS(pgoff) >> SECTOR_SHIFT; in __pmem_direct_access() 248 unsigned int num = PFN_PHYS(nr_pages) >> SECTOR_SHIFT; in __pmem_direct_access() 271 PAGE_ALIGN((first_bad - sector) << SECTOR_SHIFT)); in __pmem_direct_access() 299 PFN_PHYS(pgoff) >> SECTOR_SHIFT, in pmem_dax_zero_page_range() 336 if (!is_bad_pmem(&pmem->bb, PFN_PHYS(pgoff) >> SECTOR_SHIFT, len)) in pmem_recovery_write() 362 pmem_clear_bb(pmem, to_sect(pmem, pmem_off), cleared >> SECTOR_SHIFT); in pmem_recovery_write()
|
/linux/drivers/md/persistent-data/ |
H A D | dm-space-map-metadata.h | 13 #define DM_SM_METADATA_BLOCK_SIZE (4096 >> SECTOR_SHIFT)
|
/linux/drivers/md/dm-vdo/ |
H A D | constants.h | 84 VDO_SECTORS_PER_BLOCK = (VDO_BLOCK_SIZE >> SECTOR_SHIFT),
|
/linux/fs/zonefs/ |
H A D | super.c | 117 z->z_size >> SECTOR_SHIFT); in zonefs_zone_mgmt() 221 return (zone->wp - zone->start) << SECTOR_SHIFT; in zonefs_check_zone_condition() 383 zone.len = z->z_size >> SECTOR_SHIFT; in __zonefs_io_error() 659 inode->i_blocks = z->z_capacity >> SECTOR_SHIFT; in zonefs_get_file_inode() 998 z->z_size = zone->len << SECTOR_SHIFT; in zonefs_init_zgroup() 999 if (z->z_size > bdev_zone_sectors(sb->s_bdev) << SECTOR_SHIFT && in zonefs_init_zgroup() 1004 bdev_zone_sectors(sb->s_bdev) << SECTOR_SHIFT); in zonefs_init_zgroup() 1009 zone->capacity << SECTOR_SHIFT); in zonefs_init_zgroup()
|
/linux/fs/xfs/scrub/ |
H A D | xfile.h | 32 return file_inode(xf->file)->i_blocks << SECTOR_SHIFT; in xfile_bytes()
|