| /linux/fs/hfs/ |
| H A D | bnode.c | 21 bool is_valid = off < node->tree->node_size; in is_bnode_offset_valid() 28 node->tree->node_size, off); in is_bnode_offset_valid() 37 unsigned int node_size; in check_and_correct_requested_length() local 42 node_size = node->tree->node_size; in check_and_correct_requested_length() 44 if ((off + len) > node_size) { in check_and_correct_requested_length() 45 u32 new_len = node_size - off; in check_and_correct_requested_length() 52 node->tree->node_size, off, len, new_len); in check_and_correct_requested_length() 75 node->tree->node_size, off, len); in hfs_bnode_read() 147 node->tree->node_size, off, len); in hfs_bnode_write() 185 node->tree->node_size, off, len); in hfs_bnode_clear() [all …]
|
| H A D | brec.c | 24 dataoff = node->tree->node_size - (rec + 2) * 2; in hfs_brec_lenoff() 45 recoff = hfs_bnode_read_u16(node, node->tree->node_size - (rec + 1) * 2); in hfs_brec_keylen() 94 end_rec_off = tree->node_size - (node->num_recs + 1) * 2; in hfs_brec_insert() 117 idx_rec_off = tree->node_size - (rec + 1) * 2; in hfs_brec_insert() 187 rec_off = tree->node_size - (fd->record + 2) * 2; in hfs_brec_remove() 188 end_off = tree->node_size - (node->num_recs + 1) * 2; in hfs_brec_remove() 283 size = tree->node_size / 2 - node->num_recs * 2 - 14; in hfs_bnode_split() 284 old_rec_off = tree->node_size - 4; in hfs_bnode_split() 319 new_rec_off = tree->node_size - 2; in hfs_bnode_split() 403 rec_off = tree->node_size - (rec + 2) * 2; in hfs_brec_update_parent() [all …]
|
| H A D | btree.c | 124 tree->node_size = be16_to_cpu(head->node_size); in hfs_btree_open() 128 size = tree->node_size; in hfs_btree_open() 153 tree->pages_per_bnode = (tree->node_size + PAGE_SIZE - 1) >> PAGE_SHIFT; in hfs_btree_open() 246 hfs_bnode_clear(node, 0, tree->node_size); in hfs_bmap_new_bmap() 255 hfs_bnode_write_u16(node, tree->node_size - 2, 14); in hfs_bmap_new_bmap() 256 hfs_bnode_write_u16(node, tree->node_size - 4, tree->node_size - 6); in hfs_bmap_new_bmap()
|
| H A D | btree.h | 38 unsigned int node_size; member
|
| H A D | inode.c | 92 if (tree->node_size >= PAGE_SIZE) { in hfs_release_folio()
|
| /linux/fs/hfsplus/ |
| H A D | brec.c | 25 dataoff = node->tree->node_size - (rec + 2) * 2; in hfs_brec_lenoff() 45 node->tree->node_size - (rec + 1) * 2); in hfs_brec_keylen() 48 if (recoff > node->tree->node_size - 2) { in hfs_brec_keylen() 92 end_rec_off = tree->node_size - (node->num_recs + 1) * 2; in hfs_brec_insert() 117 idx_rec_off = tree->node_size - (rec + 1) * 2; in hfs_brec_insert() 188 rec_off = tree->node_size - (fd->record + 2) * 2; in hfs_brec_remove() 189 end_off = tree->node_size - (node->num_recs + 1) * 2; in hfs_brec_remove() 268 size = tree->node_size / 2 - node->num_recs * 2 - 14; in hfs_bnode_split() 269 old_rec_off = tree->node_size - 4; in hfs_bnode_split() 304 new_rec_off = tree->node_size - 2; in hfs_bnode_split() [all …]
|
| H A D | btree.c | 75 u32 hfsplus_calc_btree_clump_size(u32 block_size, u32 node_size, in hfsplus_calc_btree_clump_size() argument 78 u32 mod = max(node_size, block_size); in hfsplus_calc_btree_clump_size() 102 if (clump_size < (8 * node_size)) in hfsplus_calc_btree_clump_size() 103 clump_size = 8 * node_size; in hfsplus_calc_btree_clump_size() 175 tree->node_size = be16_to_cpu(head->node_size); in hfs_btree_open() 231 size = tree->node_size; in hfs_btree_open() 240 (tree->node_size + PAGE_SIZE - 1) >> in hfs_btree_open() 331 hfs_bnode_clear(node, 0, tree->node_size); in hfs_bmap_new_bmap() 340 hfs_bnode_write_u16(node, tree->node_size - 2, 14); in hfs_bmap_new_bmap() 341 hfs_bnode_write_u16(node, tree->node_size - 4, tree->node_size - 6); in hfs_bmap_new_bmap()
|
| H A D | xattr.c | 55 char *buf, u16 node_size) in hfsplus_init_header_node() argument 68 clump_size, node_size); in hfsplus_init_header_node() 71 rec_offsets = (__be16 *)(buf + node_size); in hfsplus_init_header_node() 80 head->node_size = cpu_to_be16(node_size); in hfsplus_init_header_node() 82 do_div(tmp, node_size); in hfsplus_init_header_node() 93 hdr_node_map_rec_bits = 8 * (node_size - offset - (4 * sizeof(u16))); in hfsplus_init_header_node() 99 map_node_bits = 8 * (node_size - sizeof(struct hfs_bnode_desc) - in hfsplus_init_header_node() 128 u16 node_size = HFSPLUS_ATTR_TREE_NODE_SIZE; in hfsplus_create_attributes_file() local 184 node_size, in hfsplus_create_attributes_file() 209 buf = kzalloc(node_size, GFP_NOFS); in hfsplus_create_attributes_file() [all …]
|
| H A D | hfsplus_fs.h | 54 unsigned int node_size; member 357 u32 hfsplus_calc_btree_clump_size(u32 block_size, u32 node_size, u64 sectors, 558 bool is_valid = off < node->tree->node_size; in is_bnode_offset_valid() 565 node->tree->node_size, off); in is_bnode_offset_valid() 574 unsigned int node_size; in check_and_correct_requested_length() local 579 node_size = node->tree->node_size; in check_and_correct_requested_length() 581 if ((off + len) > node_size) { in check_and_correct_requested_length() 582 u32 new_len = node_size - off; in check_and_correct_requested_length() 589 node->tree->node_size, off, len, new_len); in check_and_correct_requested_length()
|
| H A D | bnode.c | 36 node->tree->node_size, off, len); in hfs_bnode_read() 107 node->tree->node_size, off, len); in hfs_bnode_write() 149 node->tree->node_size, off, len); in hfs_bnode_clear() 360 off = node->tree->node_size - 2; in hfs_bnode_dump() 579 rec_off = tree->node_size - 2; in hfs_bnode_find() 587 next_off > tree->node_size || in hfs_bnode_find() 644 min_t(int, PAGE_SIZE, tree->node_size)); in hfs_bnode_create() 689 hfs_bnode_clear(node, 0, tree->node_size); in hfs_bnode_put()
|
| H A D | inode.c | 88 if (tree->node_size >= PAGE_SIZE) { in hfsplus_release_folio()
|
| /linux/drivers/gpu/drm/i915/ |
| H A D | i915_vma_resource.h | 127 u64 node_size; member 212 u64 node_size, in i915_vma_resource_init() argument 230 vma_res->node_size = node_size; in i915_vma_resource_init()
|
| H A D | i915_vma_resource.c | 38 #define VMA_RES_LAST(_node) ((_node)->start + (_node)->node_size + (_node)->guard - 1)
|
| /linux/fs/befs/ |
| H A D | befs_fs_types.h | 219 fs32 node_size; member 229 u32 node_size; member
|
| H A D | btree.c | 152 sup->node_size = fs32_to_cpu(sb, od_sup->node_size); in befs_bt_read_super()
|
| H A D | debug.c | 233 befs_debug(sb, " node_size %u", fs32_to_cpu(sb, super->node_size)); in befs_dump_index_entry()
|
| /linux/drivers/media/platform/st/sti/bdisp/ |
| H A D | bdisp-hw.c | 452 unsigned int i, node_size = sizeof(struct bdisp_node); in bdisp_hw_alloc_nodes() local 457 base = dma_alloc_attrs(dev, node_size * MAX_NB_NODE, &paddr, in bdisp_hw_alloc_nodes() 464 memset(base, 0, node_size * MAX_NB_NODE); in bdisp_hw_alloc_nodes() 471 base += node_size; in bdisp_hw_alloc_nodes() 472 paddr += node_size; in bdisp_hw_alloc_nodes()
|
| /linux/drivers/gpu/drm/xe/display/ |
| H A D | xe_stolen.c | 110 .node_size = xe_stolen_node_size,
|
| /linux/lib/ |
| H A D | bootconfig.c | 91 int __init xbc_get_info(int *node_size, size_t *data_size) in xbc_get_info() argument 96 if (node_size) in xbc_get_info() 97 *node_size = xbc_node_num; in xbc_get_info()
|
| /linux/tools/testing/selftests/mm/ |
| H A D | ksm_tests.c | 441 long node_size; in get_next_mem_node() local 447 node_size = numa_node_size(mem_node, NULL); in get_next_mem_node() 448 if (node_size > 0) in get_next_mem_node()
|
| /linux/drivers/net/ethernet/huawei/hinic/ |
| H A D | hinic_hw_api_cmd.c | 733 size_t node_size; in api_cmd_destroy_cell() local 739 node_size = chain->cell_size; in api_cmd_destroy_cell() 751 dma_free_coherent(&pdev->dev, node_size, node, in api_cmd_destroy_cell()
|
| /linux/include/linux/ |
| H A D | bootconfig.h | 288 int __init xbc_get_info(int *node_size, size_t *data_size);
|
| H A D | hfs_common.h | 486 __be16 node_size; /* (F) The number of bytes in a node (=512) */ member
|
| /linux/drivers/gpu/drm/i915/gt/ |
| H A D | intel_ggtt.c | 514 end += (vma_res->node_size + vma_res->guard) / I915_GTT_PAGE_SIZE; in gen8_ggtt_insert_entries() 547 end += (vma_res->node_size + vma_res->guard) / I915_GTT_PAGE_SIZE; in __gen8_ggtt_insert_entries_bind() 549 vma_res->node_size / I915_GTT_PAGE_SIZE, pte_encode)) in __gen8_ggtt_insert_entries_bind() 552 start += vma_res->node_size / I915_GTT_PAGE_SIZE; in __gen8_ggtt_insert_entries_bind() 666 end += (vma_res->node_size + vma_res->guard) / I915_GTT_PAGE_SIZE; in gen6_ggtt_insert_entries()
|
| /linux/io_uring/ |
| H A D | rsrc.c | 158 const int node_size = sizeof(struct io_rsrc_node); in io_rsrc_cache_init() local 162 node_size, 0); in io_rsrc_cache_init()
|