/linux-3.3/mm/ |
D | mempolicy.c | 121 int (*create)(struct mempolicy *pol, const nodemask_t *nodes); 136 void (*rebind)(struct mempolicy *pol, const nodemask_t *nodes, 171 static int mpol_new_interleave(struct mempolicy *pol, const nodemask_t *nodes) in mpol_new_interleave() argument 173 if (nodes_empty(*nodes)) in mpol_new_interleave() 175 pol->v.nodes = *nodes; in mpol_new_interleave() 179 static int mpol_new_preferred(struct mempolicy *pol, const nodemask_t *nodes) in mpol_new_preferred() argument 181 if (!nodes) in mpol_new_preferred() 183 else if (nodes_empty(*nodes)) in mpol_new_preferred() 186 pol->v.preferred_node = first_node(*nodes); in mpol_new_preferred() 190 static int mpol_new_bind(struct mempolicy *pol, const nodemask_t *nodes) in mpol_new_bind() argument [all …]
|
D | mmzone.c | 45 static inline int zref_in_nodemask(struct zoneref *zref, nodemask_t *nodes) in zref_in_nodemask() argument 48 return node_isset(zonelist_node_idx(zref), *nodes); in zref_in_nodemask() 57 nodemask_t *nodes, in next_zones_zonelist() argument 64 if (likely(nodes == NULL)) in next_zones_zonelist() 69 (z->zone && !zref_in_nodemask(z, nodes))) in next_zones_zonelist()
|
/linux-3.3/fs/btrfs/ |
D | ulist.c | 54 ulist->nodes = ulist->int_nodes; in ulist_init() 73 kfree(ulist->nodes); in ulist_fini() 152 if (ulist->nodes[i].val == val) in ulist_add() 166 old = ulist->nodes; in ulist_add() 177 ulist->nodes = new_nodes; in ulist_add() 180 ulist->nodes[ulist->nnodes].val = val; in ulist_add() 181 ulist->nodes[ulist->nnodes].aux = aux; in ulist_add() 212 return &ulist->nodes[0]; in ulist_next() 214 next = (prev - ulist->nodes) + 1; in ulist_next() 218 return &ulist->nodes[next]; in ulist_next()
|
D | tree-log.c | 331 u32 dst_size = btrfs_item_size_nr(path->nodes[0], in overwrite_item() 351 dst_ptr = btrfs_item_ptr_offset(path->nodes[0], path->slots[0]); in overwrite_item() 352 read_extent_buffer(path->nodes[0], dst_copy, dst_ptr, in overwrite_item() 379 found_size = btrfs_item_size_nr(path->nodes[0], in overwrite_item() 390 dst_ptr = btrfs_item_ptr_offset(path->nodes[0], in overwrite_item() 414 S_ISDIR(btrfs_inode_mode(path->nodes[0], dst_item))) { in overwrite_item() 416 saved_i_size = btrfs_inode_size(path->nodes[0], in overwrite_item() 421 copy_extent_buffer(path->nodes[0], eb, dst_ptr, in overwrite_item() 427 btrfs_set_inode_size(path->nodes[0], dst_item, saved_i_size); in overwrite_item() 434 if (btrfs_inode_generation(path->nodes[0], dst_item) == 0) { in overwrite_item() [all …]
|
D | inode-item.c | 34 leaf = path->nodes[0]; in find_name_in_backref() 116 leaf = path->nodes[0]; in btrfs_del_inode_ref() 167 old_size = btrfs_item_size_nr(path->nodes[0], path->slots[0]); in btrfs_insert_inode_ref() 169 ref = btrfs_item_ptr(path->nodes[0], path->slots[0], in btrfs_insert_inode_ref() 172 btrfs_set_inode_ref_name_len(path->nodes[0], ref, name_len); in btrfs_insert_inode_ref() 173 btrfs_set_inode_ref_index(path->nodes[0], ref, index); in btrfs_insert_inode_ref() 181 ref = btrfs_item_ptr(path->nodes[0], path->slots[0], in btrfs_insert_inode_ref() 183 btrfs_set_inode_ref_name_len(path->nodes[0], ref, name_len); in btrfs_insert_inode_ref() 184 btrfs_set_inode_ref_index(path->nodes[0], ref, index); in btrfs_insert_inode_ref() 187 write_extent_buffer(path->nodes[0], name, ptr, name_len); in btrfs_insert_inode_ref() [all …]
|
D | ctree.c | 57 if (!p->nodes[i] || !p->locks[i]) in btrfs_set_path_blocking() 59 btrfs_set_lock_blocking_rw(p->nodes[i], p->locks[i]); in btrfs_set_path_blocking() 98 if (p->nodes[i] && p->locks[i]) { in btrfs_clear_path_blocking() 99 btrfs_clear_lock_blocking_rw(p->nodes[i], p->locks[i]); in btrfs_clear_path_blocking() 134 if (!p->nodes[i]) in btrfs_release_path() 137 btrfs_tree_unlock_rw(p->nodes[i], p->locks[i]); in btrfs_release_path() 140 free_extent_buffer(p->nodes[i]); in btrfs_release_path() 141 p->nodes[i] = NULL; in btrfs_release_path() 912 mid = path->nodes[level]; in balance_level() 921 parent = path->nodes[level + 1]; in balance_level() [all …]
|
D | root-tree.c | 55 l = path->nodes[0]; in btrfs_find_last_root() 102 btrfs_print_leaf(root, path->nodes[0]); in btrfs_update_root() 109 l = path->nodes[0]; in btrfs_update_root() 113 btrfs_mark_buffer_dirty(path->nodes[0]); in btrfs_update_root() 158 leaf = path->nodes[0]; in btrfs_find_dead_roots() 165 leaf = path->nodes[0]; in btrfs_find_dead_roots() 236 leaf = path->nodes[0]; in btrfs_find_orphan_roots() 243 leaf = path->nodes[0]; in btrfs_find_orphan_roots() 295 leaf = path->nodes[0]; in btrfs_del_root() 329 leaf = path->nodes[0]; in btrfs_del_root_ref() [all …]
|
D | file-item.c | 63 leaf = path->nodes[0]; in btrfs_insert_file_extent() 103 leaf = path->nodes[0]; in btrfs_lookup_csum() 229 btrfs_item_key_to_cpu(path->nodes[0], &found_key, in __btrfs_lookup_bio_sums() 233 item_size = btrfs_item_size_nr(path->nodes[0], in __btrfs_lookup_bio_sums() 238 item = btrfs_item_ptr(path->nodes[0], path->slots[0], in __btrfs_lookup_bio_sums() 249 read_extent_buffer(path->nodes[0], &sum, in __btrfs_lookup_bio_sums() 311 leaf = path->nodes[0]; in btrfs_lookup_csums_range() 324 leaf = path->nodes[0]; in btrfs_lookup_csums_range() 331 leaf = path->nodes[0]; in btrfs_lookup_csums_range() 354 item = btrfs_item_ptr(path->nodes[0], path->slots[0], in btrfs_lookup_csums_range() [all …]
|
D | dir-item.c | 57 leaf = path->nodes[0]; in insert_with_overflow() 96 leaf = path->nodes[0]; in btrfs_insert_xattr_item() 108 btrfs_mark_buffer_dirty(path->nodes[0]); in btrfs_insert_xattr_item() 156 leaf = path->nodes[0]; in btrfs_insert_dir_item() 267 leaf = path->nodes[0]; in btrfs_search_dir_index_item() 277 leaf = path->nodes[0]; in btrfs_search_dir_index_item() 334 leaf = path->nodes[0]; in btrfs_match_dir_item_name() 372 leaf = path->nodes[0]; in btrfs_delete_one_dir_name()
|
D | tree-defrag.c | 107 if (!path->nodes[1]) { in btrfs_defrag_leaves() 111 path->slots[1] = btrfs_header_nritems(path->nodes[1]); in btrfs_defrag_leaves() 115 path->nodes[1], 0, in btrfs_defrag_leaves()
|
/linux-3.3/drivers/md/persistent-data/ |
D | dm-btree-spine.c | 131 s->nodes[0] = NULL; in init_ro_spine() 132 s->nodes[1] = NULL; in init_ro_spine() 140 int r2 = unlock_block(s->info, s->nodes[i]); in exit_ro_spine() 153 r = unlock_block(s->info, s->nodes[0]); in ro_step() 156 s->nodes[0] = s->nodes[1]; in ro_step() 160 r = bn_read_lock(s->info, new_child, s->nodes + s->count); in ro_step() 172 block = s->nodes[s->count - 1]; in ro_node() 190 int r2 = unlock_block(s->info, s->nodes[i]); in exit_shadow_spine() 204 r = unlock_block(s->info, s->nodes[0]); in shadow_step() 207 s->nodes[0] = s->nodes[1]; in shadow_step() [all …]
|
/linux-3.3/drivers/gpu/drm/nouveau/ |
D | nouveau_mm.c | 60 #define node(root, dir) ((root)->nl_entry.dir == &mm->nodes) ? NULL : \ 136 INIT_LIST_HEAD(&mm->nodes); in nouveau_mm_init() 148 list_add_tail(&node->nl_entry, &mm->nodes); in nouveau_mm_init() 158 list_first_entry(&mm->nodes, struct nouveau_mm_node, nl_entry); in nouveau_mm_fini() 159 int nodes = 0; in nouveau_mm_fini() local 161 list_for_each_entry(node, &mm->nodes, nl_entry) { in nouveau_mm_fini() 162 if (nodes++ == mm->heap_nodes) { in nouveau_mm_fini() 164 list_for_each_entry(node, &mm->nodes, nl_entry) { in nouveau_mm_fini()
|
/linux-3.3/arch/sparc/kernel/ |
D | cpumap.c | 53 struct cpuinfo_node nodes[0]; member 212 node = &new_tree->nodes[n]; in build_cpuinfo_tree() 253 node = &new_tree->nodes[level_rover[level]]; in build_cpuinfo_tree() 278 node = &new_tree->nodes[n]; in build_cpuinfo_tree() 300 struct cpuinfo_node *node = &t->nodes[node_index]; in increment_rover() 303 top_level = t->nodes[root_index].level; in increment_rover() 315 node = &t->nodes[node->parent_index]; in increment_rover() 336 for (level = t->nodes[root_index].level; level < CPUINFO_LVL_MAX; in iterate_cpu() 338 new_index = t->nodes[index].rover; in iterate_cpu() 364 for (i = 0; i < cpuinfo_tree->nodes[0].num_cpus; i++) in _cpu_map_rebuild() [all …]
|
/linux-3.3/Documentation/vm/ |
D | numa_memory_policy.txt | 13 which is an administrative mechanism for restricting the nodes from which 31 default policy will be set to interleave allocations across all nodes 140 optional set of nodes. The mode determines the behavior of the policy, 142 optional set of nodes can be viewed as the arguments to the policy 162 does not use the optional set of nodes. 164 It is an error for the set of nodes specified for this policy to 168 set of nodes specified by the policy. Memory will be allocated from 174 allocation fails, the kernel will search other nodes, in order of 193 interleaved, on a page granularity, across the nodes specified in 198 Interleave mode indexes the set of nodes specified by the policy [all …]
|
D | numa | 45 abstractions called "nodes". Linux maps the nodes onto the physical cells 47 architectures. As with physical cells, software nodes may contain 0 or more 49 "closer" nodes--nodes that map to closer cells--will generally experience 60 the emulation of additional nodes. For NUMA emulation, linux will carve up 61 the existing nodes--or the system memory for non-NUMA platforms--into multiple 62 nodes. Each emulated node will manage a fraction of the underlying cells' 72 an ordered "zonelist". A zonelist specifies the zones/nodes to visit when a 77 Because some nodes contain multiple zones containing different types of 93 nodes' zones in the selected zonelist looking for the first zone in the list 104 Thus, under sufficient imbalance, tasks can migrate between nodes, remote [all …]
|
D | hugetlbpage.txt | 91 over all the set of allowed nodes specified by the NUMA memory policy of the 92 task that modifies nr_hugepages. The default for the allowed nodes--when the 93 task has default memory policy--is all on-line nodes with memory. Allowed 94 nodes with insufficient available, contiguous memory for a huge page will be 102 some nodes in a NUMA system, it will attempt to make up the difference by 103 allocating extra pages on other nodes with sufficient available contiguous 131 across all nodes in the memory policy of the task modifying nr_hugepages. 132 Any free huge pages on the selected nodes will be freed back to the kernel's 171 nodes from which huge pages are allocated or freed are controlled by the 186 This will allocate or free abs(20 - nr_hugepages) to or from the nodes [all …]
|
/linux-3.3/arch/mips/sgi-ip27/ |
D | Kconfig | 9 The nodes of Origin, Onyx, Fuel and Tezro systems can be configured 10 in either N-Modes which allows for more nodes or M-Mode which allows 18 The nodes of Origin, Onyx, Fuel and Tezro systems can be configured 19 in either N-Modes which allows for more nodes or M-Mode which allows 39 nodes in a NUMA cluster. This trades memory for speed. 46 across multiple nodes in a NUMA cluster. This trades memory for
|
/linux-3.3/Documentation/devicetree/bindings/ |
D | marvell.txt | 6 system. In this section, we define device tree nodes to describe 47 The system-controller node contains child nodes for each system 70 2) Child nodes of /system-controller 102 of nodes. The first level describes an ethernet silicon block 103 and the second level describes up to 3 ethernet nodes within 107 shared register set, and the "ethernet" nodes describe ethernet 156 c) Marvell Discovery PHY nodes 175 d) Marvell Discovery SDMA nodes 198 e) Marvell Discovery BRG nodes 225 f) Marvell Discovery CUNIT nodes [all …]
|
/linux-3.3/Documentation/dvb/ |
D | udev.txt | 9 DVB device nodes are created automatically. 12 creating the DVB device nodes manually up to now due to the missing sysfs 16 device nodes manually. 31 1. You need to create a proper udev rule that will create the device nodes 41 If you want more control over the device nodes (for example a special group membership)
|
/linux-3.3/Documentation/x86/x86_64/ |
D | fake-numa-for-cpusets | 6 you can create fake NUMA nodes that represent contiguous chunks of memory and 14 configuring fake nodes, see Documentation/x86/x86_64/boot-options.txt. 20 you'll determine a better setup to minimize the number of nodes you have to deal 36 Documentation/cgroups/cpusets.txt, you can assign fake nodes (i.e. contiguous memory 46 Now this cpuset, 'ddset', will only allowed access to fake nodes 0 and 1 for 50 available to them according to the fake nodes assigned as mems:
|
/linux-3.3/arch/x86/kernel/cpu/ |
D | perf_event_amd_ibs.c | 136 int nodes; in setup_ibs_ctl() local 139 nodes = 0; in setup_ibs_ctl() 147 ++nodes; in setup_ibs_ctl() 159 if (!nodes) { in setup_ibs_ctl()
|
/linux-3.3/Documentation/filesystems/ |
D | ceph.txt | 10 * Seamless scaling from 1 to many thousands of nodes 12 * N-way replication of data across storage nodes 25 storage nodes run entirely as user space daemons. Storage nodes 28 across storage nodes in large chunks to distribute workload and 29 facilitate high throughputs. When storage nodes fail, data is 30 re-replicated in a distributed fashion by the storage nodes themselves 47 from a small cluster of just a few nodes to many hundreds, without 50 When the file system approaches full, new nodes can be easily added
|
/linux-3.3/fs/jffs2/ |
D | README.Locking | 26 When writing new nodes, the alloc_sem must be held until the new nodes 29 nodes to an inode may obsolete old ones, and by holding the alloc_sem 32 was written afterwards. Hence, we can ensure the newly-obsoleted nodes 79 (NB) the per-inode list of physical nodes. The latter is a special 87 Note that the per-inode list of physical nodes (f->nodes) is a special 88 case. Any changes to _valid_ nodes (i.e. ->flash_offset & 1 == 0) in 90 may remove _obsolete_ nodes from the list while holding only the 129 The latter function on NAND flash must read _obsolete_ nodes to
|
/linux-3.3/fs/dlm/ |
D | member.c | 526 if (rv->nodes[i].nodeid == nodeid) in find_config_node() 527 return &rv->nodes[i]; in find_config_node() 570 node = &rv->nodes[i]; in dlm_recover_members() 676 struct dlm_config_node *nodes; in dlm_ls_start() local 683 error = dlm_config_nodes(ls->ls_name, &nodes, &count); in dlm_ls_start() 698 rv->nodes = nodes; in dlm_ls_start() 708 kfree(rv_old->nodes); in dlm_ls_start() 717 kfree(nodes); in dlm_ls_start()
|
/linux-3.3/Documentation/devicetree/bindings/powerpc/fsl/ |
D | mpc5200.txt | 66 soc child nodes 68 Any on chip SOC devices available to Linux must appear as soc5200 child nodes. 73 Required soc5200 child nodes: 81 Recommended soc5200 child nodes; populate as needed for your board 102 fsl,mpc5200-gpt nodes 138 fsl,mpc5200-psc nodes 151 fsl,mpc5200-gpio and fsl,mpc5200-gpio-wkup nodes 158 fsl,mpc5200-fec nodes 179 The interrupts property for device nodes using the mpc5200 pic consists 196 fsl,mpc5200-mscan nodes
|