Lines Matching full:nodes

15  * interleave     Allocate memory interleaved over a set of nodes,
22 * bind Only allocate memory on a specific set of nodes,
26 * the allocation to memory nodes instead
176 int (*create)(struct mempolicy *pol, const nodemask_t *nodes);
177 void (*rebind)(struct mempolicy *pol, const nodemask_t *nodes);
193 static int mpol_new_interleave(struct mempolicy *pol, const nodemask_t *nodes) in mpol_new_interleave() argument
195 if (nodes_empty(*nodes)) in mpol_new_interleave()
197 pol->v.nodes = *nodes; in mpol_new_interleave()
201 static int mpol_new_preferred(struct mempolicy *pol, const nodemask_t *nodes) in mpol_new_preferred() argument
203 if (!nodes) in mpol_new_preferred()
205 else if (nodes_empty(*nodes)) in mpol_new_preferred()
206 return -EINVAL; /* no allowed nodes */ in mpol_new_preferred()
208 pol->v.preferred_node = first_node(*nodes); in mpol_new_preferred()
212 static int mpol_new_bind(struct mempolicy *pol, const nodemask_t *nodes) in mpol_new_bind() argument
214 if (nodes_empty(*nodes)) in mpol_new_bind()
216 pol->v.nodes = *nodes; in mpol_new_bind()
222 * any, for the new policy. mpol_new() has already validated the nodes
230 const nodemask_t *nodes, struct nodemask_scratch *nsc) in mpol_set_nodemask() argument
241 VM_BUG_ON(!nodes); in mpol_set_nodemask()
242 if (pol->mode == MPOL_PREFERRED && nodes_empty(*nodes)) in mpol_set_nodemask()
243 nodes = NULL; /* explicit local allocation */ in mpol_set_nodemask()
246 mpol_relative_nodemask(&nsc->mask2, nodes, &nsc->mask1); in mpol_set_nodemask()
248 nodes_and(nsc->mask2, *nodes, nsc->mask1); in mpol_set_nodemask()
251 pol->w.user_nodemask = *nodes; in mpol_set_nodemask()
257 if (nodes) in mpol_set_nodemask()
266 * initialization. You must invoke mpol_set_nodemask() to set nodes.
269 nodemask_t *nodes) in mpol_new() argument
273 pr_debug("setting mode %d flags %d nodes[0] %lx\n", in mpol_new()
274 mode, flags, nodes ? nodes_addr(*nodes)[0] : NUMA_NO_NODE); in mpol_new()
277 if (nodes && !nodes_empty(*nodes)) in mpol_new()
281 VM_BUG_ON(!nodes); in mpol_new()
289 if (nodes_empty(*nodes)) { in mpol_new()
295 if (!nodes_empty(*nodes) || in mpol_new()
300 } else if (nodes_empty(*nodes)) in mpol_new()
320 static void mpol_rebind_default(struct mempolicy *pol, const nodemask_t *nodes) in mpol_rebind_default() argument
324 static void mpol_rebind_nodemask(struct mempolicy *pol, const nodemask_t *nodes) in mpol_rebind_nodemask() argument
329 nodes_and(tmp, pol->w.user_nodemask, *nodes); in mpol_rebind_nodemask()
331 mpol_relative_nodemask(&tmp, &pol->w.user_nodemask, nodes); in mpol_rebind_nodemask()
333 nodes_remap(tmp, pol->v.nodes,pol->w.cpuset_mems_allowed, in mpol_rebind_nodemask()
334 *nodes); in mpol_rebind_nodemask()
335 pol->w.cpuset_mems_allowed = *nodes; in mpol_rebind_nodemask()
339 tmp = *nodes; in mpol_rebind_nodemask()
341 pol->v.nodes = tmp; in mpol_rebind_nodemask()
345 const nodemask_t *nodes) in mpol_rebind_preferred() argument
352 if (node_isset(node, *nodes)) { in mpol_rebind_preferred()
358 mpol_relative_nodemask(&tmp, &pol->w.user_nodemask, nodes); in mpol_rebind_preferred()
363 *nodes); in mpol_rebind_preferred()
364 pol->w.cpuset_mems_allowed = *nodes; in mpol_rebind_preferred()
369 * mpol_rebind_policy - Migrate a policy to a different set of nodes
729 * If pages found in a given range are on a set of nodes (determined by
730 * @nodes and @flags,) it's isolated and queued to the pagelist which is
743 nodemask_t *nodes, unsigned long flags, in queue_pages_range() argument
750 .nmask = nodes, in queue_pages_range()
863 nodemask_t *nodes) in do_set_mempolicy() argument
872 new = mpol_new(mode, flags, nodes); in do_set_mempolicy()
878 ret = mpol_set_nodemask(new, nodes, scratch); in do_set_mempolicy()
901 static void get_policy_nodemask(struct mempolicy *p, nodemask_t *nodes) in get_policy_nodemask() argument
903 nodes_clear(*nodes); in get_policy_nodemask()
910 *nodes = p->v.nodes; in get_policy_nodemask()
914 node_set(p->v.preferred_node, *nodes); in get_policy_nodemask()
1000 *policy = next_node_in(current->il_prev, pol->v.nodes); in do_get_mempolicy()
1138 * This lets us pick a pair of nodes to migrate between, such that in do_migrate_pages()
1167 * However if the number of source nodes is not equal to in do_migrate_pages()
1168 * the number of destination nodes we can not preserve in do_migrate_pages()
1188 /* dest not in remaining from nodes? */ in do_migrate_pages()
1312 pr_debug("mbind %lx-%lx mode:%d flags:%d nodes:%lx\n", in do_mbind()
1376 static int get_nodes(nodemask_t *nodes, const unsigned long __user *nmask, in get_nodes() argument
1385 nodes_clear(*nodes); in get_nodes()
1398 * When the user specified more nodes than supported just check in get_nodes()
1430 if (copy_from_user(nodes_addr(*nodes), nmask, nlongs*sizeof(unsigned long))) in get_nodes()
1432 nodes_addr(*nodes)[nlongs-1] &= endmask; in get_nodes() local
1438 nodemask_t *nodes) in copy_nodes_to_user() argument
1450 return copy_to_user(mask, nodes_addr(*nodes), copy) ? -EFAULT : 0; in copy_nodes_to_user()
1457 nodemask_t nodes; in kernel_mbind() local
1469 err = get_nodes(&nodes, nmask, maxnode); in kernel_mbind()
1472 return do_mbind(start, len, mode, mode_flags, &nodes, flags); in kernel_mbind()
1487 nodemask_t nodes; in kernel_set_mempolicy() local
1496 err = get_nodes(&nodes, nmask, maxnode); in kernel_set_mempolicy()
1499 return do_set_mempolicy(mode, flags, &nodes); in kernel_set_mempolicy()
1558 /* Is the user allowed to access the target nodes? */ in kernel_migrate_pages()
1613 nodemask_t nodes; in kernel_get_mempolicy() local
1620 err = do_get_mempolicy(&pval, &nodes, addr, flags); in kernel_get_mempolicy()
1629 err = copy_nodes_to_user(nmask, maxnode, &nodes); in kernel_get_mempolicy()
1856 * if policy->v.nodes has movable memory only, in apply_policy_zone()
1859 * policy->v.nodes is intersect with node_states[N_MEMORY]. in apply_policy_zone()
1861 * policy->v.nodes has movable memory only. in apply_policy_zone()
1863 if (!nodes_intersects(policy->v.nodes, node_states[N_HIGH_MEMORY])) in apply_policy_zone()
1870 * Return a nodemask representing a mempolicy for filtering nodes for
1878 cpuset_nodemask_valid_mems_allowed(&policy->v.nodes)) in policy_nodemask()
1879 return &policy->v.nodes; in policy_nodemask()
1907 next = next_node_in(me->il_prev, policy->v.nodes); in interleave_nodes()
1950 &policy->v.nodes); in mempolicy_slab_node()
1961 * node in pol->v.nodes (starting from n=0), wrapping around if n exceeds the
1962 * number of present nodes.
1966 unsigned nnodes = nodes_weight(pol->v.nodes); in offset_il_node()
1974 nid = first_node(pol->v.nodes); in offset_il_node()
1976 nid = next_node(nid, pol->v.nodes); in offset_il_node()
2032 *nodemask = &(*mpol)->v.nodes; in huge_node()
2074 *mask = mempolicy->v.nodes; in init_nodemask_of_mempolicy()
2112 * MPOL_PREFERRED and MPOL_F_LOCAL are only preferred nodes to in mempolicy_nodemask_intersects()
2113 * allocate from, they may fallback to other nodes when oom. in mempolicy_nodemask_intersects()
2115 * nodes in mask. in mempolicy_nodemask_intersects()
2120 ret = nodes_intersects(mempolicy->v.nodes, *mask); in mempolicy_nodemask_intersects()
2199 * node and don't fall back to other nodes, as the cost of in alloc_pages_vma()
2340 return !!nodes_equal(a->v.nodes, b->v.nodes); in __mpol_equal()
2495 * allows binding to multiple nodes. in mpol_misplaced()
2498 * If no allowed nodes, use current [!misplaced]. in mpol_misplaced()
2500 if (node_isset(curnid, pol->v.nodes)) in mpol_misplaced()
2505 &pol->v.nodes); in mpol_misplaced()
2708 npol ? nodes_addr(npol->v.nodes)[0] : NUMA_NO_NODE); in mpol_set_shared_policy()
2812 * enabled across suitably sized nodes (default is >= 16MB), or in numa_policy_init()
2878 nodemask_t nodes; in mpol_parse_str() local
2889 if (nodelist_parse(nodelist, nodes)) in mpol_parse_str()
2891 if (!nodes_subset(nodes, node_states[N_MEMORY])) in mpol_parse_str()
2894 nodes_clear(nodes); in mpol_parse_str()
2904 * we use first_node(nodes) to grab a single node, so here in mpol_parse_str()
2905 * nodelist (or nodes) cannot be empty. in mpol_parse_str()
2913 if (nodes_empty(nodes)) in mpol_parse_str()
2919 * Default to online nodes with memory if no nodelist in mpol_parse_str()
2922 nodes = node_states[N_MEMORY]; in mpol_parse_str()
2961 new = mpol_new(mode, mode_flags, &nodes); in mpol_parse_str()
2966 * Save nodes for mpol_to_str() to show the tmpfs mount options in mpol_parse_str()
2970 new->v.nodes = nodes; in mpol_parse_str()
2972 new->v.preferred_node = first_node(nodes); in mpol_parse_str()
2977 * Save nodes for contextualization: this will be used to "clone" in mpol_parse_str()
2980 new->w.user_nodemask = nodes; in mpol_parse_str()
3009 nodemask_t nodes = NODE_MASK_NONE; in mpol_to_str() local
3025 node_set(pol->v.preferred_node, nodes); in mpol_to_str()
3029 nodes = pol->v.nodes; in mpol_to_str()
3051 if (!nodes_empty(nodes)) in mpol_to_str()
3053 nodemask_pr_args(&nodes)); in mpol_to_str()