next_node 285 arch/x86/mm/numa_64.c rr = next_node(rr, node_online_map); next_node 232 fs/hfs/brec.c struct hfs_bnode *node, *new_node, *next_node; next_node 252 fs/hfs/brec.c next_node = hfs_bnode_find(tree, node->next); next_node 254 fs/hfs/brec.c next_node = NULL; next_node 256 fs/hfs/brec.c if (IS_ERR(next_node)) { next_node 259 fs/hfs/brec.c return next_node; next_node 275 fs/hfs/brec.c if (next_node) next_node 276 fs/hfs/brec.c hfs_bnode_put(next_node); next_node 331 fs/hfs/brec.c if (next_node) { next_node 332 fs/hfs/brec.c next_node->prev = new_node->this; next_node 333 fs/hfs/brec.c hfs_bnode_read(next_node, &node_desc, 0, sizeof(node_desc)); next_node 334 fs/hfs/brec.c node_desc.prev = cpu_to_be32(next_node->prev); next_node 335 fs/hfs/brec.c hfs_bnode_write(next_node, &node_desc, 0, sizeof(node_desc)); next_node 336 fs/hfs/brec.c hfs_bnode_put(next_node); next_node 208 fs/hfs/btree.c struct hfs_bnode *node, *next_node; next_node 278 fs/hfs/btree.c next_node = hfs_bmap_new_bmap(node, idx); next_node 280 fs/hfs/btree.c next_node = hfs_bnode_find(tree, nidx); next_node 282 fs/hfs/btree.c if (IS_ERR(next_node)) next_node 283 fs/hfs/btree.c return next_node; next_node 284 fs/hfs/btree.c node = next_node; next_node 306 fs/hfsplus/brec.c struct hfs_bnode *next_node = hfs_bnode_find(tree, new_node->next); next_node 307 fs/hfsplus/brec.c next_node->prev = new_node->this; next_node 308 fs/hfsplus/brec.c hfs_bnode_read(next_node, &node_desc, 0, sizeof(node_desc)); next_node 309 fs/hfsplus/brec.c node_desc.prev = cpu_to_be32(next_node->prev); next_node 310 fs/hfsplus/brec.c hfs_bnode_write(next_node, &node_desc, 0, sizeof(node_desc)); next_node 311 fs/hfsplus/brec.c hfs_bnode_put(next_node); next_node 184 fs/hfsplus/btree.c struct hfs_bnode *node, *next_node; next_node 254 fs/hfsplus/btree.c next_node = hfs_bmap_new_bmap(node, idx); next_node 256 fs/hfsplus/btree.c next_node = hfs_bnode_find(tree, nidx); next_node 258 fs/hfsplus/btree.c if (IS_ERR(next_node)) next_node 259 fs/hfsplus/btree.c return next_node; next_node 260 fs/hfsplus/btree.c node = next_node; next_node 1980 fs/ocfs2/alloc.c goto next_node; next_node 2001 fs/ocfs2/alloc.c next_node: next_node 2570 fs/ocfs2/alloc.c goto next_node; next_node 2588 fs/ocfs2/alloc.c next_node: next_node 353 include/linux/nodemask.h (node) = next_node((node), (mask))) next_node 408 include/linux/nodemask.h #define next_online_node(nid) next_node((nid), node_states[N_ONLINE]) next_node 2318 kernel/cpuset.c node = next_node(current->cpuset_mem_spread_rotor, current->mems_allowed); next_node 7033 kernel/sched.c int next_node = find_next_best_node(node, &used_nodes); next_node 7035 kernel/sched.c node_to_cpumask_ptr_next(nodemask, next_node); next_node 274 kernel/sched_fair.c struct rb_node *next_node; next_node 277 kernel/sched_fair.c next_node = rb_next(&se->run_node); next_node 278 kernel/sched_fair.c cfs_rq->rb_leftmost = next_node; next_node 280 kernel/sched_fair.c if (next_node) { next_node 281 kernel/sched_fair.c next = rb_entry(next_node, next_node 523 mm/hugetlb.c nid = next_node(nid, node_online_map); next_node 591 mm/hugetlb.c next_nid = next_node(h->hugetlb_next_nid, node_online_map); next_node 819 mm/hugetlb.c nid = next_node(nid, node_online_map); next_node 278 mm/mempolicy.c current->il_next = next_node(current->il_next, tmp); next_node 1365 mm/mempolicy.c next = next_node(nid, policy->v.nodes); next_node 1430 mm/mempolicy.c nid = next_node(nid, pol->v.nodes); next_node 904 mm/slab.c node = next_node(cpu_to_node(cpu), node_online_map); next_node 915 mm/slab.c node = next_node(node, node_online_map);