node 369 arch/x86/kernel/aperture_64.c int i, node;
node 381 arch/x86/kernel/aperture_64.c node = 0;
node 403 arch/x86/kernel/aperture_64.c node, aper_base, aper_size >> 20);
node 404 arch/x86/kernel/aperture_64.c node++;
node 186 arch/x86/kernel/cpu/amd.c int i, node;
node 189 arch/x86/kernel/cpu/amd.c node = apicid_to_node[i];
node 190 arch/x86/kernel/cpu/amd.c if (node != NUMA_NO_NODE && node_online(node))
node 191 arch/x86/kernel/cpu/amd.c return node;
node 194 arch/x86/kernel/cpu/amd.c node = apicid_to_node[i];
node 195 arch/x86/kernel/cpu/amd.c if (node != NUMA_NO_NODE && node_online(node))
node 196 arch/x86/kernel/cpu/amd.c return node;
node 224 arch/x86/kernel/cpu/amd.c int node;
node 227 arch/x86/kernel/cpu/amd.c node = c->phys_proc_id;
node 229 arch/x86/kernel/cpu/amd.c node = apicid_to_node[apicid];
node 230 arch/x86/kernel/cpu/amd.c if (!node_online(node)) {
node 245 arch/x86/kernel/cpu/amd.c node = apicid_to_node[ht_nodeid];
node 247 arch/x86/kernel/cpu/amd.c if (!node_online(node))
node 248 arch/x86/kernel/cpu/amd.c node = nearby_node(apicid);
node 250 arch/x86/kernel/cpu/amd.c numa_set_node(cpu, node);
node 252 arch/x86/kernel/cpu/amd.c printk(KERN_INFO "CPU %d/%x -> Node %d\n", cpu, apicid, node);
node 166 arch/x86/kernel/cpu/intel.c unsigned node;
node 172 arch/x86/kernel/cpu/intel.c node = apicid_to_node[apicid];
node 173 arch/x86/kernel/cpu/intel.c if (node == NUMA_NO_NODE || !node_online(node))
node 174 arch/x86/kernel/cpu/intel.c node = first_node(node_online_map);
node 175 arch/x86/kernel/cpu/intel.c numa_set_node(cpu, node);
node 177 arch/x86/kernel/cpu/intel.c printk(KERN_INFO "CPU %d/%x -> Node %d\n", cpu, apicid, node);
node 673 arch/x86/kernel/cpu/intel_cacheinfo.c for (i = 0; i <= node; i++) {
node 693 arch/x86/kernel/cpu/intel_cacheinfo.c int node = cpu_to_node(first_cpu(this_leaf->shared_cpu_map));
node 701 arch/x86/kernel/cpu/intel_cacheinfo.c dev = get_k8_northbridge(node);
node 727 arch/x86/kernel/cpu/intel_cacheinfo.c int node = cpu_to_node(first_cpu(this_leaf->shared_cpu_map));
node 744 arch/x86/kernel/cpu/intel_cacheinfo.c dev = get_k8_northbridge(node);
node 33 arch/x86/kernel/kdebugfs.c struct setup_data_node *node = file->private_data;
node 42 arch/x86/kernel/kdebugfs.c if (pos >= node->len)
node 45 arch/x86/kernel/kdebugfs.c if (count > node->len - pos)
node 46 arch/x86/kernel/kdebugfs.c count = node->len - pos;
node 47 arch/x86/kernel/kdebugfs.c pa = node->paddr + sizeof(struct setup_data) + pos;
node 95 arch/x86/kernel/kdebugfs.c type = debugfs_create_x32("type", S_IRUGO, d, &node->type);
node 100 arch/x86/kernel/kdebugfs.c data = debugfs_create_file("data", S_IRUGO, d, node, &fops_setup_data);
node 117 arch/x86/kernel/kdebugfs.c struct setup_data_node *node;
node 133 arch/x86/kernel/kdebugfs.c node = kmalloc(sizeof(*node), GFP_KERNEL);
node 134 arch/x86/kernel/kdebugfs.c if (!node) {
node 142 arch/x86/kernel/kdebugfs.c kfree(node);
node 150 arch/x86/kernel/kdebugfs.c node->paddr = pa_data;
node 151 arch/x86/kernel/kdebugfs.c node->type = data->type;
node 152 arch/x86/kernel/kdebugfs.c node->len = data->len;
node 153 arch/x86/kernel/kdebugfs.c error = create_setup_data_node(d, no, node);
node 679 arch/x86/kernel/kprobes.c struct hlist_node *node, *tmp;
node 707 arch/x86/kernel/kprobes.c hlist_for_each_entry_safe(ri, node, tmp, head, hlist) {
node 735 arch/x86/kernel/kprobes.c hlist_for_each_entry_safe(ri, node, tmp, &empty_rp, hlist) {
node 48 arch/x86/kernel/numaq_32.c int node;
node 54 arch/x86/kernel/numaq_32.c for_each_node(node) {
node 55 arch/x86/kernel/numaq_32.c if (scd->quads_present31_0 & (1 << node)) {
node 56 arch/x86/kernel/numaq_32.c node_set_online(node);
node 57 arch/x86/kernel/numaq_32.c eq = &scd->eq[node];
node 59 arch/x86/kernel/numaq_32.c node_start_pfn[node] = MB_TO_PAGES(
node 61 arch/x86/kernel/numaq_32.c node_end_pfn[node] = MB_TO_PAGES(
node 64 arch/x86/kernel/numaq_32.c e820_register_active_regions(node, node_start_pfn[node],
node 65 arch/x86/kernel/numaq_32.c node_end_pfn[node]);
node 66 arch/x86/kernel/numaq_32.c memory_present(node,
node 67 arch/x86/kernel/numaq_32.c node_start_pfn[node], node_end_pfn[node]);
node 68 arch/x86/kernel/numaq_32.c node_remap_size[node] = node_memmap_size_bytes(node,
node 69 arch/x86/kernel/numaq_32.c node_start_pfn[node],
node 70 arch/x86/kernel/numaq_32.c node_end_pfn[node]);
node 159 arch/x86/kernel/setup_percpu.c int node = early_cpu_to_node(cpu);
node 160 arch/x86/kernel/setup_percpu.c if (!node_online(node) || !NODE_DATA(node)) {
node 164 arch/x86/kernel/setup_percpu.c cpu, node);
node 170 arch/x86/kernel/setup_percpu.c ptr = alloc_bootmem_pages_node(NODE_DATA(node), size);
node 173 arch/x86/kernel/setup_percpu.c cpu, node, __pa(ptr));
node 203 arch/x86/kernel/setup_percpu.c unsigned int node, num = 0;
node 208 arch/x86/kernel/setup_percpu.c for_each_node_mask(node, node_possible_map)
node 209 arch/x86/kernel/setup_percpu.c num = node;
node 227 arch/x86/kernel/setup_percpu.c if (cpu_pda(cpu) && node != NUMA_NO_NODE)
node 228 arch/x86/kernel/setup_percpu.c cpu_pda(cpu)->nodenumber = node;
node 231 arch/x86/kernel/setup_percpu.c cpu_to_node_map[cpu] = node;
node 234 arch/x86/kernel/setup_percpu.c per_cpu(x86_cpu_to_node_map, cpu) = node;
node 264 arch/x86/kernel/setup_percpu.c int node = cpu_to_node(cpu);
node 274 arch/x86/kernel/setup_percpu.c mask = &node_to_cpumask_map[node];
node 282 arch/x86/kernel/setup_percpu.c enable? "numa_add_cpu":"numa_remove_cpu", cpu, node, buf);
node 337 arch/x86/kernel/setup_percpu.c node);
node 341 arch/x86/kernel/setup_percpu.c if (node >= nr_node_ids) {
node 344 arch/x86/kernel/setup_percpu.c node, nr_node_ids);
node 348 arch/x86/kernel/setup_percpu.c return &node_to_cpumask_map[node];
node 363 arch/x86/kernel/setup_percpu.c "node_to_cpumask(%d): no node_to_cpumask_map!\n", node);
node 367 arch/x86/kernel/setup_percpu.c if (node >= nr_node_ids) {
node 370 arch/x86/kernel/setup_percpu.c node, nr_node_ids);
node 374 arch/x86/kernel/setup_percpu.c return node_to_cpumask_map[node];
node 147 arch/x86/kernel/smpboot.c printk(KERN_INFO "Mapping cpu %d to node %d\n", cpu, node);
node 148 arch/x86/kernel/smpboot.c cpu_set(cpu, node_to_cpumask_map[node]);
node 149 arch/x86/kernel/smpboot.c cpu_to_node_map[cpu] = node;
node 155 arch/x86/kernel/smpboot.c int node;
node 158 arch/x86/kernel/smpboot.c for (node = 0; node < MAX_NUMNODES; node++)
node 159 arch/x86/kernel/smpboot.c cpu_clear(cpu, node_to_cpumask_map[node]);
node 177 arch/x86/kernel/smpboot.c int node = apicid_to_node(apicid);
node 179 arch/x86/kernel/smpboot.c if (!node_online(node))
node 180 arch/x86/kernel/smpboot.c node = first_online_node;
node 183 arch/x86/kernel/smpboot.c map_cpu_to_node(cpu, node);
node 776 arch/x86/kernel/smpboot.c int node = cpu_to_node(cpu);
node 782 arch/x86/kernel/smpboot.c newpda = kmalloc_node(size, GFP_ATOMIC, node);
node 785 arch/x86/kernel/smpboot.c "for CPU %d on node %d\n", cpu, node);
node 45 arch/x86/kernel/summit_32.c int twister = 0, node = 0;
node 61 arch/x86/kernel/summit_32.c node = scal_devs[i]->node_id;
node 101 arch/x86/kernel/summit_32.c mp_bus_id_to_node[bus] = node;
node 594 arch/x86/kernel/tlb_uv.c kmalloc_node(sizeof(struct bau_control), GFP_KERNEL, node);
node 599 arch/x86/kernel/tlb_uv.c DEST_Q_SIZE, GFP_KERNEL, node);
node 607 arch/x86/kernel/tlb_uv.c kmalloc_node(sizeof(int) * DEST_NUM_RESOURCES, GFP_KERNEL, node);
node 656 arch/x86/kernel/tlb_uv.c kmalloc_node(16384, GFP_KERNEL, node);
node 696 arch/x86/kernel/tlb_uv.c GFP_KERNEL, node);
node 730 arch/x86/kernel/tlb_uv.c bau_tablesp = uv_table_bases_init(blade, node);
node 732 arch/x86/kernel/tlb_uv.c adp = uv_activation_descriptor_init(node, pnode);
node 733 arch/x86/kernel/tlb_uv.c pqp = uv_payload_queue_init(node, pnode, bau_tablesp);
node 734 arch/x86/kernel/tlb_uv.c uv_table_bases_finish(blade, node, cur_cpu, bau_tablesp, adp);
node 754 arch/x86/kernel/tlb_uv.c int node;
node 767 arch/x86/kernel/tlb_uv.c for_each_online_node(node) {
node 768 arch/x86/kernel/tlb_uv.c blade = uv_node_to_blade_id(node);
node 779 arch/x86/kernel/tlb_uv.c for_each_online_node(node) {
node 780 arch/x86/kernel/tlb_uv.c blade = uv_node_to_blade_id(node);
node 784 arch/x86/kernel/tlb_uv.c uv_init_blade(blade, node, cur_cpu);
node 215 arch/x86/kernel/vsyscall_64.c if (node)
node 216 arch/x86/kernel/vsyscall_64.c *node = p >> 12;
node 254 arch/x86/kernel/vsyscall_64.c unsigned long node = 0;
node 256 arch/x86/kernel/vsyscall_64.c node = cpu_to_node(cpu);
node 259 arch/x86/kernel/vsyscall_64.c write_rdtscp_aux((node << 12) | cpu);
node 266 arch/x86/kernel/vsyscall_64.c d |= (node & 0xf) << 12;
node 267 arch/x86/kernel/vsyscall_64.c d |= (node >> 4) << 48;
node 791 arch/x86/kvm/mmu.c struct hlist_node *node;
node 809 arch/x86/kvm/mmu.c hlist_for_each_entry(pte_chain, node, &sp->parent_ptes, link) {
node 828 arch/x86/kvm/mmu.c struct hlist_node *node;
node 836 arch/x86/kvm/mmu.c hlist_for_each_entry(pte_chain, node, &sp->parent_ptes, link)
node 876 arch/x86/kvm/mmu.c struct hlist_node *node;
node 881 arch/x86/kvm/mmu.c hlist_for_each_entry(sp, node, bucket, hash_link)
node 904 arch/x86/kvm/mmu.c struct hlist_node *node;
node 920 arch/x86/kvm/mmu.c hlist_for_each_entry(sp, node, bucket, hash_link)
node 1069 arch/x86/kvm/mmu.c struct hlist_node *node, *n;
node 1076 arch/x86/kvm/mmu.c hlist_for_each_entry_safe(sp, node, n, bucket, hash_link)
node 1805 arch/x86/kvm/mmu.c struct hlist_node *node, *n;
node 1839 arch/x86/kvm/mmu.c hlist_for_each_entry_safe(sp, node, n, bucket, hash_link) {
node 2075 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp, *node;
node 2078 arch/x86/kvm/mmu.c list_for_each_entry_safe(sp, node, &kvm->arch.active_mmu_pages, link)
node 1207 arch/x86/kvm/vmx.c int node = cpu_to_node(cpu);
node 1211 arch/x86/kvm/vmx.c pages = alloc_pages_node(node, GFP_KERNEL, vmcs_config.order);
node 1155 arch/x86/mm/init_64.c pgd = vmemmap_pgd_populate(addr, node);
node 1159 arch/x86/mm/init_64.c pud = vmemmap_pud_populate(pgd, addr, node);
node 1165 arch/x86/mm/init_64.c pmd = vmemmap_pmd_populate(pud, addr, node);
node 1170 arch/x86/mm/init_64.c p = vmemmap_pte_populate(pmd, addr, node);
node 1184 arch/x86/mm/init_64.c p = vmemmap_alloc_block(PMD_SIZE, node);
node 1193 arch/x86/mm/init_64.c if (p_end != p || node_start != node) {
node 1198 arch/x86/mm/init_64.c node_start = node;
node 1205 arch/x86/mm/init_64.c vmemmap_verify((pte_t *)pmd, node, addr, next);
node 209 arch/x86/mm/numa_32.c int node;
node 211 arch/x86/mm/numa_32.c for_each_online_node(node) {
node 212 arch/x86/mm/numa_32.c printk(KERN_DEBUG "remap_numa_kva: node %d\n", node);
node 213 arch/x86/mm/numa_32.c for (pfn=0; pfn < node_remap_size[node]; pfn += PTRS_PER_PTE) {
node 214 arch/x86/mm/numa_32.c vaddr = node_remap_start_vaddr[node]+(pfn<<PAGE_SHIFT);
node 217 arch/x86/mm/numa_32.c node_remap_start_pfn[node] + pfn);
node 219 arch/x86/mm/numa_32.c node_remap_start_pfn[node] + pfn,
node 627 arch/x86/mm/numa_64.c int node;
node 632 arch/x86/mm/numa_64.c node = apicid_to_node[apicid];
node 633 arch/x86/mm/numa_64.c if (node == NUMA_NO_NODE)
node 635 arch/x86/mm/numa_64.c if (!node_online(node))
node 637 arch/x86/mm/numa_64.c numa_set_node(cpu, node);
node 122 arch/x86/mm/srat_64.c int pxm, node;
node 134 arch/x86/mm/srat_64.c node = setup_node(pxm);
node 135 arch/x86/mm/srat_64.c if (node < 0) {
node 145 arch/x86/mm/srat_64.c apicid_to_node[apic_id] = node;
node 148 arch/x86/mm/srat_64.c pxm, apic_id, node);
node 169 arch/x86/mm/srat_64.c struct bootnode *nd = &nodes_add[node];
node 190 arch/x86/mm/srat_64.c if (!hotadd_enough_memory(&nodes_add[node])) {
node 227 arch/x86/mm/srat_64.c int node, pxm;
node 244 arch/x86/mm/srat_64.c node = setup_node(pxm);
node 245 arch/x86/mm/srat_64.c if (node < 0) {
node 251 arch/x86/mm/srat_64.c if (i == node) {
node 263 arch/x86/mm/srat_64.c nd = &nodes[node];
node 265 arch/x86/mm/srat_64.c if (!node_test_and_set(node, nodes_parsed)) {
node 275 arch/x86/mm/srat_64.c printk(KERN_INFO "SRAT: Node %u PXM %u %lx-%lx\n", node, pxm,
node 277 arch/x86/mm/srat_64.c e820_register_active_regions(node, start >> PAGE_SHIFT,
node 279 arch/x86/mm/srat_64.c push_node_boundaries(node, nd->start >> PAGE_SHIFT,
node 283 arch/x86/mm/srat_64.c (reserve_hotadd(node, start, end) < 0)) {
node 288 arch/x86/mm/srat_64.c node_clear(node, nodes_parsed);
node 293 arch/x86/mm/srat_64.c memblk_nodeid[num_node_memblks] = node;
node 329 arch/x86/mm/srat_64.c node_clear(node, nodes_parsed);
node 331 arch/x86/mm/srat_64.c if (apicid_to_node[i] == node)
node 386 arch/x86/mm/srat_64.c int node = early_cpu_to_node(i);
node 388 arch/x86/mm/srat_64.c if (node == NUMA_NO_NODE)
node 390 arch/x86/mm/srat_64.c if (!node_isset(node, node_possible_map))
node 155 arch/x86/pci/acpi.c int node;
node 166 arch/x86/pci/acpi.c node = -1;
node 170 arch/x86/pci/acpi.c node = pxm_to_node(pxm);
node 171 arch/x86/pci/acpi.c if (node != -1)
node 172 arch/x86/pci/acpi.c set_mp_bus_to_node(busnum, node);
node 175 arch/x86/pci/acpi.c node = get_mp_bus_to_node(busnum);
node 177 arch/x86/pci/acpi.c if (node != -1 && !node_online(node))
node 178 arch/x86/pci/acpi.c node = -1;
node 191 arch/x86/pci/acpi.c sd->node = node;
node 210 arch/x86/pci/acpi.c if (bus && node != -1) {
node 214 arch/x86/pci/acpi.c busnum, pxm, node);
node 217 arch/x86/pci/acpi.c busnum, node);
node 29 arch/x86/pci/amd_bus.c mp_bus_to_node[busnum] = node;
node 34 arch/x86/pci/amd_bus.c int node = -1;
node 37 arch/x86/pci/amd_bus.c return node;
node 39 arch/x86/pci/amd_bus.c node = mp_bus_to_node[busnum];
node 45 arch/x86/pci/amd_bus.c if (node != -1 && !node_online(node))
node 46 arch/x86/pci/amd_bus.c node = -1;
node 48 arch/x86/pci/amd_bus.c return node;
node 58 arch/x86/pci/amd_bus.c mp_bus_to_node[busnum] = (unsigned char) node;
node 63 arch/x86/pci/amd_bus.c int node;
node 67 arch/x86/pci/amd_bus.c node = mp_bus_to_node[busnum];
node 68 arch/x86/pci/amd_bus.c return node;
node 88 arch/x86/pci/amd_bus.c int node;
node 283 arch/x86/pci/amd_bus.c int node;
node 338 arch/x86/pci/amd_bus.c node = (reg >> 4) & 0x07;
node 341 arch/x86/pci/amd_bus.c mp_bus_to_node[j] = (unsigned char) node;
node 348 arch/x86/pci/amd_bus.c info->node = node;
node 370 arch/x86/pci/amd_bus.c node = reg & 0x07;
node 377 arch/x86/pci/amd_bus.c if (info->node == node && info->link == link)
node 385 arch/x86/pci/amd_bus.c node, link, (u64)start, (u64)end);
node 397 arch/x86/pci/amd_bus.c if (info->node == def_node && info->link == def_link)
node 440 arch/x86/pci/amd_bus.c node = reg & 0x07;
node 449 arch/x86/pci/amd_bus.c if (info->node == node && info->link == link)
node 458 arch/x86/pci/amd_bus.c node, link, (u64)start, (u64)end);
node 523 arch/x86/pci/amd_bus.c if (info->node == def_node && info->link == def_link)
node 546 arch/x86/pci/amd_bus.c info->bus_min, info->bus_max, info->node, info->link);
node 395 arch/x86/pci/common.c sd->node = get_mp_bus_to_node(busnum);
node 564 arch/x86/pci/common.c sd->node = node;
node 122 arch/x86/pci/i386.c list_for_each_entry(bus, bus_list, node) {
node 141 arch/x86/pci/irq.c int node;
node 144 arch/x86/pci/irq.c node = get_mp_bus_to_node(i);
node 145 arch/x86/pci/irq.c if (pci_scan_bus_on_node(i, &pci_root_ops, node))
node 15 arch/x86/pci/legacy.c long node;
node 25 arch/x86/pci/legacy.c node = get_mp_bus_to_node(n);
node 31 arch/x86/pci/legacy.c pci_scan_bus_on_node(n, &pci_root_ops, node);
node 30 arch/x86/vdso/vgetcpu.c if (node)
node 31 arch/x86/vdso/vgetcpu.c *node = p >> 12;
node 229 block/as-iosched.c struct io_context *ioc = get_io_context(GFP_ATOMIC, node);
node 1196 block/as-iosched.c rq->elevator_private = as_get_io_context(q->node);
node 1333 block/as-iosched.c ioc = as_get_io_context(q->node);
node 1363 block/as-iosched.c ad = kmalloc_node(sizeof(*ad), GFP_KERNEL | __GFP_ZERO, q->node);
node 469 block/blk-core.c mempool_free_slab, request_cachep, q->node);
node 561 block/blk-core.c q->node = node_id;
node 734 block/blk-core.c ioc = current_io_context(GFP_ATOMIC, q->node);
node 849 block/blk-core.c ioc = current_io_context(GFP_NOIO, q->node);
node 91 block/blk-ioc.c ret = kmem_cache_alloc_node(iocontext_cachep, gfp_flags, node);
node 126 block/blk-ioc.c ret = alloc_io_context(gfp_flags, node);
node 151 block/blk-ioc.c ret = current_io_context(gfp_flags, node);
node 1341 block/cfq-iosched.c cfqd->queue->node);
node 1457 block/cfq-iosched.c cfqd->queue->node);
node 1463 block/cfq-iosched.c cfqd->queue->node);
node 1659 block/cfq-iosched.c ioc = get_io_context(gfp_mask, cfqd->queue->node);
node 2214 block/cfq-iosched.c cfqd = kmalloc_node(sizeof(*cfqd), GFP_KERNEL | __GFP_ZERO, q->node);
node 68 block/deadline-iosched.c struct rb_node *node = rb_next(&rq->rb_node);
node 70 block/deadline-iosched.c if (node)
node 71 block/deadline-iosched.c return rb_entry_rq(node);
node 354 block/deadline-iosched.c dd = kmalloc_node(sizeof(*dd), GFP_KERNEL | __GFP_ZERO, q->node);
node 211 block/elevator.c eq = kmalloc_node(sizeof(elevator_t), GFP_KERNEL | __GFP_ZERO, q->node);
node 221 block/elevator.c GFP_KERNEL, q->node);
node 72 block/noop-iosched.c nd = kmalloc_node(sizeof(*nd), GFP_KERNEL, q->node);
node 69 crypto/async_tx/async_tx.c list_for_each_entry_rcu(ref, &async_tx_master_list, node)
node 163 crypto/async_tx/async_tx.c INIT_LIST_HEAD(&ref->node);
node 183 crypto/async_tx/async_tx.c list_for_each_entry_rcu(ref, &async_tx_master_list, node)
node 260 crypto/async_tx/async_tx.c list_for_each_entry_rcu(ref, &async_tx_master_list, node)
node 282 crypto/async_tx/async_tx.c list_add_tail_rcu(&master_ref->node,
node 299 crypto/async_tx/async_tx.c list_for_each_entry(ref, &async_tx_master_list, node)
node 303 crypto/async_tx/async_tx.c list_del_rcu(&ref->node);
node 480 fs/afs/cmservice.c r->node[loop] = ntohl(b[loop + 5]);
node 547 fs/afs/cmservice.c reply.ia.uuid[loop + 5] = htonl((s8) afs_uuid.node[loop]);
node 426 fs/afs/internal.h u8 node[6]; /* spatially unique node ID (MAC addr) */
node 58 fs/afs/main.c ret = afs_get_MAC_address(afs_uuid.node, sizeof(afs_uuid.node));
node 83 fs/afs/main.c afs_uuid.node[0], afs_uuid.node[1], afs_uuid.node[2],
node 84 fs/afs/main.c afs_uuid.node[3], afs_uuid.node[4], afs_uuid.node[5]);
node 27 fs/afs/vnode.c if (!node)
node 30 fs/afs/vnode.c if (node->rb_left)
node 31 fs/afs/vnode.c bad = dump_tree_aux(node->rb_left, node, depth + 2, '/');
node 33 fs/afs/vnode.c vnode = rb_entry(node, struct afs_vnode, cb_promise);
node 35 fs/afs/vnode.c rb_is_red(node) ? 'R' : 'B',
node 38 fs/afs/vnode.c if (rb_parent(node) != parent) {
node 39 fs/afs/vnode.c printk("BAD: %p != %p\n", rb_parent(node), parent);
node 43 fs/afs/vnode.c if (node->rb_right)
node 44 fs/afs/vnode.c bad |= dump_tree_aux(node->rb_right, node, depth + 2, '\\');
node 200 fs/befs/btree.c if (node->bh)
node 201 fs/befs/btree.c brelse(node->bh);
node 203 fs/befs/btree.c node->bh = befs_read_datastream(sb, ds, node_off, &off);
node 204 fs/befs/btree.c if (!node->bh) {
node 211 fs/befs/btree.c node->od_node =
node 212 fs/befs/btree.c (befs_btree_nodehead *) ((void *) node->bh->b_data + off);
node 214 fs/befs/btree.c befs_dump_index_node(sb, node->od_node);
node 216 fs/befs/btree.c node->head.left = fs64_to_cpu(sb, node->od_node->left);
node 217 fs/befs/btree.c node->head.right = fs64_to_cpu(sb, node->od_node->right);
node 218 fs/befs/btree.c node->head.overflow = fs64_to_cpu(sb, node->od_node->overflow);
node 219 fs/befs/btree.c node->head.all_key_count =
node 220 fs/befs/btree.c fs16_to_cpu(sb, node->od_node->all_key_count);
node 221 fs/befs/btree.c node->head.all_key_length =
node 222 fs/befs/btree.c fs16_to_cpu(sb, node->od_node->all_key_length);
node 353 fs/befs/btree.c last = node->head.all_key_count - 1;
node 354 fs/befs/btree.c thiskey = befs_bt_get_key(sb, node, last, &keylen);
node 362 fs/befs/btree.c valarray = befs_bt_valarray(node);
node 371 fs/befs/btree.c thiskey = befs_bt_get_key(sb, node, mid, &keylen);
node 605 fs/befs/btree.c if (node->head.overflow == befs_bt_inval)
node 629 fs/befs/btree.c (sizeof (befs_btree_nodehead) + node->head.all_key_length);
node 635 fs/befs/btree.c return (fs16 *) ((void *) node->od_node + off);
node 648 fs/befs/btree.c void *keylen_index_start = (void *) befs_bt_keylen_index(node);
node 649 fs/befs/btree.c size_t keylen_index_size = node->head.all_key_count * sizeof (fs16);
node 664 fs/befs/btree.c return (char *) ((void *) node->od_node + sizeof (befs_btree_nodehead));
node 685 fs/befs/btree.c if (index < 0 || index > node->head.all_key_count) {
node 690 fs/befs/btree.c keystart = befs_bt_keydata(node);
node 691 fs/befs/btree.c keylen_index = befs_bt_keylen_index(node);
node 272 fs/befs/debug.c befs_debug(sb, " left %016LX", fs64_to_cpu(sb, node->left));
node 273 fs/befs/debug.c befs_debug(sb, " right %016LX", fs64_to_cpu(sb, node->right));
node 274 fs/befs/debug.c befs_debug(sb, " overflow %016LX", fs64_to_cpu(sb, node->overflow));
node 276 fs/befs/debug.c fs16_to_cpu(sb, node->all_key_count));
node 278 fs/befs/debug.c fs16_to_cpu(sb, node->all_key_length));
node 3981 fs/cifs/cifssmb.c struct dfs_info3_param *node = (*target_nodes)+i;
node 3983 fs/cifs/cifssmb.c node->flags = le16_to_cpu(pSMBr->DFSFlags);
node 3984 fs/cifs/cifssmb.c node->path_consumed = le16_to_cpu(pSMBr->PathConsumed);
node 3985 fs/cifs/cifssmb.c node->server_type = le16_to_cpu(ref->ServerType);
node 3986 fs/cifs/cifssmb.c node->ref_flag = le16_to_cpu(ref->ReferralEntryFlags);
node 3991 fs/cifs/cifssmb.c rc = cifs_strncpy_to_host(&(node->path_name), temp,
node 3999 fs/cifs/cifssmb.c rc = cifs_strncpy_to_host(&(node->node_name), temp,
node 1375 fs/dcache.c struct hlist_node *node;
node 1380 fs/dcache.c hlist_for_each_entry_rcu(dentry, node, head, d_hash) {
node 83 fs/dlm/dir.c uint32_t node, n = 0;
node 92 fs/dlm/dir.c node = (hash >> 16) % ls->ls_total_weight;
node 93 fs/dlm/dir.c nodeid = ls->ls_node_array[node];
node 99 fs/dlm/dir.c node = (hash >> 16) % ls->ls_num_nodes;
node 102 fs/dlm/dir.c if (n++ != node)
node 109 fs/dlm/dir.c ls->ls_num_nodes, n, node););
node 243 fs/dquot.c struct hlist_node *node;
node 246 fs/dquot.c hlist_for_each (node, dquot_hash+hashent) {
node 247 fs/dquot.c dquot = hlist_entry(node, struct dquot, dq_hash);
node 398 fs/ecryptfs/ecryptfs_kernel.h struct list_head node;
node 65 fs/ecryptfs/messaging.c *msg_ctx = list_entry(p, struct ecryptfs_msg_ctx, node);
node 85 fs/ecryptfs/messaging.c list_move(&msg_ctx->node, &ecryptfs_msg_ctx_alloc_list);
node 98 fs/ecryptfs/messaging.c list_move(&(msg_ctx->node), &ecryptfs_msg_ctx_free_list);
node 628 fs/ecryptfs/messaging.c INIT_LIST_HEAD(&ecryptfs_msg_ctx_arr[i].node);
node 637 fs/ecryptfs/messaging.c list_add_tail(&ecryptfs_msg_ctx_arr[i].node,
node 343 fs/ext2/balloc.c struct rb_node *node = &rsv->rsv_node;
node 365 fs/ext2/balloc.c rb_link_node(node, parent, p);
node 366 fs/ext2/balloc.c rb_insert_color(node, root);
node 330 fs/ext3/balloc.c struct rb_node *node = &rsv->rsv_node;
node 352 fs/ext3/balloc.c rb_link_node(node, parent, p);
node 353 fs/ext3/balloc.c rb_insert_color(node, root);
node 21 fs/hfs/bnode.c off += node->page_offset;
node 22 fs/hfs/bnode.c page = node->page[0];
node 32 fs/hfs/bnode.c hfs_bnode_read(node, &data, off, 2);
node 40 fs/hfs/bnode.c hfs_bnode_read(node, &data, off, 1);
node 49 fs/hfs/bnode.c tree = node->tree;
node 50 fs/hfs/bnode.c if (node->type == HFS_NODE_LEAF ||
node 52 fs/hfs/bnode.c key_len = hfs_bnode_read_u8(node, off) + 1;
node 56 fs/hfs/bnode.c hfs_bnode_read(node, key, off, key_len);
node 63 fs/hfs/bnode.c off += node->page_offset;
node 64 fs/hfs/bnode.c page = node->page[0];
node 75 fs/hfs/bnode.c hfs_bnode_write(node, &v, off, 2);
node 81 fs/hfs/bnode.c hfs_bnode_write(node, &data, off, 1);
node 88 fs/hfs/bnode.c off += node->page_offset;
node 89 fs/hfs/bnode.c page = node->page[0];
node 125 fs/hfs/bnode.c src += node->page_offset;
node 126 fs/hfs/bnode.c dst += node->page_offset;
node 127 fs/hfs/bnode.c page = node->page[0];
node 140 fs/hfs/bnode.c dprint(DBG_BNODE_MOD, "bnode: %d\n", node->this);
node 141 fs/hfs/bnode.c hfs_bnode_read(node, &desc, 0, sizeof(desc));
node 146 fs/hfs/bnode.c off = node->tree->node_size - 2;
node 148 fs/hfs/bnode.c key_off = hfs_bnode_read_u16(node, off);
node 150 fs/hfs/bnode.c if (i && node->type == HFS_NODE_INDEX) {
node 153 fs/hfs/bnode.c if (node->tree->attributes & HFS_TREE_VARIDXKEYS)
node 154 fs/hfs/bnode.c tmp = (hfs_bnode_read_u8(node, key_off) | 1) + 1;
node 156 fs/hfs/bnode.c tmp = node->tree->max_key_len + 1;
node 157 fs/hfs/bnode.c dprint(DBG_BNODE_MOD, " (%d,%d", tmp, hfs_bnode_read_u8(node, key_off));
node 158 fs/hfs/bnode.c hfs_bnode_read(node, &cnid, key_off + tmp, 4);
node 160 fs/hfs/bnode.c } else if (i && node->type == HFS_NODE_LEAF) {
node 163 fs/hfs/bnode.c tmp = hfs_bnode_read_u8(node, key_off);
node 176 fs/hfs/bnode.c tree = node->tree;
node 177 fs/hfs/bnode.c if (node->prev) {
node 178 fs/hfs/bnode.c tmp = hfs_bnode_find(tree, node->prev);
node 181 fs/hfs/bnode.c tmp->next = node->next;
node 185 fs/hfs/bnode.c } else if (node->type == HFS_NODE_LEAF)
node 186 fs/hfs/bnode.c tree->leaf_head = node->next;
node 188 fs/hfs/bnode.c if (node->next) {
node 189 fs/hfs/bnode.c tmp = hfs_bnode_find(tree, node->next);
node 192 fs/hfs/bnode.c tmp->prev = node->prev;
node 196 fs/hfs/bnode.c } else if (node->type == HFS_NODE_LEAF)
node 197 fs/hfs/bnode.c tree->leaf_tail = node->prev;
node 200 fs/hfs/bnode.c if (!node->prev && !node->next) {
node 203 fs/hfs/bnode.c if (!node->parent) {
node 207 fs/hfs/bnode.c set_bit(HFS_BNODE_DELETED, &node->flags);
node 219 fs/hfs/bnode.c struct hfs_bnode *node;
node 226 fs/hfs/bnode.c for (node = tree->node_hash[hfs_bnode_hash(cnid)];
node 227 fs/hfs/bnode.c node; node = node->next_hash) {
node 228 fs/hfs/bnode.c if (node->this == cnid) {
node 229 fs/hfs/bnode.c return node;
node 238 fs/hfs/bnode.c struct hfs_bnode *node, *node2;
node 252 fs/hfs/bnode.c node = kzalloc(size, GFP_KERNEL);
node 253 fs/hfs/bnode.c if (!node)
node 255 fs/hfs/bnode.c node->tree = tree;
node 256 fs/hfs/bnode.c node->this = cnid;
node 257 fs/hfs/bnode.c set_bit(HFS_BNODE_NEW, &node->flags);
node 258 fs/hfs/bnode.c atomic_set(&node->refcnt, 1);
node 260 fs/hfs/bnode.c node->tree->cnid, node->this);
node 261 fs/hfs/bnode.c init_waitqueue_head(&node->lock_wq);
node 266 fs/hfs/bnode.c node->next_hash = tree->node_hash[hash];
node 267 fs/hfs/bnode.c tree->node_hash[hash] = node;
node 271 fs/hfs/bnode.c kfree(node);
node 280 fs/hfs/bnode.c node->page_offset = off & ~PAGE_CACHE_MASK;
node 290 fs/hfs/bnode.c node->page[i] = page;
node 293 fs/hfs/bnode.c return node;
node 295 fs/hfs/bnode.c set_bit(HFS_BNODE_ERROR, &node->flags);
node 296 fs/hfs/bnode.c return node;
node 304 fs/hfs/bnode.c node->tree->cnid, node->this, atomic_read(&node->refcnt));
node 305 fs/hfs/bnode.c for (p = &node->tree->node_hash[hfs_bnode_hash(node->this)];
node 306 fs/hfs/bnode.c *p && *p != node; p = &(*p)->next_hash)
node 309 fs/hfs/bnode.c *p = node->next_hash;
node 310 fs/hfs/bnode.c node->tree->node_hash_cnt--;
node 316 fs/hfs/bnode.c struct hfs_bnode *node;
node 322 fs/hfs/bnode.c node = hfs_bnode_findhash(tree, num);
node 323 fs/hfs/bnode.c if (node) {
node 324 fs/hfs/bnode.c hfs_bnode_get(node);
node 326 fs/hfs/bnode.c wait_event(node->lock_wq, !test_bit(HFS_BNODE_NEW, &node->flags));
node 327 fs/hfs/bnode.c if (test_bit(HFS_BNODE_ERROR, &node->flags))
node 329 fs/hfs/bnode.c return node;
node 332 fs/hfs/bnode.c node = __hfs_bnode_create(tree, num);
node 333 fs/hfs/bnode.c if (!node)
node 335 fs/hfs/bnode.c if (test_bit(HFS_BNODE_ERROR, &node->flags))
node 337 fs/hfs/bnode.c if (!test_bit(HFS_BNODE_NEW, &node->flags))
node 338 fs/hfs/bnode.c return node;
node 340 fs/hfs/bnode.c desc = (struct hfs_bnode_desc *)(kmap(node->page[0]) + node->page_offset);
node 341 fs/hfs/bnode.c node->prev = be32_to_cpu(desc->prev);
node 342 fs/hfs/bnode.c node->next = be32_to_cpu(desc->next);
node 343 fs/hfs/bnode.c node->num_recs = be16_to_cpu(desc->num_recs);
node 344 fs/hfs/bnode.c node->type = desc->type;
node 345 fs/hfs/bnode.c node->height = desc->height;
node 346 fs/hfs/bnode.c kunmap(node->page[0]);
node 348 fs/hfs/bnode.c switch (node->type) {
node 351 fs/hfs/bnode.c if (node->height != 0)
node 355 fs/hfs/bnode.c if (node->height != 1)
node 359 fs/hfs/bnode.c if (node->height <= 1 || node->height > tree->depth)
node 367 fs/hfs/bnode.c off = hfs_bnode_read_u16(node, rec_off);
node 370 fs/hfs/bnode.c for (i = 1; i <= node->num_recs; off = next_off, i++) {
node 372 fs/hfs/bnode.c next_off = hfs_bnode_read_u16(node, rec_off);
node 378 fs/hfs/bnode.c if (node->type != HFS_NODE_INDEX &&
node 379 fs/hfs/bnode.c node->type != HFS_NODE_LEAF)
node 381 fs/hfs/bnode.c key_size = hfs_bnode_read_u8(node, off) + 1;
node 385 fs/hfs/bnode.c clear_bit(HFS_BNODE_NEW, &node->flags);
node 386 fs/hfs/bnode.c wake_up(&node->lock_wq);
node 387 fs/hfs/bnode.c return node;
node 390 fs/hfs/bnode.c set_bit(HFS_BNODE_ERROR, &node->flags);
node 391 fs/hfs/bnode.c clear_bit(HFS_BNODE_NEW, &node->flags);
node 392 fs/hfs/bnode.c wake_up(&node->lock_wq);
node 393 fs/hfs/bnode.c hfs_bnode_put(node);
node 404 fs/hfs/bnode.c kfree(node);
node 409 fs/hfs/bnode.c struct hfs_bnode *node;
node 414 fs/hfs/bnode.c node = hfs_bnode_findhash(tree, num);
node 416 fs/hfs/bnode.c BUG_ON(node);
node 417 fs/hfs/bnode.c node = __hfs_bnode_create(tree, num);
node 418 fs/hfs/bnode.c if (!node)
node 420 fs/hfs/bnode.c if (test_bit(HFS_BNODE_ERROR, &node->flags)) {
node 421 fs/hfs/bnode.c hfs_bnode_put(node);
node 425 fs/hfs/bnode.c pagep = node->page;
node 426 fs/hfs/bnode.c memset(kmap(*pagep) + node->page_offset, 0,
node 435 fs/hfs/bnode.c clear_bit(HFS_BNODE_NEW, &node->flags);
node 436 fs/hfs/bnode.c wake_up(&node->lock_wq);
node 438 fs/hfs/bnode.c return node;
node 443 fs/hfs/bnode.c if (node) {
node 444 fs/hfs/bnode.c atomic_inc(&node->refcnt);
node 446 fs/hfs/bnode.c node->tree->cnid, node->this, atomic_read(&node->refcnt));
node 453 fs/hfs/bnode.c if (node) {
node 454 fs/hfs/bnode.c struct hfs_btree *tree = node->tree;
node 458 fs/hfs/bnode.c node->tree->cnid, node->this, atomic_read(&node->refcnt));
node 459 fs/hfs/bnode.c BUG_ON(!atomic_read(&node->refcnt));
node 460 fs/hfs/bnode.c if (!atomic_dec_and_lock(&node->refcnt, &tree->hash_lock))
node 463 fs/hfs/bnode.c if (!node->page[i])
node 465 fs/hfs/bnode.c mark_page_accessed(node->page[i]);
node 468 fs/hfs/bnode.c if (test_bit(HFS_BNODE_DELETED, &node->flags)) {
node 469 fs/hfs/bnode.c hfs_bnode_unhash(node);
node 471 fs/hfs/bnode.c hfs_bmap_free(node);
node 472 fs/hfs/bnode.c hfs_bnode_free(node);
node 23 fs/hfs/brec.c dataoff = node->tree->node_size - (rec + 2) * 2;
node 24 fs/hfs/brec.c hfs_bnode_read(node, retval, dataoff, 4);
node 34 fs/hfs/brec.c if (node->type != HFS_NODE_INDEX && node->type != HFS_NODE_LEAF)
node 37 fs/hfs/brec.c if ((node->type == HFS_NODE_INDEX) &&
node 38 fs/hfs/brec.c !(node->tree->attributes & HFS_TREE_VARIDXKEYS)) {
node 39 fs/hfs/brec.c if (node->tree->attributes & HFS_TREE_BIGKEYS)
node 40 fs/hfs/brec.c retval = node->tree->max_key_len + 2;
node 42 fs/hfs/brec.c retval = node->tree->max_key_len + 1;
node 44 fs/hfs/brec.c recoff = hfs_bnode_read_u16(node, node->tree->node_size - (rec + 1) * 2);
node 47 fs/hfs/brec.c if (node->tree->attributes & HFS_TREE_BIGKEYS) {
node 48 fs/hfs/brec.c retval = hfs_bnode_read_u16(node, recoff) + 2;
node 49 fs/hfs/brec.c if (retval > node->tree->max_key_len + 2) {
node 55 fs/hfs/brec.c retval = (hfs_bnode_read_u8(node, recoff) | 1) + 1;
node 56 fs/hfs/brec.c if (retval > node->tree->max_key_len + 1) {
node 69 fs/hfs/brec.c struct hfs_bnode *node, *new_node;
node 91 fs/hfs/brec.c node = fd->bnode;
node 92 fs/hfs/brec.c hfs_bnode_dump(node);
node 94 fs/hfs/brec.c end_rec_off = tree->node_size - (node->num_recs + 1) * 2;
node 95 fs/hfs/brec.c end_off = hfs_bnode_read_u16(node, end_rec_off);
node 106 fs/hfs/brec.c if (node->type == HFS_NODE_LEAF) {
node 110 fs/hfs/brec.c node->num_recs++;
node 112 fs/hfs/brec.c hfs_bnode_write_u16(node, offsetof(struct hfs_bnode_desc, num_recs), node->num_recs);
node 113 fs/hfs/brec.c hfs_bnode_write_u16(node, end_rec_off, end_off + size);
node 121 fs/hfs/brec.c data_off = hfs_bnode_read_u16(node, data_rec_off + 2);
node 122 fs/hfs/brec.c hfs_bnode_write_u16(node, data_rec_off, data_off + size);
node 127 fs/hfs/brec.c hfs_bnode_move(node, data_off + size, data_off,
node 131 fs/hfs/brec.c hfs_bnode_write(node, fd->search_key, data_off, key_len);
node 132 fs/hfs/brec.c hfs_bnode_write(node, entry, data_off + key_len, entry_len);
node 133 fs/hfs/brec.c hfs_bnode_dump(node);
node 139 fs/hfs/brec.c if (!rec && new_node != node)
node 179 fs/hfs/brec.c struct hfs_bnode *node, *parent;
node 183 fs/hfs/brec.c node = fd->bnode;
node 186 fs/hfs/brec.c end_off = tree->node_size - (node->num_recs + 1) * 2;
node 188 fs/hfs/brec.c if (node->type == HFS_NODE_LEAF) {
node 192 fs/hfs/brec.c hfs_bnode_dump(node);
node 194 fs/hfs/brec.c if (!--node->num_recs) {
node 195 fs/hfs/brec.c hfs_bnode_unlink(node);
node 196 fs/hfs/brec.c if (!node->parent)
node 198 fs/hfs/brec.c parent = hfs_bnode_find(tree, node->parent);
node 201 fs/hfs/brec.c hfs_bnode_put(node);
node 202 fs/hfs/brec.c node = fd->bnode = parent;
node 204 fs/hfs/brec.c __hfs_brec_find(node, fd);
node 207 fs/hfs/brec.c hfs_bnode_write_u16(node, offsetof(struct hfs_bnode_desc, num_recs), node->num_recs);
node 214 fs/hfs/brec.c data_off = hfs_bnode_read_u16(node, rec_off);
node 215 fs/hfs/brec.c hfs_bnode_write_u16(node, rec_off + 2, data_off - size);
node 220 fs/hfs/brec.c hfs_bnode_move(node, fd->keyoffset, fd->keyoffset + size,
node 223 fs/hfs/brec.c hfs_bnode_dump(node);
node 232 fs/hfs/brec.c struct hfs_bnode *node, *new_node, *next_node;
node 238 fs/hfs/brec.c node = fd->bnode;
node 242 fs/hfs/brec.c hfs_bnode_get(node);
node 244 fs/hfs/brec.c node->this, new_node->this, node->next);
node 245 fs/hfs/brec.c new_node->next = node->next;
node 246 fs/hfs/brec.c new_node->prev = node->this;
node 247 fs/hfs/brec.c new_node->parent = node->parent;
node 248 fs/hfs/brec.c new_node->type = node->type;
node 249 fs/hfs/brec.c new_node->height = node->height;
node 251 fs/hfs/brec.c if (node->next)
node 252 fs/hfs/brec.c next_node = hfs_bnode_find(tree, node->next);
node 257 fs/hfs/brec.c hfs_bnode_put(node);
node 262 fs/hfs/brec.c size = tree->node_size / 2 - node->num_recs * 2 - 14;
node 266 fs/hfs/brec.c data_start = hfs_bnode_read_u16(node, old_rec_off);
node 270 fs/hfs/brec.c if (++num_recs < node->num_recs)
node 273 fs/hfs/brec.c hfs_bnode_put(node);
node 286 fs/hfs/brec.c data_start = hfs_bnode_read_u16(node, old_rec_off);
node 288 fs/hfs/brec.c hfs_bnode_put(node);
node 295 fs/hfs/brec.c new_node->num_recs = node->num_recs - num_recs;
node 296 fs/hfs/brec.c node->num_recs = num_recs;
node 307 fs/hfs/brec.c data_end = hfs_bnode_read_u16(node, old_rec_off);
node 312 fs/hfs/brec.c hfs_bnode_copy(new_node, 14, node, data_start, data_end - data_start);
node 324 fs/hfs/brec.c node->next = new_node->this;
node 325 fs/hfs/brec.c hfs_bnode_read(node, &node_desc, 0, sizeof(node_desc));
node 326 fs/hfs/brec.c node_desc.next = cpu_to_be32(node->next);
node 327 fs/hfs/brec.c node_desc.num_recs = cpu_to_be16(node->num_recs);
node 328 fs/hfs/brec.c hfs_bnode_write(node, &node_desc, 0, sizeof(node_desc));
node 337 fs/hfs/brec.c } else if (node->this == tree->leaf_tail) {
node 343 fs/hfs/brec.c hfs_bnode_dump(node);
node 345 fs/hfs/brec.c hfs_bnode_put(node);
node 353 fs/hfs/brec.c struct hfs_bnode *node, *new_node, *parent;
node 359 fs/hfs/brec.c node = fd->bnode;
node 361 fs/hfs/brec.c if (!node->parent)
node 365 fs/hfs/brec.c parent = hfs_bnode_find(tree, node->parent);
node 374 fs/hfs/brec.c newkeylen = (hfs_bnode_read_u8(node, 14) | 1) + 1;
node 412 fs/hfs/brec.c hfs_bnode_copy(parent, fd->keyoffset, node, 14, newkeylen);
node 417 fs/hfs/brec.c hfs_bnode_put(node);
node 418 fs/hfs/brec.c node = parent;
node 434 fs/hfs/brec.c if (new_node == node)
node 437 fs/hfs/brec.c hfs_bnode_read_key(node, fd->search_key, 14);
node 441 fs/hfs/brec.c if (!rec && node->parent)
node 444 fs/hfs/brec.c fd->bnode = node;
node 450 fs/hfs/brec.c struct hfs_bnode *node, *new_node;
node 455 fs/hfs/brec.c node = NULL;
node 457 fs/hfs/brec.c node = hfs_bnode_find(tree, tree->root);
node 458 fs/hfs/brec.c if (IS_ERR(node))
node 459 fs/hfs/brec.c return PTR_ERR(node);
node 463 fs/hfs/brec.c hfs_bnode_put(node);
node 492 fs/hfs/brec.c if (node) {
node 494 fs/hfs/brec.c node->parent = tree->root;
node 495 fs/hfs/brec.c if (node->type == HFS_NODE_LEAF ||
node 497 fs/hfs/brec.c key_size = hfs_bnode_read_u8(node, 14) + 1;
node 500 fs/hfs/brec.c hfs_bnode_copy(new_node, 14, node, 14, key_size);
node 507 fs/hfs/brec.c cnid = cpu_to_be32(node->this);
node 513 fs/hfs/brec.c hfs_bnode_put(node);
node 123 fs/hfs/btree.c struct hfs_bnode *node;
node 130 fs/hfs/btree.c while ((node = tree->node_hash[i])) {
node 131 fs/hfs/btree.c tree->node_hash[i] = node->next_hash;
node 132 fs/hfs/btree.c if (atomic_read(&node->refcnt))
node 134 fs/hfs/btree.c node->tree->cnid, node->this, atomic_read(&node->refcnt));
node 135 fs/hfs/btree.c hfs_bnode_free(node);
node 146 fs/hfs/btree.c struct hfs_bnode *node;
node 149 fs/hfs/btree.c node = hfs_bnode_find(tree, 0);
node 150 fs/hfs/btree.c if (IS_ERR(node))
node 154 fs/hfs/btree.c page = node->page[0];
node 168 fs/hfs/btree.c hfs_bnode_put(node);
node 174 fs/hfs/btree.c struct hfs_bnode *node;
node 178 fs/hfs/btree.c node = hfs_bnode_create(tree, idx);
node 179 fs/hfs/btree.c if (IS_ERR(node))
node 180 fs/hfs/btree.c return node;
node 189 fs/hfs/btree.c node->type = HFS_NODE_MAP;
node 190 fs/hfs/btree.c node->num_recs = 1;
node 191 fs/hfs/btree.c hfs_bnode_clear(node, 0, tree->node_size);
node 198 fs/hfs/btree.c hfs_bnode_write(node, &desc, 0, sizeof(desc));
node 199 fs/hfs/btree.c hfs_bnode_write_u16(node, 14, 0x8000);
node 200 fs/hfs/btree.c hfs_bnode_write_u16(node, tree->node_size - 2, 14);
node 201 fs/hfs/btree.c hfs_bnode_write_u16(node, tree->node_size - 4, tree->node_size - 6);
node 203 fs/hfs/btree.c return node;
node 208 fs/hfs/btree.c struct hfs_bnode *node, *next_node;
node 237 fs/hfs/btree.c node = hfs_bnode_find(tree, nidx);
node 238 fs/hfs/btree.c if (IS_ERR(node))
node 239 fs/hfs/btree.c return node;
node 240 fs/hfs/btree.c len = hfs_brec_lenoff(node, 2, &off16);
node 243 fs/hfs/btree.c off += node->page_offset;
node 244 fs/hfs/btree.c pagep = node->page + (off >> PAGE_CACHE_SHIFT);
node 261 fs/hfs/btree.c hfs_bnode_put(node);
node 275 fs/hfs/btree.c nidx = node->next;
node 278 fs/hfs/btree.c next_node = hfs_bmap_new_bmap(node, idx);
node 281 fs/hfs/btree.c hfs_bnode_put(node);
node 284 fs/hfs/btree.c node = next_node;
node 286 fs/hfs/btree.c len = hfs_brec_lenoff(node, 0, &off16);
node 288 fs/hfs/btree.c off += node->page_offset;
node 289 fs/hfs/btree.c pagep = node->page + (off >> PAGE_CACHE_SHIFT);
node 303 fs/hfs/btree.c dprint(DBG_BNODE_MOD, "btree_free_node: %u\n", node->this);
node 304 fs/hfs/btree.c tree = node->tree;
node 305 fs/hfs/btree.c nidx = node->this;
node 306 fs/hfs/btree.c node = hfs_bnode_find(tree, 0);
node 307 fs/hfs/btree.c if (IS_ERR(node))
node 309 fs/hfs/btree.c len = hfs_brec_lenoff(node, 2, &off);
node 314 fs/hfs/btree.c i = node->next;
node 315 fs/hfs/btree.c hfs_bnode_put(node);
node 318 fs/hfs/btree.c printk(KERN_CRIT "hfs: unable to free bnode %u. bmap not found!\n", node->this);
node 321 fs/hfs/btree.c node = hfs_bnode_find(tree, i);
node 322 fs/hfs/btree.c if (IS_ERR(node))
node 324 fs/hfs/btree.c if (node->type != HFS_NODE_MAP) {
node 326 fs/hfs/btree.c printk(KERN_CRIT "hfs: invalid bmap found! (%u,%d)\n", node->this, node->type);
node 327 fs/hfs/btree.c hfs_bnode_put(node);
node 330 fs/hfs/btree.c len = hfs_brec_lenoff(node, 0, &off);
node 332 fs/hfs/btree.c off += node->page_offset + nidx / 8;
node 333 fs/hfs/btree.c page = node->page[off >> PAGE_CACHE_SHIFT];
node 339 fs/hfs/btree.c printk(KERN_CRIT "hfs: trying to free free bnode %u(%d)\n", node->this, node->type);
node 341 fs/hfs/btree.c hfs_bnode_put(node);
node 347 fs/hfs/btree.c hfs_bnode_put(node);
node 58 fs/hfs/inode.c struct hfs_bnode *node;
node 76 fs/hfs/inode.c node = hfs_bnode_findhash(tree, nidx);
node 77 fs/hfs/inode.c if (!node)
node 79 fs/hfs/inode.c else if (atomic_read(&node->refcnt))
node 81 fs/hfs/inode.c if (res && node) {
node 82 fs/hfs/inode.c hfs_bnode_unhash(node);
node 83 fs/hfs/inode.c hfs_bnode_free(node);
node 91 fs/hfs/inode.c node = hfs_bnode_findhash(tree, nidx++);
node 92 fs/hfs/inode.c if (!node)
node 94 fs/hfs/inode.c if (atomic_read(&node->refcnt)) {
node 98 fs/hfs/inode.c hfs_bnode_unhash(node);
node 99 fs/hfs/inode.c hfs_bnode_free(node);
node 26 fs/hfsplus/bnode.c off += node->page_offset;
node 27 fs/hfsplus/bnode.c pagep = node->page + (off >> PAGE_CACHE_SHIFT);
node 46 fs/hfsplus/bnode.c hfs_bnode_read(node, &data, off, 2);
node 54 fs/hfsplus/bnode.c hfs_bnode_read(node, &data, off, 1);
node 63 fs/hfsplus/bnode.c tree = node->tree;
node 64 fs/hfsplus/bnode.c if (node->type == HFS_NODE_LEAF ||
node 66 fs/hfsplus/bnode.c key_len = hfs_bnode_read_u16(node, off) + 2;
node 70 fs/hfsplus/bnode.c hfs_bnode_read(node, key, off, key_len);
node 78 fs/hfsplus/bnode.c off += node->page_offset;
node 79 fs/hfsplus/bnode.c pagep = node->page + (off >> PAGE_CACHE_SHIFT);
node 100 fs/hfsplus/bnode.c hfs_bnode_write(node, &v, off, 2);
node 108 fs/hfsplus/bnode.c off += node->page_offset;
node 109 fs/hfsplus/bnode.c pagep = node->page + (off >> PAGE_CACHE_SHIFT);
node 193 fs/hfsplus/bnode.c src += node->page_offset;
node 194 fs/hfsplus/bnode.c dst += node->page_offset;
node 197 fs/hfsplus/bnode.c src_page = node->page + (src >> PAGE_CACHE_SHIFT);
node 200 fs/hfsplus/bnode.c dst_page = node->page + (dst >> PAGE_CACHE_SHIFT);
node 246 fs/hfsplus/bnode.c src_page = node->page + (src >> PAGE_CACHE_SHIFT);
node 248 fs/hfsplus/bnode.c dst_page = node->page + (dst >> PAGE_CACHE_SHIFT);
node 300 fs/hfsplus/bnode.c dprint(DBG_BNODE_MOD, "bnode: %d\n", node->this);
node 301 fs/hfsplus/bnode.c hfs_bnode_read(node, &desc, 0, sizeof(desc));
node 306 fs/hfsplus/bnode.c off = node->tree->node_size - 2;
node 308 fs/hfsplus/bnode.c key_off = hfs_bnode_read_u16(node, off);
node 310 fs/hfsplus/bnode.c if (i && node->type == HFS_NODE_INDEX) {
node 313 fs/hfsplus/bnode.c if (node->tree->attributes & HFS_TREE_VARIDXKEYS)
node 314 fs/hfsplus/bnode.c tmp = hfs_bnode_read_u16(node, key_off) + 2;
node 316 fs/hfsplus/bnode.c tmp = node->tree->max_key_len + 2;
node 318 fs/hfsplus/bnode.c hfs_bnode_read(node, &cnid, key_off + tmp, 4);
node 320 fs/hfsplus/bnode.c } else if (i && node->type == HFS_NODE_LEAF) {
node 323 fs/hfsplus/bnode.c tmp = hfs_bnode_read_u16(node, key_off);
node 336 fs/hfsplus/bnode.c tree = node->tree;
node 337 fs/hfsplus/bnode.c if (node->prev) {
node 338 fs/hfsplus/bnode.c tmp = hfs_bnode_find(tree, node->prev);
node 341 fs/hfsplus/bnode.c tmp->next = node->next;
node 345 fs/hfsplus/bnode.c } else if (node->type == HFS_NODE_LEAF)
node 346 fs/hfsplus/bnode.c tree->leaf_head = node->next;
node 348 fs/hfsplus/bnode.c if (node->next) {
node 349 fs/hfsplus/bnode.c tmp = hfs_bnode_find(tree, node->next);
node 352 fs/hfsplus/bnode.c tmp->prev = node->prev;
node 356 fs/hfsplus/bnode.c } else if (node->type == HFS_NODE_LEAF)
node 357 fs/hfsplus/bnode.c tree->leaf_tail = node->prev;
node 360 fs/hfsplus/bnode.c if (!node->prev && !node->next) {
node 363 fs/hfsplus/bnode.c if (!node->parent) {
node 367 fs/hfsplus/bnode.c set_bit(HFS_BNODE_DELETED, &node->flags);
node 379 fs/hfsplus/bnode.c struct hfs_bnode *node;
node 386 fs/hfsplus/bnode.c for (node = tree->node_hash[hfs_bnode_hash(cnid)];
node 387 fs/hfsplus/bnode.c node; node = node->next_hash) {
node 388 fs/hfsplus/bnode.c if (node->this == cnid) {
node 389 fs/hfsplus/bnode.c return node;
node 398 fs/hfsplus/bnode.c struct hfs_bnode *node, *node2;
node 412 fs/hfsplus/bnode.c node = kzalloc(size, GFP_KERNEL);
node 413 fs/hfsplus/bnode.c if (!node)
node 415 fs/hfsplus/bnode.c node->tree = tree;
node 416 fs/hfsplus/bnode.c node->this = cnid;
node 417 fs/hfsplus/bnode.c set_bit(HFS_BNODE_NEW, &node->flags);
node 418 fs/hfsplus/bnode.c atomic_set(&node->refcnt, 1);
node 420 fs/hfsplus/bnode.c node->tree->cnid, node->this);
node 421 fs/hfsplus/bnode.c init_waitqueue_head(&node->lock_wq);
node 426 fs/hfsplus/bnode.c node->next_hash = tree->node_hash[hash];
node 427 fs/hfsplus/bnode.c tree->node_hash[hash] = node;
node 431 fs/hfsplus/bnode.c kfree(node);
node 440 fs/hfsplus/bnode.c node->page_offset = off & ~PAGE_CACHE_MASK;
node 450 fs/hfsplus/bnode.c node->page[i] = page;
node 453 fs/hfsplus/bnode.c return node;
node 455 fs/hfsplus/bnode.c set_bit(HFS_BNODE_ERROR, &node->flags);
node 456 fs/hfsplus/bnode.c return node;
node 464 fs/hfsplus/bnode.c node->tree->cnid, node->this, atomic_read(&node->refcnt));
node 465 fs/hfsplus/bnode.c for (p = &node->tree->node_hash[hfs_bnode_hash(node->this)];
node 466 fs/hfsplus/bnode.c *p && *p != node; p = &(*p)->next_hash)
node 469 fs/hfsplus/bnode.c *p = node->next_hash;
node 470 fs/hfsplus/bnode.c node->tree->node_hash_cnt--;
node 476 fs/hfsplus/bnode.c struct hfs_bnode *node;
node 482 fs/hfsplus/bnode.c node = hfs_bnode_findhash(tree, num);
node 483 fs/hfsplus/bnode.c if (node) {
node 484 fs/hfsplus/bnode.c hfs_bnode_get(node);
node 486 fs/hfsplus/bnode.c wait_event(node->lock_wq, !test_bit(HFS_BNODE_NEW, &node->flags));
node 487 fs/hfsplus/bnode.c if (test_bit(HFS_BNODE_ERROR, &node->flags))
node 489 fs/hfsplus/bnode.c return node;
node 492 fs/hfsplus/bnode.c node = __hfs_bnode_create(tree, num);
node 493 fs/hfsplus/bnode.c if (!node)
node 495 fs/hfsplus/bnode.c if (test_bit(HFS_BNODE_ERROR, &node->flags))
node 497 fs/hfsplus/bnode.c if (!test_bit(HFS_BNODE_NEW, &node->flags))
node 498 fs/hfsplus/bnode.c return node;
node 500 fs/hfsplus/bnode.c desc = (struct hfs_bnode_desc *)(kmap(node->page[0]) + node->page_offset);
node 501 fs/hfsplus/bnode.c node->prev = be32_to_cpu(desc->prev);
node 502 fs/hfsplus/bnode.c node->next = be32_to_cpu(desc->next);
node 503 fs/hfsplus/bnode.c node->num_recs = be16_to_cpu(desc->num_recs);
node 504 fs/hfsplus/bnode.c node->type = desc->type;
node 505 fs/hfsplus/bnode.c node->height = desc->height;
node 506 fs/hfsplus/bnode.c kunmap(node->page[0]);
node 508 fs/hfsplus/bnode.c switch (node->type) {
node 511 fs/hfsplus/bnode.c if (node->height != 0)
node 515 fs/hfsplus/bnode.c if (node->height != 1)
node 519 fs/hfsplus/bnode.c if (node->height <= 1 || node->height > tree->depth)
node 527 fs/hfsplus/bnode.c off = hfs_bnode_read_u16(node, rec_off);
node 530 fs/hfsplus/bnode.c for (i = 1; i <= node->num_recs; off = next_off, i++) {
node 532 fs/hfsplus/bnode.c next_off = hfs_bnode_read_u16(node, rec_off);
node 538 fs/hfsplus/bnode.c if (node->type != HFS_NODE_INDEX &&
node 539 fs/hfsplus/bnode.c node->type != HFS_NODE_LEAF)
node 541 fs/hfsplus/bnode.c key_size = hfs_bnode_read_u16(node, off) + 2;
node 545 fs/hfsplus/bnode.c clear_bit(HFS_BNODE_NEW, &node->flags);
node 546 fs/hfsplus/bnode.c wake_up(&node->lock_wq);
node 547 fs/hfsplus/bnode.c return node;
node 550 fs/hfsplus/bnode.c set_bit(HFS_BNODE_ERROR, &node->flags);
node 551 fs/hfsplus/bnode.c clear_bit(HFS_BNODE_NEW, &node->flags);
node 552 fs/hfsplus/bnode.c wake_up(&node->lock_wq);
node 553 fs/hfsplus/bnode.c hfs_bnode_put(node);
node 564 fs/hfsplus/bnode.c kfree(node);
node 569 fs/hfsplus/bnode.c struct hfs_bnode *node;
node 574 fs/hfsplus/bnode.c node = hfs_bnode_findhash(tree, num);
node 576 fs/hfsplus/bnode.c if (node) {
node 579 fs/hfsplus/bnode.c return node;
node 581 fs/hfsplus/bnode.c node = __hfs_bnode_create(tree, num);
node 582 fs/hfsplus/bnode.c if (!node)
node 584 fs/hfsplus/bnode.c if (test_bit(HFS_BNODE_ERROR, &node->flags)) {
node 585 fs/hfsplus/bnode.c hfs_bnode_put(node);
node 589 fs/hfsplus/bnode.c pagep = node->page;
node 590 fs/hfsplus/bnode.c memset(kmap(*pagep) + node->page_offset, 0,
node 599 fs/hfsplus/bnode.c clear_bit(HFS_BNODE_NEW, &node->flags);
node 600 fs/hfsplus/bnode.c wake_up(&node->lock_wq);
node 602 fs/hfsplus/bnode.c return node;
node 607 fs/hfsplus/bnode.c if (node) {
node 608 fs/hfsplus/bnode.c atomic_inc(&node->refcnt);
node 610 fs/hfsplus/bnode.c node->tree->cnid, node->this, atomic_read(&node->refcnt));
node 617 fs/hfsplus/bnode.c if (node) {
node 618 fs/hfsplus/bnode.c struct hfs_btree *tree = node->tree;
node 622 fs/hfsplus/bnode.c node->tree->cnid, node->this, atomic_read(&node->refcnt));
node 623 fs/hfsplus/bnode.c BUG_ON(!atomic_read(&node->refcnt));
node 624 fs/hfsplus/bnode.c if (!atomic_dec_and_lock(&node->refcnt, &tree->hash_lock))
node 627 fs/hfsplus/bnode.c if (!node->page[i])
node 629 fs/hfsplus/bnode.c mark_page_accessed(node->page[i]);
node 632 fs/hfsplus/bnode.c if (test_bit(HFS_BNODE_DELETED, &node->flags)) {
node 633 fs/hfsplus/bnode.c hfs_bnode_unhash(node);
node 635 fs/hfsplus/bnode.c hfs_bmap_free(node);
node 636 fs/hfsplus/bnode.c hfs_bnode_free(node);
node 24 fs/hfsplus/brec.c dataoff = node->tree->node_size - (rec + 2) * 2;
node 25 fs/hfsplus/brec.c hfs_bnode_read(node, retval, dataoff, 4);
node 35 fs/hfsplus/brec.c if (node->type != HFS_NODE_INDEX && node->type != HFS_NODE_LEAF)
node 38 fs/hfsplus/brec.c if ((node->type == HFS_NODE_INDEX) &&
node 39 fs/hfsplus/brec.c !(node->tree->attributes & HFS_TREE_VARIDXKEYS)) {
node 40 fs/hfsplus/brec.c retval = node->tree->max_key_len + 2;
node 42 fs/hfsplus/brec.c recoff = hfs_bnode_read_u16(node, node->tree->node_size - (rec + 1) * 2);
node 45 fs/hfsplus/brec.c if (node->tree->attributes & HFS_TREE_BIGKEYS)
node 46 fs/hfsplus/brec.c retval = hfs_bnode_read_u16(node, recoff) + 2;
node 48 fs/hfsplus/brec.c retval = (hfs_bnode_read_u8(node, recoff) | 1) + 1;
node 56 fs/hfsplus/brec.c struct hfs_bnode *node, *new_node;
node 78 fs/hfsplus/brec.c node = fd->bnode;
node 79 fs/hfsplus/brec.c hfs_bnode_dump(node);
node 81 fs/hfsplus/brec.c end_rec_off = tree->node_size - (node->num_recs + 1) * 2;
node 82 fs/hfsplus/brec.c end_off = hfs_bnode_read_u16(node, end_rec_off);
node 93 fs/hfsplus/brec.c if (node->type == HFS_NODE_LEAF) {
node 97 fs/hfsplus/brec.c node->num_recs++;
node 99 fs/hfsplus/brec.c hfs_bnode_write_u16(node, offsetof(struct hfs_bnode_desc, num_recs), node->num_recs);
node 100 fs/hfsplus/brec.c hfs_bnode_write_u16(node, end_rec_off, end_off + size);
node 108 fs/hfsplus/brec.c data_off = hfs_bnode_read_u16(node, data_rec_off + 2);
node 109 fs/hfsplus/brec.c hfs_bnode_write_u16(node, data_rec_off, data_off + size);
node 114 fs/hfsplus/brec.c hfs_bnode_move(node, data_off + size, data_off,
node 118 fs/hfsplus/brec.c hfs_bnode_write(node, fd->search_key, data_off, key_len);
node 119 fs/hfsplus/brec.c hfs_bnode_write(node, entry, data_off + key_len, entry_len);
node 120 fs/hfsplus/brec.c hfs_bnode_dump(node);
node 126 fs/hfsplus/brec.c if (!rec && new_node != node)
node 166 fs/hfsplus/brec.c struct hfs_bnode *node, *parent;
node 170 fs/hfsplus/brec.c node = fd->bnode;
node 173 fs/hfsplus/brec.c end_off = tree->node_size - (node->num_recs + 1) * 2;
node 175 fs/hfsplus/brec.c if (node->type == HFS_NODE_LEAF) {
node 179 fs/hfsplus/brec.c hfs_bnode_dump(node);
node 181 fs/hfsplus/brec.c if (!--node->num_recs) {
node 182 fs/hfsplus/brec.c hfs_bnode_unlink(node);
node 183 fs/hfsplus/brec.c if (!node->parent)
node 185 fs/hfsplus/brec.c parent = hfs_bnode_find(tree, node->parent);
node 188 fs/hfsplus/brec.c hfs_bnode_put(node);
node 189 fs/hfsplus/brec.c node = fd->bnode = parent;
node 191 fs/hfsplus/brec.c __hfs_brec_find(node, fd);
node 194 fs/hfsplus/brec.c hfs_bnode_write_u16(node, offsetof(struct hfs_bnode_desc, num_recs), node->num_recs);
node 201 fs/hfsplus/brec.c data_off = hfs_bnode_read_u16(node, rec_off);
node 202 fs/hfsplus/brec.c hfs_bnode_write_u16(node, rec_off + 2, data_off - size);
node 207 fs/hfsplus/brec.c hfs_bnode_move(node, fd->keyoffset, fd->keyoffset + size,
node 210 fs/hfsplus/brec.c hfs_bnode_dump(node);
node 219 fs/hfsplus/brec.c struct hfs_bnode *node, *new_node;
node 225 fs/hfsplus/brec.c node = fd->bnode;
node 229 fs/hfsplus/brec.c hfs_bnode_get(node);
node 231 fs/hfsplus/brec.c node->this, new_node->this, node->next);
node 232 fs/hfsplus/brec.c new_node->next = node->next;
node 233 fs/hfsplus/brec.c new_node->prev = node->this;
node 234 fs/hfsplus/brec.c new_node->parent = node->parent;
node 235 fs/hfsplus/brec.c new_node->type = node->type;
node 236 fs/hfsplus/brec.c new_node->height = node->height;
node 238 fs/hfsplus/brec.c size = tree->node_size / 2 - node->num_recs * 2 - 14;
node 242 fs/hfsplus/brec.c data_start = hfs_bnode_read_u16(node, old_rec_off);
node 246 fs/hfsplus/brec.c if (++num_recs < node->num_recs)
node 249 fs/hfsplus/brec.c hfs_bnode_put(node);
node 260 fs/hfsplus/brec.c data_start = hfs_bnode_read_u16(node, old_rec_off);
node 262 fs/hfsplus/brec.c hfs_bnode_put(node);
node 269 fs/hfsplus/brec.c new_node->num_recs = node->num_recs - num_recs;
node 270 fs/hfsplus/brec.c node->num_recs = num_recs;
node 281 fs/hfsplus/brec.c data_end = hfs_bnode_read_u16(node, old_rec_off);
node 286 fs/hfsplus/brec.c hfs_bnode_copy(new_node, 14, node, data_start, data_end - data_start);
node 298 fs/hfsplus/brec.c node->next = new_node->this;
node 299 fs/hfsplus/brec.c hfs_bnode_read(node, &node_desc, 0, sizeof(node_desc));
node 300 fs/hfsplus/brec.c node_desc.next = cpu_to_be32(node->next);
node 301 fs/hfsplus/brec.c node_desc.num_recs = cpu_to_be16(node->num_recs);
node 302 fs/hfsplus/brec.c hfs_bnode_write(node, &node_desc, 0, sizeof(node_desc));
node 312 fs/hfsplus/brec.c } else if (node->this == tree->leaf_tail) {
node 318 fs/hfsplus/brec.c hfs_bnode_dump(node);
node 320 fs/hfsplus/brec.c hfs_bnode_put(node);
node 328 fs/hfsplus/brec.c struct hfs_bnode *node, *new_node, *parent;
node 334 fs/hfsplus/brec.c node = fd->bnode;
node 336 fs/hfsplus/brec.c if (!node->parent)
node 340 fs/hfsplus/brec.c parent = hfs_bnode_find(tree, node->parent);
node 349 fs/hfsplus/brec.c newkeylen = hfs_bnode_read_u16(node, 14) + 2;
node 387 fs/hfsplus/brec.c hfs_bnode_copy(parent, fd->keyoffset, node, 14, newkeylen);
node 390 fs/hfsplus/brec.c hfs_bnode_put(node);
node 391 fs/hfsplus/brec.c node = parent;
node 407 fs/hfsplus/brec.c if (new_node == node)
node 410 fs/hfsplus/brec.c hfs_bnode_read_key(node, fd->search_key, 14);
node 414 fs/hfsplus/brec.c if (!rec && node->parent)
node 417 fs/hfsplus/brec.c fd->bnode = node;
node 423 fs/hfsplus/brec.c struct hfs_bnode *node, *new_node;
node 428 fs/hfsplus/brec.c node = NULL;
node 430 fs/hfsplus/brec.c node = hfs_bnode_find(tree, tree->root);
node 431 fs/hfsplus/brec.c if (IS_ERR(node))
node 432 fs/hfsplus/brec.c return PTR_ERR(node);
node 436 fs/hfsplus/brec.c hfs_bnode_put(node);
node 465 fs/hfsplus/brec.c if (node) {
node 467 fs/hfsplus/brec.c node->parent = tree->root;
node 468 fs/hfsplus/brec.c if (node->type == HFS_NODE_LEAF ||
node 470 fs/hfsplus/brec.c key_size = hfs_bnode_read_u16(node, 14) + 2;
node 473 fs/hfsplus/brec.c hfs_bnode_copy(new_node, 14, node, 14, key_size);
node 479 fs/hfsplus/brec.c cnid = cpu_to_be32(node->this);
node 485 fs/hfsplus/brec.c hfs_bnode_put(node);
node 101 fs/hfsplus/btree.c struct hfs_bnode *node;
node 108 fs/hfsplus/btree.c while ((node = tree->node_hash[i])) {
node 109 fs/hfsplus/btree.c tree->node_hash[i] = node->next_hash;
node 110 fs/hfsplus/btree.c if (atomic_read(&node->refcnt))
node 112 fs/hfsplus/btree.c node->tree->cnid, node->this, atomic_read(&node->refcnt));
node 113 fs/hfsplus/btree.c hfs_bnode_free(node);
node 124 fs/hfsplus/btree.c struct hfs_bnode *node;
node 127 fs/hfsplus/btree.c node = hfs_bnode_find(tree, 0);
node 128 fs/hfsplus/btree.c if (IS_ERR(node))
node 132 fs/hfsplus/btree.c page = node->page[0];
node 146 fs/hfsplus/btree.c hfs_bnode_put(node);
node 152 fs/hfsplus/btree.c struct hfs_bnode *node;
node 156 fs/hfsplus/btree.c node = hfs_bnode_create(tree, idx);
node 157 fs/hfsplus/btree.c if (IS_ERR(node))
node 158 fs/hfsplus/btree.c return node;
node 165 fs/hfsplus/btree.c node->type = HFS_NODE_MAP;
node 166 fs/hfsplus/btree.c node->num_recs = 1;
node 167 fs/hfsplus/btree.c hfs_bnode_clear(node, 0, tree->node_size);
node 174 fs/hfsplus/btree.c hfs_bnode_write(node, &desc, 0, sizeof(desc));
node 175 fs/hfsplus/btree.c hfs_bnode_write_u16(node, 14, 0x8000);
node 176 fs/hfsplus/btree.c hfs_bnode_write_u16(node, tree->node_size - 2, 14);
node 177 fs/hfsplus/btree.c hfs_bnode_write_u16(node, tree->node_size - 4, tree->node_size - 6);
node 179 fs/hfsplus/btree.c return node;
node 184 fs/hfsplus/btree.c struct hfs_bnode *node, *next_node;
node 213 fs/hfsplus/btree.c node = hfs_bnode_find(tree, nidx);
node 214 fs/hfsplus/btree.c if (IS_ERR(node))
node 215 fs/hfsplus/btree.c return node;
node 216 fs/hfsplus/btree.c len = hfs_brec_lenoff(node, 2, &off16);
node 219 fs/hfsplus/btree.c off += node->page_offset;
node 220 fs/hfsplus/btree.c pagep = node->page + (off >> PAGE_CACHE_SHIFT);
node 237 fs/hfsplus/btree.c hfs_bnode_put(node);
node 251 fs/hfsplus/btree.c nidx = node->next;
node 254 fs/hfsplus/btree.c next_node = hfs_bmap_new_bmap(node, idx);
node 257 fs/hfsplus/btree.c hfs_bnode_put(node);
node 260 fs/hfsplus/btree.c node = next_node;
node 262 fs/hfsplus/btree.c len = hfs_brec_lenoff(node, 0, &off16);
node 264 fs/hfsplus/btree.c off += node->page_offset;
node 265 fs/hfsplus/btree.c pagep = node->page + (off >> PAGE_CACHE_SHIFT);
node 279 fs/hfsplus/btree.c dprint(DBG_BNODE_MOD, "btree_free_node: %u\n", node->this);
node 280 fs/hfsplus/btree.c BUG_ON(!node->this);
node 281 fs/hfsplus/btree.c tree = node->tree;
node 282 fs/hfsplus/btree.c nidx = node->this;
node 283 fs/hfsplus/btree.c node = hfs_bnode_find(tree, 0);
node 284 fs/hfsplus/btree.c if (IS_ERR(node))
node 286 fs/hfsplus/btree.c len = hfs_brec_lenoff(node, 2, &off);
node 291 fs/hfsplus/btree.c i = node->next;
node 292 fs/hfsplus/btree.c hfs_bnode_put(node);
node 295 fs/hfsplus/btree.c printk(KERN_CRIT "hfs: unable to free bnode %u. bmap not found!\n", node->this);
node 298 fs/hfsplus/btree.c node = hfs_bnode_find(tree, i);
node 299 fs/hfsplus/btree.c if (IS_ERR(node))
node 301 fs/hfsplus/btree.c if (node->type != HFS_NODE_MAP) {
node 303 fs/hfsplus/btree.c printk(KERN_CRIT "hfs: invalid bmap found! (%u,%d)\n", node->this, node->type);
node 304 fs/hfsplus/btree.c hfs_bnode_put(node);
node 307 fs/hfsplus/btree.c len = hfs_brec_lenoff(node, 0, &off);
node 309 fs/hfsplus/btree.c off += node->page_offset + nidx / 8;
node 310 fs/hfsplus/btree.c page = node->page[off >> PAGE_CACHE_SHIFT];
node 316 fs/hfsplus/btree.c printk(KERN_CRIT "hfs: trying to free free bnode %u(%d)\n", node->this, node->type);
node 318 fs/hfsplus/btree.c hfs_bnode_put(node);
node 324 fs/hfsplus/btree.c hfs_bnode_put(node);
node 50 fs/hfsplus/inode.c struct hfs_bnode *node;
node 73 fs/hfsplus/inode.c node = hfs_bnode_findhash(tree, nidx);
node 74 fs/hfsplus/inode.c if (!node)
node 76 fs/hfsplus/inode.c else if (atomic_read(&node->refcnt))
node 78 fs/hfsplus/inode.c if (res && node) {
node 79 fs/hfsplus/inode.c hfs_bnode_unhash(node);
node 80 fs/hfsplus/inode.c hfs_bnode_free(node);
node 88 fs/hfsplus/inode.c node = hfs_bnode_findhash(tree, nidx++);
node 89 fs/hfsplus/inode.c if (!node)
node 91 fs/hfsplus/inode.c if (atomic_read(&node->refcnt)) {
node 95 fs/hfsplus/inode.c hfs_bnode_unhash(node);
node 96 fs/hfsplus/inode.c hfs_bnode_free(node);
node 72 fs/hpfs/anode.c if (!(fnode = hpfs_map_fnode(s, node, &bh))) return -1;
node 75 fs/hpfs/anode.c if (!(anode = hpfs_map_anode(s, node, &bh))) return -1;
node 78 fs/hpfs/anode.c a = node;
node 100 fs/hpfs/anode.c fnod?'f':'a', node);
node 112 fs/hpfs/anode.c hpfs_error(s, "empty file %08x, trying to add sector %08x", node, fsecno);
node 116 fs/hpfs/anode.c se = !fnod ? node : (node + 16384) & ~16383;
node 124 fs/hpfs/anode.c up = a != node ? anode->up : -1;
node 130 fs/hpfs/anode.c if (a == node && fnod) {
node 131 fs/hpfs/anode.c anode->up = node;
node 162 fs/hpfs/anode.c if ((a == node && fnod) || na == -1) return se;
node 168 fs/hpfs/anode.c if (up != node || !fnod) {
node 187 fs/hpfs/anode.c anode->btree.fnode_parent = up == node && fnod;
node 193 fs/hpfs/anode.c up = up != node ? anode->up : -1;
node 217 fs/hpfs/anode.c anode->up = node;
node 223 fs/hpfs/anode.c if (!(anode = hpfs_map_anode(s, node, &bh))) {
node 229 fs/hpfs/anode.c if (!(fnode = hpfs_map_fnode(s, node, &bh))) {
node 235 fs/hpfs/anode.c ranode->up = node;
node 399 fs/hpfs/anode.c anode_secno node = f;
node 426 fs/hpfs/anode.c hpfs_error(s, "internal btree %08x doesn't end with -1", node);
node 439 fs/hpfs/anode.c node = btree->u.internal[i].down;
node 442 fs/hpfs/anode.c if (hpfs_stop_cycles(s, node, &c1, &c2, "hpfs_truncate_btree"))
node 444 fs/hpfs/anode.c if (!(anode = hpfs_map_anode(s, node, &bh))) return;
node 495 fs/inode.c struct hlist_node *node;
node 499 fs/inode.c hlist_for_each_entry(inode, node, head, i_hash) {
node 510 fs/inode.c return node ? inode : NULL;
node 519 fs/inode.c struct hlist_node *node;
node 523 fs/inode.c hlist_for_each_entry(inode, node, head, i_hash) {
node 534 fs/inode.c return node ? inode : NULL;
node 77 fs/jffs2/debug.c struct jffs2_full_dnode *fn = frag->node;
node 95 fs/jffs2/debug.c && frag_prev(frag)->size < PAGE_CACHE_SIZE && frag_prev(frag)->node) {
node 102 fs/jffs2/debug.c && frag_next(frag)->size < PAGE_CACHE_SIZE && frag_next(frag)->node) {
node 703 fs/jffs2/debug.c if (this->node)
node 705 fs/jffs2/debug.c this->ofs, this->ofs+this->size, ref_offset(this->node->raw),
node 706 fs/jffs2/debug.c ref_flags(this->node->raw), this, frag_left(this), frag_right(this),
node 767 fs/jffs2/debug.c union jffs2_node_union node;
node 775 fs/jffs2/debug.c ret = jffs2_flash_read(c, ofs, len, &retlen, (unsigned char *)&node);
node 782 fs/jffs2/debug.c printk(JFFS2_DBG "magic:\t%#04x\n", je16_to_cpu(node.u.magic));
node 783 fs/jffs2/debug.c printk(JFFS2_DBG "nodetype:\t%#04x\n", je16_to_cpu(node.u.nodetype));
node 784 fs/jffs2/debug.c printk(JFFS2_DBG "totlen:\t%#08x\n", je32_to_cpu(node.u.totlen));
node 785 fs/jffs2/debug.c printk(JFFS2_DBG "hdr_crc:\t%#08x\n", je32_to_cpu(node.u.hdr_crc));
node 787 fs/jffs2/debug.c crc = crc32(0, &node.u, sizeof(node.u) - 4);
node 788 fs/jffs2/debug.c if (crc != je32_to_cpu(node.u.hdr_crc)) {
node 793 fs/jffs2/debug.c if (je16_to_cpu(node.u.magic) != JFFS2_MAGIC_BITMASK &&
node 794 fs/jffs2/debug.c je16_to_cpu(node.u.magic) != JFFS2_OLD_MAGIC_BITMASK)
node 797 fs/jffs2/debug.c je16_to_cpu(node.u.magic), JFFS2_MAGIC_BITMASK);
node 801 fs/jffs2/debug.c switch(je16_to_cpu(node.u.nodetype)) {
node 806 fs/jffs2/debug.c printk(JFFS2_DBG "ino:\t%#08x\n", je32_to_cpu(node.i.ino));
node 807 fs/jffs2/debug.c printk(JFFS2_DBG "version:\t%#08x\n", je32_to_cpu(node.i.version));
node 808 fs/jffs2/debug.c printk(JFFS2_DBG "mode:\t%#08x\n", node.i.mode.m);
node 809 fs/jffs2/debug.c printk(JFFS2_DBG "uid:\t%#04x\n", je16_to_cpu(node.i.uid));
node 810 fs/jffs2/debug.c printk(JFFS2_DBG "gid:\t%#04x\n", je16_to_cpu(node.i.gid));
node 811 fs/jffs2/debug.c printk(JFFS2_DBG "isize:\t%#08x\n", je32_to_cpu(node.i.isize));
node 812 fs/jffs2/debug.c printk(JFFS2_DBG "atime:\t%#08x\n", je32_to_cpu(node.i.atime));
node 813 fs/jffs2/debug.c printk(JFFS2_DBG "mtime:\t%#08x\n", je32_to_cpu(node.i.mtime));
node 814 fs/jffs2/debug.c printk(JFFS2_DBG "ctime:\t%#08x\n", je32_to_cpu(node.i.ctime));
node 815 fs/jffs2/debug.c printk(JFFS2_DBG "offset:\t%#08x\n", je32_to_cpu(node.i.offset));
node 816 fs/jffs2/debug.c printk(JFFS2_DBG "csize:\t%#08x\n", je32_to_cpu(node.i.csize));
node 817 fs/jffs2/debug.c printk(JFFS2_DBG "dsize:\t%#08x\n", je32_to_cpu(node.i.dsize));
node 818 fs/jffs2/debug.c printk(JFFS2_DBG "compr:\t%#02x\n", node.i.compr);
node 819 fs/jffs2/debug.c printk(JFFS2_DBG "usercompr:\t%#02x\n", node.i.usercompr);
node 820 fs/jffs2/debug.c printk(JFFS2_DBG "flags:\t%#04x\n", je16_to_cpu(node.i.flags));
node 821 fs/jffs2/debug.c printk(JFFS2_DBG "data_crc:\t%#08x\n", je32_to_cpu(node.i.data_crc));
node 822 fs/jffs2/debug.c printk(JFFS2_DBG "node_crc:\t%#08x\n", je32_to_cpu(node.i.node_crc));
node 824 fs/jffs2/debug.c crc = crc32(0, &node.i, sizeof(node.i) - 8);
node 825 fs/jffs2/debug.c if (crc != je32_to_cpu(node.i.node_crc)) {
node 834 fs/jffs2/debug.c printk(JFFS2_DBG "pino:\t%#08x\n", je32_to_cpu(node.d.pino));
node 835 fs/jffs2/debug.c printk(JFFS2_DBG "version:\t%#08x\n", je32_to_cpu(node.d.version));
node 836 fs/jffs2/debug.c printk(JFFS2_DBG "ino:\t%#08x\n", je32_to_cpu(node.d.ino));
node 837 fs/jffs2/debug.c printk(JFFS2_DBG "mctime:\t%#08x\n", je32_to_cpu(node.d.mctime));
node 838 fs/jffs2/debug.c printk(JFFS2_DBG "nsize:\t%#02x\n", node.d.nsize);
node 839 fs/jffs2/debug.c printk(JFFS2_DBG "type:\t%#02x\n", node.d.type);
node 840 fs/jffs2/debug.c printk(JFFS2_DBG "node_crc:\t%#08x\n", je32_to_cpu(node.d.node_crc));
node 841 fs/jffs2/debug.c printk(JFFS2_DBG "name_crc:\t%#08x\n", je32_to_cpu(node.d.name_crc));
node 843 fs/jffs2/debug.c node.d.name[node.d.nsize] = '\0';
node 844 fs/jffs2/debug.c printk(JFFS2_DBG "name:\t\"%s\"\n", node.d.name);
node 846 fs/jffs2/debug.c crc = crc32(0, &node.d, sizeof(node.d) - 8);
node 847 fs/jffs2/debug.c if (crc != je32_to_cpu(node.d.node_crc)) {
node 483 fs/jffs2/gc.c if (frag->node && frag->node->raw == raw) {
node 484 fs/jffs2/gc.c fn = frag->node;
node 488 fs/jffs2/gc.c if (nrfrags == frag->node->frags)
node 497 fs/jffs2/gc.c frag->node->raw = f->inocache->nodes;
node 543 fs/jffs2/gc.c union jffs2_node_union *node;
node 571 fs/jffs2/gc.c node = kmalloc(rawlen, GFP_KERNEL);
node 572 fs/jffs2/gc.c if (!node)
node 575 fs/jffs2/gc.c ret = jffs2_flash_read(c, ref_offset(raw), rawlen, &retlen, (char *)node);
node 581 fs/jffs2/gc.c crc = crc32(0, node, sizeof(struct jffs2_unknown_node)-4);
node 582 fs/jffs2/gc.c if (je32_to_cpu(node->u.hdr_crc) != crc) {
node 584 fs/jffs2/gc.c ref_offset(raw), je32_to_cpu(node->u.hdr_crc), crc);
node 588 fs/jffs2/gc.c switch(je16_to_cpu(node->u.nodetype)) {
node 590 fs/jffs2/gc.c crc = crc32(0, node, sizeof(node->i)-8);
node 591 fs/jffs2/gc.c if (je32_to_cpu(node->i.node_crc) != crc) {
node 593 fs/jffs2/gc.c ref_offset(raw), je32_to_cpu(node->i.node_crc), crc);
node 597 fs/jffs2/gc.c if (je32_to_cpu(node->i.dsize)) {
node 598 fs/jffs2/gc.c crc = crc32(0, node->i.data, je32_to_cpu(node->i.csize));
node 599 fs/jffs2/gc.c if (je32_to_cpu(node->i.data_crc) != crc) {
node 601 fs/jffs2/gc.c ref_offset(raw), je32_to_cpu(node->i.data_crc), crc);
node 608 fs/jffs2/gc.c crc = crc32(0, node, sizeof(node->d)-8);
node 609 fs/jffs2/gc.c if (je32_to_cpu(node->d.node_crc) != crc) {
node 611 fs/jffs2/gc.c ref_offset(raw), je32_to_cpu(node->d.node_crc), crc);
node 615 fs/jffs2/gc.c if (strnlen(node->d.name, node->d.nsize) != node->d.nsize) {
node 620 fs/jffs2/gc.c if (node->d.nsize) {
node 621 fs/jffs2/gc.c crc = crc32(0, node->d.name, node->d.nsize);
node 622 fs/jffs2/gc.c if (je32_to_cpu(node->d.name_crc) != crc) {
node 624 fs/jffs2/gc.c ref_offset(raw), je32_to_cpu(node->d.name_crc), crc);
node 633 fs/jffs2/gc.c ref_offset(raw), je16_to_cpu(node->u.nodetype));
node 642 fs/jffs2/gc.c ret = jffs2_flash_write(c, phys_ofs, rawlen, &retlen, (char *)node);
node 689 fs/jffs2/gc.c kfree(node);
node 1069 fs/jffs2/gc.c if (frag->node == fn) {
node 1070 fs/jffs2/gc.c frag->node = new_fn;
node 1142 fs/jffs2/gc.c if (!frag->node || !frag->node->raw) {
node 1152 fs/jffs2/gc.c struct jffs2_raw_node_ref *raw = frag->node->raw;
node 1192 fs/jffs2/gc.c if (!frag->node || !frag->node->raw) {
node 1202 fs/jffs2/gc.c struct jffs2_raw_node_ref *raw = frag->node->raw;
node 292 fs/jffs2/malloc.c xd->node = (void *)xd;
node 311 fs/jffs2/malloc.c ref->node = (void *)ref;
node 92 fs/jffs2/nodelist.c if (frag->node && (frag->ofs & (PAGE_CACHE_SIZE - 1)) == 0) {
node 95 fs/jffs2/nodelist.c frag->node->raw->flash_offset = ref_offset(frag->node->raw) | REF_PRISTINE;
node 103 fs/jffs2/nodelist.c if (this->node) {
node 104 fs/jffs2/nodelist.c this->node->frags--;
node 105 fs/jffs2/nodelist.c if (!this->node->frags) {
node 108 fs/jffs2/nodelist.c ref_offset(this->node->raw), this->node->ofs, this->node->ofs+this->node->size);
node 109 fs/jffs2/nodelist.c jffs2_mark_node_obsolete(c, this->node->raw);
node 110 fs/jffs2/nodelist.c jffs2_free_full_dnode(this->node);
node 113 fs/jffs2/nodelist.c ref_offset(this->node->raw), this->node->ofs, this->node->ofs+this->node->size, this->node->frags);
node 114 fs/jffs2/nodelist.c mark_ref_normal(this->node->raw);
node 156 fs/jffs2/nodelist.c newfrag->node = fn;
node 172 fs/jffs2/nodelist.c if (lastend < newfrag->node->ofs) {
node 176 fs/jffs2/nodelist.c holefrag= new_fragment(NULL, lastend, newfrag->node->ofs - lastend);
node 220 fs/jffs2/nodelist.c this = jffs2_lookup_node_frag(root, newfrag->node->ofs);
node 224 fs/jffs2/nodelist.c this->ofs, this->ofs+this->size, this->node?(ref_offset(this->node->raw)):0xffffffff, this);
node 240 fs/jffs2/nodelist.c if (this->node)
node 241 fs/jffs2/nodelist.c mark_ref_normal(this->node->raw);
node 242 fs/jffs2/nodelist.c mark_ref_normal(newfrag->node->raw);
node 248 fs/jffs2/nodelist.c if (this->node)
node 251 fs/jffs2/nodelist.c ref_offset(this->node->raw), ref_flags(this->node->raw));
node 264 fs/jffs2/nodelist.c mark_ref_normal(newfrag->node->raw);
node 265 fs/jffs2/nodelist.c if (this->node)
node 266 fs/jffs2/nodelist.c mark_ref_normal(this->node->raw);
node 272 fs/jffs2/nodelist.c if (this->node)
node 274 fs/jffs2/nodelist.c this->ofs, this->ofs+this->size, ref_offset(this->node->raw));
node 280 fs/jffs2/nodelist.c newfrag2 = new_fragment(this->node, newfrag->ofs + newfrag->size,
node 284 fs/jffs2/nodelist.c if (this->node)
node 285 fs/jffs2/nodelist.c this->node->frags++;
node 351 fs/jffs2/nodelist.c if (this->node)
node 352 fs/jffs2/nodelist.c mark_ref_normal(this->node->raw);
node 353 fs/jffs2/nodelist.c mark_ref_normal(newfrag->node->raw);
node 373 fs/jffs2/nodelist.c newfrag->node->frags = 1;
node 389 fs/jffs2/nodelist.c if (prev->node)
node 390 fs/jffs2/nodelist.c mark_ref_normal(prev->node->raw);
node 398 fs/jffs2/nodelist.c if (next->node)
node 399 fs/jffs2/nodelist.c mark_ref_normal(next->node->raw);
node 585 fs/jffs2/nodelist.c if (frag->node && !(--frag->node->frags)) {
node 589 fs/jffs2/nodelist.c jffs2_mark_node_obsolete(c, frag->node->raw);
node 591 fs/jffs2/nodelist.c jffs2_free_full_dnode(frag->node);
node 268 fs/jffs2/nodelist.h struct jffs2_full_dnode *node; /* NULL for holes */
node 325 fs/jffs2/nodelist.h struct rb_node *node = rb_first(root);
node 327 fs/jffs2/nodelist.h if (!node)
node 330 fs/jffs2/nodelist.h return rb_entry(node, struct jffs2_node_frag, rb);
node 335 fs/jffs2/nodelist.h struct rb_node *node = rb_last(root);
node 337 fs/jffs2/nodelist.h if (!node)
node 340 fs/jffs2/nodelist.h return rb_entry(node, struct jffs2_node_frag, rb);
node 181 fs/jffs2/read.c } else if (unlikely(!frag->node)) {
node 197 fs/jffs2/read.c ref_offset(frag->node->raw), ref_flags(frag->node->raw)));
node 198 fs/jffs2/read.c ret = jffs2_read_dnode(c, f, frag->node, buf, fragofs + frag->ofs - frag->node->ofs, readlen);
node 384 fs/jffs2/readinode.c struct rb_node *parent = rb_parent(node);
node 388 fs/jffs2/readinode.c BUG_ON(node->rb_right);
node 392 fs/jffs2/readinode.c else if (node == parent->rb_left)
node 397 fs/jffs2/readinode.c *link = node->rb_left;
node 399 fs/jffs2/readinode.c if (node->rb_left)
node 400 fs/jffs2/readinode.c node->rb_left->rb_parent_color = node->rb_parent_color;
node 959 fs/jffs2/readinode.c union jffs2_node_union *node;
node 1030 fs/jffs2/readinode.c node = (union jffs2_node_union *)buf;
node 1033 fs/jffs2/readinode.c if (je32_to_cpu(node->u.hdr_crc) != crc32(0, node, sizeof(node->u)-4)) {
node 1035 fs/jffs2/readinode.c ref_offset(ref), je16_to_cpu(node->u.magic),
node 1036 fs/jffs2/readinode.c je16_to_cpu(node->u.nodetype),
node 1037 fs/jffs2/readinode.c je32_to_cpu(node->u.totlen),
node 1038 fs/jffs2/readinode.c je32_to_cpu(node->u.hdr_crc));
node 1043 fs/jffs2/readinode.c if (je16_to_cpu(node->u.magic) != JFFS2_MAGIC_BITMASK) {
node 1046 fs/jffs2/readinode.c je16_to_cpu(node->u.magic), ref_offset(ref));
node 1051 fs/jffs2/readinode.c switch (je16_to_cpu(node->u.nodetype)) {
node 1062 fs/jffs2/readinode.c err = read_direntry(c, ref, &node->d, retlen, rii);
node 1077 fs/jffs2/readinode.c err = read_dnode(c, ref, &node->i, len, rii);
node 1091 fs/jffs2/readinode.c err = read_unknown(c, ref, &node->u);
node 1300 fs/jffs2/readinode.c f->metadata = frag_first(&f->fragtree)->node;
node 354 fs/jffs2/scan.c raw->next_in_ino = xd->node->next_in_ino;
node 355 fs/jffs2/scan.c xd->node->next_in_ino = raw;
node 434 fs/jffs2/scan.c struct jffs2_unknown_node *node;
node 617 fs/jffs2/scan.c if (jeb->offset + c->sector_size < ofs + sizeof(*node)) {
node 619 fs/jffs2/scan.c jeb->offset, c->sector_size, ofs, sizeof(*node)));
node 625 fs/jffs2/scan.c if (buf_ofs + buf_len < ofs + sizeof(*node)) {
node 635 fs/jffs2/scan.c node = (struct jffs2_unknown_node *)&buf[ofs-buf_ofs];
node 694 fs/jffs2/scan.c if (ofs == jeb->offset && je16_to_cpu(node->magic) == KSAMTIB_CIGAM_2SFFJ) {
node 701 fs/jffs2/scan.c if (je16_to_cpu(node->magic) == JFFS2_DIRTY_BITMASK) {
node 708 fs/jffs2/scan.c if (je16_to_cpu(node->magic) == JFFS2_OLD_MAGIC_BITMASK) {
node 716 fs/jffs2/scan.c if (je16_to_cpu(node->magic) != JFFS2_MAGIC_BITMASK) {
node 720 fs/jffs2/scan.c je16_to_cpu(node->magic));
node 727 fs/jffs2/scan.c crcnode.magic = node->magic;
node 728 fs/jffs2/scan.c crcnode.nodetype = cpu_to_je16( je16_to_cpu(node->nodetype) | JFFS2_NODE_ACCURATE);
node 729 fs/jffs2/scan.c crcnode.totlen = node->totlen;
node 732 fs/jffs2/scan.c if (hdr_crc != je32_to_cpu(node->hdr_crc)) {
node 734 fs/jffs2/scan.c ofs, je16_to_cpu(node->magic),
node 735 fs/jffs2/scan.c je16_to_cpu(node->nodetype),
node 736 fs/jffs2/scan.c je32_to_cpu(node->totlen),
node 737 fs/jffs2/scan.c je32_to_cpu(node->hdr_crc),
node 745 fs/jffs2/scan.c if (ofs + je32_to_cpu(node->totlen) > jeb->offset + c->sector_size) {
node 748 fs/jffs2/scan.c ofs, je32_to_cpu(node->totlen));
node 756 fs/jffs2/scan.c if (!(je16_to_cpu(node->nodetype) & JFFS2_NODE_ACCURATE)) {
node 759 fs/jffs2/scan.c if ((err = jffs2_scan_dirty_space(c, jeb, PAD(je32_to_cpu(node->totlen)))))
node 761 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen));
node 765 fs/jffs2/scan.c switch(je16_to_cpu(node->nodetype)) {
node 775 fs/jffs2/scan.c node = (void *)buf;
node 777 fs/jffs2/scan.c err = jffs2_scan_inode_node(c, jeb, (void *)node, ofs, s);
node 779 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen));
node 783 fs/jffs2/scan.c if (buf_ofs + buf_len < ofs + je32_to_cpu(node->totlen)) {
node 786 fs/jffs2/scan.c je32_to_cpu(node->totlen), buf_len, ofs));
node 791 fs/jffs2/scan.c node = (void *)buf;
node 793 fs/jffs2/scan.c err = jffs2_scan_dirent_node(c, jeb, (void *)node, ofs, s);
node 795 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen));
node 800 fs/jffs2/scan.c if (buf_ofs + buf_len < ofs + je32_to_cpu(node->totlen)) {
node 804 fs/jffs2/scan.c je32_to_cpu(node->totlen), buf_len, ofs));
node 809 fs/jffs2/scan.c node = (void *)buf;
node 811 fs/jffs2/scan.c err = jffs2_scan_xattr_node(c, jeb, (void *)node, ofs, s);
node 814 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen));
node 817 fs/jffs2/scan.c if (buf_ofs + buf_len < ofs + je32_to_cpu(node->totlen)) {
node 821 fs/jffs2/scan.c je32_to_cpu(node->totlen), buf_len, ofs));
node 826 fs/jffs2/scan.c node = (void *)buf;
node 828 fs/jffs2/scan.c err = jffs2_scan_xref_node(c, jeb, (void *)node, ofs, s);
node 831 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen));
node 837 fs/jffs2/scan.c if (je32_to_cpu(node->totlen) != c->cleanmarker_size) {
node 839 fs/jffs2/scan.c ofs, je32_to_cpu(node->totlen), c->cleanmarker_size);
node 857 fs/jffs2/scan.c jffs2_sum_add_padding_mem(s, je32_to_cpu(node->totlen));
node 858 fs/jffs2/scan.c if ((err = jffs2_scan_dirty_space(c, jeb, PAD(je32_to_cpu(node->totlen)))))
node 860 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen));
node 864 fs/jffs2/scan.c switch (je16_to_cpu(node->nodetype) & JFFS2_COMPAT_MASK) {
node 866 fs/jffs2/scan.c printk(KERN_NOTICE "Read-only compatible feature node (0x%04x) found at offset 0x%08x\n", je16_to_cpu(node->nodetype), ofs);
node 870 fs/jffs2/scan.c if ((err = jffs2_scan_dirty_space(c, jeb, PAD(je32_to_cpu(node->totlen)))))
node 872 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen));
node 876 fs/jffs2/scan.c printk(KERN_NOTICE "Incompatible feature node (0x%04x) found at offset 0x%08x\n", je16_to_cpu(node->nodetype), ofs);
node 880 fs/jffs2/scan.c D1(printk(KERN_NOTICE "Unknown but compatible feature node (0x%04x) found at offset 0x%08x\n", je16_to_cpu(node->nodetype), ofs));
node 881 fs/jffs2/scan.c if ((err = jffs2_scan_dirty_space(c, jeb, PAD(je32_to_cpu(node->totlen)))))
node 883 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen));
node 887 fs/jffs2/scan.c D1(printk(KERN_NOTICE "Unknown but compatible feature node (0x%04x) found at offset 0x%08x\n", je16_to_cpu(node->nodetype), ofs));
node 889 fs/jffs2/scan.c jffs2_link_node_ref(c, jeb, ofs | REF_PRISTINE, PAD(je32_to_cpu(node->totlen)), NULL);
node 893 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen));
node 249 fs/jffs2/summary.c union jffs2_node_union *node;
node 257 fs/jffs2/summary.c node = invecs[0].iov_base;
node 261 fs/jffs2/summary.c switch (je16_to_cpu(node->u.nodetype)) {
node 269 fs/jffs2/summary.c temp->nodetype = node->i.nodetype;
node 270 fs/jffs2/summary.c temp->inode = node->i.ino;
node 271 fs/jffs2/summary.c temp->version = node->i.version;
node 273 fs/jffs2/summary.c temp->totlen = node->i.totlen;
node 281 fs/jffs2/summary.c kmalloc(sizeof(struct jffs2_sum_dirent_mem) + node->d.nsize, GFP_KERNEL);
node 286 fs/jffs2/summary.c temp->nodetype = node->d.nodetype;
node 287 fs/jffs2/summary.c temp->totlen = node->d.totlen;
node 289 fs/jffs2/summary.c temp->pino = node->d.pino;
node 290 fs/jffs2/summary.c temp->version = node->d.version;
node 291 fs/jffs2/summary.c temp->ino = node->d.ino;
node 292 fs/jffs2/summary.c temp->nsize = node->d.nsize;
node 293 fs/jffs2/summary.c temp->type = node->d.type;
node 298 fs/jffs2/summary.c memcpy(temp->name,node->d.name,node->d.nsize);
node 302 fs/jffs2/summary.c memcpy(temp->name,invecs[1].iov_base,node->d.nsize);
node 319 fs/jffs2/summary.c temp->nodetype = node->x.nodetype;
node 320 fs/jffs2/summary.c temp->xid = node->x.xid;
node 321 fs/jffs2/summary.c temp->version = node->x.version;
node 322 fs/jffs2/summary.c temp->totlen = node->x.totlen;
node 333 fs/jffs2/summary.c temp->nodetype = node->r.nodetype;
node 342 fs/jffs2/summary.c c->summary->sum_padded += je32_to_cpu(node->u.totlen);
node 505 fs/jffs2/summary.c raw->next_in_ino = xd->node->next_in_ino;
node 506 fs/jffs2/summary.c xd->node->next_in_ino = raw;
node 538 fs/jffs2/summary.c *pseudo_random += ref->node->flash_offset;
node 185 fs/jffs2/wbuf.c node, je16_to_cpu(node->u.magic), je16_to_cpu(node->u.nodetype));
node 187 fs/jffs2/wbuf.c BUG_ON(je16_to_cpu(node->u.magic) != 0x1985 &&
node 188 fs/jffs2/wbuf.c je16_to_cpu(node->u.magic) != 0);
node 190 fs/jffs2/wbuf.c switch (je16_to_cpu(node->u.nodetype)) {
node 196 fs/jffs2/wbuf.c frag = jffs2_lookup_node_frag(&f->fragtree, je32_to_cpu(node->i.offset));
node 199 fs/jffs2/wbuf.c while (!frag->node || frag->node->raw != raw) {
node 203 fs/jffs2/wbuf.c dbg_noderef("Will replace ->raw in full_dnode at %p\n", frag->node);
node 204 fs/jffs2/wbuf.c return &frag->node->raw;
node 217 fs/jffs2/wbuf.c je16_to_cpu(node->u.nodetype));
node 470 fs/jffs2/wbuf.c BUG_ON(xd->node != raw);
node 471 fs/jffs2/wbuf.c adjust_ref = &xd->node;
node 476 fs/jffs2/wbuf.c BUG_ON(xr->node != raw);
node 477 fs/jffs2/wbuf.c adjust_ref = &xr->node;
node 72 fs/jffs2/xattr.c for (raw=xd->node; raw != (void *)xd; raw=raw->next_in_ino) {
node 138 fs/jffs2/xattr.c offset = ref_offset(xd->node);
node 139 fs/jffs2/xattr.c if (ref_flags(xd->node) == REF_PRISTINE)
node 179 fs/jffs2/xattr.c for (raw=xd->node; raw != (void *)xd; raw=raw->next_in_ino) {
node 186 fs/jffs2/xattr.c raw->flash_offset = ref_offset(raw) | ((xd->node==raw) ? REF_PRISTINE : REF_NORMAL);
node 207 fs/jffs2/xattr.c BUG_ON(ref_flags(xd->node) != REF_PRISTINE);
node 215 fs/jffs2/xattr.c ret = jffs2_flash_read(c, ref_offset(xd->node)+sizeof(struct jffs2_raw_xattr),
node 220 fs/jffs2/xattr.c ret, length, readlen, ref_offset(xd->node));
node 230 fs/jffs2/xattr.c ref_offset(xd->node), xd->data_crc, crc);
node 405 fs/jffs2/xattr.c if (xd->node == (void *)xd) {
node 447 fs/jffs2/xattr.c if (ref_flags(ref->node) != REF_UNCHECKED)
node 449 fs/jffs2/xattr.c offset = ref_offset(ref->node);
node 483 fs/jffs2/xattr.c for (raw=ref->node; raw != (void *)ref; raw=raw->next_in_ino) {
node 490 fs/jffs2/xattr.c raw->flash_offset = ref_offset(raw) | ((ref->node==raw) ? REF_PRISTINE : REF_NORMAL);
node 495 fs/jffs2/xattr.c ref->ino, ref->xid, ref_offset(ref->node));
node 783 fs/jffs2/xattr.c if (ref_flags(ref->node) != REF_PRISTINE) {
node 785 fs/jffs2/xattr.c BUG_ON(ref->node->next_in_ino != (void *)ref);
node 786 fs/jffs2/xattr.c ref->node->next_in_ino = NULL;
node 787 fs/jffs2/xattr.c jffs2_mark_node_obsolete(c, ref->node);
node 799 fs/jffs2/xattr.c raw = ref->node;
node 802 fs/jffs2/xattr.c raw->next_in_ino = tmp->node;
node 803 fs/jffs2/xattr.c tmp->node = raw;
node 805 fs/jffs2/xattr.c raw->next_in_ino = tmp->node->next_in_ino;
node 806 fs/jffs2/xattr.c tmp->node->next_in_ino = raw;
node 1206 fs/jffs2/xattr.c if (xd->node != raw)
node 1216 fs/jffs2/xattr.c old_ofs = ref_offset(xd->node);
node 1227 fs/jffs2/xattr.c xd->xid, xd->version, old_ofs, ref_offset(xd->node));
node 1242 fs/jffs2/xattr.c BUG_ON(!ref->node);
node 1244 fs/jffs2/xattr.c if (ref->node != raw)
node 1249 fs/jffs2/xattr.c old_ofs = ref_offset(ref->node);
node 1250 fs/jffs2/xattr.c totlen = ref_totlen(c, c->gcblock, ref->node);
node 1262 fs/jffs2/xattr.c ref->ic->ino, ref->xd->xid, old_ofs, ref_offset(ref->node));
node 1285 fs/jffs2/xattr.c for (raw=xd->node; raw != (void *)xd; raw=raw->next_in_ino) {
node 1293 fs/jffs2/xattr.c | ((xd->node == (void *)raw) ? REF_PRISTINE : REF_NORMAL);
node 1306 fs/jffs2/xattr.c if (atomic_read(&xd->refcnt) || xd->node != (void *)xd)
node 1318 fs/jffs2/xattr.c if (ref->node != (void *)ref)
node 26 fs/jffs2/xattr.h struct jffs2_raw_node_ref *node;
node 48 fs/jffs2/xattr.h struct jffs2_raw_node_ref *node;
node 423 fs/ocfs2/alloc.c struct ocfs2_path_item *node;
node 429 fs/ocfs2/alloc.c node = &path->p_node[i];
node 431 fs/ocfs2/alloc.c brelse(node->bh);
node 432 fs/ocfs2/alloc.c node->bh = NULL;
node 433 fs/ocfs2/alloc.c node->el = NULL;
node 450 fs/ocfs2/cluster/heartbeat.c (f->hc_func)(node, idx, f->hc_data);
node 509 fs/ocfs2/cluster/heartbeat.c event->hn_node = node;
node 522 fs/ocfs2/cluster/heartbeat.c struct o2nm_node *node;
node 524 fs/ocfs2/cluster/heartbeat.c node = o2nm_get_node_by_num(slot->ds_node_num);
node 525 fs/ocfs2/cluster/heartbeat.c if (!node)
node 538 fs/ocfs2/cluster/heartbeat.c o2hb_queue_node_event(&event, O2HB_NODE_DOWN_CB, node,
node 546 fs/ocfs2/cluster/heartbeat.c o2nm_node_put(node);
node 555 fs/ocfs2/cluster/heartbeat.c struct o2nm_node *node;
node 565 fs/ocfs2/cluster/heartbeat.c node = o2nm_get_node_by_num(slot->ds_node_num);
node 566 fs/ocfs2/cluster/heartbeat.c if (!node)
node 638 fs/ocfs2/cluster/heartbeat.c o2hb_queue_node_event(&event, O2HB_NODE_UP_CB, node,
node 684 fs/ocfs2/cluster/heartbeat.c o2hb_queue_node_event(&event, O2HB_NODE_DOWN_CB, node,
node 705 fs/ocfs2/cluster/heartbeat.c o2nm_node_put(node);
node 714 fs/ocfs2/cluster/heartbeat.c int highest, node;
node 717 fs/ocfs2/cluster/heartbeat.c node = -1;
node 718 fs/ocfs2/cluster/heartbeat.c while ((node = find_next_bit(nodes, numbits, node + 1)) != -1) {
node 719 fs/ocfs2/cluster/heartbeat.c if (node >= numbits)
node 722 fs/ocfs2/cluster/heartbeat.c highest = node;
node 41 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node = NULL;
node 47 fs/ocfs2/cluster/nodemanager.c node = o2nm_single_cluster->cl_nodes[node_num];
node 48 fs/ocfs2/cluster/nodemanager.c if (node)
node 49 fs/ocfs2/cluster/nodemanager.c config_item_get(&node->nd_item);
node 52 fs/ocfs2/cluster/nodemanager.c return node;
node 80 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node, *ret = NULL;
node 86 fs/ocfs2/cluster/nodemanager.c node = rb_entry(parent, struct o2nm_node, nd_ip_node);
node 88 fs/ocfs2/cluster/nodemanager.c cmp = memcmp(&ip_needle, &node->nd_ipv4_address,
node 95 fs/ocfs2/cluster/nodemanager.c ret = node;
node 110 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node = NULL;
node 117 fs/ocfs2/cluster/nodemanager.c node = o2nm_node_ip_tree_lookup(cluster, addr, NULL, NULL);
node 118 fs/ocfs2/cluster/nodemanager.c if (node)
node 119 fs/ocfs2/cluster/nodemanager.c config_item_get(&node->nd_item);
node 123 fs/ocfs2/cluster/nodemanager.c return node;
node 129 fs/ocfs2/cluster/nodemanager.c config_item_put(&node->nd_item);
node 135 fs/ocfs2/cluster/nodemanager.c config_item_get(&node->nd_item);
node 167 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node = to_o2nm_node(item);
node 168 fs/ocfs2/cluster/nodemanager.c kfree(node);
node 173 fs/ocfs2/cluster/nodemanager.c return sprintf(page, "%d\n", node->nd_num);
node 180 fs/ocfs2/cluster/nodemanager.c return to_o2nm_cluster(node->nd_item.ci_parent->ci_parent);
node 193 fs/ocfs2/cluster/nodemanager.c struct o2nm_cluster *cluster = to_o2nm_cluster_from_node(node);
node 208 fs/ocfs2/cluster/nodemanager.c if (!test_bit(O2NM_NODE_ATTR_ADDRESS, &node->nd_set_attributes) ||
node 209 fs/ocfs2/cluster/nodemanager.c !test_bit(O2NM_NODE_ATTR_PORT, &node->nd_set_attributes))
node 216 fs/ocfs2/cluster/nodemanager.c cluster->cl_nodes[tmp] = node;
node 217 fs/ocfs2/cluster/nodemanager.c node->nd_num = tmp;
node 228 fs/ocfs2/cluster/nodemanager.c return sprintf(page, "%u\n", ntohs(node->nd_ipv4_port));
node 246 fs/ocfs2/cluster/nodemanager.c node->nd_ipv4_port = htons(tmp);
node 253 fs/ocfs2/cluster/nodemanager.c return sprintf(page, "%u.%u.%u.%u\n", NIPQUAD(node->nd_ipv4_address));
node 260 fs/ocfs2/cluster/nodemanager.c struct o2nm_cluster *cluster = to_o2nm_cluster_from_node(node);
node 282 fs/ocfs2/cluster/nodemanager.c rb_link_node(&node->nd_ip_node, parent, p);
node 283 fs/ocfs2/cluster/nodemanager.c rb_insert_color(&node->nd_ip_node, &cluster->cl_node_ip_tree);
node 289 fs/ocfs2/cluster/nodemanager.c memcpy(&node->nd_ipv4_address, &ipv4_addr, sizeof(ipv4_addr));
node 296 fs/ocfs2/cluster/nodemanager.c return sprintf(page, "%d\n", node->nd_local);
node 302 fs/ocfs2/cluster/nodemanager.c struct o2nm_cluster *cluster = to_o2nm_cluster_from_node(node);
node 315 fs/ocfs2/cluster/nodemanager.c if (!test_bit(O2NM_NODE_ATTR_ADDRESS, &node->nd_set_attributes) ||
node 316 fs/ocfs2/cluster/nodemanager.c !test_bit(O2NM_NODE_ATTR_NUM, &node->nd_set_attributes) ||
node 317 fs/ocfs2/cluster/nodemanager.c !test_bit(O2NM_NODE_ATTR_PORT, &node->nd_set_attributes))
node 323 fs/ocfs2/cluster/nodemanager.c cluster->cl_local_node != node->nd_num)
node 328 fs/ocfs2/cluster/nodemanager.c ret = o2net_start_listening(node);
node 334 fs/ocfs2/cluster/nodemanager.c cluster->cl_local_node == node->nd_num) {
node 335 fs/ocfs2/cluster/nodemanager.c o2net_stop_listening(node);
node 339 fs/ocfs2/cluster/nodemanager.c node->nd_local = tmp;
node 340 fs/ocfs2/cluster/nodemanager.c if (node->nd_local) {
node 342 fs/ocfs2/cluster/nodemanager.c cluster->cl_local_node = node->nd_num;
node 409 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node = to_o2nm_node(item);
node 415 fs/ocfs2/cluster/nodemanager.c ret = o2nm_node_attr->show(node, page);
node 423 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node = to_o2nm_node(item);
node 434 fs/ocfs2/cluster/nodemanager.c if (test_bit(attr_index, &node->nd_set_attributes))
node 437 fs/ocfs2/cluster/nodemanager.c ret = o2nm_node_attr->store(node, page, count);
node 441 fs/ocfs2/cluster/nodemanager.c set_bit(attr_index, &node->nd_set_attributes);
node 650 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node = NULL;
node 655 fs/ocfs2/cluster/nodemanager.c node = kzalloc(sizeof(struct o2nm_node), GFP_KERNEL);
node 656 fs/ocfs2/cluster/nodemanager.c if (node == NULL)
node 659 fs/ocfs2/cluster/nodemanager.c strcpy(node->nd_name, name); /* use item.ci_namebuf instead? */
node 660 fs/ocfs2/cluster/nodemanager.c config_item_init_type_name(&node->nd_item, name, &o2nm_node_type);
node 661 fs/ocfs2/cluster/nodemanager.c spin_lock_init(&node->nd_lock);
node 663 fs/ocfs2/cluster/nodemanager.c return &node->nd_item;
node 669 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node = to_o2nm_node(item);
node 672 fs/ocfs2/cluster/nodemanager.c o2net_disconnect_node(node);
node 675 fs/ocfs2/cluster/nodemanager.c (cluster->cl_local_node == node->nd_num)) {
node 678 fs/ocfs2/cluster/nodemanager.c o2net_stop_listening(node);
node 686 fs/ocfs2/cluster/nodemanager.c if (node->nd_ipv4_address)
node 687 fs/ocfs2/cluster/nodemanager.c rb_erase(&node->nd_ip_node, &cluster->cl_node_ip_tree);
node 690 fs/ocfs2/cluster/nodemanager.c if (cluster->cl_nodes[node->nd_num] == node) {
node 691 fs/ocfs2/cluster/nodemanager.c cluster->cl_nodes[node->nd_num] = NULL;
node 692 fs/ocfs2/cluster/nodemanager.c clear_bit(node->nd_num, cluster->cl_nodes_bitmap);
node 161 fs/ocfs2/cluster/quorum.c if (!test_and_set_bit(node, qs->qs_hold_bm)) {
node 164 fs/ocfs2/cluster/quorum.c "node %u\n", node);
node 165 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_holds);
node 173 fs/ocfs2/cluster/quorum.c if (test_and_clear_bit(node, qs->qs_hold_bm)) {
node 174 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_holds - 1);
node 182 fs/ocfs2/cluster/quorum.c node, qs->qs_holds);
node 198 fs/ocfs2/cluster/quorum.c "node %u\n", node);
node 199 fs/ocfs2/cluster/quorum.c mlog_bug_on_msg(test_bit(node, qs->qs_hb_bm), "node %u\n", node);
node 200 fs/ocfs2/cluster/quorum.c set_bit(node, qs->qs_hb_bm);
node 202 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_heartbeating);
node 204 fs/ocfs2/cluster/quorum.c if (!test_bit(node, qs->qs_conn_bm))
node 205 fs/ocfs2/cluster/quorum.c o2quo_set_hold(qs, node);
node 207 fs/ocfs2/cluster/quorum.c o2quo_clear_hold(qs, node);
node 223 fs/ocfs2/cluster/quorum.c node, qs->qs_heartbeating);
node 224 fs/ocfs2/cluster/quorum.c mlog_bug_on_msg(!test_bit(node, qs->qs_hb_bm), "node %u\n", node);
node 225 fs/ocfs2/cluster/quorum.c clear_bit(node, qs->qs_hb_bm);
node 227 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_heartbeating);
node 229 fs/ocfs2/cluster/quorum.c o2quo_clear_hold(qs, node);
node 245 fs/ocfs2/cluster/quorum.c mlog(0, "node %u\n", node);
node 248 fs/ocfs2/cluster/quorum.c o2quo_clear_hold(qs, node);
node 266 fs/ocfs2/cluster/quorum.c "node %u\n", node);
node 267 fs/ocfs2/cluster/quorum.c mlog_bug_on_msg(test_bit(node, qs->qs_conn_bm), "node %u\n", node);
node 268 fs/ocfs2/cluster/quorum.c set_bit(node, qs->qs_conn_bm);
node 270 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_connected);
node 272 fs/ocfs2/cluster/quorum.c if (!test_bit(node, qs->qs_hb_bm))
node 273 fs/ocfs2/cluster/quorum.c o2quo_set_hold(qs, node);
node 275 fs/ocfs2/cluster/quorum.c o2quo_clear_hold(qs, node);
node 290 fs/ocfs2/cluster/quorum.c if (test_bit(node, qs->qs_conn_bm)) {
node 294 fs/ocfs2/cluster/quorum.c node, qs->qs_connected);
node 296 fs/ocfs2/cluster/quorum.c clear_bit(node, qs->qs_conn_bm);
node 299 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_connected);
node 301 fs/ocfs2/cluster/quorum.c if (test_bit(node, qs->qs_hb_bm))
node 302 fs/ocfs2/cluster/quorum.c o2quo_set_hold(qs, node);
node 153 fs/ocfs2/cluster/tcp.c nst->st_node = node;
node 386 fs/ocfs2/cluster/tcp.c o2nm_node_get(node);
node 387 fs/ocfs2/cluster/tcp.c sc->sc_node = node;
node 1536 fs/ocfs2/cluster/tcp.c struct o2nm_node *node = NULL, *mynode = NULL;
node 1547 fs/ocfs2/cluster/tcp.c node = o2nm_get_node_by_num(o2net_num_from_nn(nn));
node 1548 fs/ocfs2/cluster/tcp.c if (node == NULL) {
node 1577 fs/ocfs2/cluster/tcp.c sc = sc_alloc(node);
node 1619 fs/ocfs2/cluster/tcp.c remoteaddr.sin_addr.s_addr = node->nd_ipv4_address;
node 1620 fs/ocfs2/cluster/tcp.c remoteaddr.sin_port = node->nd_ipv4_port;
node 1640 fs/ocfs2/cluster/tcp.c if (node)
node 1641 fs/ocfs2/cluster/tcp.c o2nm_node_put(node);
node 1678 fs/ocfs2/cluster/tcp.c struct o2net_node *nn = o2net_nn_from_num(node->nd_num);
node 1700 fs/ocfs2/cluster/tcp.c o2net_disconnect_node(node);
node 1761 fs/ocfs2/cluster/tcp.c struct o2nm_node *node = NULL;
node 1791 fs/ocfs2/cluster/tcp.c node = o2nm_get_node_by_ip(sin.sin_addr.s_addr);
node 1792 fs/ocfs2/cluster/tcp.c if (node == NULL) {
node 1800 fs/ocfs2/cluster/tcp.c if (o2nm_this_node() > node->nd_num) {
node 1803 fs/ocfs2/cluster/tcp.c node->nd_name, NIPQUAD(sin.sin_addr.s_addr),
node 1804 fs/ocfs2/cluster/tcp.c ntohs(sin.sin_port), node->nd_num);
node 1811 fs/ocfs2/cluster/tcp.c if (!o2hb_check_node_heartbeating_from_callback(node->nd_num)) {
node 1814 fs/ocfs2/cluster/tcp.c node->nd_name, NIPQUAD(sin.sin_addr.s_addr),
node 1820 fs/ocfs2/cluster/tcp.c nn = o2net_nn_from_num(node->nd_num);
node 1831 fs/ocfs2/cluster/tcp.c node->nd_name, NIPQUAD(sin.sin_addr.s_addr),
node 1836 fs/ocfs2/cluster/tcp.c sc = sc_alloc(node);
node 1859 fs/ocfs2/cluster/tcp.c if (node)
node 1860 fs/ocfs2/cluster/tcp.c o2nm_node_put(node);
node 1968 fs/ocfs2/cluster/tcp.c ret = o2net_open_listening_sock(node->nd_ipv4_address,
node 1969 fs/ocfs2/cluster/tcp.c node->nd_ipv4_port);
node 1974 fs/ocfs2/cluster/tcp.c o2quo_conn_up(node->nd_num);
node 1996 fs/ocfs2/cluster/tcp.c struct o2nm_node *node = o2nm_get_node_by_num(i);
node 1997 fs/ocfs2/cluster/tcp.c if (node) {
node 1998 fs/ocfs2/cluster/tcp.c o2net_disconnect_node(node);
node 1999 fs/ocfs2/cluster/tcp.c o2nm_node_put(node);
node 2011 fs/ocfs2/cluster/tcp.c o2quo_conn_err(node->nd_num);
node 196 fs/ocfs2/dlm/dlmast.c lock->ml.node == dlm->node_num ? "master" :
node 224 fs/ocfs2/dlm/dlmast.c BUG_ON(lock->ml.node != dlm->node_num);
node 241 fs/ocfs2/dlm/dlmast.c BUG_ON(lock->ml.node == dlm->node_num);
node 258 fs/ocfs2/dlm/dlmast.c BUG_ON(lock->ml.node != dlm->node_num);
node 432 fs/ocfs2/dlm/dlmast.c res->lockname.len, res->lockname.name, lock->ml.node,
node 454 fs/ocfs2/dlm/dlmast.c lock->ml.node, &status);
node 460 fs/ocfs2/dlm/dlmast.c "node is dead!\n", lock->ml.node);
node 464 fs/ocfs2/dlm/dlmast.c "DLM_MIGRATING!\n", lock->ml.node);
node 468 fs/ocfs2/dlm/dlmast.c lock->ml.node, status);
node 267 fs/ocfs2/dlm/dlmcommon.h dlm->joining_node = node;
node 343 fs/ocfs2/dlm/dlmcommon.h u8 node;
node 214 fs/ocfs2/dlm/dlmconvert.c if (lock->ml.node == dlm->node_num)
node 477 fs/ocfs2/dlm/dlmconvert.c lock->ml.node == cnv->node_idx) {
node 81 fs/ocfs2/dlm/dlmdebug.c lock->ml.type, lock->ml.convert_type, lock->ml.node,
node 562 fs/ocfs2/dlm/dlmdebug.c lock->ml.node,
node 762 fs/ocfs2/dlm/dlmdebug.c struct dlm_reco_node_data *node;
node 857 fs/ocfs2/dlm/dlmdebug.c list_for_each_entry(node, &dlm->reco.node_data, list) {
node 858 fs/ocfs2/dlm/dlmdebug.c switch (node->state) {
node 885 fs/ocfs2/dlm/dlmdebug.c node->node_num, state);
node 508 fs/ocfs2/dlm/dlmdomain.c int node = -1;
node 514 fs/ocfs2/dlm/dlmdomain.c while ((node = find_next_bit(dlm->domain_map, O2NM_MAX_NODES,
node 515 fs/ocfs2/dlm/dlmdomain.c node + 1)) < O2NM_MAX_NODES) {
node 516 fs/ocfs2/dlm/dlmdomain.c printk("%d ", node);
node 525 fs/ocfs2/dlm/dlmdomain.c unsigned int node;
node 533 fs/ocfs2/dlm/dlmdomain.c node = exit_msg->node_idx;
node 535 fs/ocfs2/dlm/dlmdomain.c printk(KERN_INFO "ocfs2_dlm: Node %u leaves domain %s\n", node, dlm->name);
node 538 fs/ocfs2/dlm/dlmdomain.c clear_bit(node, dlm->domain_map);
node 542 fs/ocfs2/dlm/dlmdomain.c dlm_hb_event_notify_attached(dlm, node, 0);
node 558 fs/ocfs2/dlm/dlmdomain.c node, dlm->name, dlm->node_num);
node 564 fs/ocfs2/dlm/dlmdomain.c &leave_msg, sizeof(leave_msg), node,
node 575 fs/ocfs2/dlm/dlmdomain.c int node, clear_node, status;
node 585 fs/ocfs2/dlm/dlmdomain.c while ((node = find_next_bit(dlm->domain_map, O2NM_MAX_NODES,
node 597 fs/ocfs2/dlm/dlmdomain.c status = dlm_send_one_domain_exit(dlm, node);
node 602 fs/ocfs2/dlm/dlmdomain.c "to node %d\n", status, node);
node 615 fs/ocfs2/dlm/dlmdomain.c clear_bit(node, dlm->domain_map);
node 709 fs/ocfs2/dlm/dlmdomain.c node, proto_type,
node 719 fs/ocfs2/dlm/dlmdomain.c node, proto_type,
node 960 fs/ocfs2/dlm/dlmdomain.c &cancel_msg, sizeof(cancel_msg), node,
node 977 fs/ocfs2/dlm/dlmdomain.c unsigned int node;
node 988 fs/ocfs2/dlm/dlmdomain.c node = -1;
node 989 fs/ocfs2/dlm/dlmdomain.c while ((node = find_next_bit(node_map, O2NM_MAX_NODES,
node 990 fs/ocfs2/dlm/dlmdomain.c node + 1)) < O2NM_MAX_NODES) {
node 991 fs/ocfs2/dlm/dlmdomain.c if (node == dlm->node_num)
node 994 fs/ocfs2/dlm/dlmdomain.c tmpstat = dlm_send_one_join_cancel(dlm, node);
node 997 fs/ocfs2/dlm/dlmdomain.c "node %d\n", tmpstat, node);
node 1017 fs/ocfs2/dlm/dlmdomain.c mlog(0, "querying node %d\n", node);
node 1030 fs/ocfs2/dlm/dlmdomain.c sizeof(join_msg), node,
node 1058 fs/ocfs2/dlm/dlmdomain.c node);
node 1069 fs/ocfs2/dlm/dlmdomain.c node,
node 1077 fs/ocfs2/dlm/dlmdomain.c packet.code, node);
node 1080 fs/ocfs2/dlm/dlmdomain.c mlog(0, "status %d, node %d response is %d\n", status, node,
node 1093 fs/ocfs2/dlm/dlmdomain.c mlog(0, "Sending join assert to node %u\n", node);
node 1101 fs/ocfs2/dlm/dlmdomain.c &assert_msg, sizeof(assert_msg), node,
node 1112 fs/ocfs2/dlm/dlmdomain.c int status, node, live;
node 1115 fs/ocfs2/dlm/dlmdomain.c node = -1;
node 1116 fs/ocfs2/dlm/dlmdomain.c while ((node = find_next_bit(node_map, O2NM_MAX_NODES,
node 1117 fs/ocfs2/dlm/dlmdomain.c node + 1)) < O2NM_MAX_NODES) {
node 1118 fs/ocfs2/dlm/dlmdomain.c if (node == dlm->node_num)
node 1125 fs/ocfs2/dlm/dlmdomain.c status = dlm_send_one_join_assert(dlm, node);
node 1128 fs/ocfs2/dlm/dlmdomain.c live = test_bit(node, dlm->live_nodes_map);
node 1133 fs/ocfs2/dlm/dlmdomain.c "join on node %d\n", status, node);
node 1174 fs/ocfs2/dlm/dlmdomain.c int status = 0, tmpstat, node;
node 1199 fs/ocfs2/dlm/dlmdomain.c node = -1;
node 1200 fs/ocfs2/dlm/dlmdomain.c while ((node = find_next_bit(ctxt->live_map, O2NM_MAX_NODES,
node 1201 fs/ocfs2/dlm/dlmdomain.c node + 1)) < O2NM_MAX_NODES) {
node 1202 fs/ocfs2/dlm/dlmdomain.c if (node == dlm->node_num)
node 1205 fs/ocfs2/dlm/dlmdomain.c status = dlm_request_join(dlm, node, &response);
node 1214 fs/ocfs2/dlm/dlmdomain.c set_bit(node, ctxt->yes_resp_map);
node 136 fs/ocfs2/dlm/dlmlock.c lock->ml.node != dlm->node_num) {
node 165 fs/ocfs2/dlm/dlmlock.c lock->ml.node);
node 176 fs/ocfs2/dlm/dlmlock.c lock->ml.node);
node 186 fs/ocfs2/dlm/dlmlock.c if (lock->ml.node == dlm->node_num)
node 408 fs/ocfs2/dlm/dlmlock.c newlock->ml.node = node;
node 447 fs/ocfs2/dlm/dlmlock.c dlm_init_lock(lock, type, node, cookie);
node 1196 fs/ocfs2/dlm/dlmmaster.c int node;
node 1205 fs/ocfs2/dlm/dlmmaster.c node = dlm_bitmap_diff_iter_next(&bdi, &sc);
node 1206 fs/ocfs2/dlm/dlmmaster.c while (node >= 0) {
node 1211 fs/ocfs2/dlm/dlmmaster.c mlog(ML_NOTICE, "node %d up while restarting\n", node);
node 1215 fs/ocfs2/dlm/dlmmaster.c clear_bit(node, mle->response_map);
node 1216 fs/ocfs2/dlm/dlmmaster.c set_bit(node, mle->vote_map);
node 1218 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "node down! %d\n", node);
node 1224 fs/ocfs2/dlm/dlmmaster.c clear_bit(node, mle->maybe_map);
node 1226 fs/ocfs2/dlm/dlmmaster.c if (node == lowest) {
node 1229 fs/ocfs2/dlm/dlmmaster.c "waiting on it!\n", node);
node 1273 fs/ocfs2/dlm/dlmmaster.c node = dlm_bitmap_diff_iter_next(&bdi, &sc);
node 2233 fs/ocfs2/dlm/dlmmaster.c u8 node;
node 2244 fs/ocfs2/dlm/dlmmaster.c node = deref->node_idx;
node 2251 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "Invalid node number: %u\n", node);
node 2272 fs/ocfs2/dlm/dlmmaster.c if (test_bit(node, res->refmap)) {
node 2273 fs/ocfs2/dlm/dlmmaster.c dlm_lockres_clear_refmap_bit(node, res);
node 2285 fs/ocfs2/dlm/dlmmaster.c res->lockname.len, res->lockname.name, node);
node 2301 fs/ocfs2/dlm/dlmmaster.c item->u.dl.deref_node = node;
node 2322 fs/ocfs2/dlm/dlmmaster.c u8 node;
node 2327 fs/ocfs2/dlm/dlmmaster.c node = item->u.dl.deref_node;
node 2331 fs/ocfs2/dlm/dlmmaster.c if (test_bit(node, res->refmap)) {
node 2333 fs/ocfs2/dlm/dlmmaster.c dlm_lockres_clear_refmap_bit(node, res);
node 2340 fs/ocfs2/dlm/dlmmaster.c dlm->name, res->lockname.len, res->lockname.name, node);
node 2345 fs/ocfs2/dlm/dlmmaster.c res->lockname.len, res->lockname.name, node);
node 2383 fs/ocfs2/dlm/dlmmaster.c if (lock->ml.node == dlm->node_num) {
node 2860 fs/ocfs2/dlm/dlmmaster.c if (lock->ml.node != dlm->node_num) {
node 2862 fs/ocfs2/dlm/dlmmaster.c lock->ml.node);
node 2868 fs/ocfs2/dlm/dlmmaster.c dlm_lockres_clear_refmap_bit(lock->ml.node, res);
node 2914 fs/ocfs2/dlm/dlmmaster.c if (lock->ml.node != dlm->node_num) {
node 2916 fs/ocfs2/dlm/dlmmaster.c return lock->ml.node;
node 349 fs/ocfs2/dlm/dlmrecovery.c dead = !test_bit(node, dlm->domain_map);
node 360 fs/ocfs2/dlm/dlmrecovery.c recovered = !test_bit(node, dlm->recovery_map);
node 370 fs/ocfs2/dlm/dlmrecovery.c "death of node %u\n", dlm->name, timeout, node);
node 372 fs/ocfs2/dlm/dlmrecovery.c dlm_is_node_dead(dlm, node),
node 376 fs/ocfs2/dlm/dlmrecovery.c "of death of node %u\n", dlm->name, node);
node 378 fs/ocfs2/dlm/dlmrecovery.c dlm_is_node_dead(dlm, node));
node 388 fs/ocfs2/dlm/dlmrecovery.c "recovery of node %u\n", dlm->name, timeout, node);
node 390 fs/ocfs2/dlm/dlmrecovery.c dlm_is_node_recovered(dlm, node),
node 394 fs/ocfs2/dlm/dlmrecovery.c "of recovery of node %u\n", dlm->name, node);
node 396 fs/ocfs2/dlm/dlmrecovery.c dlm_is_node_recovered(dlm, node));
node 1051 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == dead_node) {
node 1200 fs/ocfs2/dlm/dlmrecovery.c ml->node = lock->ml.node;
node 1218 fs/ocfs2/dlm/dlmrecovery.c dummy.ml.node = dlm->node_num;
node 1231 fs/ocfs2/dlm/dlmrecovery.c *nodenum = ml->node;
node 1761 fs/ocfs2/dlm/dlmrecovery.c if (ml->node == dlm->node_num) {
node 1789 fs/ocfs2/dlm/dlmrecovery.c BUG_ON(lock->ml.node != ml->node);
node 1812 fs/ocfs2/dlm/dlmrecovery.c newlock = dlm_new_lock(ml->type, ml->node,
node 1899 fs/ocfs2/dlm/dlmrecovery.c ml->type, ml->convert_type, ml->node,
node 1914 fs/ocfs2/dlm/dlmrecovery.c res->lockname.len, res->lockname.name, ml->node);
node 1915 fs/ocfs2/dlm/dlmrecovery.c dlm_lockres_set_refmap_bit(ml->node, res);
node 2131 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == search_node) {
node 2166 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == dead_node) {
node 2175 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == dead_node) {
node 2184 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == dead_node) {
node 2252 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == dead_node) {
node 59 fs/ocfs2/dlm/dlmthread.c #define dlm_lock_is_remote(dlm, lock) ((lock)->ml.node != (dlm)->node_num)
node 356 fs/ocfs2/dlm/dlmthread.c target->ml.convert_type, target->ml.node);
node 418 fs/ocfs2/dlm/dlmthread.c target->ml.type, target->ml.node);
node 533 fs/ocfs2/dlm/dlmthread.c if (lock->ml.node != dlm->node_num) {
node 581 fs/ocfs2/dlm/dlmthread.c if (lock->ml.node != dlm->node_num) {
node 465 fs/ocfs2/dlm/dlmunlock.c lock->ml.node == unlock->node_idx) {
node 346 fs/ocfs2/stack_o2cb.c *node = node_num;
node 439 fs/ocfs2/stackglue.c return active_stack->sp_ops->this_node(node);
node 150 fs/ocfs2/stackglue.h int (*this_node)(unsigned int *node);
node 92 fs/ocfs2/uptodate.c struct rb_node *node;
node 95 fs/ocfs2/uptodate.c while ((node = rb_last(root)) != NULL) {
node 96 fs/ocfs2/uptodate.c item = rb_entry(node, struct ocfs2_meta_cache_item, c_node);
node 31 fs/openpromfs/inode.c struct device_node *node;
node 192 fs/openpromfs/inode.c dp = oi->u.node;
node 206 fs/openpromfs/inode.c ent_data.node = child;
node 269 fs/openpromfs/inode.c struct device_node *dp = oi->u.node;
node 405 fs/openpromfs/inode.c oi->u.node = of_find_node_by_path("/");
node 255 fs/ubifs/debug.c const struct ubifs_ch *ch = node;
node 264 fs/ubifs/debug.c (void *)node, UBIFS_CH_SZ, 1);
node 269 fs/ubifs/debug.c dump_ch(node);
node 274 fs/ubifs/debug.c const struct ubifs_pad_node *pad = node;
node 282 fs/ubifs/debug.c const struct ubifs_sb_node *sup = node;
node 337 fs/ubifs/debug.c const struct ubifs_mst_node *mst = node;
node 399 fs/ubifs/debug.c const struct ubifs_ref_node *ref = node;
node 411 fs/ubifs/debug.c const struct ubifs_ino_node *ino = node;
node 453 fs/ubifs/debug.c const struct ubifs_dent_node *dent = node;
node 477 fs/ubifs/debug.c const struct ubifs_data_node *dn = node;
node 495 fs/ubifs/debug.c const struct ubifs_trun_node *trun = node;
node 507 fs/ubifs/debug.c const struct ubifs_idx_node *idx = node;
node 530 fs/ubifs/debug.c const struct ubifs_orph_node *orph = node;
node 673 fs/ubifs/debug.c dbg_dump_node(c, snod->node);
node 1682 fs/ubifs/debug.c void *node;
node 1693 fs/ubifs/debug.c node = kmalloc(zbr->len, GFP_NOFS);
node 1694 fs/ubifs/debug.c if (!node)
node 1697 fs/ubifs/debug.c err = ubifs_tnc_read_node(c, zbr, node);
node 1706 fs/ubifs/debug.c fscki = add_inode(c, priv, node);
node 1723 fs/ubifs/debug.c ch = node;
node 1733 fs/ubifs/debug.c struct ubifs_data_node *dn = node;
node 1761 fs/ubifs/debug.c struct ubifs_dent_node *dent = node;
node 1808 fs/ubifs/debug.c kfree(node);
node 1813 fs/ubifs/debug.c dbg_dump_node(c, node);
node 1815 fs/ubifs/debug.c kfree(node);
node 197 fs/ubifs/gc.c err = ubifs_wbuf_write_nolock(wbuf, snod->node,
node 299 fs/ubifs/gc.c struct ubifs_idx_node *idx = snod->node;
node 228 fs/ubifs/io.c struct ubifs_ch *ch = node;
node 238 fs/ubifs/io.c crc = crc32(UBIFS_CRC32_INIT, node + 8, len - 8);
node 244 fs/ubifs/io.c ubifs_pad(c, node + len, pad);
node 261 fs/ubifs/io.c struct ubifs_ch *ch = node;
node 274 fs/ubifs/io.c crc = crc32(UBIFS_CRC32_INIT, node + 8, len - 8);
node 260 fs/ubifs/journal.c ubifs_prepare_node(c, node, len, 0);
node 262 fs/ubifs/journal.c return ubifs_wbuf_write_nolock(wbuf, node, len);
node 655 fs/ubifs/log.c struct ubifs_ch *ch = node;
node 668 fs/ubifs/log.c memcpy(buf + *offs, node, len);
node 707 fs/ubifs/log.c struct ubifs_ref_node *ref = snod->node;
node 715 fs/ubifs/log.c &offs, snod->node);
node 725 fs/ubifs/log.c snod->node);
node 1166 fs/ubifs/lprops.c struct ubifs_idx_node *idx = snod->node;
node 52 fs/ubifs/master.c memcpy(c->mst_node, snod->node, snod->len);
node 72 fs/ubifs/master.c (void *)snod->node + UBIFS_CH_SZ,
node 310 fs/ubifs/misc.h return ubifs_tnc_locate(c, key, node, NULL, NULL);
node 570 fs/ubifs/orphan.c dbg_dump_node(c, snod->node);
node 574 fs/ubifs/orphan.c orph = snod->node;
node 598 fs/ubifs/orphan.c dbg_dump_node(c, snod->node);
node 735 fs/ubifs/orphan.c struct ubifs_ino_node *node;
node 847 fs/ubifs/orphan.c err = ubifs_tnc_read_node(c, zbr, ci->node);
node 852 fs/ubifs/orphan.c if (ci->node->nlink == 0)
node 875 fs/ubifs/orphan.c orph = snod->node;
node 926 fs/ubifs/orphan.c ci.node = kmalloc(UBIFS_MAX_INO_NODE_SZ, GFP_NOFS);
node 927 fs/ubifs/orphan.c if (!ci.node) {
node 954 fs/ubifs/orphan.c kfree(ci.node);
node 570 fs/ubifs/recovery.c ch = snod->node;
node 551 fs/ubifs/replay.c struct ubifs_ino_node *ino = snod->node;
node 563 fs/ubifs/replay.c struct ubifs_data_node *dn = snod->node;
node 576 fs/ubifs/replay.c struct ubifs_dent_node *dent = snod->node;
node 590 fs/ubifs/replay.c struct ubifs_trun_node *trun = snod->node;
node 644 fs/ubifs/replay.c dbg_dump_node(c, snod->node);
node 836 fs/ubifs/replay.c const struct ubifs_cs_node *node;
node 852 fs/ubifs/replay.c node = sleb->buf;
node 868 fs/ubifs/replay.c if (le64_to_cpu(node->cmt_no) != c->cmt_no) {
node 872 fs/ubifs/replay.c (unsigned long long)le64_to_cpu(node->cmt_no),
node 877 fs/ubifs/replay.c c->cs_sqnum = le64_to_cpu(node->ch.sqnum);
node 919 fs/ubifs/replay.c const struct ubifs_ref_node *ref = snod->node;
node 962 fs/ubifs/replay.c dbg_dump_node(c, snod->node);
node 209 fs/ubifs/scan.c snod->node = buf;
node 352 fs/ubifs/scan.c struct ubifs_scan_node *node;
node 357 fs/ubifs/scan.c node = list_entry(head->next, struct ubifs_scan_node, list);
node 358 fs/ubifs/scan.c list_del(&node->list);
node 359 fs/ubifs/scan.c kfree(node);
node 333 fs/ubifs/tnc.c const struct ubifs_dent_node *dent = node;
node 351 fs/ubifs/tnc.c memcpy(lnc_node, node, zbr->len);
node 373 fs/ubifs/tnc.c err = ubifs_validate_entry(c, node);
node 376 fs/ubifs/tnc.c dbg_dump_node(c, node);
node 380 fs/ubifs/tnc.c zbr->leaf = node;
node 418 fs/ubifs/tnc.c memcpy(node, zbr->leaf, zbr->len);
node 422 fs/ubifs/tnc.c err = ubifs_tnc_read_node(c, zbr, node);
node 427 fs/ubifs/tnc.c err = lnc_add(c, zbr, node);
node 498 fs/ubifs/tnc.c ret = try_read_node(c, node, key_type(c, key), zbr->len, zbr->lnum,
node 502 fs/ubifs/tnc.c struct ubifs_dent_node *dent = node;
node 1460 fs/ubifs/tnc.c err = tnc_read_node_nm(c, zt, node);
node 1464 fs/ubifs/tnc.c err = ubifs_tnc_read_node(c, zt, node);
node 1474 fs/ubifs/tnc.c err = ubifs_tnc_read_node(c, &zbr, node);
node 1478 fs/ubifs/tnc.c err = fallible_read_node(c, key, &zbr, node);
node 1535 fs/ubifs/tnc.c err = tnc_read_node_nm(c, &znode->zbranch[n], node);
node 1559 fs/ubifs/tnc.c const struct ubifs_dent_node *dent = node;
node 1565 fs/ubifs/tnc.c err = ubifs_tnc_lookup(c, key, node);
node 1577 fs/ubifs/tnc.c return do_lookup_nm(c, key, node, nm);
node 258 fs/ubifs/tnc_commit.c idx = snod->node;
node 471 fs/ubifs/tnc_misc.c err = ubifs_read_node_wbuf(wbuf, node, type, zbr->len,
node 474 fs/ubifs/tnc_misc.c err = ubifs_read_node(c, node, type, zbr->len, zbr->lnum,
node 484 fs/ubifs/tnc_misc.c if (memcmp(node + UBIFS_KEY_OFFSET, &key1, c->key_len)) {
node 489 fs/ubifs/tnc_misc.c dbg_dump_node(c, node);
node 282 fs/ubifs/ubifs.h void *node;
node 1828 fs/xfs/xfs_attr.c xfs_da_intnode_t *node;
node 1848 fs/xfs/xfs_attr.c node = bp->data;
node 1849 fs/xfs/xfs_attr.c switch (be16_to_cpu(node->hdr.info.magic)) {
node 1851 fs/xfs/xfs_attr.c xfs_attr_trace_l_cn("wrong blk", context, node);
node 1898 fs/xfs/xfs_attr.c node = bp->data;
node 1899 fs/xfs/xfs_attr.c if (be16_to_cpu(node->hdr.info.magic)
node 1902 fs/xfs/xfs_attr.c if (unlikely(be16_to_cpu(node->hdr.info.magic)
node 1907 fs/xfs/xfs_attr.c node);
node 1911 fs/xfs/xfs_attr.c btree = node->btree;
node 1912 fs/xfs/xfs_attr.c for (i = 0; i < be16_to_cpu(node->hdr.count);
node 1922 fs/xfs/xfs_attr.c if (i == be16_to_cpu(node->hdr.count)) {
node 2288 fs/xfs/xfs_attr.c (__psunsigned_t)be16_to_cpu(node->hdr.count),
node 2289 fs/xfs/xfs_attr.c (__psunsigned_t)be32_to_cpu(node->btree[0].hashval),
node 2290 fs/xfs/xfs_attr.c (__psunsigned_t)be32_to_cpu(node->btree[
node 2291 fs/xfs/xfs_attr.c be16_to_cpu(node->hdr.count)-1].hashval));
node 849 fs/xfs/xfs_attr_leaf.c xfs_da_intnode_t *node;
node 882 fs/xfs/xfs_attr_leaf.c node = bp1->data;
node 886 fs/xfs/xfs_attr_leaf.c node->btree[0].hashval =
node 888 fs/xfs/xfs_attr_leaf.c node->btree[0].before = cpu_to_be32(blkno);
node 889 fs/xfs/xfs_attr_leaf.c node->hdr.count = cpu_to_be16(1);
node 2736 fs/xfs/xfs_attr_leaf.c xfs_da_intnode_t *node;
node 2750 fs/xfs/xfs_attr_leaf.c node = bp->data;
node 2751 fs/xfs/xfs_attr_leaf.c ASSERT(be16_to_cpu(node->hdr.info.magic) == XFS_DA_NODE_MAGIC);
node 2753 fs/xfs/xfs_attr_leaf.c count = be16_to_cpu(node->hdr.count);
node 2758 fs/xfs/xfs_attr_leaf.c child_fsb = be32_to_cpu(node->btree[0].before);
node 2818 fs/xfs/xfs_attr_leaf.c child_fsb = be32_to_cpu(node->btree[i+1].before);
node 113 fs/xfs/xfs_da_btree.c xfs_da_intnode_t *node;
node 123 fs/xfs/xfs_da_btree.c node = bp->data;
node 124 fs/xfs/xfs_da_btree.c node->hdr.info.forw = 0;
node 125 fs/xfs/xfs_da_btree.c node->hdr.info.back = 0;
node 126 fs/xfs/xfs_da_btree.c node->hdr.info.magic = cpu_to_be16(XFS_DA_NODE_MAGIC);
node 127 fs/xfs/xfs_da_btree.c node->hdr.info.pad = 0;
node 128 fs/xfs/xfs_da_btree.c node->hdr.count = 0;
node 129 fs/xfs/xfs_da_btree.c node->hdr.level = cpu_to_be16(level);
node 132 fs/xfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, &node->hdr, sizeof(node->hdr)));
node 146 fs/xfs/xfs_da_btree.c xfs_da_intnode_t *node;
node 257 fs/xfs/xfs_da_btree.c node = oldblk->bp->data;
node 258 fs/xfs/xfs_da_btree.c if (node->hdr.info.forw) {
node 259 fs/xfs/xfs_da_btree.c if (be32_to_cpu(node->hdr.info.forw) == addblk->blkno) {
node 265 fs/xfs/xfs_da_btree.c node = bp->data;
node 266 fs/xfs/xfs_da_btree.c node->hdr.info.back = cpu_to_be32(oldblk->blkno);
node 268 fs/xfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, &node->hdr.info,
node 269 fs/xfs/xfs_da_btree.c sizeof(node->hdr.info)));
node 271 fs/xfs/xfs_da_btree.c node = oldblk->bp->data;
node 272 fs/xfs/xfs_da_btree.c if (node->hdr.info.back) {
node 273 fs/xfs/xfs_da_btree.c if (be32_to_cpu(node->hdr.info.back) == addblk->blkno) {
node 279 fs/xfs/xfs_da_btree.c node = bp->data;
node 280 fs/xfs/xfs_da_btree.c node->hdr.info.forw = cpu_to_be32(oldblk->blkno);
node 282 fs/xfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, &node->hdr.info,
node 283 fs/xfs/xfs_da_btree.c sizeof(node->hdr.info)));
node 300 fs/xfs/xfs_da_btree.c xfs_da_intnode_t *node, *oldroot;
node 326 fs/xfs/xfs_da_btree.c node = bp->data;
node 337 fs/xfs/xfs_da_btree.c memcpy(node, oldroot, size);
node 348 fs/xfs/xfs_da_btree.c be16_to_cpu(node->hdr.level) + 1, &bp, args->whichfork);
node 351 fs/xfs/xfs_da_btree.c node = bp->data;
node 352 fs/xfs/xfs_da_btree.c node->btree[0].hashval = cpu_to_be32(blk1->hashval);
node 353 fs/xfs/xfs_da_btree.c node->btree[0].before = cpu_to_be32(blk1->blkno);
node 354 fs/xfs/xfs_da_btree.c node->btree[1].hashval = cpu_to_be32(blk2->hashval);
node 355 fs/xfs/xfs_da_btree.c node->btree[1].before = cpu_to_be32(blk2->blkno);
node 356 fs/xfs/xfs_da_btree.c node->hdr.count = cpu_to_be16(2);
node 369 fs/xfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, node->btree,
node 385 fs/xfs/xfs_da_btree.c xfs_da_intnode_t *node;
node 390 fs/xfs/xfs_da_btree.c node = oldblk->bp->data;
node 391 fs/xfs/xfs_da_btree.c ASSERT(be16_to_cpu(node->hdr.info.magic) == XFS_DA_NODE_MAGIC);
node 401 fs/xfs/xfs_da_btree.c if ((be16_to_cpu(node->hdr.count) + newcount) > state->node_ents) {
node 437 fs/xfs/xfs_da_btree.c node = oldblk->bp->data;
node 438 fs/xfs/xfs_da_btree.c if (oldblk->index <= be16_to_cpu(node->hdr.count)) {
node 580 fs/xfs/xfs_da_btree.c xfs_da_intnode_t *node;
node 585 fs/xfs/xfs_da_btree.c node = oldblk->bp->data;
node 587 fs/xfs/xfs_da_btree.c ASSERT(be16_to_cpu(node->hdr.info.magic) == XFS_DA_NODE_MAGIC);
node 588 fs/xfs/xfs_da_btree.c ASSERT((oldblk->index >= 0) && (oldblk->index <= be16_to_cpu(node->hdr.count)));
node 598 fs/xfs/xfs_da_btree.c btree = &node->btree[ oldblk->index ];
node 599 fs/xfs/xfs_da_btree.c if (oldblk->index < be16_to_cpu(node->hdr.count)) {
node 600 fs/xfs/xfs_da_btree.c tmp = (be16_to_cpu(node->hdr.count) - oldblk->index) * (uint)sizeof(*btree);
node 606 fs/xfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, btree, tmp + sizeof(*btree)));
node 607 fs/xfs/xfs_da_btree.c be16_add_cpu(&node->hdr.count, 1);
node 609 fs/xfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, &node->hdr, sizeof(node->hdr)));
node 614 fs/xfs/xfs_da_btree.c oldblk->hashval = be32_to_cpu(node->btree[be16_to_cpu(node->hdr.count)-1 ].hashval);
node 771 fs/xfs/xfs_da_btree.c xfs_da_intnode_t *node;
node 786 fs/xfs/xfs_da_btree.c node = (xfs_da_intnode_t *)info;
node 787 fs/xfs/xfs_da_btree.c count = be16_to_cpu(node->hdr.count);
node 840 fs/xfs/xfs_da_btree.c node = (xfs_da_intnode_t *)info;
node 843 fs/xfs/xfs_da_btree.c count -= be16_to_cpu(node->hdr.count);
node 844 fs/xfs/xfs_da_btree.c node = bp->data;
node 845 fs/xfs/xfs_da_btree.c ASSERT(be16_to_cpu(node->hdr.info.magic) == XFS_DA_NODE_MAGIC);
node 846 fs/xfs/xfs_da_btree.c count -= be16_to_cpu(node->hdr.count);
node 894 fs/xfs/xfs_da_btree.c xfs_da_intnode_t *node;
node 919 fs/xfs/xfs_da_btree.c node = blk->bp->data;
node 920 fs/xfs/xfs_da_btree.c ASSERT(be16_to_cpu(node->hdr.info.magic) == XFS_DA_NODE_MAGIC);
node 921 fs/xfs/xfs_da_btree.c btree = &node->btree[ blk->index ];
node 927 fs/xfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, btree, sizeof(*btree)));
node 929 fs/xfs/xfs_da_btree.c lasthash = be32_to_cpu(node->btree[be16_to_cpu(node->hdr.count)-1].hashval);
node 939 fs/xfs/xfs_da_btree.c xfs_da_intnode_t *node;
node 943 fs/xfs/xfs_da_btree.c node = drop_blk->bp->data;
node 944 fs/xfs/xfs_da_btree.c ASSERT(drop_blk->index < be16_to_cpu(node->hdr.count));
node 950 fs/xfs/xfs_da_btree.c btree = &node->btree[drop_blk->index];
node 951 fs/xfs/xfs_da_btree.c if (drop_blk->index < (be16_to_cpu(node->hdr.count)-1)) {
node 952 fs/xfs/xfs_da_btree.c tmp = be16_to_cpu(node->hdr.count) - drop_blk->index - 1;
node 956 fs/xfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, btree, tmp));
node 957 fs/xfs/xfs_da_btree.c btree = &node->btree[be16_to_cpu(node->hdr.count)-1];
node 961 fs/xfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, btree, sizeof(*btree)));
node 962 fs/xfs/xfs_da_btree.c be16_add_cpu(&node->hdr.count, -1);
node 964 fs/xfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, &node->hdr, sizeof(node->hdr)));
node 1053 fs/xfs/xfs_da_btree.c xfs_da_intnode_t *node;
node 1091 fs/xfs/xfs_da_btree.c node = blk->bp->data;
node 1092 fs/xfs/xfs_da_btree.c max = be16_to_cpu(node->hdr.count);
node 1093 fs/xfs/xfs_da_btree.c blk->hashval = be32_to_cpu(node->btree[max-1].hashval);
node 1100 fs/xfs/xfs_da_btree.c for (btree = &node->btree[probe]; span > 4;
node 1101 fs/xfs/xfs_da_btree.c btree = &node->btree[probe]) {
node 1132 fs/xfs/xfs_da_btree.c blkno = be32_to_cpu(node->btree[max-1].before);
node 1305 fs/xfs/xfs_da_btree.c xfs_da_intnode_t *node;
node 1307 fs/xfs/xfs_da_btree.c node = bp->data;
node 1308 fs/xfs/xfs_da_btree.c ASSERT(be16_to_cpu(node->hdr.info.magic) == XFS_DA_NODE_MAGIC);
node 1310 fs/xfs/xfs_da_btree.c *count = be16_to_cpu(node->hdr.count);
node 1311 fs/xfs/xfs_da_btree.c if (!node->hdr.count)
node 1313 fs/xfs/xfs_da_btree.c return be32_to_cpu(node->btree[be16_to_cpu(node->hdr.count)-1].hashval);
node 1403 fs/xfs/xfs_da_btree.c xfs_da_intnode_t *node;
node 1420 fs/xfs/xfs_da_btree.c node = blk->bp->data;
node 1421 fs/xfs/xfs_da_btree.c ASSERT(be16_to_cpu(node->hdr.info.magic) == XFS_DA_NODE_MAGIC);
node 1422 fs/xfs/xfs_da_btree.c if (forward && (blk->index < be16_to_cpu(node->hdr.count)-1)) {
node 1424 fs/xfs/xfs_da_btree.c blkno = be32_to_cpu(node->btree[blk->index].before);
node 1428 fs/xfs/xfs_da_btree.c blkno = be32_to_cpu(node->btree[blk->index].before);
node 1465 fs/xfs/xfs_da_btree.c node = (xfs_da_intnode_t *)info;
node 1466 fs/xfs/xfs_da_btree.c blk->hashval = be32_to_cpu(node->btree[be16_to_cpu(node->hdr.count)-1].hashval);
node 1470 fs/xfs/xfs_da_btree.c blk->index = be16_to_cpu(node->hdr.count)-1;
node 1471 fs/xfs/xfs_da_btree.c blkno = be32_to_cpu(node->btree[blk->index].before);
node 265 include/acpi/aclocal.h struct acpi_namespace_node *node;
node 388 include/acpi/aclocal.h struct acpi_namespace_node *node;
node 498 include/acpi/aclocal.h ACPI_STATE_COMMON struct acpi_namespace_node *node;
node 540 include/acpi/aclocal.h ACPI_STATE_COMMON struct acpi_namespace_node *node;
node 604 include/acpi/aclocal.h struct acpi_namespace_node *node; /* For use by interpreter */\
node 134 include/acpi/acobject.h struct acpi_namespace_node *node; /* Link back to parent node */
node 138 include/acpi/acobject.h ACPI_OBJECT_COMMON_HEADER struct acpi_namespace_node *node; /* Link back to parent node */
node 163 include/acpi/acobject.h struct acpi_namespace_node *node; /* Containing namespace node */
node 169 include/acpi/acobject.h struct acpi_namespace_node *node; /* Containing namespace node */
node 241 include/acpi/acobject.h struct acpi_namespace_node *node; /* Link back to parent node */\
node 284 include/acpi/acobject.h ACPI_OBJECT_COMMON_HEADER struct acpi_namespace_node *node; /* Parent device */
node 293 include/acpi/acobject.h struct acpi_namespace_node *node; /* Parent device */
node 318 include/acpi/acobject.h struct acpi_namespace_node *node;
node 417 include/acpi/acobject.h struct acpi_namespace_node node;
node 285 include/acpi/acpi_bus.h struct list_head node;
node 316 include/acpi/acpi_bus.h struct list_head node;
node 38 include/asm-generic/topology.h #define parent_node(node) ((void)(node),0)
node 41 include/asm-generic/topology.h #define node_to_cpumask(node) ((void)node, cpu_online_map)
node 44 include/asm-generic/topology.h #define node_to_first_cpu(node) ((void)(node),0)
node 63 include/asm-generic/topology.h cpumask_t _##v = node_to_cpumask(node); \
node 67 include/asm-generic/topology.h _##v = node_to_cpumask(node)
node 44 include/asm-m32r/mmzone.h int node;
node 46 include/asm-m32r/mmzone.h for (node = 0 ; node < MAX_NUMNODES ; node++)
node 47 include/asm-m32r/mmzone.h if (pfn >= node_start_pfn(node) && pfn <= node_end_pfn(node))
node 50 include/asm-m32r/mmzone.h return node;
node 115 include/asm-m68k/dvma.h int node; /* Prom node for this DMA device */
node 277 include/asm-m68k/openprom.h int (*no_nextnode)(int node);
node 278 include/asm-m68k/openprom.h int (*no_child)(int node);
node 279 include/asm-m68k/openprom.h int (*no_proplen)(int node, char *name);
node 280 include/asm-m68k/openprom.h int (*no_getprop)(int node, char *name, char *val);
node 281 include/asm-m68k/openprom.h int (*no_setprop)(int node, char *name, char *val, int len);
node 282 include/asm-m68k/openprom.h char * (*no_nextprop)(int node, char *name);
node 93 include/asm-x86/numaq/apic.h int node = apicid_to_node(logical_apicid);
node 96 include/asm-x86/numaq/apic.h return physid_mask_of_physid(cpu + 4*node);
node 15 include/asm-x86/pci.h int node; /* NUMA node */
node 105 include/asm-x86/pci.h return sd->node;
node 68 include/asm-x86/topology.h return node_to_cpumask_map[node];
node 108 include/asm-x86/topology.h return &node_to_cpumask_map[node];
node 114 include/asm-x86/topology.h return node_to_cpumask_map[node];
node 121 include/asm-x86/topology.h const cpumask_t *v = _node_to_cpumask_ptr(node)
node 124 include/asm-x86/topology.h v = _node_to_cpumask_ptr(node)
node 132 include/asm-x86/topology.h #define parent_node(node) (node)
node 204 include/asm-x86/topology.h const cpumask_t *v = _node_to_cpumask_ptr(node)
node 207 include/asm-x86/topology.h v = _node_to_cpumask_ptr(node)
node 216 include/asm-x86/topology.h node_to_cpumask_ptr(mask, node);
node 309 include/asm-x86/uv/uv_bau.h return constant_test_bit(node, &dstp->bits[0]);
node 313 include/asm-x86/uv/uv_bau.h __set_bit(node, &dstp->bits[0]);
node 143 include/linux/acpi.h struct list_head node;
node 35 include/linux/async_tx.h struct list_head node;
node 59 include/linux/atmel_tc.h struct list_head node;
node 17 include/linux/attribute_container.h struct list_head node;
node 413 include/linux/blkdev.h int node;
node 13 include/linux/dca.h struct list_head node;
node 27 include/linux/debugobjects.h struct hlist_node node;
node 250 include/linux/device.h struct list_head node;
node 438 include/linux/device.h dev->numa_node = node;
node 86 include/linux/dio.h struct list_head node;
node 174 include/linux/elevator.h #define rb_entry_rq(node) rb_entry((node), struct request, rb_node)
node 97 include/linux/enclosure.h struct list_head node;
node 809 include/linux/fb.h int node;
node 56 include/linux/ftrace.h struct hlist_node node;
node 50 include/linux/gameport.h struct list_head node;
node 67 include/linux/hidraw.h struct list_head node;
node 123 include/linux/hrtimer.h struct rb_node node;
node 75 include/linux/init_task.h { .first = &init_task.pids[PIDTYPE_PID].node }, \
node 76 include/linux/init_task.h { .first = &init_task.pids[PIDTYPE_PGID].node }, \
node 77 include/linux/init_task.h { .first = &init_task.pids[PIDTYPE_SID].node }, \
node 90 include/linux/init_task.h .node = { \
node 1091 include/linux/input.h struct list_head node;
node 1193 include/linux/input.h struct list_head node;
node 50 include/linux/leds.h struct list_head node; /* LED Device list */
node 511 include/linux/mm.h return zone->node;
node 547 include/linux/mm.h page->flags |= (node & NODES_MASK) << NODES_PGSHIFT;
node 560 include/linux/mm.h set_page_node(page, node);
node 135 include/linux/mmc/card.h #define mmc_list_to_card(l) container_of(l, struct mmc_card, node)
node 221 include/linux/mmzone.h int node;
node 739 include/linux/mmzone.h return zoneref->zone->node;
node 29 include/linux/node.h extern struct node node_devices[];
node 57 include/linux/node.h #define to_node(sys_device) container_of(sys_device, struct node, sysdev)
node 95 include/linux/nodemask.h #define node_set(node, dst) __node_set((node), &(dst))
node 98 include/linux/nodemask.h set_bit(node, dstp->bits);
node 101 include/linux/nodemask.h #define node_clear(node, dst) __node_clear((node), &(dst))
node 104 include/linux/nodemask.h clear_bit(node, dstp->bits);
node 120 include/linux/nodemask.h #define node_isset(node, nodemask) test_bit((node), (nodemask).bits)
node 123 include/linux/nodemask.h __node_test_and_set((node), &(nodemask))
node 126 include/linux/nodemask.h return test_and_set_bit(node, addr->bits);
node 246 include/linux/nodemask.h m.bits[0] = 1UL<<(node); \
node 249 include/linux/nodemask.h node_set((node), m); \
node 351 include/linux/nodemask.h for ((node) = first_node(mask); \
node 352 include/linux/nodemask.h (node) < MAX_NUMNODES; \
node 353 include/linux/nodemask.h (node) = next_node((node), (mask)))
node 357 include/linux/nodemask.h for ((node) = 0; (node) < 1; (node)++)
node 386 include/linux/nodemask.h return node_isset(node, node_states[state]);
node 391 include/linux/nodemask.h __node_set(node, &node_states[state]);
node 396 include/linux/nodemask.h __node_clear(node, &node_states[state]);
node 415 include/linux/nodemask.h return node == 0;
node 432 include/linux/nodemask.h for ( (node) = 0; (node) == 0; (node) = 1)
node 445 include/linux/nodemask.h int node; \
node 446 include/linux/nodemask.h for_each_node_mask(node, (mask)) \
node 447 include/linux/nodemask.h if (node_online(node)) \
node 449 include/linux/nodemask.h node; \
node 454 include/linux/nodemask.h #define node_online(node) node_state((node), N_ONLINE)
node 455 include/linux/nodemask.h #define node_possible(node) node_state((node), N_POSSIBLE)
node 457 include/linux/nodemask.h #define node_set_online(node) node_set_state((node), N_ONLINE)
node 458 include/linux/nodemask.h #define node_set_offline(node) node_clear_state((node), N_ONLINE)
node 460 include/linux/nodemask.h #define for_each_node(node) for_each_node_state(node, N_POSSIBLE)
node 461 include/linux/nodemask.h #define for_each_online_node(node) for_each_node_state(node, N_ONLINE)
node 283 include/linux/pci.h struct list_head node; /* node in list of buses */
node 312 include/linux/pci.h #define pci_bus_b(n) list_entry(n, struct pci_bus, node)
node 402 include/linux/pci.h struct list_head node;
node 71 include/linux/pid.h struct hlist_node node;
node 152 include/linux/pid.h &pid->tasks[type], pids[type].node) {
node 119 include/linux/plist.h .plist = PLIST_HEAD_INIT((node).plist, NULL), \
node 144 include/linux/plist.h node->prio = prio;
node 145 include/linux/plist.h plist_head_init(&node->plist, NULL);
node 206 include/linux/plist.h return plist_head_empty(&node->plist);
node 60 include/linux/posix-timers.h unsigned int node;
node 91 include/linux/prio_tree.h return node->parent == node;
node 96 include/linux/prio_tree.h return node->left == node;
node 101 include/linux/prio_tree.h return node->right == node;
node 114 include/linux/prio_tree.h (struct prio_tree_node *) (node))
node 116 include/linux/prio_tree.h prio_tree_insert(root, (struct prio_tree_node *) (node))
node 118 include/linux/prio_tree.h prio_tree_remove(root, (struct prio_tree_node *) (node))
node 136 include/linux/rbtree.h #define RB_EMPTY_NODE(node) (rb_parent(node) == node)
node 137 include/linux/rbtree.h #define RB_CLEAR_NODE(node) (rb_set_parent(node, node))
node 155 include/linux/rbtree.h node->rb_parent_color = (unsigned long )parent;
node 156 include/linux/rbtree.h node->rb_left = node->rb_right = NULL;
node 158 include/linux/rbtree.h *rb_link = node;
node 110 include/linux/rfkill.h struct list_head node;
node 140 include/linux/rio.h struct list_head node;
node 171 include/linux/rio.h struct list_head node; /* node in global list of ports */
node 200 include/linux/rio.h struct list_head node; /* node in list of networks */
node 218 include/linux/rio.h struct list_head node;
node 276 include/linux/rio.h struct list_head node;
node 52 include/linux/serio.h struct list_head node;
node 248 include/linux/slab.h __kmalloc_node_track_caller(size, flags, node, \
node 252 include/linux/slab.h __kmalloc_node(size, flags, node)
node 288 include/linux/slab.h return kmalloc_node(size, flags | __GFP_ZERO, node);
node 88 include/linux/slab_def.h flags, node);
node 91 include/linux/slab_def.h flags, node);
node 93 include/linux/slab_def.h return __kmalloc_node(size, flags, node);
node 15 include/linux/slob_def.h return __kmalloc_node(size, flags, node);
node 38 include/linux/slub_def.h int node; /* The node of the page (or -1 for debug) */
node 103 include/linux/slub_def.h struct kmem_cache_node *node[MAX_NUMNODES];
node 243 include/linux/slub_def.h return kmem_cache_alloc_node(s, flags, node);
node 245 include/linux/slub_def.h return __kmalloc_node(size, flags, node);
node 81 include/linux/tc.h struct list_head node; /* Node in list of all TC devices. */
node 105 include/linux/tc.h struct list_head node;
node 62 include/linux/thermal.h struct list_head node;
node 76 include/linux/thermal.h struct list_head node;
node 95 include/linux/thermal.h struct list_head node;
node 48 include/linux/tipc.h __u32 node;
node 66 include/linux/tipc.h return (zone << 24) | (cluster << 12) | node;
node 43 include/linux/topology.h node_to_cpumask_ptr(__tmp__, node); \
node 49 include/linux/topology.h for_each_online_node(node) \
node 50 include/linux/topology.h if (nr_cpus_node(node))
node 925 include/linux/usb.h struct list_head node;
node 171 include/linux/vmstat.h struct zone *zones = NODE_DATA(node)->node_zones;
node 172 include/linux/zorro.h struct list_head node;
node 164 include/net/ax25.h hlist_for_each_entry(__ax25, node, list, uid_node)
node 250 include/net/ax25.h hlist_for_each_entry(__ax25, node, list, ax25_node)
node 15 include/net/datalink.h struct list_head node;
node 79 include/net/garp.h struct rb_node node;
node 83 include/net/inet_hashtables.h struct hlist_node node;
node 88 include/net/inet_hashtables.h hlist_for_each_entry(tb, node, head, node)
node 178 include/net/inet_timewait_sock.h hlist_for_each_entry(tw, node, head, tw_node)
node 181 include/net/inet_timewait_sock.h hlist_for_each_entry(tw, node, jail, tw_death_node)
node 184 include/net/inet_timewait_sock.h hlist_for_each_entry_safe(tw, node, safe, jail, tw_death_node)
node 127 include/net/ip6_fib.h struct fib6_node *root, *node;
node 19 include/net/ipx.h __u8 node[IPX_NODE_LEN];
node 70 include/net/ipx.h struct list_head node; /* node in ipx_interfaces list */
node 78 include/net/ipx.h struct list_head node; /* node in ipx_routes list */
node 102 include/net/ipx.h unsigned char node[IPX_NODE_LEN];
node 82 include/net/lapb.h struct list_head node;
node 55 include/net/llc.h struct list_head node;
node 159 include/net/netrom.h hlist_for_each_entry(__nr_neigh, node, list, neigh_node)
node 162 include/net/netrom.h hlist_for_each_entry_safe(__nr_neigh, node, node2, list, neigh_node)
node 165 include/net/netrom.h hlist_for_each_entry(__nr_node, node, list, node_node)
node 168 include/net/netrom.h hlist_for_each_entry_safe(__nr_node, node, node2, list, node_node)
node 15 include/net/pkt_cls.h int (*fn)(struct tcf_proto *, unsigned long node, struct tcf_walker *);
node 653 include/net/sctp/sctp.h hlist_for_each_entry(epb, node, head, node)
node 103 include/net/sctp/structs.h struct hlist_node node;
node 1259 include/net/sctp/structs.h struct hlist_node node;
node 315 include/net/sock.h node->pprev = NULL;
node 387 include/net/sock.h hlist_for_each_entry(__sk, node, list, sk_node)
node 389 include/net/sock.h if (__sk && ({ node = &(__sk)->sk_node; 1; })) \
node 390 include/net/sock.h hlist_for_each_entry_from(__sk, node, sk_node)
node 392 include/net/sock.h if (__sk && ({ node = &(__sk)->sk_node; 1; })) \
node 393 include/net/sock.h hlist_for_each_entry_continue(__sk, node, sk_node)
node 395 include/net/sock.h hlist_for_each_entry_safe(__sk, node, tmp, list, sk_node)
node 397 include/net/sock.h hlist_for_each_entry(__sk, node, list, sk_bind_node)
node 608 include/net/sock.h struct list_head node;
node 124 include/net/x25.h struct list_head node;
node 132 include/net/x25.h struct list_head node;
node 165 include/net/x25.h struct list_head node;
node 61 include/scsi/scsi_device.h struct list_head node;
node 53 include/xen/xenbus.h const char *node;
node 116 kernel/audit_tree.c size = offsetof(struct audit_chunk, owners) + count * sizeof(struct node);
node 323 kernel/audit_tree.c struct node *p;
node 433 kernel/audit_tree.c struct node *p;
node 436 kernel/audit_tree.c p = list_entry(victim->chunks.next, struct node, list);
node 462 kernel/audit_tree.c struct node *node = list_entry(p, struct node, list);
node 464 kernel/audit_tree.c if (node->index & (1U<<31)) {
node 471 kernel/audit_tree.c struct node *node;
node 474 kernel/audit_tree.c node = list_entry(tree->chunks.next, struct node, list);
node 477 kernel/audit_tree.c if (!(node->index & (1U<<31)))
node 480 kernel/audit_tree.c chunk = find_chunk(node);
node 484 kernel/audit_tree.c untag_chunk(chunk, node);
node 537 kernel/audit_tree.c struct node *node;
node 558 kernel/audit_tree.c list_for_each_entry(node, &tree->chunks, list) {
node 559 kernel/audit_tree.c struct audit_chunk *chunk = find_chunk(node);
node 562 kernel/audit_tree.c node->index |= 1U<<31;
node 565 kernel/audit_tree.c node->index &= ~(1U<<31);
node 662 kernel/audit_tree.c struct node *node;
node 664 kernel/audit_tree.c list_for_each_entry(node, &tree->chunks, list)
node 665 kernel/audit_tree.c node->index &= ~(1U<<31);
node 786 kernel/audit_tree.c struct node *node;
node 788 kernel/audit_tree.c list_for_each_entry(node, &tree->chunks, list)
node 789 kernel/audit_tree.c node->index &= ~(1U<<31);
node 328 kernel/cgroup.c struct hlist_node *node;
node 347 kernel/cgroup.c hlist_for_each_entry(cg, node, hhead, hlist) {
node 645 kernel/cgroup.c struct list_head *node;
node 649 kernel/cgroup.c node = dentry->d_subdirs.next;
node 650 kernel/cgroup.c while (node != &dentry->d_subdirs) {
node 651 kernel/cgroup.c struct dentry *d = list_entry(node, struct dentry, d_u.d_child);
node 652 kernel/cgroup.c list_del_init(node);
node 664 kernel/cgroup.c node = dentry->d_subdirs.next;
node 1041 kernel/cgroup.c struct hlist_node *node;
node 1044 kernel/cgroup.c hlist_for_each_entry(cg, node, hhead, hlist) {
node 2185 kernel/cpuset.c int node; /* node that zone z is on */
node 2191 kernel/cpuset.c node = zone_to_nid(z);
node 2193 kernel/cpuset.c if (node_isset(node, current->mems_allowed))
node 2214 kernel/cpuset.c allowed = node_isset(node, cs->mems_allowed);
node 2245 kernel/cpuset.c int node; /* node that zone z is on */
node 2249 kernel/cpuset.c node = zone_to_nid(z);
node 2250 kernel/cpuset.c if (node_isset(node, current->mems_allowed))
node 2316 kernel/cpuset.c int node;
node 2318 kernel/cpuset.c node = next_node(current->cpuset_mem_spread_rotor, current->mems_allowed);
node 2319 kernel/cpuset.c if (node == MAX_NUMNODES)
node 2320 kernel/cpuset.c node = first_node(current->mems_allowed);
node 2321 kernel/cpuset.c current->cpuset_mem_spread_rotor = node;
node 2322 kernel/cpuset.c return node;
node 519 kernel/hrtimer.c timer = rb_entry(base->first, struct hrtimer, node);
node 846 kernel/hrtimer.c entry = rb_entry(parent, struct hrtimer, node);
node 875 kernel/hrtimer.c base->first = &timer->node;
node 878 kernel/hrtimer.c rb_link_node(&timer->node, parent, link);
node 879 kernel/hrtimer.c rb_insert_color(&timer->node, &base->active);
node 909 kernel/hrtimer.c if (base->first == &timer->node) {
node 910 kernel/hrtimer.c base->first = rb_next(&timer->node);
node 915 kernel/hrtimer.c rb_erase(&timer->node, &base->active);
node 1111 kernel/hrtimer.c timer = rb_entry(base->first, struct hrtimer, node);
node 1221 kernel/hrtimer.c if (base->first == &timer->node &&
node 1302 kernel/hrtimer.c struct rb_node *node;
node 1308 kernel/hrtimer.c while ((node = base->first)) {
node 1311 kernel/hrtimer.c timer = rb_entry(node, struct hrtimer, node);
node 1392 kernel/hrtimer.c struct rb_node *node;
node 1415 kernel/hrtimer.c while ((node = base->first)) {
node 1418 kernel/hrtimer.c timer = rb_entry(node, struct hrtimer, node);
node 1600 kernel/hrtimer.c struct rb_node *node;
node 1603 kernel/hrtimer.c while ((node = rb_first(&old_base->active))) {
node 1604 kernel/hrtimer.c timer = rb_entry(node, struct hrtimer, node);
node 293 kernel/kprobes.c struct hlist_node *node;
node 297 kernel/kprobes.c hlist_for_each_entry_rcu(p, node, head, hlist) {
node 438 kernel/kprobes.c struct hlist_node *node, *tmp;
node 448 kernel/kprobes.c hlist_for_each_entry_safe(ri, node, tmp, head, hlist) {
node 454 kernel/kprobes.c hlist_for_each_entry_safe(ri, node, tmp, &empty_rp, hlist) {
node 1150 kernel/kprobes.c struct hlist_node *node;
node 1159 kernel/kprobes.c hlist_for_each_entry_rcu(p, node, head, hlist) {
node 1194 kernel/kprobes.c struct hlist_node *node;
node 1206 kernel/kprobes.c hlist_for_each_entry_rcu(p, node, head, hlist)
node 1221 kernel/kprobes.c struct hlist_node *node;
node 1235 kernel/kprobes.c hlist_for_each_entry_rcu(p, node, head, hlist) {
node 358 kernel/marker.c struct hlist_node *node;
node 363 kernel/marker.c hlist_for_each_entry(e, node, head, hlist) {
node 377 kernel/marker.c struct hlist_node *node;
node 386 kernel/marker.c hlist_for_each_entry(e, node, head, hlist) {
node 432 kernel/marker.c struct hlist_node *node;
node 439 kernel/marker.c hlist_for_each_entry(e, node, head, hlist) {
node 739 kernel/marker.c struct hlist_node *node;
node 743 kernel/marker.c hlist_for_each_entry(entry, node, head, hlist) {
node 827 kernel/marker.c struct hlist_node *node;
node 834 kernel/marker.c hlist_for_each_entry(e, node, head, hlist) {
node 322 kernel/pid.c hlist_add_head_rcu(&link->node, &pid->tasks[type]);
node 335 kernel/pid.c hlist_del_rcu(&link->node);
node 362 kernel/pid.c hlist_replace_rcu(&old->pids[type].node, &new->pids[type].node);
node 372 kernel/pid.c result = hlist_entry(first, struct task_struct, pids[(type)].node);
node 142 kernel/pm_qos_params.c struct requirement_list *node;
node 148 kernel/pm_qos_params.c list_for_each_entry(node,
node 151 kernel/pm_qos_params.c extreme_value, node->value);
node 254 kernel/pm_qos_params.c struct requirement_list *node;
node 258 kernel/pm_qos_params.c list_for_each_entry(node,
node 260 kernel/pm_qos_params.c if (strcmp(node->name, name) == 0) {
node 262 kernel/pm_qos_params.c node->value =
node 265 kernel/pm_qos_params.c node->value = new_value;
node 289 kernel/pm_qos_params.c struct requirement_list *node;
node 293 kernel/pm_qos_params.c list_for_each_entry(node,
node 295 kernel/pm_qos_params.c if (strcmp(node->name, name) == 0) {
node 296 kernel/pm_qos_params.c kfree(node->name);
node 297 kernel/pm_qos_params.c list_del(&node->list);
node 298 kernel/pm_qos_params.c kfree(node);
node 73 kernel/power/swsusp.c struct rb_node node;
node 88 kernel/power/swsusp.c ext = container_of(*new, struct swsusp_extent, node);
node 116 kernel/power/swsusp.c rb_link_node(&ext->node, parent, new);
node 117 kernel/power/swsusp.c rb_insert_color(&ext->node, &swsusp_extents);
node 148 kernel/power/swsusp.c struct rb_node *node;
node 150 kernel/power/swsusp.c while ((node = swsusp_extents.rb_node)) {
node 154 kernel/power/swsusp.c ext = container_of(node, struct swsusp_extent, node);
node 155 kernel/power/swsusp.c rb_erase(node, &swsusp_extents);
node 336 kernel/profile.c int node, cpu = (unsigned long)__cpu;
node 342 kernel/profile.c node = cpu_to_node(cpu);
node 345 kernel/profile.c page = alloc_pages_node(node,
node 353 kernel/profile.c page = alloc_pages_node(node,
node 535 kernel/profile.c int node = cpu_to_node(cpu);
node 538 kernel/profile.c page = alloc_pages_node(node,
node 545 kernel/profile.c page = alloc_pages_node(node,
node 2473 kernel/sched.c struct hlist_node *node;
node 2475 kernel/sched.c hlist_for_each_entry(notifier, node, &curr->preempt_notifiers, link)
node 2484 kernel/sched.c struct hlist_node *node;
node 2486 kernel/sched.c hlist_for_each_entry(notifier, node, &curr->preempt_notifiers, link)
node 6989 kernel/sched.c n = (node + i) % nr_node_ids;
node 6999 kernel/sched.c val = node_distance(node, n);
node 7023 kernel/sched.c node_to_cpumask_ptr(nodemask, node);
node 7030 kernel/sched.c node_set(node, used_nodes);
node 7033 kernel/sched.c int next_node = find_next_best_node(node, &used_nodes);
node 230 kernel/time/clockevents.c struct list_head *node, *tmp;
node 241 kernel/time/clockevents.c list_for_each_safe(node, tmp, &clockevents_released)
node 242 kernel/time/clockevents.c list_del(node);
node 98 kernel/time/timer_list.c timer = rb_entry(curr, struct hrtimer, node);
node 240 kernel/trace/ftrace.c hlist_for_each_entry_rcu(rec, t, head, node) {
node 273 kernel/trace/ftrace.c hlist_for_each_entry_rcu(p, t, &ftrace_hash[key], node) {
node 286 kernel/trace/ftrace.c hlist_add_head_rcu(&node->node, &ftrace_hash[key]);
node 292 kernel/trace/ftrace.c hlist_del(&node->node);
node 337 kernel/trace/ftrace.c struct dyn_ftrace *node;
node 379 kernel/trace/ftrace.c node = ftrace_alloc_dyn_node(ip);
node 380 kernel/trace/ftrace.c if (!node)
node 383 kernel/trace/ftrace.c node->ip = ip;
node 385 kernel/trace/ftrace.c ftrace_add_hash(node, key);
node 724 kernel/trace/ftrace.c hlist_for_each_entry_safe(p, t, n, head, node) {
node 745 kernel/trace/ftrace.c INIT_HLIST_NODE(&p->node);
node 746 kernel/trace/ftrace.c hlist_add_head(&p->node, &temp_list);
node 766 kernel/trace/ftrace.c hlist_for_each_entry_safe(p, t, n, &temp_list, node) {
node 767 kernel/trace/ftrace.c hlist_del(&p->node);
node 768 kernel/trace/ftrace.c INIT_HLIST_NODE(&p->node);
node 769 kernel/trace/ftrace.c hlist_add_head(&p->node, head);
node 86 lib/debugobjects.c hlist_add_head(&new->node, &obj_pool);
node 98 lib/debugobjects.c struct hlist_node *node;
node 102 lib/debugobjects.c hlist_for_each_entry(obj, node, &b->list, node) {
node 124 lib/debugobjects.c obj = hlist_entry(obj_pool.first, typeof(*obj), node);
node 129 lib/debugobjects.c hlist_del(&obj->node);
node 131 lib/debugobjects.c hlist_add_head(&obj->node, &b->list);
node 156 lib/debugobjects.c hlist_add_head(&obj->node, &obj_pool);
node 175 lib/debugobjects.c struct hlist_node *node, *tmp;
node 189 lib/debugobjects.c hlist_for_each_entry_safe(obj, node, tmp, &freelist, node) {
node 190 lib/debugobjects.c hlist_del(&obj->node);
node 506 lib/debugobjects.c hlist_del(&obj->node);
node 519 lib/debugobjects.c struct hlist_node *node, *tmp;
node 539 lib/debugobjects.c hlist_for_each_entry_safe(obj, node, tmp, &db->list, node) {
node 555 lib/debugobjects.c hlist_del(&obj->node);
node 556 lib/debugobjects.c hlist_add_head(&obj->node, &freelist);
node 563 lib/debugobjects.c hlist_for_each_entry_safe(obj, node, tmp, &freelist, node) {
node 564 lib/debugobjects.c hlist_del(&obj->node);
node 873 lib/debugobjects.c hlist_add_head(&obj_static_pool[i].node, &obj_pool);
node 79 lib/plist.c WARN_ON(!plist_node_empty(node));
node 82 lib/plist.c if (node->prio < iter->prio)
node 84 lib/plist.c else if (node->prio == iter->prio) {
node 92 lib/plist.c list_add_tail(&node->plist.prio_list, &iter->plist.prio_list);
node 94 lib/plist.c list_add_tail(&node->plist.node_list, &iter->plist.node_list);
node 109 lib/plist.c if (!list_empty(&node->plist.prio_list)) {
node 110 lib/plist.c struct plist_node *next = plist_first(&node->plist);
node 112 lib/plist.c list_move_tail(&next->plist.prio_list, &node->plist.prio_list);
node 113 lib/plist.c list_del_init(&node->plist.prio_list);
node 116 lib/plist.c list_del_init(&node->plist.node_list);
node 58 lib/prio_tree.c node, struct vm_area_struct, shared.prio_tree_node);
node 64 lib/prio_tree.c *radix = node->start;
node 65 lib/prio_tree.c *heap = node->last;
node 123 lib/prio_tree.c INIT_PRIO_TREE_NODE(node);
node 126 lib/prio_tree.c node->left = first;
node 127 lib/prio_tree.c first->parent = node;
node 129 lib/prio_tree.c last = node;
node 136 lib/prio_tree.c root->prio_tree_node = node;
node 137 lib/prio_tree.c return node;
node 146 lib/prio_tree.c INIT_PRIO_TREE_NODE(node);
node 154 lib/prio_tree.c node->parent = node;
node 155 lib/prio_tree.c root->prio_tree_node = node;
node 157 lib/prio_tree.c node->parent = old->parent;
node 159 lib/prio_tree.c old->parent->left = node;
node 161 lib/prio_tree.c old->parent->right = node;
node 165 lib/prio_tree.c node->left = old->left;
node 166 lib/prio_tree.c old->left->parent = node;
node 170 lib/prio_tree.c node->right = old->right;
node 171 lib/prio_tree.c old->right->parent = node;
node 190 lib/prio_tree.c struct prio_tree_node *cur, *res = node;
node 195 lib/prio_tree.c get_index(root, node, &radix_index, &heap_index);
node 199 lib/prio_tree.c return prio_tree_expand(root, node, heap_index);
node 212 lib/prio_tree.c struct prio_tree_node *tmp = node;
node 213 lib/prio_tree.c node = prio_tree_replace(root, cur, node);
node 231 lib/prio_tree.c INIT_PRIO_TREE_NODE(node);
node 232 lib/prio_tree.c cur->right = node;
node 233 lib/prio_tree.c node->parent = cur;
node 239 lib/prio_tree.c INIT_PRIO_TREE_NODE(node);
node 240 lib/prio_tree.c cur->left = node;
node 241 lib/prio_tree.c node->parent = cur;
node 269 lib/prio_tree.c cur = node;
node 304 lib/prio_tree.c while (cur != node)
node 58 lib/radix-tree.c struct radix_tree_node *node;
node 94 lib/radix-tree.c __set_bit(offset, node->tags[tag]);
node 100 lib/radix-tree.c __clear_bit(offset, node->tags[tag]);
node 106 lib/radix-tree.c return test_bit(offset, node->tags[tag]);
node 137 lib/radix-tree.c if (node->tags[tag][idx])
node 176 lib/radix-tree.c struct radix_tree_node *node =
node 184 lib/radix-tree.c tag_clear(node, 0, 0);
node 185 lib/radix-tree.c tag_clear(node, 1, 0);
node 186 lib/radix-tree.c node->slots[0] = NULL;
node 187 lib/radix-tree.c node->count = 0;
node 189 lib/radix-tree.c kmem_cache_free(radix_tree_node_cachep, node);
node 195 lib/radix-tree.c call_rcu(&node->rcu_head, radix_tree_node_rcu_free);
node 207 lib/radix-tree.c struct radix_tree_node *node;
node 214 lib/radix-tree.c node = kmem_cache_alloc(radix_tree_node_cachep, gfp_mask);
node 215 lib/radix-tree.c if (node == NULL)
node 220 lib/radix-tree.c rtp->nodes[rtp->nr++] = node;
node 222 lib/radix-tree.c kmem_cache_free(radix_tree_node_cachep, node);
node 244 lib/radix-tree.c struct radix_tree_node *node;
node 260 lib/radix-tree.c if (!(node = radix_tree_node_alloc(root)))
node 264 lib/radix-tree.c node->slots[0] = radix_tree_indirect_to_ptr(root->rnode);
node 269 lib/radix-tree.c tag_set(node, tag, 0);
node 273 lib/radix-tree.c node->height = newheight;
node 274 lib/radix-tree.c node->count = 1;
node 275 lib/radix-tree.c node = radix_tree_ptr_to_indirect(node);
node 276 lib/radix-tree.c rcu_assign_pointer(root->rnode, node);
node 294 lib/radix-tree.c struct radix_tree_node *node = NULL, *slot;
node 320 lib/radix-tree.c if (node) {
node 321 lib/radix-tree.c rcu_assign_pointer(node->slots[offset], slot);
node 322 lib/radix-tree.c node->count++;
node 330 lib/radix-tree.c node = slot;
node 331 lib/radix-tree.c slot = node->slots[offset];
node 339 lib/radix-tree.c if (node) {
node 340 lib/radix-tree.c node->count++;
node 341 lib/radix-tree.c rcu_assign_pointer(node->slots[offset], item);
node 342 lib/radix-tree.c BUG_ON(tag_get(node, 0, offset));
node 343 lib/radix-tree.c BUG_ON(tag_get(node, 1, offset));
node 370 lib/radix-tree.c struct radix_tree_node *node, **slot;
node 372 lib/radix-tree.c node = rcu_dereference(root->rnode);
node 373 lib/radix-tree.c if (node == NULL)
node 376 lib/radix-tree.c if (!radix_tree_is_indirect_ptr(node)) {
node 381 lib/radix-tree.c node = radix_tree_indirect_to_ptr(node);
node 383 lib/radix-tree.c height = node->height;
node 391 lib/radix-tree.c (node->slots + ((index>>shift) & RADIX_TREE_MAP_MASK));
node 392 lib/radix-tree.c node = rcu_dereference(*slot);
node 393 lib/radix-tree.c if (node == NULL)
node 419 lib/radix-tree.c struct radix_tree_node *node, **slot;
node 421 lib/radix-tree.c node = rcu_dereference(root->rnode);
node 422 lib/radix-tree.c if (node == NULL)
node 425 lib/radix-tree.c if (!radix_tree_is_indirect_ptr(node)) {
node 428 lib/radix-tree.c return node;
node 430 lib/radix-tree.c node = radix_tree_indirect_to_ptr(node);
node 432 lib/radix-tree.c height = node->height;
node 440 lib/radix-tree.c (node->slots + ((index>>shift) & RADIX_TREE_MAP_MASK));
node 441 lib/radix-tree.c node = rcu_dereference(*slot);
node 442 lib/radix-tree.c if (node == NULL)
node 449 lib/radix-tree.c return node;
node 528 lib/radix-tree.c pathp->node = NULL;
node 539 lib/radix-tree.c pathp[1].node = slot;
node 549 lib/radix-tree.c while (pathp->node) {
node 550 lib/radix-tree.c if (!tag_get(pathp->node, tag, pathp->offset))
node 552 lib/radix-tree.c tag_clear(pathp->node, tag, pathp->offset);
node 553 lib/radix-tree.c if (any_tag_set(pathp->node, tag))
node 583 lib/radix-tree.c struct radix_tree_node *node;
node 590 lib/radix-tree.c node = rcu_dereference(root->rnode);
node 591 lib/radix-tree.c if (node == NULL)
node 594 lib/radix-tree.c if (!radix_tree_is_indirect_ptr(node))
node 596 lib/radix-tree.c node = radix_tree_indirect_to_ptr(node);
node 598 lib/radix-tree.c height = node->height;
node 607 lib/radix-tree.c if (node == NULL)
node 616 lib/radix-tree.c if (!tag_get(node, tag, offset))
node 619 lib/radix-tree.c int ret = tag_get(node, tag, offset);
node 624 lib/radix-tree.c node = rcu_dereference(node->slots[offset]);
node 739 lib/radix-tree.c struct radix_tree_node *node;
node 743 lib/radix-tree.c node = rcu_dereference(root->rnode);
node 744 lib/radix-tree.c if (!node)
node 747 lib/radix-tree.c if (!radix_tree_is_indirect_ptr(node)) {
node 750 lib/radix-tree.c results[0] = node;
node 753 lib/radix-tree.c node = radix_tree_indirect_to_ptr(node);
node 755 lib/radix-tree.c max_index = radix_tree_maxindex(node->height);
node 764 lib/radix-tree.c slots_found = __lookup(node, (void ***)results + ret, cur_index,
node 807 lib/radix-tree.c struct radix_tree_node *node;
node 811 lib/radix-tree.c node = rcu_dereference(root->rnode);
node 812 lib/radix-tree.c if (!node)
node 815 lib/radix-tree.c if (!radix_tree_is_indirect_ptr(node)) {
node 821 lib/radix-tree.c node = radix_tree_indirect_to_ptr(node);
node 823 lib/radix-tree.c max_index = radix_tree_maxindex(node->height);
node 832 lib/radix-tree.c slots_found = __lookup(node, results + ret, cur_index,
node 927 lib/radix-tree.c struct radix_tree_node *node;
node 936 lib/radix-tree.c node = rcu_dereference(root->rnode);
node 937 lib/radix-tree.c if (!node)
node 940 lib/radix-tree.c if (!radix_tree_is_indirect_ptr(node)) {
node 943 lib/radix-tree.c results[0] = node;
node 946 lib/radix-tree.c node = radix_tree_indirect_to_ptr(node);
node 948 lib/radix-tree.c max_index = radix_tree_maxindex(node->height);
node 957 lib/radix-tree.c slots_found = __lookup_tag(node, (void ***)results + ret,
node 996 lib/radix-tree.c struct radix_tree_node *node;
node 1005 lib/radix-tree.c node = rcu_dereference(root->rnode);
node 1006 lib/radix-tree.c if (!node)
node 1009 lib/radix-tree.c if (!radix_tree_is_indirect_ptr(node)) {
node 1015 lib/radix-tree.c node = radix_tree_indirect_to_ptr(node);
node 1017 lib/radix-tree.c max_index = radix_tree_maxindex(node->height);
node 1026 lib/radix-tree.c slots_found = __lookup_tag(node, results + ret,
node 1113 lib/radix-tree.c pathp->node = NULL;
node 1122 lib/radix-tree.c pathp->node = slot;
node 1135 lib/radix-tree.c if (tag_get(pathp->node, tag, pathp->offset))
node 1141 lib/radix-tree.c while (pathp->node) {
node 1142 lib/radix-tree.c pathp->node->slots[pathp->offset] = NULL;
node 1143 lib/radix-tree.c pathp->node->count--;
node 1151 lib/radix-tree.c if (pathp->node->count) {
node 1152 lib/radix-tree.c if (pathp->node ==
node 1159 lib/radix-tree.c to_free = pathp->node;
node 1188 lib/radix-tree.c memset(node, 0, sizeof(struct radix_tree_node));
node 28 lib/rbtree.c struct rb_node *right = node->rb_right;
node 29 lib/rbtree.c struct rb_node *parent = rb_parent(node);
node 31 lib/rbtree.c if ((node->rb_right = right->rb_left))
node 32 lib/rbtree.c rb_set_parent(right->rb_left, node);
node 33 lib/rbtree.c right->rb_left = node;
node 39 lib/rbtree.c if (node == parent->rb_left)
node 46 lib/rbtree.c rb_set_parent(node, right);
node 51 lib/rbtree.c struct rb_node *left = node->rb_left;
node 52 lib/rbtree.c struct rb_node *parent = rb_parent(node);
node 54 lib/rbtree.c if ((node->rb_left = left->rb_right))
node 55 lib/rbtree.c rb_set_parent(left->rb_right, node);
node 56 lib/rbtree.c left->rb_right = node;
node 62 lib/rbtree.c if (node == parent->rb_right)
node 69 lib/rbtree.c rb_set_parent(node, left);
node 76 lib/rbtree.c while ((parent = rb_parent(node)) && rb_is_red(parent))
node 89 lib/rbtree.c node = gparent;
node 94 lib/rbtree.c if (parent->rb_right == node)
node 99 lib/rbtree.c parent = node;
node 100 lib/rbtree.c node = tmp;
node 114 lib/rbtree.c node = gparent;
node 119 lib/rbtree.c if (parent->rb_left == node)
node 124 lib/rbtree.c parent = node;
node 125 lib/rbtree.c node = tmp;
node 143 lib/rbtree.c while ((!node || rb_is_black(node)) && node != root->rb_node)
node 145 lib/rbtree.c if (parent->rb_left == node)
node 159 lib/rbtree.c node = parent;
node 160 lib/rbtree.c parent = rb_parent(node);
node 178 lib/rbtree.c node = root->rb_node;
node 196 lib/rbtree.c node = parent;
node 197 lib/rbtree.c parent = rb_parent(node);
node 215 lib/rbtree.c node = root->rb_node;
node 220 lib/rbtree.c if (node)
node 221 lib/rbtree.c rb_set_black(node);
node 229 lib/rbtree.c if (!node->rb_left)
node 230 lib/rbtree.c child = node->rb_right;
node 231 lib/rbtree.c else if (!node->rb_right)
node 232 lib/rbtree.c child = node->rb_left;
node 235 lib/rbtree.c struct rb_node *old = node, *left;
node 237 lib/rbtree.c node = node->rb_right;
node 238 lib/rbtree.c while ((left = node->rb_left) != NULL)
node 239 lib/rbtree.c node = left;
node 240 lib/rbtree.c child = node->rb_right;
node 241 lib/rbtree.c parent = rb_parent(node);
node 242 lib/rbtree.c color = rb_color(node);
node 248 lib/rbtree.c parent = node;
node 252 lib/rbtree.c node->rb_parent_color = old->rb_parent_color;
node 253 lib/rbtree.c node->rb_right = old->rb_right;
node 254 lib/rbtree.c node->rb_left = old->rb_left;
node 259 lib/rbtree.c rb_parent(old)->rb_left = node;
node 261 lib/rbtree.c rb_parent(old)->rb_right = node;
node 263 lib/rbtree.c root->rb_node = node;
node 265 lib/rbtree.c rb_set_parent(old->rb_left, node);
node 267 lib/rbtree.c rb_set_parent(old->rb_right, node);
node 271 lib/rbtree.c parent = rb_parent(node);
node 272 lib/rbtree.c color = rb_color(node);
node 278 lib/rbtree.c if (parent->rb_left == node)
node 325 lib/rbtree.c if (rb_parent(node) == node)
node 330 lib/rbtree.c if (node->rb_right) {
node 331 lib/rbtree.c node = node->rb_right;
node 332 lib/rbtree.c while (node->rb_left)
node 333 lib/rbtree.c node=node->rb_left;
node 334 lib/rbtree.c return node;
node 343 lib/rbtree.c while ((parent = rb_parent(node)) && node == parent->rb_right)
node 344 lib/rbtree.c node = parent;
node 354 lib/rbtree.c if (rb_parent(node) == node)
node 359 lib/rbtree.c if (node->rb_left) {
node 360 lib/rbtree.c node = node->rb_left;
node 361 lib/rbtree.c while (node->rb_right)
node 362 lib/rbtree.c node=node->rb_right;
node 363 lib/rbtree.c return node;
node 368 lib/rbtree.c while ((parent = rb_parent(node)) && node == parent->rb_left)
node 369 lib/rbtree.c node = parent;
node 548 lib/zlib_deflate/deftree.c int node; /* new node being created */
node 571 lib/zlib_deflate/deftree.c node = s->heap[++(s->heap_len)] = (max_code < 2 ? ++max_code : 0);
node 572 lib/zlib_deflate/deftree.c tree[node].Freq = 1;
node 573 lib/zlib_deflate/deftree.c s->depth[node] = 0;
node 574 lib/zlib_deflate/deftree.c s->opt_len--; if (stree) s->static_len -= stree[node].Len;
node 587 lib/zlib_deflate/deftree.c node = elems; /* next internal node of the tree */
node 596 lib/zlib_deflate/deftree.c tree[node].Freq = tree[n].Freq + tree[m].Freq;
node 597 lib/zlib_deflate/deftree.c s->depth[node] = (uch) (max(s->depth[n], s->depth[m]) + 1);
node 598 lib/zlib_deflate/deftree.c tree[n].Dad = tree[m].Dad = (ush)node;
node 602 lib/zlib_deflate/deftree.c node, tree[node].Freq, n, tree[n].Freq, m, tree[m].Freq);
node 606 lib/zlib_deflate/deftree.c s->heap[SMALLEST] = node++;
node 58 mm/allocpercpu.c int node = cpu_to_node(cpu);
node 66 mm/allocpercpu.c if (node_online(node))
node 67 mm/allocpercpu.c pdata->ptrs[cpu] = kmalloc_node(size, gfp|__GFP_ZERO, node);
node 1391 mm/hugetlb.c int node;
node 1394 mm/hugetlb.c for_each_node_mask(node, cpuset_current_mems_allowed)
node 1395 mm/hugetlb.c nr += array[node];
node 911 mm/memcontrol.c int node, zid;
node 922 mm/memcontrol.c for_each_node_state(node, N_POSSIBLE)
node 925 mm/memcontrol.c mz = mem_cgroup_zoneinfo(mem, node, zid);
node 1065 mm/memcontrol.c int zone, tmp = node;
node 1074 mm/memcontrol.c if (!node_state(node, N_NORMAL_MEMORY))
node 1080 mm/memcontrol.c mem->info.nodeinfo[node] = pn;
node 1094 mm/memcontrol.c kfree(mem->info.nodeinfo[node]);
node 1124 mm/memcontrol.c int node;
node 1137 mm/memcontrol.c for_each_node_state(node, N_POSSIBLE)
node 1138 mm/memcontrol.c if (alloc_mem_cgroup_per_zone_info(mem, node))
node 1143 mm/memcontrol.c for_each_node_state(node, N_POSSIBLE)
node 1144 mm/memcontrol.c free_mem_cgroup_per_zone_info(mem, node);
node 1160 mm/memcontrol.c int node;
node 1163 mm/memcontrol.c for_each_node_state(node, N_POSSIBLE)
node 1164 mm/memcontrol.c free_mem_cgroup_per_zone_info(mem, node);
node 129 mm/memory_hotplug.c int node = pgdat->node_id;
node 137 mm/memory_hotplug.c get_page_bootmem(node, page, NODE_INFO);
node 148 mm/memory_hotplug.c get_page_bootmem(node, page, NODE_INFO);
node 743 mm/memory_hotplug.c int ret, drain, retry_max, node;
node 759 mm/memory_hotplug.c node = zone_to_nid(zone);
node 770 mm/memory_hotplug.c if (nr_pages >= node_present_pages(node))
node 771 mm/memory_hotplug.c arg.status_change_nid = node;
node 292 mm/mempolicy.c int node = first_node(pol->w.user_nodemask);
node 294 mm/mempolicy.c if (node_isset(node, *nodes)) {
node 295 mm/mempolicy.c pol->v.preferred_node = node;
node 771 mm/mempolicy.c return alloc_pages_node(node, GFP_HIGHUSER_MOVABLE, 0);
node 1408 mm/mempolicy.c return zone->node;
node 2185 mm/mempolicy.c unsigned long node[MAX_NUMNODES];
node 2212 mm/mempolicy.c md->node[page_to_nid(page)]++;
node 2322 mm/mempolicy.c if (md->node[n])
node 2323 mm/mempolicy.c seq_printf(m, " N%d=%lu", n, md->node[n]);
node 836 mm/migrate.c int node;
node 845 mm/migrate.c while (pm->node != MAX_NUMNODES && pm->page != p)
node 848 mm/migrate.c if (pm->node == MAX_NUMNODES)
node 853 mm/migrate.c return alloc_pages_node(pm->node,
node 875 mm/migrate.c for (pp = pm; pp->node != MAX_NUMNODES; pp++) {
node 906 mm/migrate.c if (err == pp->node)
node 948 mm/migrate.c for ( ; pm->node != MAX_NUMNODES; pm++) {
node 1059 mm/migrate.c int node;
node 1061 mm/migrate.c if (get_user(node, nodes + i))
node 1065 mm/migrate.c if (!node_state(node, N_HIGH_MEMORY))
node 1069 mm/migrate.c if (!node_isset(node, task_nodes))
node 1072 mm/migrate.c pm[i].node = node;
node 1074 mm/migrate.c pm[i].node = 0; /* anything to not match MAX_NUMNODES */
node 1077 mm/migrate.c pm[nr_pages].node = MAX_NUMNODES;
node 56 mm/mm_init.c zone->node, zone->name);
node 325 mm/page-writeback.c int node;
node 328 mm/page-writeback.c for_each_node_state(node, N_HIGH_MEMORY) {
node 330 mm/page-writeback.c &NODE_DATA(node)->node_zones[ZONE_HIGHMEM];
node 2089 mm/page_alloc.c if (!node_isset(node, *used_node_mask)) {
node 2090 mm/page_alloc.c node_set(node, *used_node_mask);
node 2091 mm/page_alloc.c return node;
node 2101 mm/page_alloc.c val = node_distance(node, n);
node 2104 mm/page_alloc.c val += (n < node);
node 2141 mm/page_alloc.c j = build_zonelists_node(NODE_DATA(node), zonelist, j,
node 2171 mm/page_alloc.c int pos, j, node;
node 2180 mm/page_alloc.c node = node_order[j];
node 2181 mm/page_alloc.c z = &NODE_DATA(node)->node_zones[zone_type];
node 2257 mm/page_alloc.c int j, node, load;
node 2281 mm/page_alloc.c while ((node = find_next_best_node(local_node, &used_mask)) >= 0) {
node 2282 mm/page_alloc.c int distance = node_distance(local_node, node);
node 2297 mm/page_alloc.c node_load[node] = load;
node 2299 mm/page_alloc.c prev_node = node;
node 2302 mm/page_alloc.c build_zonelists_in_node_order(pgdat, node);
node 2304 mm/page_alloc.c node_order[j++] = node; /* remember order */
node 2339 mm/page_alloc.c int node, local_node;
node 2356 mm/page_alloc.c for (node = local_node + 1; node < MAX_NUMNODES; node++) {
node 2357 mm/page_alloc.c if (!node_online(node))
node 2359 mm/page_alloc.c j = build_zonelists_node(NODE_DATA(node), zonelist, j,
node 2362 mm/page_alloc.c for (node = 0; node < local_node; node++) {
node 2363 mm/page_alloc.c if (!node_online(node))
node 2365 mm/page_alloc.c j = build_zonelists_node(NODE_DATA(node), zonelist, j,
node 2724 mm/page_alloc.c int node = cpu_to_node(cpu);
node 2726 mm/page_alloc.c node_set_state(node, N_CPU); /* this node has a cpu */
node 2734 mm/page_alloc.c GFP_KERNEL, node);
node 3454 mm/page_alloc.c zone->node = nid;
node 3555 mm/page_alloc.c unsigned int node;
node 3558 mm/page_alloc.c for_each_node_mask(node, node_possible_map)
node 3559 mm/page_alloc.c highest = node;
node 116 mm/prio_tree.c struct vm_area_struct *node, *head, *new_head;
node 144 mm/prio_tree.c node = vma->shared.vm_set.head;
node 151 mm/prio_tree.c node->shared.vm_set.head = new_head;
node 152 mm/prio_tree.c new_head->shared.vm_set.head = node;
node 154 mm/prio_tree.c node->shared.vm_set.head = NULL;
node 29 mm/quicklist.c int node = numa_node_id();
node 30 mm/quicklist.c struct zone *zones = NODE_DATA(node)->node_zones;
node 32 mm/quicklist.c node_to_cpumask_ptr(cpumask_on_node, node);
node 902 mm/slab.c int node;
node 904 mm/slab.c node = next_node(cpu_to_node(cpu), node_online_map);
node 905 mm/slab.c if (node == MAX_NUMNODES)
node 906 mm/slab.c node = first_node(node_online_map);
node 908 mm/slab.c per_cpu(reap_node, cpu) = node;
node 913 mm/slab.c int node = __get_cpu_var(reap_node);
node 915 mm/slab.c node = next_node(node, node_online_map);
node 916 mm/slab.c if (unlikely(node >= MAX_NUMNODES))
node 917 mm/slab.c node = first_node(node_online_map);
node 918 mm/slab.c __get_cpu_var(reap_node) = node;
node 956 mm/slab.c nc = kmalloc_node(memsize, GFP_KERNEL, node);
node 1035 mm/slab.c ac_ptr = kmalloc_node(memsize, GFP_KERNEL, node);
node 1038 mm/slab.c if (i == node || !node_online(i)) {
node 1042 mm/slab.c ac_ptr[i] = alloc_arraycache(node, limit, 0xbaadf00d);
node 1068 mm/slab.c struct kmem_list3 *rl3 = cachep->nodelists[node];
node 1080 mm/slab.c free_block(cachep, ac->entry, ac->avail, node);
node 1091 mm/slab.c int node = __get_cpu_var(reap_node);
node 1094 mm/slab.c struct array_cache *ac = l3->alien[node];
node 1097 mm/slab.c __drain_alien_cache(cachep, ac, node);
node 1126 mm/slab.c int node;
node 1128 mm/slab.c node = numa_node_id();
node 1134 mm/slab.c if (likely(slabp->nodeid == node))
node 1137 mm/slab.c l3 = cachep->nodelists[node];
node 1161 mm/slab.c int node = cpu_to_node(cpu);
node 1162 mm/slab.c node_to_cpumask_ptr(mask, node);
node 1172 mm/slab.c l3 = cachep->nodelists[node];
node 1182 mm/slab.c free_block(cachep, nc->entry, nc->avail, node);
node 1192 mm/slab.c shared->avail, node);
node 1215 mm/slab.c l3 = cachep->nodelists[node];
node 1226 mm/slab.c int node = cpu_to_node(cpu);
node 1242 mm/slab.c if (!cachep->nodelists[node]) {
node 1243 mm/slab.c l3 = kmalloc_node(memsize, GFP_KERNEL, node);
node 1255 mm/slab.c cachep->nodelists[node] = l3;
node 1258 mm/slab.c spin_lock_irq(&cachep->nodelists[node]->list_lock);
node 1259 mm/slab.c cachep->nodelists[node]->free_limit =
node 1260 mm/slab.c (1 + nr_cpus_node(node)) *
node 1262 mm/slab.c spin_unlock_irq(&cachep->nodelists[node]->list_lock);
node 1274 mm/slab.c nc = alloc_arraycache(node, cachep->limit,
node 1279 mm/slab.c shared = alloc_arraycache(node,
node 1288 mm/slab.c alien = alloc_alien_cache(node, cachep->limit);
node 1296 mm/slab.c l3 = cachep->nodelists[node];
node 1413 mm/slab.c int node;
node 1415 mm/slab.c for_each_online_node(node) {
node 1416 mm/slab.c cachep->nodelists[node] = &initkmem_list3[index + node];
node 1417 mm/slab.c cachep->nodelists[node]->next_reap = jiffies +
node 1434 mm/slab.c int node;
node 1475 mm/slab.c node = numa_node_id();
node 1482 mm/slab.c cache_cache.nodelists[node] = &initkmem_list3[CACHE_CACHE + node];
node 2088 mm/slab.c int node;
node 2089 mm/slab.c for_each_online_node(node) {
node 2090 mm/slab.c cachep->nodelists[node] =
node 2092 mm/slab.c GFP_KERNEL, node);
node 2093 mm/slab.c BUG_ON(!cachep->nodelists[node]);
node 2094 mm/slab.c kmem_list3_init(cachep->nodelists[node]);
node 2414 mm/slab.c assert_spin_locked(&cachep->nodelists[node]->list_lock);
node 2433 mm/slab.c int node = numa_node_id();
node 2437 mm/slab.c spin_lock(&cachep->nodelists[node]->list_lock);
node 2438 mm/slab.c free_block(cachep, ac->entry, ac->avail, node);
node 2439 mm/slab.c spin_unlock(&cachep->nodelists[node]->list_lock);
node 2446 mm/slab.c int node;
node 2450 mm/slab.c for_each_online_node(node) {
node 2451 mm/slab.c l3 = cachep->nodelists[node];
node 2456 mm/slab.c for_each_online_node(node) {
node 2457 mm/slab.c l3 = cachep->nodelists[node];
node 2459 mm/slab.c drain_array(cachep, l3, l3->shared, 1, node);
node 2954 mm/slab.c int node;
node 2958 mm/slab.c node = numa_node_id();
node 2969 mm/slab.c l3 = cachep->nodelists[node];
node 3007 mm/slab.c node);
node 3026 mm/slab.c x = cache_grow(cachep, flags | GFP_THISNODE, node, NULL);
node 3489 mm/slab.c l3 = cachep->nodelists[node];
node 3491 mm/slab.c check_spinlock_acquired_node(cachep, node);
node 3493 mm/slab.c slab_put_obj(cachep, slabp, objp, node);
node 3526 mm/slab.c int node = numa_node_id();
node 3533 mm/slab.c l3 = cachep->nodelists[node];
node 3548 mm/slab.c free_block(cachep, ac->entry, batchcount, node);
node 3676 mm/slab.c return kmem_cache_alloc_node(cachep, flags, node);
node 3682 mm/slab.c return __do_kmalloc_node(size, flags, node,
node 3690 mm/slab.c return __do_kmalloc_node(size, flags, node, caller);
node 3696 mm/slab.c return __do_kmalloc_node(size, flags, node, NULL);
node 3810 mm/slab.c int node;
node 3815 mm/slab.c for_each_online_node(node) {
node 3818 mm/slab.c new_alien = alloc_alien_cache(node, cachep->limit);
node 3825 mm/slab.c new_shared = alloc_arraycache(node,
node 3834 mm/slab.c l3 = cachep->nodelists[node];
node 3842 mm/slab.c shared->avail, node);
node 3849 mm/slab.c l3->free_limit = (1 + nr_cpus_node(node)) *
node 3856 mm/slab.c l3 = kmalloc_node(sizeof(struct kmem_list3), GFP_KERNEL, node);
node 3868 mm/slab.c l3->free_limit = (1 + nr_cpus_node(node)) *
node 3870 mm/slab.c cachep->nodelists[node] = l3;
node 3877 mm/slab.c node--;
node 3878 mm/slab.c while (node >= 0) {
node 3879 mm/slab.c if (cachep->nodelists[node]) {
node 3880 mm/slab.c l3 = cachep->nodelists[node];
node 3885 mm/slab.c cachep->nodelists[node] = NULL;
node 3887 mm/slab.c node--;
node 4027 mm/slab.c free_block(cachep, ac->entry, tofree, node);
node 4052 mm/slab.c int node = numa_node_id();
node 4068 mm/slab.c l3 = searchp->nodelists[node];
node 4072 mm/slab.c drain_array(searchp, l3, cpu_cache_get(searchp), 0, node);
node 4083 mm/slab.c drain_array(searchp, l3, l3->shared, 0, node);
node 4161 mm/slab.c int node;
node 4166 mm/slab.c for_each_online_node(node) {
node 4167 mm/slab.c l3 = cachep->nodelists[node];
node 4393 mm/slab.c int node;
node 4405 mm/slab.c for_each_online_node(node) {
node 4406 mm/slab.c l3 = cachep->nodelists[node];
node 238 mm/slob.c if (node != -1)
node 239 mm/slob.c page = alloc_pages_node(node, gfp, order);
node 327 mm/slob.c if (node != -1 && page_to_nid(&sp->page) != node)
node 352 mm/slob.c b = slob_new_page(gfp & ~__GFP_ZERO, 0, node);
node 471 mm/slob.c m = slob_alloc(size + align, gfp, align, node);
node 479 mm/slob.c ret = slob_new_page(gfp | __GFP_COMP, get_order(size), node);
node 573 mm/slob.c b = slob_alloc(c->size, flags, c->align, node);
node 575 mm/slob.c b = slob_new_page(flags, get_order(c->size), node);
node 223 mm/slub.c return s->node[node];
node 828 mm/slub.c struct kmem_cache_node *n = get_node(s, node);
node 835 mm/slub.c struct kmem_cache_node *n = get_node(s, node);
node 850 mm/slub.c struct kmem_cache_node *n = get_node(s, node);
node 1065 mm/slub.c if (node == -1)
node 1068 mm/slub.c return alloc_pages_node(node, flags, order);
node 1078 mm/slub.c page = alloc_slab_page(flags | __GFP_NOWARN | __GFP_NORETRY, node,
node 1086 mm/slub.c page = alloc_slab_page(flags, node, oo);
node 1119 mm/slub.c flags & (GFP_RECLAIM_MASK | GFP_CONSTRAINT_MASK), node);
node 1348 mm/slub.c int searchnode = (node == -1) ? numa_node_id() : node;
node 1477 mm/slub.c if (node != -1 && c->node != node)
node 1514 mm/slub.c if (unlikely(!node_match(c, node)))
node 1529 mm/slub.c c->node = page_to_nid(c->page);
node 1539 mm/slub.c new = get_partial(s, gfpflags, node);
node 1549 mm/slub.c new = new_slab(s, gfpflags, node);
node 1571 mm/slub.c c->node = -1;
node 1596 mm/slub.c if (unlikely(!c->freelist || !node_match(c, node)))
node 1598 mm/slub.c object = __slab_alloc(s, gfpflags, node, addr, c);
node 1622 mm/slub.c return slab_alloc(s, gfpflags, node, __builtin_return_address(0));
node 1717 mm/slub.c if (likely(page == c->page && c->node >= 0)) {
node 1908 mm/slub.c c->node = 0;
node 2084 mm/slub.c page = new_slab(kmalloc_caches, gfpflags, node);
node 2087 mm/slub.c if (page_to_nid(page) != node) {
node 2089 mm/slub.c "node %d\n", node);
node 2098 mm/slub.c kmalloc_caches->node[node] = n;
node 2104 mm/slub.c inc_slabs_node(kmalloc_caches, node, page->objects);
node 2119 mm/slub.c int node;
node 2121 mm/slub.c for_each_node_state(node, N_NORMAL_MEMORY) {
node 2122 mm/slub.c struct kmem_cache_node *n = s->node[node];
node 2125 mm/slub.c s->node[node] = NULL;
node 2131 mm/slub.c int node;
node 2139 mm/slub.c for_each_node_state(node, N_NORMAL_MEMORY) {
node 2142 mm/slub.c if (local_node == node)
node 2147 mm/slub.c node);
node 2151 mm/slub.c gfpflags, node);
node 2159 mm/slub.c s->node[node] = n;
node 2427 mm/slub.c int node;
node 2433 mm/slub.c for_each_node_state(node, N_NORMAL_MEMORY) {
node 2434 mm/slub.c struct kmem_cache_node *n = get_node(s, node);
node 2437 mm/slub.c if (n->nr_partial || slabs_node(s, node))
node 2667 mm/slub.c struct page *page = alloc_pages_node(node, flags | __GFP_COMP,
node 2682 mm/slub.c return kmalloc_large_node(size, flags, node);
node 2689 mm/slub.c return slab_alloc(s, flags, node, __builtin_return_address(0));
node 2762 mm/slub.c int node;
node 2776 mm/slub.c for_each_node_state(node, N_NORMAL_MEMORY) {
node 2777 mm/slub.c n = get_node(s, node);
node 2866 mm/slub.c s->node[offline_node] = NULL;
node 2906 mm/slub.c s->node[nid] = n;
node 3223 mm/slub.c return kmalloc_large_node(size, gfpflags, node);
node 3230 mm/slub.c return slab_alloc(s, gfpflags, node, caller);
node 3346 mm/slub.c int node;
node 3355 mm/slub.c for_each_node_state(node, N_NORMAL_MEMORY) {
node 3356 mm/slub.c struct kmem_cache_node *n = get_node(s, node);
node 3569 mm/slub.c int node;
node 3578 mm/slub.c for_each_node_state(node, N_NORMAL_MEMORY) {
node 3579 mm/slub.c struct kmem_cache_node *n = get_node(s, node);
node 3663 mm/slub.c int node;
node 3679 mm/slub.c if (!c || c->node < 0)
node 3691 mm/slub.c nodes[c->node] += x;
node 3693 mm/slub.c per_cpu[c->node]++;
node 3698 mm/slub.c for_each_node_state(node, N_NORMAL_MEMORY) {
node 3699 mm/slub.c struct kmem_cache_node *n = get_node(s, node);
node 3710 mm/slub.c nodes[node] += x;
node 3714 mm/slub.c for_each_node_state(node, N_NORMAL_MEMORY) {
node 3715 mm/slub.c struct kmem_cache_node *n = get_node(s, node);
node 3724 mm/slub.c nodes[node] += x;
node 3729 mm/slub.c for_each_node_state(node, N_NORMAL_MEMORY)
node 3730 mm/slub.c if (nodes[node])
node 3732 mm/slub.c node, nodes[node]);
node 3740 mm/slub.c int node;
node 3742 mm/slub.c for_each_online_node(node) {
node 3743 mm/slub.c struct kmem_cache_node *n = get_node(s, node);
node 4467 mm/slub.c int node;
node 4471 mm/slub.c for_each_online_node(node) {
node 4472 mm/slub.c struct kmem_cache_node *n = get_node(s, node);
node 43 mm/sparse-vmemmap.c return __alloc_bootmem_node(NODE_DATA(node), size, align, goal);
node 51 mm/sparse-vmemmap.c struct page *page = alloc_pages_node(node,
node 57 mm/sparse-vmemmap.c return __earlyonly_bootmem_alloc(node, size, size,
node 67 mm/sparse-vmemmap.c if (actual_node != node)
node 77 mm/sparse-vmemmap.c void *p = vmemmap_alloc_block(PAGE_SIZE, node);
node 90 mm/sparse-vmemmap.c void *p = vmemmap_alloc_block(PAGE_SIZE, node);
node 102 mm/sparse-vmemmap.c void *p = vmemmap_alloc_block(PAGE_SIZE, node);
node 114 mm/sparse-vmemmap.c void *p = vmemmap_alloc_block(PAGE_SIZE, node);
node 133 mm/sparse-vmemmap.c pgd = vmemmap_pgd_populate(addr, node);
node 136 mm/sparse-vmemmap.c pud = vmemmap_pud_populate(pgd, addr, node);
node 139 mm/sparse-vmemmap.c pmd = vmemmap_pmd_populate(pud, addr, node);
node 142 mm/sparse-vmemmap.c pte = vmemmap_pte_populate(pmd, addr, node);
node 145 mm/sparse-vmemmap.c vmemmap_verify(pte, node, addr, addr + PAGE_SIZE);
node 240 mm/vmalloc.c area = kmalloc_node(sizeof(*area), gfp_mask & GFP_RECLAIM_MASK, node);
node 327 mm/vmalloc.c return __get_vm_area_node(size, flags, VMALLOC_START, VMALLOC_END, node,
node 503 mm/vmalloc.c PAGE_KERNEL, node, caller);
node 508 mm/vmalloc.c node);
node 521 mm/vmalloc.c if (node < 0)
node 524 mm/vmalloc.c page = alloc_pages_node(node, gfp_mask, 0);
node 571 mm/vmalloc.c node, gfp_mask, caller);
node 576 mm/vmalloc.c return __vmalloc_area_node(area, gfp_mask, prot, node, caller);
node 639 mm/vmalloc.c node, __builtin_return_address(0));
node 379 mm/vmstat.c if (z->node == numa_node_id())
node 401 mm/vmstat.c loff_t node = *pos;
node 403 mm/vmstat.c pgdat && node;
node 405 mm/vmstat.c --node;
node 156 net/802/garp.c attr = rb_entry(parent, struct garp_attr, node);
node 176 net/802/garp.c attr = rb_entry(parent, struct garp_attr, node);
node 183 net/802/garp.c rb_link_node(&new->node, parent, p);
node 184 net/802/garp.c rb_insert_color(&new->node, &app->gid);
node 205 net/802/garp.c rb_erase(&attr->node, &app->gid);
node 385 net/802/garp.c struct rb_node *node, *next;
node 388 net/802/garp.c for (node = rb_first(&app->gid);
node 389 net/802/garp.c next = node ? rb_next(node) : NULL, node != NULL;
node 390 net/802/garp.c node = next) {
node 391 net/802/garp.c attr = rb_entry(node, struct garp_attr, node);
node 36 net/802/psnap.c list_for_each_entry_rcu(p, &snap_list, node) {
node 143 net/802/psnap.c list_add_rcu(&proto->node, &snap_list);
node 158 net/802/psnap.c list_del_rcu(&proto->node);
node 150 net/appletalk/atalk_proc.c struct hlist_node *node;
node 152 net/appletalk/atalk_proc.c sk_for_each(s, node, &atalk_sockets)
node 94 net/appletalk/ddp.c struct hlist_node *node;
node 97 net/appletalk/ddp.c sk_for_each(s, node, &atalk_sockets) {
node 142 net/appletalk/ddp.c struct hlist_node *node;
node 146 net/appletalk/ddp.c sk_for_each(s, node, &atalk_sockets) {
node 390 net/appletalk/ddp.c if (node != ATADDR_BCAST &&
node 391 net/appletalk/ddp.c iface->address.s_node != node &&
node 392 net/appletalk/ddp.c node != ATADDR_ANYNODE)
node 408 net/appletalk/ddp.c if ((node == ATADDR_BCAST ||
node 409 net/appletalk/ddp.c node == ATADDR_ANYNODE ||
node 410 net/appletalk/ddp.c iface->address.s_node == node) &&
node 416 net/appletalk/ddp.c if (node == ATADDR_ANYNODE && net != ATADDR_ANYNET &&
node 1092 net/appletalk/ddp.c struct hlist_node *node;
node 1094 net/appletalk/ddp.c sk_for_each(s, node, &atalk_sockets) {
node 223 net/atm/common.c struct hlist_node *node, *tmp;
node 227 net/atm/common.c sk_for_each_safe(s, node, tmp, head) {
node 269 net/atm/common.c struct hlist_node *node;
node 273 net/atm/common.c sk_for_each(s, node, head) {
node 1010 net/atm/lec.c struct hlist_node *node;
node 1020 net/atm/lec.c struct hlist_node *e = state->node;
node 1034 net/atm/lec.c state->node = e;
node 1127 net/atm/lec.c state->node = SEQ_START_TOKEN;
node 1163 net/atm/lec.c struct lec_arp_table *entry = hlist_entry(state->node, struct lec_arp_table, next);
node 1500 net/atm/lec.c struct hlist_node *node;
node 1517 net/atm/lec.c hlist_for_each_entry(entry, node, &priv->lec_arp_tables[i], next) {
node 1558 net/atm/lec.c struct hlist_node *node;
node 1565 net/atm/lec.c hlist_for_each_entry(rulla, node, &priv->lec_arp_tables[i], next) {
node 1600 net/atm/lec.c hlist_for_each_entry(rulla, node, &priv->lec_no_forward, next) {
node 1629 net/atm/lec.c hlist_for_each_entry(rulla, node, &priv->lec_arp_empty_ones, next) {
node 1658 net/atm/lec.c hlist_for_each_entry(rulla, node, &priv->mcast_fwds, next) {
node 1696 net/atm/lec.c struct hlist_node *node, *next;
node 1708 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_tables[i], next) {
node 1715 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_empty_ones, next) {
node 1723 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->lec_no_forward, next) {
node 1731 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->mcast_fwds, next) {
node 1748 net/atm/lec.c struct hlist_node *node;
node 1757 net/atm/lec.c hlist_for_each_entry(entry, node, head, next) {
node 1851 net/atm/lec.c struct hlist_node *node, *next;
node 1862 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_tables[i], next) {
node 2022 net/atm/lec.c struct hlist_node *node, *next;
node 2029 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_tables[i], next) {
node 2053 net/atm/lec.c struct hlist_node *node, *next;
node 2070 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_empty_ones, next) {
node 2114 net/atm/lec.c hlist_for_each_entry(tmp, node, &priv->lec_arp_tables[i], next) {
node 2154 net/atm/lec.c struct hlist_node *node;
node 2229 net/atm/lec.c hlist_for_each_entry(entry, node, &priv->lec_arp_tables[i], next) {
node 2305 net/atm/lec.c struct hlist_node *node;
node 2313 net/atm/lec.c hlist_for_each_entry(entry, node, &priv->lec_arp_tables[i], next) {
node 2340 net/atm/lec.c struct hlist_node *node;
node 2346 net/atm/lec.c hlist_for_each_entry(entry, node, &priv->lec_arp_tables[i], next) {
node 2396 net/atm/lec.c struct hlist_node *node, *next;
node 2406 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_tables[i], next) {
node 2417 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_empty_ones, next) {
node 2426 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->lec_no_forward, next) {
node 2435 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->mcast_fwds, next) {
node 2453 net/atm/lec.c struct hlist_node *node, *next;
node 2467 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_empty_ones, next) {
node 218 net/atm/signaling.c struct hlist_node *node;
node 232 net/atm/signaling.c sk_for_each(s, node, head) {
node 84 net/ax25/af_ax25.c struct hlist_node *node;
node 91 net/ax25/af_ax25.c ax25_for_each(s, node, &ax25_list) {
node 161 net/ax25/af_ax25.c struct hlist_node *node;
node 164 net/ax25/af_ax25.c ax25_for_each(s, node, &ax25_list) {
node 190 net/ax25/af_ax25.c struct hlist_node *node;
node 193 net/ax25/af_ax25.c ax25_for_each(s, node, &ax25_list) {
node 216 net/ax25/af_ax25.c struct hlist_node *node;
node 219 net/ax25/af_ax25.c ax25_for_each(s, node, &ax25_list) {
node 251 net/ax25/af_ax25.c struct hlist_node *node;
node 254 net/ax25/af_ax25.c ax25_for_each(s, node, &ax25_list) {
node 1861 net/ax25/af_ax25.c struct hlist_node *node;
node 1865 net/ax25/af_ax25.c ax25_for_each(ax25, node, &ax25_list) {
node 42 net/ax25/ax25_ds_subr.c struct hlist_node *node;
node 83 net/ax25/ax25_ds_subr.c ax25_for_each(ax25o, node, &ax25_list) {
node 162 net/ax25/ax25_ds_subr.c struct hlist_node *node;
node 165 net/ax25/ax25_ds_subr.c ax25_for_each(ax25, node, &ax25_list)
node 74 net/ax25/ax25_ds_timer.c struct hlist_node *node;
node 85 net/ax25/ax25_ds_timer.c ax25_for_each(ax25, node, &ax25_list) {
node 200 net/ax25/ax25_iface.c struct hlist_node *node;
node 203 net/ax25/ax25_iface.c hlist_for_each_entry(lf, node, &ax25_linkfail_list, lf_node)
node 56 net/ax25/ax25_uid.c struct hlist_node *node;
node 59 net/ax25/ax25_uid.c ax25_uid_for_each(ax25_uid, node, &ax25_uid_list) {
node 76 net/ax25/ax25_uid.c struct hlist_node *node;
node 84 net/ax25/ax25_uid.c ax25_uid_for_each(ax25_uid, node, &ax25_uid_list) {
node 123 net/ax25/ax25_uid.c ax25_uid_for_each(ax25_uid, node, &ax25_uid_list) {
node 150 net/ax25/ax25_uid.c struct hlist_node *node;
node 154 net/ax25/ax25_uid.c ax25_uid_for_each(pt, node, &ax25_uid_list) {
node 218 net/ax25/ax25_uid.c struct hlist_node *node;
node 222 net/ax25/ax25_uid.c ax25_uid_for_each(ax25_uid, node, &ax25_uid_list) {
node 94 net/bluetooth/hci_sock.c struct hlist_node *node;
node 99 net/bluetooth/hci_sock.c sk_for_each(sk, node, &hci_sk_list.head) {
node 671 net/bluetooth/hci_sock.c struct hlist_node *node;
node 675 net/bluetooth/hci_sock.c sk_for_each(sk, node, &hci_sk_list.head) {
node 521 net/bluetooth/l2cap.c struct hlist_node *node;
node 522 net/bluetooth/l2cap.c sk_for_each(sk, node, &l2cap_sk_list.head)
node 536 net/bluetooth/l2cap.c struct hlist_node *node;
node 538 net/bluetooth/l2cap.c sk_for_each(sk, node, &l2cap_sk_list.head) {
node 552 net/bluetooth/l2cap.c return node ? sk : sk1;
node 2137 net/bluetooth/l2cap.c struct hlist_node *node;
node 2146 net/bluetooth/l2cap.c sk_for_each(sk, node, &l2cap_sk_list.head) {
node 2424 net/bluetooth/l2cap.c struct hlist_node *node;
node 2429 net/bluetooth/l2cap.c sk_for_each(sk, node, &l2cap_sk_list.head) {
node 128 net/bluetooth/rfcomm/sock.c struct hlist_node *node;
node 130 net/bluetooth/rfcomm/sock.c sk_for_each(sk, node, &rfcomm_sk_list.head) {
node 136 net/bluetooth/rfcomm/sock.c return node ? sk : NULL;
node 145 net/bluetooth/rfcomm/sock.c struct hlist_node *node;
node 147 net/bluetooth/rfcomm/sock.c sk_for_each(sk, node, &rfcomm_sk_list.head) {
node 161 net/bluetooth/rfcomm/sock.c return node ? sk : sk1;
node 902 net/bluetooth/rfcomm/sock.c struct hlist_node *node;
node 907 net/bluetooth/rfcomm/sock.c sk_for_each(sk, node, &rfcomm_sk_list.head) {
node 289 net/bluetooth/sco.c struct hlist_node *node;
node 291 net/bluetooth/sco.c sk_for_each(sk, node, &sco_sk_list.head)
node 305 net/bluetooth/sco.c struct hlist_node *node;
node 309 net/bluetooth/sco.c sk_for_each(sk, node, &sco_sk_list.head) {
node 324 net/bluetooth/sco.c return node ? sk : sk1;
node 899 net/bluetooth/sco.c struct hlist_node *node;
node 904 net/bluetooth/sco.c sk_for_each(sk, node, &sco_sk_list.head) {
node 190 net/core/skbuff.c skb = kmem_cache_alloc_node(cache, gfp_mask & ~__GFP_DMA, node);
node 196 net/core/skbuff.c gfp_mask, node);
node 255 net/core/skbuff.c int node = dev->dev.parent ? dev_to_node(dev->dev.parent) : -1;
node 258 net/core/skbuff.c skb = __alloc_skb(length + NET_SKB_PAD, gfp_mask, 0, node);
node 268 net/core/skbuff.c int node = dev->dev.parent ? dev_to_node(dev->dev.parent) : -1;
node 271 net/core/skbuff.c page = alloc_pages_node(node, gfp_mask, 0);
node 2092 net/core/sock.c list_add(&prot->node, &proto_list);
node 2119 net/core/sock.c list_del(&prot->node);
node 2217 net/core/sock.c proto_seq_printf(seq, list_entry(v, struct proto, node));
node 179 net/decnet/af_decnet.c struct hlist_node *node;
node 184 net/decnet/af_decnet.c sk_for_each(sk, node, &dn_sk_hash[dn_ntohs(port) & DN_SK_HASH_MASK]) {
node 378 net/decnet/af_decnet.c struct hlist_node *node;
node 382 net/decnet/af_decnet.c sk_for_each(sk, node, list) {
node 418 net/decnet/af_decnet.c struct hlist_node *node;
node 422 net/decnet/af_decnet.c sk_for_each(sk, node, &dn_sk_hash[dn_ntohs(cb->dst_port) & DN_SK_HASH_MASK]) {
node 667 net/decnet/af_decnet.c unsigned short node, area;
node 669 net/decnet/af_decnet.c node = addr & 0x03ff;
node 671 net/decnet/af_decnet.c sprintf(buf, "%hd.%hd", area, node);
node 470 net/decnet/dn_table.c struct hlist_node *node;
node 485 net/decnet/dn_table.c hlist_for_each_entry(tb, node, &dn_fib_table_hash[h], hlist) {
node 814 net/decnet/dn_table.c struct hlist_node *node;
node 825 net/decnet/dn_table.c hlist_for_each_entry_rcu(t, node, &dn_fib_table_hash[h], hlist) {
node 871 net/decnet/dn_table.c struct hlist_node *node;
node 875 net/decnet/dn_table.c hlist_for_each_entry(tb, node, &dn_fib_table_hash[h], hlist)
node 894 net/decnet/dn_table.c struct hlist_node *node, *next;
node 899 net/decnet/dn_table.c hlist_for_each_entry_safe(t, node, next, &dn_fib_table_hash[h],
node 90 net/decnet/sysctl_net_decnet.c __u16 area, node;
node 109 net/decnet/sysctl_net_decnet.c node = *str++ - '0';
node 111 net/decnet/sysctl_net_decnet.c node *= 10;
node 112 net/decnet/sysctl_net_decnet.c node += (*str++ - '0');
node 115 net/decnet/sysctl_net_decnet.c node *= 10;
node 116 net/decnet/sysctl_net_decnet.c node += (*str++ - '0');
node 119 net/decnet/sysctl_net_decnet.c node *= 10;
node 120 net/decnet/sysctl_net_decnet.c node += (*str++ - '0');
node 123 net/decnet/sysctl_net_decnet.c if ((node > 1023) || (area > 63))
node 129 net/decnet/sysctl_net_decnet.c *addr = dn_htons((area << 10) | node);
node 782 net/econet/af_econet.c struct hlist_node *node;
node 784 net/econet/af_econet.c sk_for_each(sk, node, &econet_sklist) {
node 96 net/ipv4/fib_frontend.c struct hlist_node *node;
node 106 net/ipv4/fib_frontend.c hlist_for_each_entry_rcu(tb, node, head, tb_hlist) {
node 136 net/ipv4/fib_frontend.c struct hlist_node *node;
node 142 net/ipv4/fib_frontend.c hlist_for_each_entry(tb, node, head, tb_hlist)
node 631 net/ipv4/fib_frontend.c struct hlist_node *node;
node 645 net/ipv4/fib_frontend.c hlist_for_each_entry(tb, node, head, tb_hlist) {
node 1008 net/ipv4/fib_frontend.c struct hlist_node *node, *tmp;
node 1011 net/ipv4/fib_frontend.c hlist_for_each_entry_safe(tb, node, tmp, head, tb_hlist) {
node 1012 net/ipv4/fib_frontend.c hlist_del(node);
node 119 net/ipv4/fib_hash.c struct hlist_node *node, *n;
node 122 net/ipv4/fib_hash.c hlist_for_each_entry_safe(f, node, n, &old_ht[i], fn_hash) {
node 255 net/ipv4/fib_hash.c struct hlist_node *node;
node 260 net/ipv4/fib_hash.c hlist_for_each_entry(f, node, head, fn_hash) {
node 282 net/ipv4/fib_hash.c struct hlist_node *node;
node 297 net/ipv4/fib_hash.c hlist_for_each_entry(f, node, &fz->fz_hash[0], fn_hash) {
node 359 net/ipv4/fib_hash.c struct hlist_node *node;
node 362 net/ipv4/fib_hash.c hlist_for_each_entry(f, node, head, fn_hash) {
node 632 net/ipv4/fib_hash.c struct hlist_node *node, *n;
node 636 net/ipv4/fib_hash.c hlist_for_each_entry_safe(f, node, n, head, fn_hash) {
node 688 net/ipv4/fib_hash.c struct hlist_node *node;
node 694 net/ipv4/fib_hash.c hlist_for_each_entry(f, node, head, fn_hash) {
node 845 net/ipv4/fib_hash.c struct hlist_node *node;
node 848 net/ipv4/fib_hash.c hlist_for_each_entry(fn,node,iter->hash_head,fn_hash) {
node 884 net/ipv4/fib_hash.c struct hlist_node *node = &fn->fn_hash;
node 885 net/ipv4/fib_hash.c hlist_for_each_entry_continue(fn, node, fn_hash) {
node 902 net/ipv4/fib_hash.c struct hlist_node *node;
node 910 net/ipv4/fib_hash.c hlist_for_each_entry(fn, node, iter->hash_head, fn_hash) {
node 927 net/ipv4/fib_hash.c hlist_for_each_entry(fn, node, iter->hash_head, fn_hash) {
node 223 net/ipv4/fib_semantics.c struct hlist_node *node;
node 230 net/ipv4/fib_semantics.c hlist_for_each_entry(fi, node, head, fib_hash) {
node 255 net/ipv4/fib_semantics.c struct hlist_node *node;
node 263 net/ipv4/fib_semantics.c hlist_for_each_entry(nh, node, head, nh_hash) {
node 644 net/ipv4/fib_semantics.c struct hlist_node *node, *n;
node 647 net/ipv4/fib_semantics.c hlist_for_each_entry_safe(fi, node, n, head, fib_hash) {
node 662 net/ipv4/fib_semantics.c struct hlist_node *node, *n;
node 665 net/ipv4/fib_semantics.c hlist_for_each_entry_safe(fi, node, n, lhead, fib_lhash) {
node 1043 net/ipv4/fib_semantics.c struct hlist_node *node;
node 1049 net/ipv4/fib_semantics.c hlist_for_each_entry(fi, node, head, fib_lhash) {
node 1067 net/ipv4/fib_semantics.c struct hlist_node *node;
node 1073 net/ipv4/fib_semantics.c hlist_for_each_entry(nh, node, head, nh_hash) {
node 1124 net/ipv4/fib_semantics.c struct hlist_node *node;
node 1136 net/ipv4/fib_semantics.c hlist_for_each_entry(nh, node, head, nh_hash) {
node 92 net/ipv4/fib_trie.c #define NODE_TYPE(node) ((node)->parent & NODE_TYPE_MASK)
node 127 net/ipv4/fib_trie.c struct node *child[0];
node 152 net/ipv4/fib_trie.c struct node *trie;
node 161 net/ipv4/fib_trie.c static struct node *resize(struct trie *t, struct tnode *tn);
node 170 net/ipv4/fib_trie.c return (struct tnode *)(node->parent & ~NODE_TYPE_MASK);
node 175 net/ipv4/fib_trie.c struct tnode *ret = node_parent(node);
node 186 net/ipv4/fib_trie.c node->parent = (unsigned long)ptr | NODE_TYPE(node);
node 189 net/ipv4/fib_trie.c static inline struct node *tnode_get_child(struct tnode *tn, unsigned int i)
node 196 net/ipv4/fib_trie.c static inline struct node *tnode_get_child_rcu(struct tnode *tn, unsigned int i)
node 198 net/ipv4/fib_trie.c struct node *ret = tnode_get_child(tn, i);
node 370 net/ipv4/fib_trie.c (sizeof(struct node *) << tn->bits);
node 410 net/ipv4/fib_trie.c size_t sz = sizeof(struct tnode) + (sizeof(struct node *) << bits);
node 423 net/ipv4/fib_trie.c (unsigned long) (sizeof(struct node) << bits));
node 454 net/ipv4/fib_trie.c struct node *chi = tn->child[i];
node 481 net/ipv4/fib_trie.c static struct node *resize(struct trie *t, struct tnode *tn)
node 504 net/ipv4/fib_trie.c struct node *n;
node 664 net/ipv4/fib_trie.c struct node *n;
node 677 net/ipv4/fib_trie.c return (struct node *) tn;
node 724 net/ipv4/fib_trie.c put_child(t, tn, 2*i, (struct node *) left);
node 725 net/ipv4/fib_trie.c put_child(t, tn, 2*i+1, (struct node *) right);
node 731 net/ipv4/fib_trie.c struct node *node = tnode_get_child(oldtnode, i);
node 736 net/ipv4/fib_trie.c if (node == NULL)
node 741 net/ipv4/fib_trie.c if (IS_LEAF(node) || ((struct tnode *) node)->pos >
node 743 net/ipv4/fib_trie.c if (tkey_extract_bits(node->key,
node 746 net/ipv4/fib_trie.c put_child(t, tn, 2*i, node);
node 748 net/ipv4/fib_trie.c put_child(t, tn, 2*i+1, node);
node 753 net/ipv4/fib_trie.c inode = (struct tnode *) node;
node 826 net/ipv4/fib_trie.c struct node *left, *right;
node 857 net/ipv4/fib_trie.c put_child(t, tn, i/2, (struct node *)newn);
node 911 net/ipv4/fib_trie.c struct hlist_node *node;
node 914 net/ipv4/fib_trie.c hlist_for_each_entry_rcu(li, node, head, hlist)
node 934 net/ipv4/fib_trie.c struct hlist_node *node;
node 939 net/ipv4/fib_trie.c hlist_for_each_entry(li, node, head, hlist) {
node 959 net/ipv4/fib_trie.c struct node *n;
node 986 net/ipv4/fib_trie.c static struct node *trie_rebalance(struct trie *t, struct tnode *tn)
node 992 net/ipv4/fib_trie.c while (tn != NULL && (tp = node_parent((struct node *)tn)) != NULL) {
node 998 net/ipv4/fib_trie.c (struct node *)tn, wasfull);
node 1000 net/ipv4/fib_trie.c tp = node_parent((struct node *) tn);
node 1010 net/ipv4/fib_trie.c return (struct node *)tn;
node 1019 net/ipv4/fib_trie.c struct node *n;
node 1105 net/ipv4/fib_trie.c node_set_parent((struct node *)l, tp);
node 1108 net/ipv4/fib_trie.c put_child(t, (struct tnode *)tp, cindex, (struct node *)l);
node 1135 net/ipv4/fib_trie.c node_set_parent((struct node *)tn, tp);
node 1138 net/ipv4/fib_trie.c put_child(t, tn, missbit, (struct node *)l);
node 1144 net/ipv4/fib_trie.c (struct node *)tn);
node 1146 net/ipv4/fib_trie.c rcu_assign_pointer(t->trie, (struct node *)tn);
node 1340 net/ipv4/fib_trie.c struct hlist_node *node;
node 1342 net/ipv4/fib_trie.c hlist_for_each_entry_rcu(li, node, hhead, hlist) {
node 1371 net/ipv4/fib_trie.c struct node *n;
node 1544 net/ipv4/fib_trie.c struct tnode *parent = node_parent((struct node *) pn);
node 1571 net/ipv4/fib_trie.c struct tnode *tp = node_parent((struct node *) l);
node 1691 net/ipv4/fib_trie.c struct hlist_node *node, *tmp;
node 1694 net/ipv4/fib_trie.c hlist_for_each_entry_safe(li, node, tmp, lih, hlist) {
node 1735 net/ipv4/fib_trie.c c = (struct node *) p;
node 1756 net/ipv4/fib_trie.c struct node *c = (struct node *) l;
node 1915 net/ipv4/fib_trie.c struct hlist_node *node;
node 1922 net/ipv4/fib_trie.c hlist_for_each_entry_rcu(li, node, &l->list, hlist) {
node 2039 net/ipv4/fib_trie.c static struct node *fib_trie_get_next(struct fib_trie_iter *iter)
node 2053 net/ipv4/fib_trie.c struct node *n = tnode_get_child_rcu(tn, cindex);
node 2072 net/ipv4/fib_trie.c p = node_parent_rcu((struct node *)tn);
node 2084 net/ipv4/fib_trie.c static struct node *fib_trie_get_first(struct fib_trie_iter *iter,
node 2087 net/ipv4/fib_trie.c struct node *n;
node 2111 net/ipv4/fib_trie.c struct node *n;
node 2184 net/ipv4/fib_trie.c bytes += sizeof(struct node *) * pointers;
node 2229 net/ipv4/fib_trie.c struct hlist_node *node;
node 2232 net/ipv4/fib_trie.c hlist_for_each_entry_rcu(tb, node, head, tb_hlist) {
node 2265 net/ipv4/fib_trie.c static struct node *fib_trie_get_idx(struct seq_file *seq, loff_t pos)
node 2274 net/ipv4/fib_trie.c struct hlist_node *node;
node 2277 net/ipv4/fib_trie.c hlist_for_each_entry_rcu(tb, node, head, tb_hlist) {
node 2278 net/ipv4/fib_trie.c struct node *n;
node 2307 net/ipv4/fib_trie.c struct node *n;
node 2392 net/ipv4/fib_trie.c struct node *n = v;
node 2409 net/ipv4/fib_trie.c struct hlist_node *node;
node 2415 net/ipv4/fib_trie.c hlist_for_each_entry_rcu(li, node, &l->list, hlist) {
node 2561 net/ipv4/fib_trie.c struct hlist_node *node;
node 2570 net/ipv4/fib_trie.c hlist_for_each_entry_rcu(li, node, &l->list, hlist) {
node 57 net/ipv4/inet_connection_sock.c struct hlist_node *node;
node 67 net/ipv4/inet_connection_sock.c sk_for_each_bound(sk2, node, &tb->owners) {
node 82 net/ipv4/inet_connection_sock.c return node != NULL;
node 94 net/ipv4/inet_connection_sock.c struct hlist_node *node;
node 111 net/ipv4/inet_connection_sock.c inet_bind_bucket_for_each(tb, node, &head->chain)
node 139 net/ipv4/inet_connection_sock.c inet_bind_bucket_for_each(tb, node, &head->chain)
node 726 net/ipv4/inet_diag.c struct hlist_node *node;
node 729 net/ipv4/inet_diag.c sk_for_each(sk, node, &hashinfo->listening_hash[i]) {
node 783 net/ipv4/inet_diag.c struct hlist_node *node;
node 794 net/ipv4/inet_diag.c sk_for_each(sk, node, &head->chain) {
node 818 net/ipv4/inet_diag.c inet_twsk_for_each(tw, node,
node 42 net/ipv4/inet_hashtables.c hlist_add_head(&tb->node, &head->chain);
node 53 net/ipv4/inet_hashtables.c __hlist_del(&tb->node);
node 155 net/ipv4/inet_hashtables.c const struct hlist_node *node;
node 158 net/ipv4/inet_hashtables.c sk_for_each(sk, node, head) {
node 226 net/ipv4/inet_hashtables.c const struct hlist_node *node;
node 236 net/ipv4/inet_hashtables.c sk_for_each(sk, node, &head->chain) {
node 243 net/ipv4/inet_hashtables.c sk_for_each(sk, node, &head->twchain) {
node 275 net/ipv4/inet_hashtables.c const struct hlist_node *node;
node 282 net/ipv4/inet_hashtables.c sk_for_each(sk2, node, &head->twchain) {
node 296 net/ipv4/inet_hashtables.c sk_for_each(sk2, node, &head->chain) {
node 434 net/ipv4/inet_hashtables.c struct hlist_node *node;
node 451 net/ipv4/inet_hashtables.c inet_bind_bucket_for_each(tb, node, &head->chain) {
node 147 net/ipv4/inet_timewait_sock.c struct hlist_node *node;
node 160 net/ipv4/inet_timewait_sock.c inet_twsk_for_each_inmate(tw, node, &twdr->cells[slot]) {
node 373 net/ipv4/inet_timewait_sock.c struct hlist_node *node, *safe;
node 376 net/ipv4/inet_timewait_sock.c inet_twsk_for_each_inmate_safe(tw, node, safe,
node 419 net/ipv4/inet_timewait_sock.c struct hlist_node *node;
node 429 net/ipv4/inet_timewait_sock.c sk_for_each(sk, node, &head->twchain) {
node 188 net/ipv4/inetpeer.c struct inet_peer **nodep, *node, *l, *r;
node 193 net/ipv4/inetpeer.c node = *nodep;
node 194 net/ipv4/inetpeer.c l = node->avl_left;
node 195 net/ipv4/inetpeer.c r = node->avl_right;
node 205 net/ipv4/inetpeer.c node->avl_left = lr; /* lr: RH or RH+1 */
node 206 net/ipv4/inetpeer.c node->avl_right = r; /* r: RH */
node 207 net/ipv4/inetpeer.c node->avl_height = lrh + 1; /* RH+1 or RH+2 */
node 209 net/ipv4/inetpeer.c l->avl_right = node; /* node: RH+1 or RH+2 */
node 210 net/ipv4/inetpeer.c l->avl_height = node->avl_height + 1;
node 215 net/ipv4/inetpeer.c node->avl_left = lrr; /* lrr: RH or RH-1 */
node 216 net/ipv4/inetpeer.c node->avl_right = r; /* r: RH */
node 217 net/ipv4/inetpeer.c node->avl_height = rh + 1; /* node: RH+1 */
node 222 net/ipv4/inetpeer.c lr->avl_right = node; /* node: RH+1 */
node 233 net/ipv4/inetpeer.c node->avl_right = rl; /* rl: LH or LH+1 */
node 234 net/ipv4/inetpeer.c node->avl_left = l; /* l: LH */
node 235 net/ipv4/inetpeer.c node->avl_height = rlh + 1; /* LH+1 or LH+2 */
node 237 net/ipv4/inetpeer.c r->avl_left = node; /* node: LH+1 or LH+2 */
node 238 net/ipv4/inetpeer.c r->avl_height = node->avl_height + 1;
node 243 net/ipv4/inetpeer.c node->avl_right = rll; /* rll: LH or LH-1 */
node 244 net/ipv4/inetpeer.c node->avl_left = l; /* l: LH */
node 245 net/ipv4/inetpeer.c node->avl_height = lh + 1; /* node: LH+1 */
node 250 net/ipv4/inetpeer.c rl->avl_left = node; /* node: LH+1 */
node 255 net/ipv4/inetpeer.c node->avl_height = (lh > rh ? lh : rh) + 1;
node 113 net/ipv4/raw.c struct hlist_node *node;
node 115 net/ipv4/raw.c sk_for_each_from(sk, node) {
node 858 net/ipv4/raw.c struct hlist_node *node;
node 860 net/ipv4/raw.c sk_for_each(sk, node, &state->h->ht[state->bucket])
node 1878 net/ipv4/tcp_ipv4.c struct hlist_node *node;
node 1921 net/ipv4/tcp_ipv4.c sk_for_each_from(sk, node) {
node 1972 net/ipv4/tcp_ipv4.c struct hlist_node *node;
node 1981 net/ipv4/tcp_ipv4.c sk_for_each(sk, node, &tcp_hashinfo.ehash[st->bucket].chain) {
node 1990 net/ipv4/tcp_ipv4.c inet_twsk_for_each(tw, node,
node 2010 net/ipv4/tcp_ipv4.c struct hlist_node *node;
node 2042 net/ipv4/tcp_ipv4.c sk_for_each_from(sk, node) {
node 132 net/ipv4/udp.c struct hlist_node *node;
node 134 net/ipv4/udp.c sk_for_each(sk2, node, &udptable[udp_hashfn(net, num)])
node 219 net/ipv4/udp.c struct hlist_node *node;
node 224 net/ipv4/udp.c sk_for_each(sk, node, &udptable[udp_hashfn(net, hnum)]) {
node 292 net/ipv4/udp.c struct hlist_node *node;
node 296 net/ipv4/udp.c sk_for_each_from(s, node) {
node 1508 net/ipv4/udp.c struct hlist_node *node;
node 1509 net/ipv4/udp.c sk_for_each(sk, node, state->hashtable + state->bucket) {
node 33 net/ipv6/inet6_connection_sock.c const struct hlist_node *node;
node 40 net/ipv6/inet6_connection_sock.c sk_for_each_bound(sk2, node, &tb->owners) {
node 51 net/ipv6/inet6_connection_sock.c return node != NULL;
node 66 net/ipv6/inet6_hashtables.c const struct hlist_node *node;
node 77 net/ipv6/inet6_hashtables.c sk_for_each(sk, node, &head->chain) {
node 83 net/ipv6/inet6_hashtables.c sk_for_each(sk, node, &head->twchain) {
node 102 net/ipv6/inet6_hashtables.c const struct hlist_node *node;
node 107 net/ipv6/inet6_hashtables.c sk_for_each(sk, node,
node 175 net/ipv6/inet6_hashtables.c const struct hlist_node *node;
node 182 net/ipv6/inet6_hashtables.c sk_for_each(sk2, node, &head->twchain) {
node 195 net/ipv6/inet6_hashtables.c sk_for_each(sk2, node, &head->chain) {
node 229 net/ipv6/ip6_fib.c struct hlist_node *node;
node 237 net/ipv6/ip6_fib.c hlist_for_each_entry_rcu(tb, node, head, tb6_hlist) {
node 353 net/ipv6/ip6_fib.c struct hlist_node *node;
node 387 net/ipv6/ip6_fib.c hlist_for_each_entry(tb, node, head, tb6_hlist) {
node 1052 net/ipv6/ip6_fib.c w->root = w->node = NULL;
node 1054 net/ipv6/ip6_fib.c } else if (w->node == fn) {
node 1056 net/ipv6/ip6_fib.c w->node = pn;
node 1064 net/ipv6/ip6_fib.c if (w->node == fn) {
node 1065 net/ipv6/ip6_fib.c w->node = child;
node 1223 net/ipv6/ip6_fib.c fn = w->node;
node 1236 net/ipv6/ip6_fib.c w->node = FIB6_SUBTREE(fn);
node 1243 net/ipv6/ip6_fib.c w->node = fn->left;
node 1250 net/ipv6/ip6_fib.c w->node = fn->right;
node 1268 net/ipv6/ip6_fib.c w->node = pn;
node 1282 net/ipv6/ip6_fib.c w->leaf = w->node->leaf;
node 1297 net/ipv6/ip6_fib.c w->node = w->root;
node 1365 net/ipv6/ip6_fib.c struct hlist_node *node;
node 1372 net/ipv6/ip6_fib.c hlist_for_each_entry_rcu(table, node, head, tb6_hlist) {
node 71 net/ipv6/raw.c struct hlist_node *node;
node 74 net/ipv6/raw.c sk_for_each_from(sk, node)
node 63 net/ipv6/udp.c struct hlist_node *node;
node 68 net/ipv6/udp.c sk_for_each(sk, node, &udptable[udp_hashfn(net, hnum)]) {
node 336 net/ipv6/udp.c struct hlist_node *node;
node 340 net/ipv6/udp.c sk_for_each_from(s, node) {
node 101 net/ipx/af_ipx.c struct ipx_interface, node);
node 172 net/ipx/af_ipx.c list_for_each_entry(i, &ipx_interfaces, node)
node 199 net/ipx/af_ipx.c list_for_each_entry(i, &ipx_interfaces, node)
node 231 net/ipx/af_ipx.c struct hlist_node *node;
node 233 net/ipx/af_ipx.c sk_for_each(s, node, &intrfc->if_sklist)
node 262 net/ipx/af_ipx.c struct hlist_node *node;
node 267 net/ipx/af_ipx.c sk_for_each(s, node, &intrfc->if_sklist) {
node 271 net/ipx/af_ipx.c !memcmp(ipx_node, ipxs->node, IPX_NODE_LEN))
node 285 net/ipx/af_ipx.c struct hlist_node *node, *t;
node 292 net/ipx/af_ipx.c sk_for_each_safe(s, node, t, &intrfc->if_sklist) {
node 306 net/ipx/af_ipx.c list_del(&intrfc->node);
node 345 net/ipx/af_ipx.c list_for_each_entry_safe(i, tmp, &ipx_interfaces, node)
node 363 net/ipx/af_ipx.c list_for_each_entry_safe(i, tmp, &ipx_interfaces, node)
node 385 net/ipx/af_ipx.c int is_broadcast = !memcmp(ipx->ipx_dest.node, ipx_broadcast_node,
node 388 net/ipx/af_ipx.c struct hlist_node *node;
node 393 net/ipx/af_ipx.c sk_for_each(s, node, &intrfc->if_sklist) {
node 397 net/ipx/af_ipx.c (is_broadcast || !memcmp(ipx->ipx_dest.node,
node 398 net/ipx/af_ipx.c ipxs->node, IPX_NODE_LEN))) {
node 449 net/ipx/af_ipx.c struct hlist_node *node;
node 454 net/ipx/af_ipx.c sk_for_each(sk, node, &intrfc->if_sklist)
node 625 net/ipx/af_ipx.c !memcmp(intrfc->if_node, node, IPX_NODE_LEN)) {
node 634 net/ipx/af_ipx.c if (!memcmp(ipx_broadcast_node, node, IPX_NODE_LEN)) {
node 667 net/ipx/af_ipx.c if (!memcmp(ipx_broadcast_node, node, IPX_NODE_LEN))
node 670 net/ipx/af_ipx.c memcpy(dest_node, &(node[IPX_NODE_LEN-addr_len]), addr_len);
node 736 net/ipx/af_ipx.c if (!memcmp(ipx_broadcast_node, ipx->ipx_dest.node, IPX_NODE_LEN) ||
node 737 net/ipx/af_ipx.c !memcmp(intrfc->if_node, ipx->ipx_dest.node, IPX_NODE_LEN)) {
node 843 net/ipx/af_ipx.c list_for_each_entry(ifcs, &ipx_interfaces, node) {
node 874 net/ipx/af_ipx.c list_add_tail(&intrfc->node, &ipx_interfaces);
node 1468 net/ipx/af_ipx.c memcpy(ipxs->node, intrfc->if_node, IPX_NODE_LEN);
node 1470 net/ipx/af_ipx.c memcpy(ipxs->node, addr->sipx_node, IPX_NODE_LEN);
node 1473 net/ipx/af_ipx.c if (ipxitf_find_internal_socket(intrfc, ipxs->node,
node 1486 net/ipx/af_ipx.c memcpy(ipxs->node, intrfc->if_node, IPX_NODE_LEN);
node 1567 net/ipx/af_ipx.c memcpy(ipxs->dest_addr.node, addr->sipx_node, IPX_NODE_LEN);
node 1602 net/ipx/af_ipx.c memcpy(sipx.sipx_node, addr->node, IPX_NODE_LEN);
node 1607 net/ipx/af_ipx.c memcpy(sipx.sipx_node, ipxs->node, IPX_NODE_LEN);
node 1745 net/ipx/af_ipx.c memcpy(usipx->sipx_node, ipxs->dest_addr.node, IPX_NODE_LEN);
node 1815 net/ipx/af_ipx.c memcpy(sipx->sipx_node, ipx->ipx_source.node, IPX_NODE_LEN);
node 20 net/ipx/ipx_proc.c list_for_each_entry(i, &ipx_interfaces, node)
node 32 net/ipx/ipx_proc.c if (i->node.next != &ipx_interfaces)
node 33 net/ipx/ipx_proc.c rc = list_entry(i->node.next, struct ipx_interface, node);
node 97 net/ipx/ipx_proc.c rc = list_entry(ipx_routes.next, struct ipx_route, node);
node 105 net/ipx/ipx_proc.c if (r->node.next != &ipx_routes)
node 106 net/ipx/ipx_proc.c rc = list_entry(r->node.next, struct ipx_route, node);
node 114 net/ipx/ipx_proc.c list_for_each_entry(r, &ipx_routes, node)
node 171 net/ipx/ipx_proc.c struct hlist_node *node;
node 174 net/ipx/ipx_proc.c list_for_each_entry(i, &ipx_interfaces, node) {
node 176 net/ipx/ipx_proc.c sk_for_each(s, node, &i->if_sklist) {
node 183 net/ipx/ipx_proc.c if (node)
node 265 net/ipx/ipx_proc.c ipxs->node[0], ipxs->node[1], ipxs->node[2], ipxs->node[3],
node 266 net/ipx/ipx_proc.c ipxs->node[4], ipxs->node[5], ntohs(ipxs->port));
node 276 net/ipx/ipx_proc.c ipxs->dest_addr.node[0], ipxs->dest_addr.node[1],
node 277 net/ipx/ipx_proc.c ipxs->dest_addr.node[2], ipxs->dest_addr.node[3],
node 278 net/ipx/ipx_proc.c ipxs->dest_addr.node[4], ipxs->dest_addr.node[5],
node 37 net/ipx/ipx_route.c list_for_each_entry(r, &ipx_routes, node)
node 68 net/ipx/ipx_route.c list_add(&rt->node, &ipx_routes);
node 78 net/ipx/ipx_route.c if (!node) {
node 82 net/ipx/ipx_route.c memcpy(rt->ir_router_node, node, IPX_NODE_LEN);
node 98 net/ipx/ipx_route.c list_for_each_entry_safe(r, tmp, &ipx_routes, node)
node 100 net/ipx/ipx_route.c list_del(&r->node);
node 127 net/ipx/ipx_route.c list_for_each_entry_safe(r, tmp, &ipx_routes, node)
node 133 net/ipx/ipx_route.c list_del(&r->node);
node 160 net/ipx/ipx_route.c r->ir_router_node : ipx->ipx_dest.node);
node 217 net/ipx/ipx_route.c memcpy(ipx->ipx_source.node, ipxs->node, IPX_NODE_LEN);
node 223 net/ipx/ipx_route.c memcpy(ipx->ipx_source.node, intrfc->if_node, IPX_NODE_LEN);
node 226 net/ipx/ipx_route.c memcpy(ipx->ipx_source.node, ipxs->intrfc->if_node,
node 232 net/ipx/ipx_route.c memcpy(ipx->ipx_dest.node, usipx->sipx_node, IPX_NODE_LEN);
node 248 net/ipx/ipx_route.c rt->ir_router_node : ipx->ipx_dest.node);
node 62 net/irda/discovery.c discovery_t *discovery, *node;
node 78 net/irda/discovery.c node = discovery;
node 83 net/irda/discovery.c if ((node->data.saddr == new->data.saddr) &&
node 84 net/irda/discovery.c ((node->data.daddr == new->data.daddr) ||
node 85 net/irda/discovery.c (strcmp(node->data.info, new->data.info) == 0)))
node 90 net/irda/discovery.c hashbin_remove_this(cachelog, (irda_queue_t *) node);
node 92 net/irda/discovery.c if (get_unaligned((__u16 *)node->data.hints) == get_unaligned((__u16 *)new->data.hints))
node 94 net/irda/discovery.c new->firststamp = node->firststamp;
node 95 net/irda/discovery.c kfree(node);
node 128 net/irda/irias_object.c struct ias_object *node;
node 134 net/irda/irias_object.c node = hashbin_remove_this(irias_objects, (irda_queue_t *) obj);
node 135 net/irda/irias_object.c if (!node)
node 156 net/irda/irias_object.c struct ias_attrib *node;
node 163 net/irda/irias_object.c node = hashbin_remove_this(obj->attribs, (irda_queue_t *) attrib);
node 164 net/irda/irias_object.c if (!node)
node 168 net/irda/irias_object.c __irias_delete_attrib(node);
node 175 net/irda/irias_object.c node = (struct ias_attrib *) hashbin_get_first(obj->attribs);
node 176 net/irda/irias_object.c if (cleanobject && !node)
node 100 net/iucv/af_iucv.c struct hlist_node *node;
node 102 net/iucv/af_iucv.c sk_for_each(sk, node, &iucv_sk_list.head)
node 978 net/iucv/af_iucv.c struct hlist_node *node;
node 989 net/iucv/af_iucv.c sk_for_each(sk, node, &iucv_sk_list.head)
node 261 net/key/af_key.c struct hlist_node *node;
node 272 net/key/af_key.c sk_for_each(sk, node, &pfkey_table) {
node 3649 net/key/af_key.c struct hlist_node *node;
node 3656 net/key/af_key.c sk_for_each(s, node, &pfkey_table)
node 69 net/lapb/lapb_iface.c if (lapb->node.next) {
node 70 net/lapb/lapb_iface.c list_del(&lapb->node);
node 80 net/lapb/lapb_iface.c list_add(&lapb->node, &lapb_list);
node 90 net/lapb/lapb_iface.c lapb = list_entry(entry, struct lapb_cb, node);
node 492 net/llc/llc_conn.c struct hlist_node *node;
node 495 net/llc/llc_conn.c sk_for_each(rc, node, &sap->sk_list.list) {
node 538 net/llc/llc_conn.c struct hlist_node *node;
node 541 net/llc/llc_conn.c sk_for_each(rc, node, &sap->sk_list.list) {
node 54 net/llc/llc_core.c list_add_tail(&sap->node, &llc_sap_list);
node 66 net/llc/llc_core.c list_del(&sap->node);
node 74 net/llc/llc_core.c list_for_each_entry(sap, &llc_sap_list, node)
node 38 net/llc/llc_proc.c struct hlist_node *node;
node 42 net/llc/llc_proc.c sap = list_entry(sap_entry, struct llc_sap, node);
node 45 net/llc/llc_proc.c sk_for_each(sk, node, &sap->sk_list.list) {
node 87 net/llc/llc_proc.c if (sap->node.next == &llc_sap_list)
node 89 net/llc/llc_proc.c sap = list_entry(sap->node.next, struct llc_sap, node);
node 312 net/llc/llc_sap.c struct hlist_node *node;
node 315 net/llc/llc_sap.c sk_for_each(rc, node, &sap->sk_list.list) {
node 344 net/llc/llc_sap.c struct hlist_node *node;
node 347 net/llc/llc_sap.c sk_for_each(sk, node, &sap->sk_list.list) {
node 80 net/llc/llc_station.c struct list_head node; /* node in station->ev_q.list */
node 298 net/mac80211/mesh.h hlist_for_each_entry_rcu(node, p, &x->hash_buckets[i], list)
node 77 net/mac80211/mesh_pathtbl.c struct mpath_node *node;
node 82 net/mac80211/mesh_pathtbl.c hlist_for_each_entry_rcu(node, n, bucket, list) {
node 83 net/mac80211/mesh_pathtbl.c mpath = node->mpath;
node 104 net/mac80211/mesh_pathtbl.c struct mpath_node *node;
node 109 net/mac80211/mesh_pathtbl.c hlist_for_each_entry_rcu(node, n, bucket, list) {
node 110 net/mac80211/mesh_pathtbl.c mpath = node->mpath;
node 137 net/mac80211/mesh_pathtbl.c struct mpath_node *node;
node 142 net/mac80211/mesh_pathtbl.c for_each_mesh_entry(mesh_paths, p, node, i) {
node 143 net/mac80211/mesh_pathtbl.c if (sdata && node->mpath->sdata != sdata)
node 146 net/mac80211/mesh_pathtbl.c if (MPATH_EXPIRED(node->mpath)) {
node 147 net/mac80211/mesh_pathtbl.c spin_lock_bh(&node->mpath->state_lock);
node 148 net/mac80211/mesh_pathtbl.c if (MPATH_EXPIRED(node->mpath))
node 149 net/mac80211/mesh_pathtbl.c node->mpath->flags &= ~MESH_PATH_ACTIVE;
node 150 net/mac80211/mesh_pathtbl.c spin_unlock_bh(&node->mpath->state_lock);
node 152 net/mac80211/mesh_pathtbl.c return node->mpath;
node 171 net/mac80211/mesh_pathtbl.c struct mpath_node *node, *new_node;
node 215 net/mac80211/mesh_pathtbl.c hlist_for_each_entry(node, n, bucket, list) {
node 216 net/mac80211/mesh_pathtbl.c mpath = node->mpath;
node 261 net/mac80211/mesh_pathtbl.c struct mpath_node *node, *new_node;
node 301 net/mac80211/mesh_pathtbl.c hlist_for_each_entry(node, n, bucket, list) {
node 302 net/mac80211/mesh_pathtbl.c mpath = node->mpath;
node 354 net/mac80211/mesh_pathtbl.c struct mpath_node *node;
node 360 net/mac80211/mesh_pathtbl.c for_each_mesh_entry(mesh_paths, p, node, i) {
node 361 net/mac80211/mesh_pathtbl.c mpath = node->mpath;
node 392 net/mac80211/mesh_pathtbl.c struct mpath_node *node;
node 396 net/mac80211/mesh_pathtbl.c for_each_mesh_entry(mesh_paths, p, node, i) {
node 397 net/mac80211/mesh_pathtbl.c mpath = node->mpath;
node 406 net/mac80211/mesh_pathtbl.c struct mpath_node *node;
node 410 net/mac80211/mesh_pathtbl.c for_each_mesh_entry(mesh_paths, p, node, i) {
node 411 net/mac80211/mesh_pathtbl.c mpath = node->mpath;
node 419 net/mac80211/mesh_pathtbl.c struct mpath_node *node = container_of(rp, struct mpath_node, rcu);
node 420 net/mac80211/mesh_pathtbl.c struct ieee80211_sub_if_data *sdata = node->mpath->sdata;
node 422 net/mac80211/mesh_pathtbl.c del_timer_sync(&node->mpath->timer);
node 424 net/mac80211/mesh_pathtbl.c kfree(node->mpath);
node 425 net/mac80211/mesh_pathtbl.c kfree(node);
node 439 net/mac80211/mesh_pathtbl.c struct mpath_node *node;
node 450 net/mac80211/mesh_pathtbl.c hlist_for_each_entry(node, n, bucket, list) {
node 451 net/mac80211/mesh_pathtbl.c mpath = node->mpath;
node 456 net/mac80211/mesh_pathtbl.c hlist_del_rcu(&node->list);
node 457 net/mac80211/mesh_pathtbl.c call_rcu(&node->rcu, mesh_path_node_reclaim);
node 562 net/mac80211/mesh_pathtbl.c struct mpath_node *node = hlist_entry(p, struct mpath_node, list);
node 563 net/mac80211/mesh_pathtbl.c mpath = node->mpath;
node 567 net/mac80211/mesh_pathtbl.c kfree(node);
node 573 net/mac80211/mesh_pathtbl.c struct mpath_node *node, *new_node;
node 580 net/mac80211/mesh_pathtbl.c node = hlist_entry(p, struct mpath_node, list);
node 581 net/mac80211/mesh_pathtbl.c mpath = node->mpath;
node 613 net/mac80211/mesh_pathtbl.c struct mpath_node *node;
node 618 net/mac80211/mesh_pathtbl.c for_each_mesh_entry(mesh_paths, p, node, i) {
node 619 net/mac80211/mesh_pathtbl.c if (node->mpath->sdata != sdata)
node 621 net/mac80211/mesh_pathtbl.c mpath = node->mpath;
node 68 net/netfilter/xt_hashlimit.c struct hlist_node node;
node 81 net/netfilter/xt_hashlimit.c struct hlist_node node; /* global list of all htables */
node 135 net/netfilter/xt_hashlimit.c hlist_for_each_entry(ent, pos, &ht->hash[hash], node)
node 174 net/netfilter/xt_hashlimit.c hlist_add_head(&ent->node, &ht->hash[hash_dst(ht, dst)]);
node 182 net/netfilter/xt_hashlimit.c hlist_del(&ent->node);
node 254 net/netfilter/xt_hashlimit.c hlist_add_head(&hinfo->node, &hashlimit_htables);
node 316 net/netfilter/xt_hashlimit.c hlist_add_head(&hinfo->node, &hashlimit_htables);
node 345 net/netfilter/xt_hashlimit.c hlist_for_each_entry_safe(dh, pos, n, &ht->hash[i], node) {
node 384 net/netfilter/xt_hashlimit.c hlist_for_each_entry(hinfo, pos, &hashlimit_htables, node) {
node 400 net/netfilter/xt_hashlimit.c hlist_del(&hinfo->node);
node 932 net/netfilter/xt_hashlimit.c hlist_for_each_entry(ent, pos, &htable->hash[*bucket], node)
node 226 net/netlink/af_netlink.c struct hlist_node *node;
node 230 net/netlink/af_netlink.c sk_for_each(sk, node, head) {
node 290 net/netlink/af_netlink.c struct hlist_node *node, *tmp;
node 292 net/netlink/af_netlink.c sk_for_each_safe(sk, node, tmp, &otable[i])
node 322 net/netlink/af_netlink.c struct hlist_node *node;
node 328 net/netlink/af_netlink.c sk_for_each_bound(sk, node, &tbl->mc_list) {
node 344 net/netlink/af_netlink.c struct hlist_node *node;
node 350 net/netlink/af_netlink.c sk_for_each(osk, node, head) {
node 355 net/netlink/af_netlink.c if (node)
node 525 net/netlink/af_netlink.c struct hlist_node *node;
node 534 net/netlink/af_netlink.c sk_for_each(osk, node, head) {
node 1011 net/netlink/af_netlink.c struct hlist_node *node;
node 1031 net/netlink/af_netlink.c sk_for_each_bound(sk, node, &nl_table[ssk->sk_protocol].mc_list)
node 1083 net/netlink/af_netlink.c struct hlist_node *node;
node 1093 net/netlink/af_netlink.c sk_for_each_bound(sk, node, &nl_table[ssk->sk_protocol].mc_list)
node 1492 net/netlink/af_netlink.c struct hlist_node *node;
node 1497 net/netlink/af_netlink.c sk_for_each_bound(sk, node, &tbl->mc_list)
node 1754 net/netlink/af_netlink.c struct hlist_node *node;
node 1761 net/netlink/af_netlink.c sk_for_each(s, node, &hash->table[j]) {
node 107 net/netrom/af_netrom.c struct hlist_node *node;
node 110 net/netrom/af_netrom.c sk_for_each(s, node, &nr_list)
node 152 net/netrom/af_netrom.c struct hlist_node *node;
node 155 net/netrom/af_netrom.c sk_for_each(s, node, &nr_list)
node 173 net/netrom/af_netrom.c struct hlist_node *node;
node 176 net/netrom/af_netrom.c sk_for_each(s, node, &nr_list) {
node 197 net/netrom/af_netrom.c struct hlist_node *node;
node 200 net/netrom/af_netrom.c sk_for_each(s, node, &nr_list) {
node 1264 net/netrom/af_netrom.c struct hlist_node *node;
node 1271 net/netrom/af_netrom.c sk_for_each(s, node, &nr_list) {
node 51 net/netrom/nr_route.c struct hlist_node *node;
node 54 net/netrom/nr_route.c nr_node_for_each(nr_node, node, &nr_node_list)
node 69 net/netrom/nr_route.c struct hlist_node *node;
node 72 net/netrom/nr_route.c nr_neigh_for_each(nr_neigh, node, &nr_neigh_list)
node 116 net/netrom/nr_route.c struct hlist_node *node;
node 119 net/netrom/nr_route.c nr_node_for_each(nr_nodet, node, &nr_node_list) {
node 484 net/netrom/nr_route.c struct hlist_node *node, *nodet;
node 488 net/netrom/nr_route.c nr_node_for_each_safe(s, node, nodet, &nr_node_list) {
node 538 net/netrom/nr_route.c struct hlist_node *node, *nodet, *node2, *node2t;
node 543 net/netrom/nr_route.c nr_neigh_for_each_safe(s, node, nodet, &nr_neigh_list) {
node 728 net/netrom/nr_route.c struct hlist_node *node;
node 732 net/netrom/nr_route.c nr_neigh_for_each(s, node, &nr_neigh_list) {
node 752 net/netrom/nr_route.c nr_node_for_each(nr_node, node, &nr_node_list) {
node 862 net/netrom/nr_route.c struct hlist_node *node;
node 869 net/netrom/nr_route.c nr_node_for_each(nr_node, node, &nr_node_list) {
node 880 net/netrom/nr_route.c struct hlist_node *node;
node 883 net/netrom/nr_route.c node = (v == SEQ_START_TOKEN)
node 887 net/netrom/nr_route.c return hlist_entry(node, struct nr_node, node_node);
node 948 net/netrom/nr_route.c struct hlist_node *node;
node 955 net/netrom/nr_route.c nr_neigh_for_each(nr_neigh, node, &nr_neigh_list) {
node 964 net/netrom/nr_route.c struct hlist_node *node;
node 967 net/netrom/nr_route.c node = (v == SEQ_START_TOKEN)
node 971 net/netrom/nr_route.c return hlist_entry(node, struct nr_neigh, neigh_node);
node 1038 net/netrom/nr_route.c struct hlist_node *node, *nodet;
node 1042 net/netrom/nr_route.c nr_node_for_each_safe(t, node, nodet, &nr_node_list) {
node 1047 net/netrom/nr_route.c nr_neigh_for_each_safe(s, node, nodet, &nr_neigh_list) {
node 1588 net/packet/af_packet.c struct hlist_node *node;
node 1593 net/packet/af_packet.c sk_for_each(sk, node, &net->packet.sklist) {
node 2004 net/packet/af_packet.c struct hlist_node *node;
node 2006 net/packet/af_packet.c sk_for_each(s, node, &net->packet.sklist) {
node 527 net/phonet/pep.c struct hlist_node *node;
node 531 net/phonet/pep.c sk_for_each(sknode, node, hlist) {
node 62 net/phonet/socket.c struct hlist_node *node;
node 70 net/phonet/socket.c sk_for_each(sknode, node, &pnsocks.hlist) {
node 251 net/rfkill/rfkill.c list_for_each_entry(rfkill, &rfkill_list, node) {
node 292 net/rfkill/rfkill.c list_for_each_entry(rfkill, &rfkill_list, node) {
node 597 net/rfkill/rfkill.c list_for_each_entry(p, &rfkill_list, node) {
node 630 net/rfkill/rfkill.c list_add_tail(&rfkill->node, &rfkill_list);
node 642 net/rfkill/rfkill.c list_del_init(&rfkill->node);
node 680 net/rfkill/rfkill.c INIT_LIST_HEAD(&rfkill->node);
node 170 net/rose/af_rose.c struct hlist_node *node;
node 173 net/rose/af_rose.c sk_for_each(s, node, &rose_list) {
node 191 net/rose/af_rose.c struct hlist_node *node;
node 194 net/rose/af_rose.c sk_for_each(s, node, &rose_list) {
node 251 net/rose/af_rose.c struct hlist_node *node;
node 254 net/rose/af_rose.c sk_for_each(s, node, &rose_list) {
node 263 net/rose/af_rose.c sk_for_each(s, node, &rose_list) {
node 283 net/rose/af_rose.c struct hlist_node *node;
node 286 net/rose/af_rose.c sk_for_each(s, node, &rose_list) {
node 1405 net/rose/af_rose.c struct hlist_node *node;
node 1412 net/rose/af_rose.c sk_for_each(s, node, &rose_list) {
node 671 net/rose/rose_route.c struct rose_node *node;
node 676 net/rose/rose_route.c for (node = rose_node_list; node != NULL; node = node->next) {
node 677 net/rose/rose_route.c if (rosecmpm(addr, &node->address, node->mask) == 0) {
node 678 net/rose/rose_route.c for (i = 0; i < node->count; i++) {
node 680 net/rose/rose_route.c if (node->neighbour[i]->restarted) {
node 681 net/rose/rose_route.c res = node->neighbour[i];
node 686 net/rose/rose_route.c if (!rose_ftimer_running(node->neighbour[i])) {
node 687 net/rose/rose_route.c res = node->neighbour[i];
node 87 net/rxrpc/ar-connection.c bundle = rb_entry(p, struct rxrpc_conn_bundle, node);
node 116 net/rxrpc/ar-connection.c bundle = rb_entry(parent, struct rxrpc_conn_bundle, node);
node 130 net/rxrpc/ar-connection.c rb_link_node(&bundle->node, parent, pp);
node 131 net/rxrpc/ar-connection.c rb_insert_color(&bundle->node, &trans->bundles);
node 177 net/rxrpc/ar-connection.c rb_erase(&bundle->node, &trans->bundles);
node 247 net/rxrpc/ar-connection.c xconn = rb_entry(parent, struct rxrpc_connection, node);
node 263 net/rxrpc/ar-connection.c rb_link_node(&conn->node, parent, p);
node 264 net/rxrpc/ar-connection.c rb_insert_color(&conn->node, &conn->trans->client_conns);
node 287 net/rxrpc/ar-connection.c xconn = rb_entry(parent, struct rxrpc_connection, node);
node 627 net/rxrpc/ar-connection.c conn = rb_entry(p, struct rxrpc_connection, node);
node 670 net/rxrpc/ar-connection.c conn = rb_entry(p, struct rxrpc_connection, node);
node 687 net/rxrpc/ar-connection.c rb_link_node(&conn->node, p, pp);
node 688 net/rxrpc/ar-connection.c rb_insert_color(&conn->node, &trans->server_conns);
node 757 net/rxrpc/ar-connection.c conn = rb_entry(p, struct rxrpc_connection, node);
node 859 net/rxrpc/ar-connection.c rb_erase(&conn->node,
node 862 net/rxrpc/ar-connection.c rb_erase(&conn->node,
node 222 net/rxrpc/ar-internal.h struct rb_node node; /* node in transport's lookup tree */
node 244 net/rxrpc/ar-internal.h struct rb_node node; /* node in transport's lookup tree */
node 106 net/sched/sch_htb.c struct rb_node node[TC_HTB_NUMPRIO]; /* node for self or feed tree */
node 262 net/sched/sch_htb.c c = rb_entry(parent, struct htb_class, node[prio]);
node 269 net/sched/sch_htb.c rb_link_node(&cl->node[prio], parent, p);
node 270 net/sched/sch_htb.c rb_insert_color(&cl->node[prio], root);
node 361 net/sched/sch_htb.c if (q->ptr[cl->level][prio] == cl->node + prio)
node 364 net/sched/sch_htb.c htb_safe_rb_erase(cl->node + prio, q->row[cl->level] + prio);
node 424 net/sched/sch_htb.c if (p->un.inner.ptr[prio] == cl->node + prio) {
node 432 net/sched/sch_htb.c htb_safe_rb_erase(cl->node + prio, p->un.inner.feed + prio);
node 735 net/sched/sch_htb.c rb_entry(n, struct htb_class, node[prio]);
node 791 net/sched/sch_htb.c cl = rb_entry(*sp->pptr, struct htb_class, node[prio]);
node 1383 net/sched/sch_htb.c RB_CLEAR_NODE(&cl->node[prio]);
node 331 net/sctp/endpointola.c struct hlist_node *node;
node 341 net/sctp/endpointola.c sctp_for_each_hentry(epb, node, &head->chain) {
node 693 net/sctp/input.c hlist_add_head(&epb->node, &head->chain);
node 713 net/sctp/input.c if (hlist_unhashed(&epb->node))
node 721 net/sctp/input.c __hlist_del(&epb->node);
node 739 net/sctp/input.c struct hlist_node *node;
node 745 net/sctp/input.c sctp_for_each_hentry(epb, node, &head->chain) {
node 773 net/sctp/input.c hlist_add_head(&epb->node, &head->chain);
node 802 net/sctp/input.c __hlist_del(&epb->node);
node 827 net/sctp/input.c struct hlist_node *node;
node 836 net/sctp/input.c sctp_for_each_hentry(epb, node, &head->chain) {
node 204 net/sctp/proc.c struct hlist_node *node;
node 213 net/sctp/proc.c sctp_for_each_hentry(epb, node, &head->chain) {
node 308 net/sctp/proc.c struct hlist_node *node;
node 317 net/sctp/proc.c sctp_for_each_hentry(epb, node, &head->chain) {
node 420 net/sctp/proc.c struct hlist_node *node;
node 430 net/sctp/proc.c sctp_for_each_hentry(epb, node, &head->chain) {
node 5551 net/sctp/socket.c struct hlist_node *node;
node 5576 net/sctp/socket.c sctp_for_each_hentry(pp, node, &head->chain)
node 5603 net/sctp/socket.c sctp_for_each_hentry(pp, node, &head->chain) {
node 5619 net/sctp/socket.c struct hlist_node *node;
node 5636 net/sctp/socket.c sk_for_each_bound(sk2, node, &pp->owner) {
node 5954 net/sctp/socket.c hlist_add_head(&pp->node, &head->chain);
node 5963 net/sctp/socket.c __hlist_del(&pp->node);
node 125 net/sunrpc/svc.c unsigned int node;
node 135 net/sunrpc/svc.c node = any_online_node(node_online_map);
node 136 net/sunrpc/svc.c if (nr_cpus_node(node) > 2) {
node 209 net/sunrpc/svc.c unsigned int node;
node 216 net/sunrpc/svc.c for_each_node_with_cpus(node) {
node 219 net/sunrpc/svc.c m->to_pool[node] = pidx;
node 220 net/sunrpc/svc.c m->pool_to[pidx] = node;
node 304 net/sunrpc/svc.c unsigned int node = m->pool_to[pidx];
node 315 net/sunrpc/svc.c set_cpus_allowed_ptr(task, &cpumask_of_cpu(node));
node 320 net/sunrpc/svc.c node_to_cpumask_ptr(nodecpumask, node);
node 113 net/tipc/bcast.c struct tipc_node node;
node 421 net/tipc/bcast.c struct tipc_node* node = tipc_node_find(msg_prevnode(msg));
node 428 net/tipc/bcast.c if (unlikely(!node || !tipc_node_is_up(node) || !node->bclink.supported ||
node 437 net/tipc/bcast.c tipc_node_lock(node);
node 438 net/tipc/bcast.c tipc_bclink_acknowledge(node, msg_bcast_ack(msg));
node 439 net/tipc/bcast.c tipc_node_unlock(node);
node 442 net/tipc/bcast.c bcl->owner->next = node; /* remember requestor */
node 465 net/tipc/bcast.c tipc_node_lock(node);
node 467 net/tipc/bcast.c deferred = node->bclink.deferred_head;
node 468 net/tipc/bcast.c next_in = mod(node->bclink.last_in + 1);
node 473 net/tipc/bcast.c node->bclink.last_in++;
node 474 net/tipc/bcast.c bclink_set_gap(node);
node 476 net/tipc/bcast.c bclink_send_ack(node);
node 480 net/tipc/bcast.c tipc_node_unlock(node);
node 485 net/tipc/bcast.c tipc_node_unlock(node);
node 489 net/tipc/bcast.c if (tipc_link_recv_fragment(&node->bclink.defragm,
node 492 net/tipc/bcast.c tipc_node_unlock(node);
node 495 net/tipc/bcast.c tipc_node_unlock(node);
node 499 net/tipc/bcast.c tipc_node_lock(node);
node 502 net/tipc/bcast.c node->bclink.deferred_head = deferred->next;
node 507 net/tipc/bcast.c u32 gap_after = node->bclink.gap_after;
node 508 net/tipc/bcast.c u32 gap_to = node->bclink.gap_to;
node 510 net/tipc/bcast.c if (tipc_link_defer_pkt(&node->bclink.deferred_head,
node 511 net/tipc/bcast.c &node->bclink.deferred_tail,
node 513 net/tipc/bcast.c node->bclink.nack_sync++;
node 516 net/tipc/bcast.c node->bclink.gap_after = seqno;
node 518 net/tipc/bcast.c node->bclink.gap_to = seqno;
node 520 net/tipc/bcast.c if (bclink_ack_allowed(node->bclink.nack_sync)) {
node 522 net/tipc/bcast.c bclink_send_nack(node);
node 523 net/tipc/bcast.c bclink_set_gap(node);
node 529 net/tipc/bcast.c tipc_node_unlock(node);
node 797 net/tipc/bcast.c spin_lock_init(&bclink->node.lock);
node 798 net/tipc/bcast.c bcl->owner = &bclink->node;
node 82 net/tipc/bcast.h int n = tipc_node(node);
node 98 net/tipc/bcast.h int n = tipc_node(node);
node 645 net/tipc/config.c rep_buf = tipc_cfg_do_cmd(orig->node,
node 1273 net/tipc/link.c struct tipc_node *node;
node 1287 net/tipc/link.c node = tipc_node_select(destaddr, selector);
node 1288 net/tipc/link.c if (likely(node)) {
node 1289 net/tipc/link.c tipc_node_lock(node);
node 1290 net/tipc/link.c l_ptr = node->active_links[selector];
node 1298 net/tipc/link.c tipc_node_unlock(node);
node 1323 net/tipc/link.c tipc_node_unlock(node);
node 1333 net/tipc/link.c tipc_node_unlock(node);
node 1367 net/tipc/link.c struct tipc_node *node;
node 1477 net/tipc/link.c node = tipc_node_select(destaddr, sender->publ.ref & 1);
node 1478 net/tipc/link.c if (likely(node)) {
node 1479 net/tipc/link.c tipc_node_lock(node);
node 1480 net/tipc/link.c l_ptr = node->active_links[sender->publ.ref & 1];
node 1482 net/tipc/link.c tipc_node_unlock(node);
node 1487 net/tipc/link.c tipc_node_unlock(node);
node 1525 net/tipc/link.c tipc_node_unlock(node);
node 2951 net/tipc/link.c *node = tipc_node_find(link_name_parts.addr_peer);
node 2952 net/tipc/link.c if (!*node)
node 2955 net/tipc/link.c l_ptr = (*node)->links[b_ptr->identity];
node 2968 net/tipc/link.c struct tipc_node *node;
node 2986 net/tipc/link.c l_ptr = link_find_link(args->name, &node);
node 2992 net/tipc/link.c tipc_node_lock(node);
node 3021 net/tipc/link.c tipc_node_unlock(node);
node 3046 net/tipc/link.c struct tipc_node *node;
node 3059 net/tipc/link.c l_ptr = link_find_link(link_name, &node);
node 3065 net/tipc/link.c tipc_node_lock(node);
node 3067 net/tipc/link.c tipc_node_unlock(node);
node 3094 net/tipc/link.c struct tipc_node *node;
node 3104 net/tipc/link.c l_ptr = link_find_link(name, &node);
node 3109 net/tipc/link.c tipc_node_lock(node);
node 3169 net/tipc/link.c tipc_node_unlock(node);
node 3210 net/tipc/link.c struct tipc_node * node;
node 3215 net/tipc/link.c node = tipc_node_find(a);
node 3216 net/tipc/link.c if (node) {
node 3217 net/tipc/link.c tipc_node_lock(node);
node 3218 net/tipc/link.c l_ptr = node->links[bearer_id];
node 3235 net/tipc/link.c tipc_node_unlock(node);
node 182 net/tipc/name_distr.c buf = named_prepare_buf(PUBLICATION, left, node);
node 193 net/tipc/name_distr.c msg_set_link_selector(buf_msg(buf), node);
node 195 net/tipc/name_distr.c "<%u.%u.%u>\n", tipc_zone(node),
node 196 net/tipc/name_distr.c tipc_cluster(node), tipc_node(node));
node 197 net/tipc/name_distr.c tipc_link_send(buf, node, node);
node 225 net/tipc/name_distr.c publ->node, publ->ref, publ->key);
node 231 net/tipc/name_distr.c publ->type, publ->lower, publ->node, publ->ref, publ->key);
node 313 net/tipc/name_distr.c if (publ->node == tipc_own_addr)
node 315 net/tipc/name_distr.c publ->node = tipc_own_addr;
node 140 net/tipc/name_table.c publ->node = node;
node 319 net/tipc/name_table.c type, lower, upper, node, port, sseq,
node 324 net/tipc/name_table.c publ = publ_create(type, lower, upper, scope, node, port, key);
node 328 net/tipc/name_table.c publ, node, publ->node, publ->subscr.node);
node 338 net/tipc/name_table.c if (in_own_cluster(node)) {
node 349 net/tipc/name_table.c if (node == tipc_own_addr) {
node 369 net/tipc/name_table.c publ->node,
node 409 net/tipc/name_table.c (publ->node && (publ->node != node))) {
node 431 net/tipc/name_table.c if (in_own_cluster(node)) {
node 443 net/tipc/name_table.c publ->type, publ->lower, publ->node,
node 462 net/tipc/name_table.c if (node == tipc_own_addr) {
node 474 net/tipc/name_table.c publ->type, publ->lower, publ->node,
node 507 net/tipc/name_table.c publ->node,
node 541 net/tipc/name_table.c crs->node,
node 583 net/tipc/name_table.c dbg("Publishing {%u,%u,%u} from 0x%x\n", type, lower, upper, node);
node 592 net/tipc/name_table.c scope, node, port, key);
node 604 net/tipc/name_table.c dbg("Withdrawing {%u,%u} from 0x%x\n", type, lower, node);
node 605 net/tipc/name_table.c publ = tipc_nameseq_remove_publ(seq, lower, node, ref, key);
node 649 net/tipc/name_table.c *destnode = publ->node;
node 892 net/tipc/name_table.c tipc_zone(publ->node), tipc_cluster(publ->node),
node 893 net/tipc/name_table.c tipc_node(publ->node), publ->ref);
node 896 net/tipc/name_table.c if (publ->node != tipc_own_addr)
node 76 net/tipc/name_table.h u32 node;
node 456 net/tipc/node.c ns->node = NULL;
node 51 net/tipc/node_subscr.c node_sub->node = NULL;
node 55 net/tipc/node_subscr.c node_sub->node = tipc_node_find(addr);
node 56 net/tipc/node_subscr.c if (!node_sub->node) {
node 63 net/tipc/node_subscr.c tipc_node_lock(node_sub->node);
node 64 net/tipc/node_subscr.c list_add_tail(&node_sub->nodesub_list, &node_sub->node->nsub);
node 65 net/tipc/node_subscr.c tipc_node_unlock(node_sub->node);
node 74 net/tipc/node_subscr.c if (!node_sub->node)
node 77 net/tipc/node_subscr.c tipc_node_lock(node_sub->node);
node 79 net/tipc/node_subscr.c tipc_node_unlock(node_sub->node);
node 53 net/tipc/node_subscr.h struct tipc_node *node;
node 821 net/tipc/port.c orig.node = msg_orignode(msg);
node 1077 net/tipc/port.c id->node = tipc_own_addr;
node 1202 net/tipc/port.c msg_set_destnode(msg, peer->node);
node 1208 net/tipc/port.c if (!may_route(peer->node))
node 1218 net/tipc/port.c tipc_nodesub_subscribe(&p_ptr->subscription,peer->node,
node 1224 net/tipc/port.c p_ptr->publ.max_pkt = tipc_link_get_max_pkt(peer->node, ref);
node 1321 net/tipc/port.c peer->node = port_peernode(p_ptr);
node 1465 net/tipc/port.c msg_set_orignode(msg, orig->node);
node 1508 net/tipc/port.c orig.node = tipc_own_addr;
node 1539 net/tipc/port.c msg_set_orignode(msg, orig->node);
node 1582 net/tipc/port.c orig.node = tipc_own_addr;
node 1608 net/tipc/port.c msg_set_orignode(msg, orig->node);
node 1610 net/tipc/port.c msg_set_destnode(msg, dest->node);
node 1616 net/tipc/port.c if (dest->node == tipc_own_addr)
node 1618 net/tipc/port.c res = tipc_link_send_sections_fast(p_ptr, msg_sect, num_sect, dest->node);
node 1640 net/tipc/port.c orig.node = tipc_own_addr;
node 1665 net/tipc/port.c msg_set_orignode(msg, orig->node);
node 1667 net/tipc/port.c msg_set_destnode(msg, dest->node);
node 1680 net/tipc/port.c if (dest->node == tipc_own_addr)
node 1682 net/tipc/port.c res = tipc_send_buf_fast(buf, dest->node);
node 1702 net/tipc/port.c orig.node = tipc_own_addr;
node 401 net/tipc/socket.c addr->addr.id.node = tsock->peer_name.node;
node 780 net/tipc/socket.c tsock->peer_name.node = msg_orignode(msg);
node 803 net/tipc/socket.c addr->addr.id.node = msg_orignode(msg);
node 1551 net/tipc/socket.c new_tsock->peer_name.node = msg_orignode(msg);
node 114 net/tipc/subscr.c sub->evt.port.node = htohl(node, sub->swap);
node 157 net/tipc/subscr.c sub->event_cb(sub, found_lower, found_upper, event, port_ref, node);
node 253 net/unix/af_unix.c struct hlist_node *node;
node 255 net/unix/af_unix.c sk_for_each(s, node, &unix_socket_table[hash ^ type]) {
node 288 net/unix/af_unix.c struct hlist_node *node;
node 291 net/unix/af_unix.c sk_for_each(s, node,
node 174 net/x25/af_x25.c struct hlist_node *node;
node 178 net/x25/af_x25.c sk_for_each(s, node, &x25_list)
node 246 net/x25/af_x25.c struct hlist_node *node;
node 251 net/x25/af_x25.c sk_for_each(s, node, &x25_list)
node 288 net/x25/af_x25.c struct hlist_node *node;
node 290 net/x25/af_x25.c sk_for_each(s, node, &x25_list)
node 1623 net/x25/af_x25.c struct hlist_node *node;
node 1627 net/x25/af_x25.c sk_for_each(s, node, &x25_list)
node 51 net/x25/x25_forward.c x25_frwd = list_entry(entry, struct x25_forward, node);
node 70 net/x25/x25_forward.c list_add(&new_frwd->node, &x25_forward_list);
node 104 net/x25/x25_forward.c frwd = list_entry(entry, struct x25_forward, node);
node 141 net/x25/x25_forward.c fwd = list_entry(entry, struct x25_forward, node);
node 143 net/x25/x25_forward.c list_del(&fwd->node);
node 159 net/x25/x25_forward.c fwd = list_entry(entry, struct x25_forward, node);
node 161 net/x25/x25_forward.c list_del(&fwd->node);
node 267 net/x25/x25_link.c list_add(&nb->node, &x25_neigh_list);
node 283 net/x25/x25_link.c if (nb->node.next) {
node 284 net/x25/x25_link.c list_del(&nb->node);
node 300 net/x25/x25_link.c nb = list_entry(entry, struct x25_neigh, node);
node 321 net/x25/x25_link.c nb = list_entry(entry, struct x25_neigh, node);
node 394 net/x25/x25_link.c nb = list_entry(entry, struct x25_neigh, node);
node 34 net/x25/x25_proc.c rt = list_entry(route_entry, struct x25_route, node);
node 61 net/x25/x25_proc.c struct x25_route, node);
node 65 net/x25/x25_proc.c if (rt->node.next != &x25_route_list)
node 66 net/x25/x25_proc.c rt = list_entry(rt->node.next, struct x25_route, node);
node 99 net/x25/x25_proc.c struct hlist_node *node;
node 101 net/x25/x25_proc.c sk_for_each(s, node, &x25_list)
node 179 net/x25/x25_proc.c f = list_entry(entry, struct x25_forward, node);
node 207 net/x25/x25_proc.c struct x25_forward, node);
node 211 net/x25/x25_proc.c if (f->node.next != &x25_forward_list)
node 212 net/x25/x25_proc.c f = list_entry(f->node.next, struct x25_forward, node);
node 40 net/x25/x25_route.c rt = list_entry(entry, struct x25_route, node);
node 59 net/x25/x25_route.c list_add(&rt->node, &x25_route_list);
node 75 net/x25/x25_route.c if (rt->node.next) {
node 76 net/x25/x25_route.c list_del(&rt->node);
node 91 net/x25/x25_route.c rt = list_entry(entry, struct x25_route, node);
node 116 net/x25/x25_route.c rt = list_entry(entry, struct x25_route, node);
node 159 net/x25/x25_route.c rt = list_entry(entry, struct x25_route, node);
node 219 net/x25/x25_route.c rt = list_entry(entry, struct x25_route, node);
node 203 scripts/genksyms/genksyms.c free(node->string);
node 204 scripts/genksyms/genksyms.c free(node);
node 221 scripts/genksyms/genksyms.c newnode->string = xstrdup(node->string);
node 222 scripts/genksyms/genksyms.c newnode->tag = node->tag;
node 40 scripts/genksyms/parse.y struct string_list *node = *p;
node 41 scripts/genksyms/parse.y *p = node->next;
node 42 scripts/genksyms/parse.y free_node(node);
node 1295 scripts/kconfig/gconf.c gtk_tree_store_set(tree, node,
node 1321 scripts/kconfig/gconf.c GtkTreeIter *node = parents[indent];
node 1323 scripts/kconfig/gconf.c gtk_tree_store_append(tree, node, parent);
node 1324 scripts/kconfig/gconf.c set_node(node, menu, row);
node 174 scripts/kconfig/lxdialog/dialog.h struct dialog_item node;
node 561 scripts/kconfig/lxdialog/util.c vsnprintf(item_cur->node.str, sizeof(item_cur->node.str), fmt, ap);
node 570 scripts/kconfig/lxdialog/util.c avail = sizeof(item_cur->node.str) - strlen(item_cur->node.str);
node 573 scripts/kconfig/lxdialog/util.c vsnprintf(item_cur->node.str + strlen(item_cur->node.str),
node 575 scripts/kconfig/lxdialog/util.c item_cur->node.str[sizeof(item_cur->node.str) - 1] = '\0';
node 581 scripts/kconfig/lxdialog/util.c item_cur->node.tag = tag;
node 585 scripts/kconfig/lxdialog/util.c item_cur->node.data = ptr;
node 590 scripts/kconfig/lxdialog/util.c item_cur->node.selected = val;
node 603 scripts/kconfig/lxdialog/util.c return item_cur->node.data;
node 608 scripts/kconfig/lxdialog/util.c return item_cur->node.tag;
node 644 scripts/kconfig/lxdialog/util.c return item_cur->node.str;
node 649 scripts/kconfig/lxdialog/util.c return (item_cur->node.selected != 0);
node 654 scripts/kconfig/lxdialog/util.c return (item_cur->node.tag == tag);
node 49 security/keys/internal.h struct rb_node node;
node 76 security/keys/key.c user = rb_entry(parent, struct key_user, node);
node 114 security/keys/key.c rb_link_node(&candidate->node, parent, p);
node 115 security/keys/key.c rb_insert_color(&candidate->node, &key_user_tree);
node 137 security/keys/key.c rb_erase(&user->node, &key_user_tree);
node 999 security/keys/key.c rb_link_node(&root_key_user.node,
node 1003 security/keys/key.c rb_insert_color(&root_key_user.node,
node 244 security/keys/proc.c struct key_user *user = rb_entry(_p, struct key_user, node);
node 250 security/selinux/avc.c struct avc_node *node;
node 260 security/selinux/avc.c list_for_each_entry_rcu(node, &avc_cache.slots[i], list)
node 277 security/selinux/avc.c struct avc_node *node = container_of(rhead, struct avc_node, rhead);
node 278 security/selinux/avc.c kmem_cache_free(avc_node_cachep, node);
node 284 security/selinux/avc.c list_del_rcu(&node->list);
node 285 security/selinux/avc.c call_rcu(&node->rhead, avc_node_free);
node 291 security/selinux/avc.c kmem_cache_free(avc_node_cachep, node);
node 305 security/selinux/avc.c struct avc_node *node;
node 316 security/selinux/avc.c list_for_each_entry(node, &avc_cache.slots[hvalue], list) {
node 317 security/selinux/avc.c if (atomic_dec_and_test(&node->ae.used)) {
node 319 security/selinux/avc.c avc_node_delete(node);
node 338 security/selinux/avc.c struct avc_node *node;
node 340 security/selinux/avc.c node = kmem_cache_zalloc(avc_node_cachep, GFP_ATOMIC);
node 341 security/selinux/avc.c if (!node)
node 344 security/selinux/avc.c INIT_RCU_HEAD(&node->rhead);
node 345 security/selinux/avc.c INIT_LIST_HEAD(&node->list);
node 346 security/selinux/avc.c atomic_set(&node->ae.used, 1);
node 353 security/selinux/avc.c return node;
node 358 security/selinux/avc.c node->ae.ssid = ssid;
node 359 security/selinux/avc.c node->ae.tsid = tsid;
node 360 security/selinux/avc.c node->ae.tclass = tclass;
node 361 security/selinux/avc.c memcpy(&node->ae.avd, &ae->avd, sizeof(node->ae.avd));
node 366 security/selinux/avc.c struct avc_node *node, *ret = NULL;
node 370 security/selinux/avc.c list_for_each_entry_rcu(node, &avc_cache.slots[hvalue], list) {
node 371 security/selinux/avc.c if (ssid == node->ae.ssid &&
node 372 security/selinux/avc.c tclass == node->ae.tclass &&
node 373 security/selinux/avc.c tsid == node->ae.tsid) {
node 374 security/selinux/avc.c ret = node;
node 407 security/selinux/avc.c struct avc_node *node;
node 410 security/selinux/avc.c node = avc_search_node(ssid, tsid, tclass);
node 412 security/selinux/avc.c if (node && ((node->ae.avd.decided & requested) == requested)) {
node 417 security/selinux/avc.c node = NULL;
node 420 security/selinux/avc.c return node;
node 464 security/selinux/avc.c struct avc_node *pos, *node = NULL;
node 471 security/selinux/avc.c node = avc_alloc_node();
node 472 security/selinux/avc.c if (node) {
node 474 security/selinux/avc.c avc_node_populate(node, ssid, tsid, tclass, ae);
node 481 security/selinux/avc.c avc_node_replace(node, pos);
node 485 security/selinux/avc.c list_add_rcu(&node->list, &avc_cache.slots[hvalue]);
node 490 security/selinux/avc.c return node;
node 753 security/selinux/avc.c struct avc_node *pos, *node, *orig = NULL;
node 755 security/selinux/avc.c node = avc_alloc_node();
node 756 security/selinux/avc.c if (!node) {
node 776 security/selinux/avc.c avc_node_kill(node);
node 784 security/selinux/avc.c avc_node_populate(node, ssid, tsid, tclass, &orig->ae);
node 788 security/selinux/avc.c node->ae.avd.allowed |= perms;
node 792 security/selinux/avc.c node->ae.avd.allowed &= ~perms;
node 795 security/selinux/avc.c node->ae.avd.auditallow |= perms;
node 798 security/selinux/avc.c node->ae.avd.auditallow &= ~perms;
node 801 security/selinux/avc.c node->ae.avd.auditdeny |= perms;
node 804 security/selinux/avc.c node->ae.avd.auditdeny &= ~perms;
node 807 security/selinux/avc.c avc_node_replace(node, orig);
node 823 security/selinux/avc.c struct avc_node *node;
node 832 security/selinux/avc.c list_for_each_entry(node, &avc_cache.slots[i], list)
node 833 security/selinux/avc.c avc_node_delete(node);
node 878 security/selinux/avc.c struct avc_node *node;
node 887 security/selinux/avc.c node = avc_lookup(ssid, tsid, tclass, requested);
node 888 security/selinux/avc.c if (!node) {
node 894 security/selinux/avc.c node = avc_insert(ssid, tsid, tclass, &entry);
node 897 security/selinux/avc.c p_ae = node ? &node->ae : &entry;
node 82 security/selinux/netnode.c struct sel_netnode *node = container_of(p, struct sel_netnode, rcu);
node 83 security/selinux/netnode.c kfree(node);
node 132 security/selinux/netnode.c struct sel_netnode *node;
node 145 security/selinux/netnode.c list_for_each_entry_rcu(node, &sel_netnode_hash[idx].list, list)
node 146 security/selinux/netnode.c if (node->nsec.family == family)
node 149 security/selinux/netnode.c if (node->nsec.addr.ipv4 == *(__be32 *)addr)
node 150 security/selinux/netnode.c return node;
node 153 security/selinux/netnode.c if (ipv6_addr_equal(&node->nsec.addr.ipv6,
node 155 security/selinux/netnode.c return node;
node 174 security/selinux/netnode.c switch (node->nsec.family) {
node 176 security/selinux/netnode.c idx = sel_netnode_hashfn_ipv4(node->nsec.addr.ipv4);
node 179 security/selinux/netnode.c idx = sel_netnode_hashfn_ipv6(&node->nsec.addr.ipv6);
node 185 security/selinux/netnode.c INIT_RCU_HEAD(&node->rcu);
node 189 security/selinux/netnode.c list_add_rcu(&node->list, &sel_netnode_hash[idx].list);
node 217 security/selinux/netnode.c struct sel_netnode *node;
node 221 security/selinux/netnode.c node = sel_netnode_find(addr, family);
node 222 security/selinux/netnode.c if (node != NULL) {
node 223 security/selinux/netnode.c *sid = node->nsec.sid;
node 278 security/selinux/netnode.c struct sel_netnode *node;
node 281 security/selinux/netnode.c node = sel_netnode_find(addr, family);
node 282 security/selinux/netnode.c if (node != NULL) {
node 283 security/selinux/netnode.c *sid = node->nsec.sid;
node 302 security/selinux/netnode.c struct sel_netnode *node, *node_tmp;
node 306 security/selinux/netnode.c list_for_each_entry_safe(node, node_tmp,
node 308 security/selinux/netnode.c list_del_rcu(&node->list);
node 309 security/selinux/netnode.c call_rcu(&node->rcu, sel_netnode_free);
node 1012 security/selinux/selinuxfs.c struct list_head *node;
node 1015 security/selinux/selinuxfs.c node = de->d_subdirs.next;
node 1016 security/selinux/selinuxfs.c while (node != &de->d_subdirs) {
node 1017 security/selinux/selinuxfs.c struct dentry *d = list_entry(node, struct dentry, d_u.d_child);
node 1018 security/selinux/selinuxfs.c list_del_init(node);
node 1028 security/selinux/selinuxfs.c node = de->d_subdirs.next;
node 198 security/selinux/ss/avtab.c if (!node)
node 202 security/selinux/ss/avtab.c for (cur = node->next; cur; cur = cur->next) {
node 203 security/selinux/ss/avtab.c if (node->key.source_type == cur->key.source_type &&
node 204 security/selinux/ss/avtab.c node->key.target_type == cur->key.target_type &&
node 205 security/selinux/ss/avtab.c node->key.target_class == cur->key.target_class &&
node 209 security/selinux/ss/avtab.c if (node->key.source_type < cur->key.source_type)
node 211 security/selinux/ss/avtab.c if (node->key.source_type == cur->key.source_type &&
node 212 security/selinux/ss/avtab.c node->key.target_type < cur->key.target_type)
node 214 security/selinux/ss/avtab.c if (node->key.source_type == cur->key.source_type &&
node 215 security/selinux/ss/avtab.c node->key.target_type == cur->key.target_type &&
node 216 security/selinux/ss/avtab.c node->key.target_class < cur->key.target_class)
node 94 security/selinux/ss/conditional.c new_state = cond_evaluate_expr(p, node->expr);
node 95 security/selinux/ss/conditional.c if (new_state != node->cur_state) {
node 96 security/selinux/ss/conditional.c node->cur_state = new_state;
node 100 security/selinux/ss/conditional.c for (cur = node->true_list; cur; cur = cur->next) {
node 102 security/selinux/ss/conditional.c cur->node->key.specified &= ~AVTAB_ENABLED;
node 104 security/selinux/ss/conditional.c cur->node->key.specified |= AVTAB_ENABLED;
node 107 security/selinux/ss/conditional.c for (cur = node->false_list; cur; cur = cur->next) {
node 110 security/selinux/ss/conditional.c cur->node->key.specified &= ~AVTAB_ENABLED;
node 112 security/selinux/ss/conditional.c cur->node->key.specified |= AVTAB_ENABLED;
node 142 security/selinux/ss/conditional.c for (cur_expr = node->expr; cur_expr; cur_expr = next_expr) {
node 146 security/selinux/ss/conditional.c cond_av_list_destroy(node->true_list);
node 147 security/selinux/ss/conditional.c cond_av_list_destroy(node->false_list);
node 148 security/selinux/ss/conditional.c kfree(node);
node 295 security/selinux/ss/conditional.c if (cur->node == node_ptr) {
node 323 security/selinux/ss/conditional.c list->node = node_ptr;
node 396 security/selinux/ss/conditional.c node->cur_state = le32_to_cpu(buf[0]);
node 424 security/selinux/ss/conditional.c node->expr = expr;
node 430 security/selinux/ss/conditional.c if (cond_read_av_list(p, fp, &node->true_list, NULL) != 0)
node 432 security/selinux/ss/conditional.c if (cond_read_av_list(p, fp, &node->false_list, node->true_list) != 0)
node 436 security/selinux/ss/conditional.c cond_node_destroy(node);
node 442 security/selinux/ss/conditional.c struct cond_node *node, *last = NULL;
node 458 security/selinux/ss/conditional.c node = kzalloc(sizeof(struct cond_node), GFP_KERNEL);
node 459 security/selinux/ss/conditional.c if (!node)
node 462 security/selinux/ss/conditional.c if (cond_read_node(p, node, fp) != 0)
node 466 security/selinux/ss/conditional.c p->cond_list = node;
node 468 security/selinux/ss/conditional.c last->next = node;
node 469 security/selinux/ss/conditional.c last = node;
node 483 security/selinux/ss/conditional.c struct avtab_node *node;
node 488 security/selinux/ss/conditional.c for (node = avtab_search_node(ctab, key); node;
node 489 security/selinux/ss/conditional.c node = avtab_search_node_next(node, key->specified)) {
node 491 security/selinux/ss/conditional.c (node->key.specified & (AVTAB_ALLOWED|AVTAB_ENABLED)))
node 492 security/selinux/ss/conditional.c avd->allowed |= node->datum.data;
node 494 security/selinux/ss/conditional.c (node->key.specified & (AVTAB_AUDITDENY|AVTAB_ENABLED)))
node 500 security/selinux/ss/conditional.c avd->auditdeny &= node->datum.data;
node 502 security/selinux/ss/conditional.c (node->key.specified & (AVTAB_AUDITALLOW|AVTAB_ENABLED)))
node 503 security/selinux/ss/conditional.c avd->auditallow |= node->datum.data;
node 43 security/selinux/ss/conditional.h struct avtab_node *node;
node 32 security/selinux/ss/ebitmap.c n1 = e1->node;
node 33 security/selinux/ss/ebitmap.c n2 = e2->node;
node 52 security/selinux/ss/ebitmap.c n = src->node;
node 66 security/selinux/ss/ebitmap.c dst->node = new;
node 89 security/selinux/ss/ebitmap.c struct ebitmap_node *e_iter = ebmap->node;
node 186 security/selinux/ss/ebitmap.c ebmap->node = e_iter;
node 222 security/selinux/ss/ebitmap.c n1 = e1->node;
node 223 security/selinux/ss/ebitmap.c n2 = e2->node;
node 251 security/selinux/ss/ebitmap.c n = e->node;
node 266 security/selinux/ss/ebitmap.c n = e->node;
node 295 security/selinux/ss/ebitmap.c e->node = n->next;
node 322 security/selinux/ss/ebitmap.c new->next = e->node;
node 323 security/selinux/ss/ebitmap.c e->node = new;
node 336 security/selinux/ss/ebitmap.c n = e->node;
node 344 security/selinux/ss/ebitmap.c e->node = NULL;
node 378 security/selinux/ss/ebitmap.c e->node = NULL;
node 417 security/selinux/ss/ebitmap.c e->node = tmp;
node 34 security/selinux/ss/ebitmap.h struct ebitmap_node *node; /* first node in the bitmap */
node 39 security/selinux/ss/ebitmap.h #define ebitmap_startbit(e) ((e)->node ? (e)->node->startbit : 0)
node 46 security/selinux/ss/ebitmap.h for (*n = e->node; *n; *n = (*n)->next) {
node 78 security/selinux/ss/ebitmap.h (((bit) - (node)->startbit) / EBITMAP_UNIT_SIZE)
node 80 security/selinux/ss/ebitmap.h (((bit) - (node)->startbit) % EBITMAP_UNIT_SIZE)
node 40 security/selinux/ss/mls.c struct ebitmap_node *node;
node 54 security/selinux/ss/mls.c ebitmap_for_each_positive_bit(e, node, i) {
node 94 security/selinux/ss/mls.c struct ebitmap_node *node;
node 113 security/selinux/ss/mls.c ebitmap_for_each_positive_bit(e, node, i) {
node 163 security/selinux/ss/mls.c struct ebitmap_node *node;
node 173 security/selinux/ss/mls.c ebitmap_for_each_positive_bit(&l->cat, node, i) {
node 477 security/selinux/ss/mls.c struct ebitmap_node *node;
node 492 security/selinux/ss/mls.c ebitmap_for_each_positive_bit(&c->range.level[l].cat, node, i) {
node 1522 security/selinux/ss/policydb.c struct ebitmap_node *node;
node 1533 security/selinux/ss/policydb.c ebitmap_for_each_positive_bit(&user->roles, node, bit) {
node 1559 security/selinux/ss/policydb.c struct ebitmap_node *node;
node 1570 security/selinux/ss/policydb.c ebitmap_for_each_positive_bit(&role->types, node, bit) {
node 1929 security/selinux/ss/policydb.c c->u.node.addr = nodebuf[0]; /* network order */
node 1930 security/selinux/ss/policydb.c c->u.node.mask = nodebuf[1]; /* network order */
node 146 security/selinux/ss/policydb.h } node; /* node information */
node 388 security/selinux/ss/services.c struct avtab_node *node;
node 457 security/selinux/ss/services.c for (node = avtab_search_node(&policydb.te_avtab, &avkey);
node 458 security/selinux/ss/services.c node;
node 459 security/selinux/ss/services.c node = avtab_search_node_next(node, avkey.specified)) {
node 460 security/selinux/ss/services.c if (node->key.specified == AVTAB_ALLOWED)
node 461 security/selinux/ss/services.c avd->allowed |= node->datum.data;
node 462 security/selinux/ss/services.c else if (node->key.specified == AVTAB_AUDITALLOW)
node 463 security/selinux/ss/services.c avd->auditallow |= node->datum.data;
node 464 security/selinux/ss/services.c else if (node->key.specified == AVTAB_AUDITDENY)
node 465 security/selinux/ss/services.c avd->auditdeny &= node->datum.data;
node 1139 security/selinux/ss/services.c struct avtab_node *node;
node 1209 security/selinux/ss/services.c node = avtab_search_node(&policydb.te_cond_avtab, &avkey);
node 1210 security/selinux/ss/services.c for (; node; node = avtab_search_node_next(node, specified)) {
node 1211 security/selinux/ss/services.c if (node->key.specified & AVTAB_ENABLED) {
node 1212 security/selinux/ss/services.c avdatum = &node->datum;
node 1872 security/selinux/ss/services.c if (c->u.node.addr == (addr & c->u.node.mask))
node 71 sound/aoa/aoa-gpio.h struct device_node *node;
node 36 sound/aoa/aoa.h struct device_node *node;
node 1036 sound/aoa/codecs/snd-aoa-codec-onyx.c onyx->codec.node = of_node_get(node);
node 1091 sound/aoa/codecs/snd-aoa-codec-onyx.c of_node_put(onyx->codec.node);
node 915 sound/aoa/codecs/snd-aoa-codec-tas.c tas->codec.node = of_node_get(node);
node 922 sound/aoa/codecs/snd-aoa-codec-tas.c addr, node->full_name);
node 982 sound/aoa/codecs/snd-aoa-codec-tas.c of_node_put(tas->codec.node);
node 20 sound/aoa/core/snd-aoa-gpio-pmf.c rc = pmf_call_function(rt->node, #name "-mute", &args); \
node 43 sound/aoa/core/snd-aoa-gpio-pmf.c rc = pmf_call_function(rt->node, "hw-reset", &args);
node 188 sound/aoa/core/snd-aoa-gpio-pmf.c err = pmf_register_irq_client(rt->node,
node 229 sound/aoa/core/snd-aoa-gpio-pmf.c err = pmf_call_function(rt->node, name, &args);
node 724 sound/aoa/fabrics/snd-aoa-fabric-layout.c if (codec->node && (strcmp(codec->node->name, "codec") == 0)) {
node 733 sound/aoa/fabrics/snd-aoa-fabric-layout.c if (*ref != codec->node->linux_phandle) {
node 951 sound/aoa/fabrics/snd-aoa-fabric-layout.c while ((sound = of_get_next_child(sdev->ofdev.node, sound))) {
node 977 sound/aoa/fabrics/snd-aoa-fabric-layout.c ldev->gpio.node = sound->parent;
node 77 sound/aoa/soundbus/core.c retval = add_uevent_var(env, "OF_NAME=%s", of->node->name);
node 81 sound/aoa/soundbus/core.c retval = add_uevent_var(env, "OF_TYPE=%s", of->node->type);
node 89 sound/aoa/soundbus/core.c compat = of_get_property(of->node, "compatible", &cplen);
node 172 sound/aoa/soundbus/core.c !dev->ofdev.node ||
node 44 sound/aoa/soundbus/i2sbus/i2sbus-control.c np = i2sdev->sound.ofdev.node;
node 211 sound/aoa/soundbus/i2sbus/i2sbus-core.c dev->sound.ofdev.node = np;
node 336 sound/aoa/soundbus/i2sbus/i2sbus-core.c while ((np = of_get_next_child(dev->ofdev.node, np))) {
node 12 sound/aoa/soundbus/sysfs.c return sprintf (buf, format_string, mdev->ofdev.node->field); \
node 28 sound/aoa/soundbus/sysfs.c of->node->name, of->node->type);
node 50 sound/pci/hda/hda_generic.c struct hda_gnode *node; /* Node for PCM volume */
node 84 sound/pci/hda/hda_generic.c #define defcfg_type(node) (((node)->def_cfg & AC_DEFCFG_DEVICE) >> \
node 86 sound/pci/hda/hda_generic.c #define defcfg_location(node) (((node)->def_cfg & AC_DEFCFG_LOCATION) >> \
node 88 sound/pci/hda/hda_generic.c #define defcfg_port_conn(node) (((node)->def_cfg & AC_DEFCFG_PORT_CONN) >> \
node 97 sound/pci/hda/hda_generic.c struct hda_gnode *node, *n;
node 102 sound/pci/hda/hda_generic.c list_for_each_entry_safe(node, n, &spec->nid_list, list) {
node 103 sound/pci/hda/hda_generic.c if (node->conn_list != node->slist)
node 104 sound/pci/hda/hda_generic.c kfree(node->conn_list);
node 105 sound/pci/hda/hda_generic.c kfree(node);
node 116 sound/pci/hda/hda_generic.c struct hda_gnode *node;
node 120 sound/pci/hda/hda_generic.c node = kzalloc(sizeof(*node), GFP_KERNEL);
node 121 sound/pci/hda/hda_generic.c if (node == NULL)
node 123 sound/pci/hda/hda_generic.c node->nid = nid;
node 127 sound/pci/hda/hda_generic.c kfree(node);
node 130 sound/pci/hda/hda_generic.c if (nconns <= ARRAY_SIZE(node->slist))
node 131 sound/pci/hda/hda_generic.c node->conn_list = node->slist;
node 133 sound/pci/hda/hda_generic.c node->conn_list = kmalloc(sizeof(hda_nid_t) * nconns,
node 135 sound/pci/hda/hda_generic.c if (! node->conn_list) {
node 137 sound/pci/hda/hda_generic.c kfree(node);
node 141 sound/pci/hda/hda_generic.c memcpy(node->conn_list, conn_list, nconns * sizeof(hda_nid_t));
node 142 sound/pci/hda/hda_generic.c node->nconns = nconns;
node 143 sound/pci/hda/hda_generic.c node->wid_caps = get_wcaps(codec, nid);
node 144 sound/pci/hda/hda_generic.c node->type = (node->wid_caps & AC_WCAP_TYPE) >> AC_WCAP_TYPE_SHIFT;
node 146 sound/pci/hda/hda_generic.c if (node->type == AC_WID_PIN) {
node 147 sound/pci/hda/hda_generic.c node->pin_caps = snd_hda_param_read(codec, node->nid, AC_PAR_PIN_CAP);
node 148 sound/pci/hda/hda_generic.c node->pin_ctl = snd_hda_codec_read(codec, node->nid, 0, AC_VERB_GET_PIN_WIDGET_CONTROL, 0);
node 149 sound/pci/hda/hda_generic.c node->def_cfg = snd_hda_codec_read(codec, node->nid, 0, AC_VERB_GET_CONFIG_DEFAULT, 0);
node 152 sound/pci/hda/hda_generic.c if (node->wid_caps & AC_WCAP_OUT_AMP) {
node 153 sound/pci/hda/hda_generic.c if (node->wid_caps & AC_WCAP_AMP_OVRD)
node 154 sound/pci/hda/hda_generic.c node->amp_out_caps = snd_hda_param_read(codec, node->nid, AC_PAR_AMP_OUT_CAP);
node 155 sound/pci/hda/hda_generic.c if (! node->amp_out_caps)
node 156 sound/pci/hda/hda_generic.c node->amp_out_caps = spec->def_amp_out_caps;
node 158 sound/pci/hda/hda_generic.c if (node->wid_caps & AC_WCAP_IN_AMP) {
node 159 sound/pci/hda/hda_generic.c if (node->wid_caps & AC_WCAP_AMP_OVRD)
node 160 sound/pci/hda/hda_generic.c node->amp_in_caps = snd_hda_param_read(codec, node->nid, AC_PAR_AMP_IN_CAP);
node 161 sound/pci/hda/hda_generic.c if (! node->amp_in_caps)
node 162 sound/pci/hda/hda_generic.c node->amp_in_caps = spec->def_amp_in_caps;
node 164 sound/pci/hda/hda_generic.c list_add_tail(&node->list, &spec->nid_list);
node 205 sound/pci/hda/hda_generic.c struct hda_gnode *node;
node 207 sound/pci/hda/hda_generic.c list_for_each_entry(node, &spec->nid_list, list) {
node 208 sound/pci/hda/hda_generic.c if (node->nid == nid)
node 209 sound/pci/hda/hda_generic.c return node;
node 220 sound/pci/hda/hda_generic.c snd_printdd("UNMUTE OUT: NID=0x%x\n", node->nid);
node 221 sound/pci/hda/hda_generic.c val = (node->amp_out_caps & AC_AMPCAP_NUM_STEPS) >> AC_AMPCAP_NUM_STEPS_SHIFT;
node 222 sound/pci/hda/hda_generic.c ofs = (node->amp_out_caps & AC_AMPCAP_OFFSET) >> AC_AMPCAP_OFFSET_SHIFT;
node 225 sound/pci/hda/hda_generic.c snd_hda_codec_amp_stereo(codec, node->nid, HDA_OUTPUT, 0, 0xff, val);
node 235 sound/pci/hda/hda_generic.c snd_printdd("UNMUTE IN: NID=0x%x IDX=0x%x\n", node->nid, index);
node 236 sound/pci/hda/hda_generic.c val = (node->amp_in_caps & AC_AMPCAP_NUM_STEPS) >> AC_AMPCAP_NUM_STEPS_SHIFT;
node 237 sound/pci/hda/hda_generic.c ofs = (node->amp_in_caps & AC_AMPCAP_OFFSET) >> AC_AMPCAP_OFFSET_SHIFT;
node 240 sound/pci/hda/hda_generic.c snd_hda_codec_amp_stereo(codec, node->nid, HDA_INPUT, index, 0xff, val);
node 250 sound/pci/hda/hda_generic.c snd_printdd("CONNECT: NID=0x%x IDX=0x%x\n", node->nid, index);
node 251 sound/pci/hda/hda_generic.c return snd_hda_codec_write_cache(codec, node->nid, 0,
node 260 sound/pci/hda/hda_generic.c struct hda_gnode *node;
node 262 sound/pci/hda/hda_generic.c list_for_each_entry(node, &spec->nid_list, list) {
node 263 sound/pci/hda/hda_generic.c node->checked = 0;
node 278 sound/pci/hda/hda_generic.c if (node->checked)
node 281 sound/pci/hda/hda_generic.c node->checked = 1;
node 282 sound/pci/hda/hda_generic.c if (node->type == AC_WID_AUD_OUT) {
node 283 sound/pci/hda/hda_generic.c if (node->wid_caps & AC_WCAP_DIGITAL) {
node 284 sound/pci/hda/hda_generic.c snd_printdd("Skip Digital OUT node %x\n", node->nid);
node 287 sound/pci/hda/hda_generic.c snd_printdd("AUD_OUT found %x\n", node->nid);
node 290 sound/pci/hda/hda_generic.c return node == spec->dac_node[dac_idx];
node 292 sound/pci/hda/hda_generic.c spec->dac_node[dac_idx] = node;
node 293 sound/pci/hda/hda_generic.c if ((node->wid_caps & AC_WCAP_OUT_AMP) &&
node 295 sound/pci/hda/hda_generic.c spec->pcm_vol[spec->pcm_vol_nodes].node = node;
node 302 sound/pci/hda/hda_generic.c for (i = 0; i < node->nconns; i++) {
node 303 sound/pci/hda/hda_generic.c child = hda_get_node(spec, node->conn_list[i]);
node 313 sound/pci/hda/hda_generic.c if (node->nconns > 1)
node 314 sound/pci/hda/hda_generic.c select_input_connection(codec, node, i);
node 315 sound/pci/hda/hda_generic.c unmute_input(codec, node, i);
node 316 sound/pci/hda/hda_generic.c unmute_output(codec, node);
node 321 sound/pci/hda/hda_generic.c if ((node->wid_caps & AC_WCAP_IN_AMP) ||
node 322 sound/pci/hda/hda_generic.c (node->wid_caps & AC_WCAP_OUT_AMP)) {
node 324 sound/pci/hda/hda_generic.c spec->pcm_vol[n].node = node;
node 345 sound/pci/hda/hda_generic.c struct hda_gnode *node;
node 348 sound/pci/hda/hda_generic.c list_for_each_entry(node, &spec->nid_list, list) {
node 349 sound/pci/hda/hda_generic.c if (node->type != AC_WID_PIN)
node 352 sound/pci/hda/hda_generic.c if (! (node->pin_caps & AC_PINCAP_OUT))
node 354 sound/pci/hda/hda_generic.c if (defcfg_port_conn(node) == AC_JACK_PORT_NONE)
node 357 sound/pci/hda/hda_generic.c if (jack_type != defcfg_type(node))
node 359 sound/pci/hda/hda_generic.c if (node->wid_caps & AC_WCAP_DIGITAL)
node 363 sound/pci/hda/hda_generic.c if (! (node->pin_ctl & AC_PINCTL_OUT_EN))
node 367 sound/pci/hda/hda_generic.c err = parse_output_path(codec, spec, node, 0);
node 371 sound/pci/hda/hda_generic.c err = parse_output_path(codec, spec, node, 1);
node 377 sound/pci/hda/hda_generic.c unmute_output(codec, node);
node 379 sound/pci/hda/hda_generic.c snd_hda_codec_write_cache(codec, node->nid, 0,
node 382 sound/pci/hda/hda_generic.c ((node->pin_caps & AC_PINCAP_HP_DRV) ?
node 384 sound/pci/hda/hda_generic.c return node;
node 397 sound/pci/hda/hda_generic.c struct hda_gnode *node;
node 403 sound/pci/hda/hda_generic.c node = parse_output_jack(codec, spec, AC_JACK_LINE_OUT);
node 404 sound/pci/hda/hda_generic.c if (node) /* found, remember the PIN node */
node 405 sound/pci/hda/hda_generic.c spec->out_pin_node[0] = node;
node 408 sound/pci/hda/hda_generic.c node = parse_output_jack(codec, spec, AC_JACK_SPEAKER);
node 409 sound/pci/hda/hda_generic.c if (node)
node 410 sound/pci/hda/hda_generic.c spec->out_pin_node[0] = node;
node 413 sound/pci/hda/hda_generic.c node = parse_output_jack(codec, spec, AC_JACK_HP_OUT);
node 414 sound/pci/hda/hda_generic.c if (node) {
node 416 sound/pci/hda/hda_generic.c spec->out_pin_node[0] = node;
node 418 sound/pci/hda/hda_generic.c spec->out_pin_node[1] = node;
node 467 sound/pci/hda/hda_generic.c unsigned int location = defcfg_location(node);
node 468 sound/pci/hda/hda_generic.c switch (defcfg_type(node)) {
node 485 sound/pci/hda/hda_generic.c (node->pin_caps &
node 512 sound/pci/hda/hda_generic.c if (node->checked)
node 515 sound/pci/hda/hda_generic.c node->checked = 1;
node 516 sound/pci/hda/hda_generic.c if (node->type != AC_WID_PIN) {
node 517 sound/pci/hda/hda_generic.c for (i = 0; i < node->nconns; i++) {
node 519 sound/pci/hda/hda_generic.c child = hda_get_node(spec, node->conn_list[i]);
node 529 sound/pci/hda/hda_generic.c if (node->nconns > 1)
node 530 sound/pci/hda/hda_generic.c select_input_connection(codec, node, i);
node 531 sound/pci/hda/hda_generic.c unmute_input(codec, node, i);
node 532 sound/pci/hda/hda_generic.c unmute_output(codec, node);
node 540 sound/pci/hda/hda_generic.c if (! (node->pin_caps & AC_PINCAP_IN))
node 543 sound/pci/hda/hda_generic.c if (defcfg_port_conn(node) == AC_JACK_PORT_NONE)
node 546 sound/pci/hda/hda_generic.c if (node->wid_caps & AC_WCAP_DIGITAL)
node 556 sound/pci/hda/hda_generic.c type = get_input_type(node, &pinctl);
node 559 sound/pci/hda/hda_generic.c if (! (node->pin_ctl & AC_PINCTL_IN_EN))
node 568 sound/pci/hda/hda_generic.c unmute_input(codec, node, 0); /* index = 0? */
node 570 sound/pci/hda/hda_generic.c snd_hda_codec_write_cache(codec, node->nid, 0,
node 605 sound/pci/hda/hda_generic.c struct hda_gnode *node;
node 621 sound/pci/hda/hda_generic.c node = hda_get_node(spec, adc_node->conn_list[i]);
node 622 sound/pci/hda/hda_generic.c if (node && node->type == AC_WID_PIN) {
node 623 sound/pci/hda/hda_generic.c err = parse_adc_sub_nodes(codec, spec, node);
node 632 sound/pci/hda/hda_generic.c node = hda_get_node(spec, adc_node->conn_list[i]);
node 633 sound/pci/hda/hda_generic.c if (node && node->type != AC_WID_PIN) {
node 634 sound/pci/hda/hda_generic.c err = parse_adc_sub_nodes(codec, spec, node);
node 660 sound/pci/hda/hda_generic.c struct hda_gnode *node;
node 668 sound/pci/hda/hda_generic.c list_for_each_entry(node, &spec->nid_list, list) {
node 669 sound/pci/hda/hda_generic.c if (node->wid_caps & AC_WCAP_DIGITAL)
node 671 sound/pci/hda/hda_generic.c if (node->type == AC_WID_AUD_IN) {
node 672 sound/pci/hda/hda_generic.c err = parse_input_path(codec, node);
node 720 sound/pci/hda/hda_generic.c if ((node->wid_caps & AC_WCAP_IN_AMP) &&
node 721 sound/pci/hda/hda_generic.c (node->amp_in_caps & AC_AMPCAP_MUTE)) {
node 722 sound/pci/hda/hda_generic.c knew = (struct snd_kcontrol_new)HDA_CODEC_MUTE(name, node->nid, index, HDA_INPUT);
node 724 sound/pci/hda/hda_generic.c add_input_loopback(codec, node->nid, HDA_INPUT, index);
node 725 sound/pci/hda/hda_generic.c snd_printdd("[%s] NID=0x%x, DIR=IN, IDX=0x%x\n", name, node->nid, index);
node 729 sound/pci/hda/hda_generic.c } else if ((node->wid_caps & AC_WCAP_OUT_AMP) &&
node 730 sound/pci/hda/hda_generic.c (node->amp_out_caps & AC_AMPCAP_MUTE)) {
node 731 sound/pci/hda/hda_generic.c knew = (struct snd_kcontrol_new)HDA_CODEC_MUTE(name, node->nid, 0, HDA_OUTPUT);
node 733 sound/pci/hda/hda_generic.c add_input_loopback(codec, node->nid, HDA_OUTPUT, 0);
node 734 sound/pci/hda/hda_generic.c snd_printdd("[%s] NID=0x%x, DIR=OUT\n", name, node->nid);
node 744 sound/pci/hda/hda_generic.c if ((node->wid_caps & AC_WCAP_IN_AMP) &&
node 745 sound/pci/hda/hda_generic.c (node->amp_in_caps & AC_AMPCAP_NUM_STEPS)) {
node 746 sound/pci/hda/hda_generic.c knew = (struct snd_kcontrol_new)HDA_CODEC_VOLUME(name, node->nid, index, HDA_INPUT);
node 747 sound/pci/hda/hda_generic.c snd_printdd("[%s] NID=0x%x, DIR=IN, IDX=0x%x\n", name, node->nid, index);
node 751 sound/pci/hda/hda_generic.c } else if ((node->wid_caps & AC_WCAP_OUT_AMP) &&
node 752 sound/pci/hda/hda_generic.c (node->amp_out_caps & AC_AMPCAP_NUM_STEPS)) {
node 753 sound/pci/hda/hda_generic.c knew = (struct snd_kcontrol_new)HDA_CODEC_VOLUME(name, node->nid, 0, HDA_OUTPUT);
node 754 sound/pci/hda/hda_generic.c snd_printdd("[%s] NID=0x%x, DIR=OUT\n", name, node->nid);
node 790 sound/pci/hda/hda_generic.c err = create_mixer(codec, spec->pcm_vol[i].node,
node 807 sound/pci/hda/hda_generic.c return create_mixer(codec, spec->pcm_vol[0].node,
node 893 sound/pci/hda/hda_generic.c if (node->checked)
node 896 sound/pci/hda/hda_generic.c node->checked = 1;
node 897 sound/pci/hda/hda_generic.c if (node == dest_node) {
node 902 sound/pci/hda/hda_generic.c for (i = 0; i < node->nconns; i++) {
node 903 sound/pci/hda/hda_generic.c struct hda_gnode *child = hda_get_node(spec, node->conn_list[i]);
node 911 sound/pci/hda/hda_generic.c err = create_mixer(codec, node, i, type,
node 921 sound/pci/hda/hda_generic.c if (node->nconns > 1)
node 922 sound/pci/hda/hda_generic.c select_input_connection(codec, node, i);
node 923 sound/pci/hda/hda_generic.c unmute_input(codec, node, i);
node 924 sound/pci/hda/hda_generic.c unmute_output(codec, node);
node 937 sound/pci/hda/hda_generic.c struct hda_gnode *node;
node 944 sound/pci/hda/hda_generic.c list_for_each_entry(node, &spec->nid_list, list) {
node 945 sound/pci/hda/hda_generic.c if (node->type != AC_WID_PIN)
node 948 sound/pci/hda/hda_generic.c if (! (node->pin_caps & AC_PINCAP_IN))
node 950 sound/pci/hda/hda_generic.c type = get_input_type(node, NULL);
node 957 sound/pci/hda/hda_generic.c node, type);
node 235 sound/pcmcia/pdaudiocf/pdaudiocf.c link->dev_node = &pdacf->node;
node 120 sound/pcmcia/pdaudiocf/pdaudiocf.h dev_node_t node;
node 244 sound/pcmcia/vx/vxpocket.c link->dev_node = &vxp->node;
node 46 sound/pcmcia/vx/vxpocket.h dev_node_t node;
node 832 sound/ppc/pmac.c ppc_md.feature_call(PMAC_FTR_SOUND_CHIP_ENABLE, chip->node, 0, enable);
node 848 sound/ppc/pmac.c if (chip->node)
node 879 sound/ppc/pmac.c if (chip->node) {
node 891 sound/ppc/pmac.c of_node_put(chip->node);
node 916 sound/ppc/pmac.c for (mio = chip->node->parent; mio; mio = mio->parent) {
node 968 sound/ppc/pmac.c chip->node = of_find_node_by_name(NULL, "awacs");
node 969 sound/ppc/pmac.c sound = of_node_get(chip->node);
node 975 sound/ppc/pmac.c if (!chip->node)
node 976 sound/ppc/pmac.c chip->node = of_find_node_by_name(NULL, "davbus");
node 981 sound/ppc/pmac.c if (! chip->node) {
node 982 sound/ppc/pmac.c chip->node = of_find_node_by_name(NULL, "i2s-a");
node 983 sound/ppc/pmac.c if (chip->node && chip->node->parent &&
node 984 sound/ppc/pmac.c chip->node->parent->parent) {
node 985 sound/ppc/pmac.c if (of_device_is_compatible(chip->node->parent->parent,
node 990 sound/ppc/pmac.c if (! chip->node)
node 995 sound/ppc/pmac.c while (sound && sound->parent != chip->node)
node 999 sound/ppc/pmac.c of_node_put(chip->node);
node 1000 sound/ppc/pmac.c chip->node = NULL;
node 1014 sound/ppc/pmac.c of_node_put(chip->node);
node 1015 sound/ppc/pmac.c chip->node = NULL;
node 1060 sound/ppc/pmac.c macio = macio_find(chip->node, macio_unknown);
node 1212 sound/ppc/pmac.c np = chip->node;
node 1322 sound/ppc/pmac.c for (mio = chip->node->parent; mio; mio = mio->parent) {
node 95 sound/ppc/pmac.h struct device_node *node;
node 1096 sound/ppc/tumbler.c struct device_node *node;
node 1102 sound/ppc/tumbler.c node = find_compatible_audio_device(device);
node 1104 sound/ppc/tumbler.c node = find_audio_device(device);
node 1105 sound/ppc/tumbler.c if (! node) {
node 1111 sound/ppc/tumbler.c base = of_get_property(node, "AAPL,address", NULL);
node 1113 sound/ppc/tumbler.c base = of_get_property(node, "reg", NULL);
node 1117 sound/ppc/tumbler.c of_node_put(node);
node 1128 sound/ppc/tumbler.c base = of_get_property(node, "audio-gpio-active-state", NULL);
node 1143 sound/ppc/tumbler.c prop = of_get_property(node, platform, NULL);
node 1159 sound/ppc/tumbler.c ret = irq_of_parse_and_map(node, 0);
node 1160 sound/ppc/tumbler.c of_node_put(node);
node 1360 sound/ppc/tumbler.c for (np = chip->node->child; np; np = np->sibling) {
node 724 sound/soc/fsl/mpc5200_psc_i2s.c prop = of_get_property(op->node, "cell-index", &size);
node 730 sound/soc/fsl/mpc5200_psc_i2s.c irq = irq_of_parse_and_map(op->node, 0);
node 731 sound/soc/fsl/mpc5200_psc_i2s.c if (of_address_to_resource(op->node, 0, &res)) {
node 787 sound/soc/fsl/mpc5200_psc_i2s.c if (of_get_property(op->node, "fsl,cellslave", NULL))
node 795 sound/soc/fsl/mpc5200_psc_i2s.c if (!of_get_property(op->node, "codec-handle", NULL))
node 828 sound/soc/fsl/mpc5200_psc_i2s.c of_snd_soc_register_platform(&psc_i2s_pcm_soc_platform, op->node,
node 215 sound/soc/fsl/mpc8610_hpcd.c struct device_node *np = ofdev->node;
node 108 sound/soc/fsl/soc-of-simple.c pr_info("registering ASoC codec driver: %s\n", node->full_name);
node 111 sound/soc/fsl/soc-of-simple.c of_soc = of_snd_soc_get_device(node);
node 120 sound/soc/fsl/soc-of-simple.c of_soc->dai_link.name = (char *)node->name;
node 121 sound/soc/fsl/soc-of-simple.c of_soc->dai_link.stream_name = (char *)node->name;
node 142 sound/soc/fsl/soc-of-simple.c pr_info("registering ASoC platform driver: %s\n", node->full_name);
node 144 sound/soc/fsl/soc-of-simple.c handle = of_get_property(node, "codec-handle", &len);
node 159 sound/soc/fsl/soc-of-simple.c of_soc->platform_node = node;
node 2078 sound/sparc/cs4231.c if (!strcmp(op->node->parent->name, "ebus"))
node 2082 sound/sparc/cs4231.c if (!strcmp(op->node->parent->name, "sbus") ||
node 2083 sound/sparc/cs4231.c !strcmp(op->node->parent->name, "sbi"))
node 2650 sound/sparc/dbri.c dbri->irq, op->node->name[9], dbri->mm.version);