node 369 arch/x86/kernel/aperture_64.c int i, node; node 381 arch/x86/kernel/aperture_64.c node = 0; node 403 arch/x86/kernel/aperture_64.c node, aper_base, aper_size >> 20); node 404 arch/x86/kernel/aperture_64.c node++; node 186 arch/x86/kernel/cpu/amd.c int i, node; node 189 arch/x86/kernel/cpu/amd.c node = apicid_to_node[i]; node 190 arch/x86/kernel/cpu/amd.c if (node != NUMA_NO_NODE && node_online(node)) node 191 arch/x86/kernel/cpu/amd.c return node; node 194 arch/x86/kernel/cpu/amd.c node = apicid_to_node[i]; node 195 arch/x86/kernel/cpu/amd.c if (node != NUMA_NO_NODE && node_online(node)) node 196 arch/x86/kernel/cpu/amd.c return node; node 224 arch/x86/kernel/cpu/amd.c int node; node 227 arch/x86/kernel/cpu/amd.c node = c->phys_proc_id; node 229 arch/x86/kernel/cpu/amd.c node = apicid_to_node[apicid]; node 230 arch/x86/kernel/cpu/amd.c if (!node_online(node)) { node 245 arch/x86/kernel/cpu/amd.c node = apicid_to_node[ht_nodeid]; node 247 arch/x86/kernel/cpu/amd.c if (!node_online(node)) node 248 arch/x86/kernel/cpu/amd.c node = nearby_node(apicid); node 250 arch/x86/kernel/cpu/amd.c numa_set_node(cpu, node); node 252 arch/x86/kernel/cpu/amd.c printk(KERN_INFO "CPU %d/%x -> Node %d\n", cpu, apicid, node); node 166 arch/x86/kernel/cpu/intel.c unsigned node; node 172 arch/x86/kernel/cpu/intel.c node = apicid_to_node[apicid]; node 173 arch/x86/kernel/cpu/intel.c if (node == NUMA_NO_NODE || !node_online(node)) node 174 arch/x86/kernel/cpu/intel.c node = first_node(node_online_map); node 175 arch/x86/kernel/cpu/intel.c numa_set_node(cpu, node); node 177 arch/x86/kernel/cpu/intel.c printk(KERN_INFO "CPU %d/%x -> Node %d\n", cpu, apicid, node); node 673 arch/x86/kernel/cpu/intel_cacheinfo.c for (i = 0; i <= node; i++) { node 693 arch/x86/kernel/cpu/intel_cacheinfo.c int node = cpu_to_node(first_cpu(this_leaf->shared_cpu_map)); node 701 arch/x86/kernel/cpu/intel_cacheinfo.c dev = get_k8_northbridge(node); node 727 arch/x86/kernel/cpu/intel_cacheinfo.c int node = cpu_to_node(first_cpu(this_leaf->shared_cpu_map)); node 744 arch/x86/kernel/cpu/intel_cacheinfo.c dev = get_k8_northbridge(node); node 33 arch/x86/kernel/kdebugfs.c struct setup_data_node *node = file->private_data; node 42 arch/x86/kernel/kdebugfs.c if (pos >= node->len) node 45 arch/x86/kernel/kdebugfs.c if (count > node->len - pos) node 46 arch/x86/kernel/kdebugfs.c count = node->len - pos; node 47 arch/x86/kernel/kdebugfs.c pa = node->paddr + sizeof(struct setup_data) + pos; node 95 arch/x86/kernel/kdebugfs.c type = debugfs_create_x32("type", S_IRUGO, d, &node->type); node 100 arch/x86/kernel/kdebugfs.c data = debugfs_create_file("data", S_IRUGO, d, node, &fops_setup_data); node 117 arch/x86/kernel/kdebugfs.c struct setup_data_node *node; node 133 arch/x86/kernel/kdebugfs.c node = kmalloc(sizeof(*node), GFP_KERNEL); node 134 arch/x86/kernel/kdebugfs.c if (!node) { node 142 arch/x86/kernel/kdebugfs.c kfree(node); node 150 arch/x86/kernel/kdebugfs.c node->paddr = pa_data; node 151 arch/x86/kernel/kdebugfs.c node->type = data->type; node 152 arch/x86/kernel/kdebugfs.c node->len = data->len; node 153 arch/x86/kernel/kdebugfs.c error = create_setup_data_node(d, no, node); node 679 arch/x86/kernel/kprobes.c struct hlist_node *node, *tmp; node 707 arch/x86/kernel/kprobes.c hlist_for_each_entry_safe(ri, node, tmp, head, hlist) { node 735 arch/x86/kernel/kprobes.c hlist_for_each_entry_safe(ri, node, tmp, &empty_rp, hlist) { node 48 arch/x86/kernel/numaq_32.c int node; node 54 arch/x86/kernel/numaq_32.c for_each_node(node) { node 55 arch/x86/kernel/numaq_32.c if (scd->quads_present31_0 & (1 << node)) { node 56 arch/x86/kernel/numaq_32.c node_set_online(node); node 57 arch/x86/kernel/numaq_32.c eq = &scd->eq[node]; node 59 arch/x86/kernel/numaq_32.c node_start_pfn[node] = MB_TO_PAGES( node 61 arch/x86/kernel/numaq_32.c node_end_pfn[node] = MB_TO_PAGES( node 64 arch/x86/kernel/numaq_32.c e820_register_active_regions(node, node_start_pfn[node], node 65 arch/x86/kernel/numaq_32.c node_end_pfn[node]); node 66 arch/x86/kernel/numaq_32.c memory_present(node, node 67 arch/x86/kernel/numaq_32.c node_start_pfn[node], node_end_pfn[node]); node 68 arch/x86/kernel/numaq_32.c node_remap_size[node] = node_memmap_size_bytes(node, node 69 arch/x86/kernel/numaq_32.c node_start_pfn[node], node 70 arch/x86/kernel/numaq_32.c node_end_pfn[node]); node 159 arch/x86/kernel/setup_percpu.c int node = early_cpu_to_node(cpu); node 160 arch/x86/kernel/setup_percpu.c if (!node_online(node) || !NODE_DATA(node)) { node 164 arch/x86/kernel/setup_percpu.c cpu, node); node 170 arch/x86/kernel/setup_percpu.c ptr = alloc_bootmem_pages_node(NODE_DATA(node), size); node 173 arch/x86/kernel/setup_percpu.c cpu, node, __pa(ptr)); node 203 arch/x86/kernel/setup_percpu.c unsigned int node, num = 0; node 208 arch/x86/kernel/setup_percpu.c for_each_node_mask(node, node_possible_map) node 209 arch/x86/kernel/setup_percpu.c num = node; node 227 arch/x86/kernel/setup_percpu.c if (cpu_pda(cpu) && node != NUMA_NO_NODE) node 228 arch/x86/kernel/setup_percpu.c cpu_pda(cpu)->nodenumber = node; node 231 arch/x86/kernel/setup_percpu.c cpu_to_node_map[cpu] = node; node 234 arch/x86/kernel/setup_percpu.c per_cpu(x86_cpu_to_node_map, cpu) = node; node 264 arch/x86/kernel/setup_percpu.c int node = cpu_to_node(cpu); node 274 arch/x86/kernel/setup_percpu.c mask = &node_to_cpumask_map[node]; node 282 arch/x86/kernel/setup_percpu.c enable? "numa_add_cpu":"numa_remove_cpu", cpu, node, buf); node 337 arch/x86/kernel/setup_percpu.c node); node 341 arch/x86/kernel/setup_percpu.c if (node >= nr_node_ids) { node 344 arch/x86/kernel/setup_percpu.c node, nr_node_ids); node 348 arch/x86/kernel/setup_percpu.c return &node_to_cpumask_map[node]; node 363 arch/x86/kernel/setup_percpu.c "node_to_cpumask(%d): no node_to_cpumask_map!\n", node); node 367 arch/x86/kernel/setup_percpu.c if (node >= nr_node_ids) { node 370 arch/x86/kernel/setup_percpu.c node, nr_node_ids); node 374 arch/x86/kernel/setup_percpu.c return node_to_cpumask_map[node]; node 147 arch/x86/kernel/smpboot.c printk(KERN_INFO "Mapping cpu %d to node %d\n", cpu, node); node 148 arch/x86/kernel/smpboot.c cpu_set(cpu, node_to_cpumask_map[node]); node 149 arch/x86/kernel/smpboot.c cpu_to_node_map[cpu] = node; node 155 arch/x86/kernel/smpboot.c int node; node 158 arch/x86/kernel/smpboot.c for (node = 0; node < MAX_NUMNODES; node++) node 159 arch/x86/kernel/smpboot.c cpu_clear(cpu, node_to_cpumask_map[node]); node 177 arch/x86/kernel/smpboot.c int node = apicid_to_node(apicid); node 179 arch/x86/kernel/smpboot.c if (!node_online(node)) node 180 arch/x86/kernel/smpboot.c node = first_online_node; node 183 arch/x86/kernel/smpboot.c map_cpu_to_node(cpu, node); node 776 arch/x86/kernel/smpboot.c int node = cpu_to_node(cpu); node 782 arch/x86/kernel/smpboot.c newpda = kmalloc_node(size, GFP_ATOMIC, node); node 785 arch/x86/kernel/smpboot.c "for CPU %d on node %d\n", cpu, node); node 45 arch/x86/kernel/summit_32.c int twister = 0, node = 0; node 61 arch/x86/kernel/summit_32.c node = scal_devs[i]->node_id; node 101 arch/x86/kernel/summit_32.c mp_bus_id_to_node[bus] = node; node 594 arch/x86/kernel/tlb_uv.c kmalloc_node(sizeof(struct bau_control), GFP_KERNEL, node); node 599 arch/x86/kernel/tlb_uv.c DEST_Q_SIZE, GFP_KERNEL, node); node 607 arch/x86/kernel/tlb_uv.c kmalloc_node(sizeof(int) * DEST_NUM_RESOURCES, GFP_KERNEL, node); node 656 arch/x86/kernel/tlb_uv.c kmalloc_node(16384, GFP_KERNEL, node); node 696 arch/x86/kernel/tlb_uv.c GFP_KERNEL, node); node 730 arch/x86/kernel/tlb_uv.c bau_tablesp = uv_table_bases_init(blade, node); node 732 arch/x86/kernel/tlb_uv.c adp = uv_activation_descriptor_init(node, pnode); node 733 arch/x86/kernel/tlb_uv.c pqp = uv_payload_queue_init(node, pnode, bau_tablesp); node 734 arch/x86/kernel/tlb_uv.c uv_table_bases_finish(blade, node, cur_cpu, bau_tablesp, adp); node 754 arch/x86/kernel/tlb_uv.c int node; node 767 arch/x86/kernel/tlb_uv.c for_each_online_node(node) { node 768 arch/x86/kernel/tlb_uv.c blade = uv_node_to_blade_id(node); node 779 arch/x86/kernel/tlb_uv.c for_each_online_node(node) { node 780 arch/x86/kernel/tlb_uv.c blade = uv_node_to_blade_id(node); node 784 arch/x86/kernel/tlb_uv.c uv_init_blade(blade, node, cur_cpu); node 215 arch/x86/kernel/vsyscall_64.c if (node) node 216 arch/x86/kernel/vsyscall_64.c *node = p >> 12; node 254 arch/x86/kernel/vsyscall_64.c unsigned long node = 0; node 256 arch/x86/kernel/vsyscall_64.c node = cpu_to_node(cpu); node 259 arch/x86/kernel/vsyscall_64.c write_rdtscp_aux((node << 12) | cpu); node 266 arch/x86/kernel/vsyscall_64.c d |= (node & 0xf) << 12; node 267 arch/x86/kernel/vsyscall_64.c d |= (node >> 4) << 48; node 791 arch/x86/kvm/mmu.c struct hlist_node *node; node 809 arch/x86/kvm/mmu.c hlist_for_each_entry(pte_chain, node, &sp->parent_ptes, link) { node 828 arch/x86/kvm/mmu.c struct hlist_node *node; node 836 arch/x86/kvm/mmu.c hlist_for_each_entry(pte_chain, node, &sp->parent_ptes, link) node 876 arch/x86/kvm/mmu.c struct hlist_node *node; node 881 arch/x86/kvm/mmu.c hlist_for_each_entry(sp, node, bucket, hash_link) node 904 arch/x86/kvm/mmu.c struct hlist_node *node; node 920 arch/x86/kvm/mmu.c hlist_for_each_entry(sp, node, bucket, hash_link) node 1069 arch/x86/kvm/mmu.c struct hlist_node *node, *n; node 1076 arch/x86/kvm/mmu.c hlist_for_each_entry_safe(sp, node, n, bucket, hash_link) node 1805 arch/x86/kvm/mmu.c struct hlist_node *node, *n; node 1839 arch/x86/kvm/mmu.c hlist_for_each_entry_safe(sp, node, n, bucket, hash_link) { node 2075 arch/x86/kvm/mmu.c struct kvm_mmu_page *sp, *node; node 2078 arch/x86/kvm/mmu.c list_for_each_entry_safe(sp, node, &kvm->arch.active_mmu_pages, link) node 1207 arch/x86/kvm/vmx.c int node = cpu_to_node(cpu); node 1211 arch/x86/kvm/vmx.c pages = alloc_pages_node(node, GFP_KERNEL, vmcs_config.order); node 1155 arch/x86/mm/init_64.c pgd = vmemmap_pgd_populate(addr, node); node 1159 arch/x86/mm/init_64.c pud = vmemmap_pud_populate(pgd, addr, node); node 1165 arch/x86/mm/init_64.c pmd = vmemmap_pmd_populate(pud, addr, node); node 1170 arch/x86/mm/init_64.c p = vmemmap_pte_populate(pmd, addr, node); node 1184 arch/x86/mm/init_64.c p = vmemmap_alloc_block(PMD_SIZE, node); node 1193 arch/x86/mm/init_64.c if (p_end != p || node_start != node) { node 1198 arch/x86/mm/init_64.c node_start = node; node 1205 arch/x86/mm/init_64.c vmemmap_verify((pte_t *)pmd, node, addr, next); node 209 arch/x86/mm/numa_32.c int node; node 211 arch/x86/mm/numa_32.c for_each_online_node(node) { node 212 arch/x86/mm/numa_32.c printk(KERN_DEBUG "remap_numa_kva: node %d\n", node); node 213 arch/x86/mm/numa_32.c for (pfn=0; pfn < node_remap_size[node]; pfn += PTRS_PER_PTE) { node 214 arch/x86/mm/numa_32.c vaddr = node_remap_start_vaddr[node]+(pfn<<PAGE_SHIFT); node 217 arch/x86/mm/numa_32.c node_remap_start_pfn[node] + pfn); node 219 arch/x86/mm/numa_32.c node_remap_start_pfn[node] + pfn, node 627 arch/x86/mm/numa_64.c int node; node 632 arch/x86/mm/numa_64.c node = apicid_to_node[apicid]; node 633 arch/x86/mm/numa_64.c if (node == NUMA_NO_NODE) node 635 arch/x86/mm/numa_64.c if (!node_online(node)) node 637 arch/x86/mm/numa_64.c numa_set_node(cpu, node); node 122 arch/x86/mm/srat_64.c int pxm, node; node 134 arch/x86/mm/srat_64.c node = setup_node(pxm); node 135 arch/x86/mm/srat_64.c if (node < 0) { node 145 arch/x86/mm/srat_64.c apicid_to_node[apic_id] = node; node 148 arch/x86/mm/srat_64.c pxm, apic_id, node); node 169 arch/x86/mm/srat_64.c struct bootnode *nd = &nodes_add[node]; node 190 arch/x86/mm/srat_64.c if (!hotadd_enough_memory(&nodes_add[node])) { node 227 arch/x86/mm/srat_64.c int node, pxm; node 244 arch/x86/mm/srat_64.c node = setup_node(pxm); node 245 arch/x86/mm/srat_64.c if (node < 0) { node 251 arch/x86/mm/srat_64.c if (i == node) { node 263 arch/x86/mm/srat_64.c nd = &nodes[node]; node 265 arch/x86/mm/srat_64.c if (!node_test_and_set(node, nodes_parsed)) { node 275 arch/x86/mm/srat_64.c printk(KERN_INFO "SRAT: Node %u PXM %u %lx-%lx\n", node, pxm, node 277 arch/x86/mm/srat_64.c e820_register_active_regions(node, start >> PAGE_SHIFT, node 279 arch/x86/mm/srat_64.c push_node_boundaries(node, nd->start >> PAGE_SHIFT, node 283 arch/x86/mm/srat_64.c (reserve_hotadd(node, start, end) < 0)) { node 288 arch/x86/mm/srat_64.c node_clear(node, nodes_parsed); node 293 arch/x86/mm/srat_64.c memblk_nodeid[num_node_memblks] = node; node 329 arch/x86/mm/srat_64.c node_clear(node, nodes_parsed); node 331 arch/x86/mm/srat_64.c if (apicid_to_node[i] == node) node 386 arch/x86/mm/srat_64.c int node = early_cpu_to_node(i); node 388 arch/x86/mm/srat_64.c if (node == NUMA_NO_NODE) node 390 arch/x86/mm/srat_64.c if (!node_isset(node, node_possible_map)) node 155 arch/x86/pci/acpi.c int node; node 166 arch/x86/pci/acpi.c node = -1; node 170 arch/x86/pci/acpi.c node = pxm_to_node(pxm); node 171 arch/x86/pci/acpi.c if (node != -1) node 172 arch/x86/pci/acpi.c set_mp_bus_to_node(busnum, node); node 175 arch/x86/pci/acpi.c node = get_mp_bus_to_node(busnum); node 177 arch/x86/pci/acpi.c if (node != -1 && !node_online(node)) node 178 arch/x86/pci/acpi.c node = -1; node 191 arch/x86/pci/acpi.c sd->node = node; node 210 arch/x86/pci/acpi.c if (bus && node != -1) { node 214 arch/x86/pci/acpi.c busnum, pxm, node); node 217 arch/x86/pci/acpi.c busnum, node); node 29 arch/x86/pci/amd_bus.c mp_bus_to_node[busnum] = node; node 34 arch/x86/pci/amd_bus.c int node = -1; node 37 arch/x86/pci/amd_bus.c return node; node 39 arch/x86/pci/amd_bus.c node = mp_bus_to_node[busnum]; node 45 arch/x86/pci/amd_bus.c if (node != -1 && !node_online(node)) node 46 arch/x86/pci/amd_bus.c node = -1; node 48 arch/x86/pci/amd_bus.c return node; node 58 arch/x86/pci/amd_bus.c mp_bus_to_node[busnum] = (unsigned char) node; node 63 arch/x86/pci/amd_bus.c int node; node 67 arch/x86/pci/amd_bus.c node = mp_bus_to_node[busnum]; node 68 arch/x86/pci/amd_bus.c return node; node 88 arch/x86/pci/amd_bus.c int node; node 283 arch/x86/pci/amd_bus.c int node; node 338 arch/x86/pci/amd_bus.c node = (reg >> 4) & 0x07; node 341 arch/x86/pci/amd_bus.c mp_bus_to_node[j] = (unsigned char) node; node 348 arch/x86/pci/amd_bus.c info->node = node; node 370 arch/x86/pci/amd_bus.c node = reg & 0x07; node 377 arch/x86/pci/amd_bus.c if (info->node == node && info->link == link) node 385 arch/x86/pci/amd_bus.c node, link, (u64)start, (u64)end); node 397 arch/x86/pci/amd_bus.c if (info->node == def_node && info->link == def_link) node 440 arch/x86/pci/amd_bus.c node = reg & 0x07; node 449 arch/x86/pci/amd_bus.c if (info->node == node && info->link == link) node 458 arch/x86/pci/amd_bus.c node, link, (u64)start, (u64)end); node 523 arch/x86/pci/amd_bus.c if (info->node == def_node && info->link == def_link) node 546 arch/x86/pci/amd_bus.c info->bus_min, info->bus_max, info->node, info->link); node 395 arch/x86/pci/common.c sd->node = get_mp_bus_to_node(busnum); node 564 arch/x86/pci/common.c sd->node = node; node 122 arch/x86/pci/i386.c list_for_each_entry(bus, bus_list, node) { node 141 arch/x86/pci/irq.c int node; node 144 arch/x86/pci/irq.c node = get_mp_bus_to_node(i); node 145 arch/x86/pci/irq.c if (pci_scan_bus_on_node(i, &pci_root_ops, node)) node 15 arch/x86/pci/legacy.c long node; node 25 arch/x86/pci/legacy.c node = get_mp_bus_to_node(n); node 31 arch/x86/pci/legacy.c pci_scan_bus_on_node(n, &pci_root_ops, node); node 30 arch/x86/vdso/vgetcpu.c if (node) node 31 arch/x86/vdso/vgetcpu.c *node = p >> 12; node 229 block/as-iosched.c struct io_context *ioc = get_io_context(GFP_ATOMIC, node); node 1196 block/as-iosched.c rq->elevator_private = as_get_io_context(q->node); node 1333 block/as-iosched.c ioc = as_get_io_context(q->node); node 1363 block/as-iosched.c ad = kmalloc_node(sizeof(*ad), GFP_KERNEL | __GFP_ZERO, q->node); node 469 block/blk-core.c mempool_free_slab, request_cachep, q->node); node 561 block/blk-core.c q->node = node_id; node 734 block/blk-core.c ioc = current_io_context(GFP_ATOMIC, q->node); node 849 block/blk-core.c ioc = current_io_context(GFP_NOIO, q->node); node 91 block/blk-ioc.c ret = kmem_cache_alloc_node(iocontext_cachep, gfp_flags, node); node 126 block/blk-ioc.c ret = alloc_io_context(gfp_flags, node); node 151 block/blk-ioc.c ret = current_io_context(gfp_flags, node); node 1341 block/cfq-iosched.c cfqd->queue->node); node 1457 block/cfq-iosched.c cfqd->queue->node); node 1463 block/cfq-iosched.c cfqd->queue->node); node 1659 block/cfq-iosched.c ioc = get_io_context(gfp_mask, cfqd->queue->node); node 2214 block/cfq-iosched.c cfqd = kmalloc_node(sizeof(*cfqd), GFP_KERNEL | __GFP_ZERO, q->node); node 68 block/deadline-iosched.c struct rb_node *node = rb_next(&rq->rb_node); node 70 block/deadline-iosched.c if (node) node 71 block/deadline-iosched.c return rb_entry_rq(node); node 354 block/deadline-iosched.c dd = kmalloc_node(sizeof(*dd), GFP_KERNEL | __GFP_ZERO, q->node); node 211 block/elevator.c eq = kmalloc_node(sizeof(elevator_t), GFP_KERNEL | __GFP_ZERO, q->node); node 221 block/elevator.c GFP_KERNEL, q->node); node 72 block/noop-iosched.c nd = kmalloc_node(sizeof(*nd), GFP_KERNEL, q->node); node 69 crypto/async_tx/async_tx.c list_for_each_entry_rcu(ref, &async_tx_master_list, node) node 163 crypto/async_tx/async_tx.c INIT_LIST_HEAD(&ref->node); node 183 crypto/async_tx/async_tx.c list_for_each_entry_rcu(ref, &async_tx_master_list, node) node 260 crypto/async_tx/async_tx.c list_for_each_entry_rcu(ref, &async_tx_master_list, node) node 282 crypto/async_tx/async_tx.c list_add_tail_rcu(&master_ref->node, node 299 crypto/async_tx/async_tx.c list_for_each_entry(ref, &async_tx_master_list, node) node 303 crypto/async_tx/async_tx.c list_del_rcu(&ref->node); node 480 fs/afs/cmservice.c r->node[loop] = ntohl(b[loop + 5]); node 547 fs/afs/cmservice.c reply.ia.uuid[loop + 5] = htonl((s8) afs_uuid.node[loop]); node 426 fs/afs/internal.h u8 node[6]; /* spatially unique node ID (MAC addr) */ node 58 fs/afs/main.c ret = afs_get_MAC_address(afs_uuid.node, sizeof(afs_uuid.node)); node 83 fs/afs/main.c afs_uuid.node[0], afs_uuid.node[1], afs_uuid.node[2], node 84 fs/afs/main.c afs_uuid.node[3], afs_uuid.node[4], afs_uuid.node[5]); node 27 fs/afs/vnode.c if (!node) node 30 fs/afs/vnode.c if (node->rb_left) node 31 fs/afs/vnode.c bad = dump_tree_aux(node->rb_left, node, depth + 2, '/'); node 33 fs/afs/vnode.c vnode = rb_entry(node, struct afs_vnode, cb_promise); node 35 fs/afs/vnode.c rb_is_red(node) ? 'R' : 'B', node 38 fs/afs/vnode.c if (rb_parent(node) != parent) { node 39 fs/afs/vnode.c printk("BAD: %p != %p\n", rb_parent(node), parent); node 43 fs/afs/vnode.c if (node->rb_right) node 44 fs/afs/vnode.c bad |= dump_tree_aux(node->rb_right, node, depth + 2, '\\'); node 200 fs/befs/btree.c if (node->bh) node 201 fs/befs/btree.c brelse(node->bh); node 203 fs/befs/btree.c node->bh = befs_read_datastream(sb, ds, node_off, &off); node 204 fs/befs/btree.c if (!node->bh) { node 211 fs/befs/btree.c node->od_node = node 212 fs/befs/btree.c (befs_btree_nodehead *) ((void *) node->bh->b_data + off); node 214 fs/befs/btree.c befs_dump_index_node(sb, node->od_node); node 216 fs/befs/btree.c node->head.left = fs64_to_cpu(sb, node->od_node->left); node 217 fs/befs/btree.c node->head.right = fs64_to_cpu(sb, node->od_node->right); node 218 fs/befs/btree.c node->head.overflow = fs64_to_cpu(sb, node->od_node->overflow); node 219 fs/befs/btree.c node->head.all_key_count = node 220 fs/befs/btree.c fs16_to_cpu(sb, node->od_node->all_key_count); node 221 fs/befs/btree.c node->head.all_key_length = node 222 fs/befs/btree.c fs16_to_cpu(sb, node->od_node->all_key_length); node 353 fs/befs/btree.c last = node->head.all_key_count - 1; node 354 fs/befs/btree.c thiskey = befs_bt_get_key(sb, node, last, &keylen); node 362 fs/befs/btree.c valarray = befs_bt_valarray(node); node 371 fs/befs/btree.c thiskey = befs_bt_get_key(sb, node, mid, &keylen); node 605 fs/befs/btree.c if (node->head.overflow == befs_bt_inval) node 629 fs/befs/btree.c (sizeof (befs_btree_nodehead) + node->head.all_key_length); node 635 fs/befs/btree.c return (fs16 *) ((void *) node->od_node + off); node 648 fs/befs/btree.c void *keylen_index_start = (void *) befs_bt_keylen_index(node); node 649 fs/befs/btree.c size_t keylen_index_size = node->head.all_key_count * sizeof (fs16); node 664 fs/befs/btree.c return (char *) ((void *) node->od_node + sizeof (befs_btree_nodehead)); node 685 fs/befs/btree.c if (index < 0 || index > node->head.all_key_count) { node 690 fs/befs/btree.c keystart = befs_bt_keydata(node); node 691 fs/befs/btree.c keylen_index = befs_bt_keylen_index(node); node 272 fs/befs/debug.c befs_debug(sb, " left %016LX", fs64_to_cpu(sb, node->left)); node 273 fs/befs/debug.c befs_debug(sb, " right %016LX", fs64_to_cpu(sb, node->right)); node 274 fs/befs/debug.c befs_debug(sb, " overflow %016LX", fs64_to_cpu(sb, node->overflow)); node 276 fs/befs/debug.c fs16_to_cpu(sb, node->all_key_count)); node 278 fs/befs/debug.c fs16_to_cpu(sb, node->all_key_length)); node 3981 fs/cifs/cifssmb.c struct dfs_info3_param *node = (*target_nodes)+i; node 3983 fs/cifs/cifssmb.c node->flags = le16_to_cpu(pSMBr->DFSFlags); node 3984 fs/cifs/cifssmb.c node->path_consumed = le16_to_cpu(pSMBr->PathConsumed); node 3985 fs/cifs/cifssmb.c node->server_type = le16_to_cpu(ref->ServerType); node 3986 fs/cifs/cifssmb.c node->ref_flag = le16_to_cpu(ref->ReferralEntryFlags); node 3991 fs/cifs/cifssmb.c rc = cifs_strncpy_to_host(&(node->path_name), temp, node 3999 fs/cifs/cifssmb.c rc = cifs_strncpy_to_host(&(node->node_name), temp, node 1375 fs/dcache.c struct hlist_node *node; node 1380 fs/dcache.c hlist_for_each_entry_rcu(dentry, node, head, d_hash) { node 83 fs/dlm/dir.c uint32_t node, n = 0; node 92 fs/dlm/dir.c node = (hash >> 16) % ls->ls_total_weight; node 93 fs/dlm/dir.c nodeid = ls->ls_node_array[node]; node 99 fs/dlm/dir.c node = (hash >> 16) % ls->ls_num_nodes; node 102 fs/dlm/dir.c if (n++ != node) node 109 fs/dlm/dir.c ls->ls_num_nodes, n, node);); node 243 fs/dquot.c struct hlist_node *node; node 246 fs/dquot.c hlist_for_each (node, dquot_hash+hashent) { node 247 fs/dquot.c dquot = hlist_entry(node, struct dquot, dq_hash); node 398 fs/ecryptfs/ecryptfs_kernel.h struct list_head node; node 65 fs/ecryptfs/messaging.c *msg_ctx = list_entry(p, struct ecryptfs_msg_ctx, node); node 85 fs/ecryptfs/messaging.c list_move(&msg_ctx->node, &ecryptfs_msg_ctx_alloc_list); node 98 fs/ecryptfs/messaging.c list_move(&(msg_ctx->node), &ecryptfs_msg_ctx_free_list); node 628 fs/ecryptfs/messaging.c INIT_LIST_HEAD(&ecryptfs_msg_ctx_arr[i].node); node 637 fs/ecryptfs/messaging.c list_add_tail(&ecryptfs_msg_ctx_arr[i].node, node 343 fs/ext2/balloc.c struct rb_node *node = &rsv->rsv_node; node 365 fs/ext2/balloc.c rb_link_node(node, parent, p); node 366 fs/ext2/balloc.c rb_insert_color(node, root); node 330 fs/ext3/balloc.c struct rb_node *node = &rsv->rsv_node; node 352 fs/ext3/balloc.c rb_link_node(node, parent, p); node 353 fs/ext3/balloc.c rb_insert_color(node, root); node 21 fs/hfs/bnode.c off += node->page_offset; node 22 fs/hfs/bnode.c page = node->page[0]; node 32 fs/hfs/bnode.c hfs_bnode_read(node, &data, off, 2); node 40 fs/hfs/bnode.c hfs_bnode_read(node, &data, off, 1); node 49 fs/hfs/bnode.c tree = node->tree; node 50 fs/hfs/bnode.c if (node->type == HFS_NODE_LEAF || node 52 fs/hfs/bnode.c key_len = hfs_bnode_read_u8(node, off) + 1; node 56 fs/hfs/bnode.c hfs_bnode_read(node, key, off, key_len); node 63 fs/hfs/bnode.c off += node->page_offset; node 64 fs/hfs/bnode.c page = node->page[0]; node 75 fs/hfs/bnode.c hfs_bnode_write(node, &v, off, 2); node 81 fs/hfs/bnode.c hfs_bnode_write(node, &data, off, 1); node 88 fs/hfs/bnode.c off += node->page_offset; node 89 fs/hfs/bnode.c page = node->page[0]; node 125 fs/hfs/bnode.c src += node->page_offset; node 126 fs/hfs/bnode.c dst += node->page_offset; node 127 fs/hfs/bnode.c page = node->page[0]; node 140 fs/hfs/bnode.c dprint(DBG_BNODE_MOD, "bnode: %d\n", node->this); node 141 fs/hfs/bnode.c hfs_bnode_read(node, &desc, 0, sizeof(desc)); node 146 fs/hfs/bnode.c off = node->tree->node_size - 2; node 148 fs/hfs/bnode.c key_off = hfs_bnode_read_u16(node, off); node 150 fs/hfs/bnode.c if (i && node->type == HFS_NODE_INDEX) { node 153 fs/hfs/bnode.c if (node->tree->attributes & HFS_TREE_VARIDXKEYS) node 154 fs/hfs/bnode.c tmp = (hfs_bnode_read_u8(node, key_off) | 1) + 1; node 156 fs/hfs/bnode.c tmp = node->tree->max_key_len + 1; node 157 fs/hfs/bnode.c dprint(DBG_BNODE_MOD, " (%d,%d", tmp, hfs_bnode_read_u8(node, key_off)); node 158 fs/hfs/bnode.c hfs_bnode_read(node, &cnid, key_off + tmp, 4); node 160 fs/hfs/bnode.c } else if (i && node->type == HFS_NODE_LEAF) { node 163 fs/hfs/bnode.c tmp = hfs_bnode_read_u8(node, key_off); node 176 fs/hfs/bnode.c tree = node->tree; node 177 fs/hfs/bnode.c if (node->prev) { node 178 fs/hfs/bnode.c tmp = hfs_bnode_find(tree, node->prev); node 181 fs/hfs/bnode.c tmp->next = node->next; node 185 fs/hfs/bnode.c } else if (node->type == HFS_NODE_LEAF) node 186 fs/hfs/bnode.c tree->leaf_head = node->next; node 188 fs/hfs/bnode.c if (node->next) { node 189 fs/hfs/bnode.c tmp = hfs_bnode_find(tree, node->next); node 192 fs/hfs/bnode.c tmp->prev = node->prev; node 196 fs/hfs/bnode.c } else if (node->type == HFS_NODE_LEAF) node 197 fs/hfs/bnode.c tree->leaf_tail = node->prev; node 200 fs/hfs/bnode.c if (!node->prev && !node->next) { node 203 fs/hfs/bnode.c if (!node->parent) { node 207 fs/hfs/bnode.c set_bit(HFS_BNODE_DELETED, &node->flags); node 219 fs/hfs/bnode.c struct hfs_bnode *node; node 226 fs/hfs/bnode.c for (node = tree->node_hash[hfs_bnode_hash(cnid)]; node 227 fs/hfs/bnode.c node; node = node->next_hash) { node 228 fs/hfs/bnode.c if (node->this == cnid) { node 229 fs/hfs/bnode.c return node; node 238 fs/hfs/bnode.c struct hfs_bnode *node, *node2; node 252 fs/hfs/bnode.c node = kzalloc(size, GFP_KERNEL); node 253 fs/hfs/bnode.c if (!node) node 255 fs/hfs/bnode.c node->tree = tree; node 256 fs/hfs/bnode.c node->this = cnid; node 257 fs/hfs/bnode.c set_bit(HFS_BNODE_NEW, &node->flags); node 258 fs/hfs/bnode.c atomic_set(&node->refcnt, 1); node 260 fs/hfs/bnode.c node->tree->cnid, node->this); node 261 fs/hfs/bnode.c init_waitqueue_head(&node->lock_wq); node 266 fs/hfs/bnode.c node->next_hash = tree->node_hash[hash]; node 267 fs/hfs/bnode.c tree->node_hash[hash] = node; node 271 fs/hfs/bnode.c kfree(node); node 280 fs/hfs/bnode.c node->page_offset = off & ~PAGE_CACHE_MASK; node 290 fs/hfs/bnode.c node->page[i] = page; node 293 fs/hfs/bnode.c return node; node 295 fs/hfs/bnode.c set_bit(HFS_BNODE_ERROR, &node->flags); node 296 fs/hfs/bnode.c return node; node 304 fs/hfs/bnode.c node->tree->cnid, node->this, atomic_read(&node->refcnt)); node 305 fs/hfs/bnode.c for (p = &node->tree->node_hash[hfs_bnode_hash(node->this)]; node 306 fs/hfs/bnode.c *p && *p != node; p = &(*p)->next_hash) node 309 fs/hfs/bnode.c *p = node->next_hash; node 310 fs/hfs/bnode.c node->tree->node_hash_cnt--; node 316 fs/hfs/bnode.c struct hfs_bnode *node; node 322 fs/hfs/bnode.c node = hfs_bnode_findhash(tree, num); node 323 fs/hfs/bnode.c if (node) { node 324 fs/hfs/bnode.c hfs_bnode_get(node); node 326 fs/hfs/bnode.c wait_event(node->lock_wq, !test_bit(HFS_BNODE_NEW, &node->flags)); node 327 fs/hfs/bnode.c if (test_bit(HFS_BNODE_ERROR, &node->flags)) node 329 fs/hfs/bnode.c return node; node 332 fs/hfs/bnode.c node = __hfs_bnode_create(tree, num); node 333 fs/hfs/bnode.c if (!node) node 335 fs/hfs/bnode.c if (test_bit(HFS_BNODE_ERROR, &node->flags)) node 337 fs/hfs/bnode.c if (!test_bit(HFS_BNODE_NEW, &node->flags)) node 338 fs/hfs/bnode.c return node; node 340 fs/hfs/bnode.c desc = (struct hfs_bnode_desc *)(kmap(node->page[0]) + node->page_offset); node 341 fs/hfs/bnode.c node->prev = be32_to_cpu(desc->prev); node 342 fs/hfs/bnode.c node->next = be32_to_cpu(desc->next); node 343 fs/hfs/bnode.c node->num_recs = be16_to_cpu(desc->num_recs); node 344 fs/hfs/bnode.c node->type = desc->type; node 345 fs/hfs/bnode.c node->height = desc->height; node 346 fs/hfs/bnode.c kunmap(node->page[0]); node 348 fs/hfs/bnode.c switch (node->type) { node 351 fs/hfs/bnode.c if (node->height != 0) node 355 fs/hfs/bnode.c if (node->height != 1) node 359 fs/hfs/bnode.c if (node->height <= 1 || node->height > tree->depth) node 367 fs/hfs/bnode.c off = hfs_bnode_read_u16(node, rec_off); node 370 fs/hfs/bnode.c for (i = 1; i <= node->num_recs; off = next_off, i++) { node 372 fs/hfs/bnode.c next_off = hfs_bnode_read_u16(node, rec_off); node 378 fs/hfs/bnode.c if (node->type != HFS_NODE_INDEX && node 379 fs/hfs/bnode.c node->type != HFS_NODE_LEAF) node 381 fs/hfs/bnode.c key_size = hfs_bnode_read_u8(node, off) + 1; node 385 fs/hfs/bnode.c clear_bit(HFS_BNODE_NEW, &node->flags); node 386 fs/hfs/bnode.c wake_up(&node->lock_wq); node 387 fs/hfs/bnode.c return node; node 390 fs/hfs/bnode.c set_bit(HFS_BNODE_ERROR, &node->flags); node 391 fs/hfs/bnode.c clear_bit(HFS_BNODE_NEW, &node->flags); node 392 fs/hfs/bnode.c wake_up(&node->lock_wq); node 393 fs/hfs/bnode.c hfs_bnode_put(node); node 404 fs/hfs/bnode.c kfree(node); node 409 fs/hfs/bnode.c struct hfs_bnode *node; node 414 fs/hfs/bnode.c node = hfs_bnode_findhash(tree, num); node 416 fs/hfs/bnode.c BUG_ON(node); node 417 fs/hfs/bnode.c node = __hfs_bnode_create(tree, num); node 418 fs/hfs/bnode.c if (!node) node 420 fs/hfs/bnode.c if (test_bit(HFS_BNODE_ERROR, &node->flags)) { node 421 fs/hfs/bnode.c hfs_bnode_put(node); node 425 fs/hfs/bnode.c pagep = node->page; node 426 fs/hfs/bnode.c memset(kmap(*pagep) + node->page_offset, 0, node 435 fs/hfs/bnode.c clear_bit(HFS_BNODE_NEW, &node->flags); node 436 fs/hfs/bnode.c wake_up(&node->lock_wq); node 438 fs/hfs/bnode.c return node; node 443 fs/hfs/bnode.c if (node) { node 444 fs/hfs/bnode.c atomic_inc(&node->refcnt); node 446 fs/hfs/bnode.c node->tree->cnid, node->this, atomic_read(&node->refcnt)); node 453 fs/hfs/bnode.c if (node) { node 454 fs/hfs/bnode.c struct hfs_btree *tree = node->tree; node 458 fs/hfs/bnode.c node->tree->cnid, node->this, atomic_read(&node->refcnt)); node 459 fs/hfs/bnode.c BUG_ON(!atomic_read(&node->refcnt)); node 460 fs/hfs/bnode.c if (!atomic_dec_and_lock(&node->refcnt, &tree->hash_lock)) node 463 fs/hfs/bnode.c if (!node->page[i]) node 465 fs/hfs/bnode.c mark_page_accessed(node->page[i]); node 468 fs/hfs/bnode.c if (test_bit(HFS_BNODE_DELETED, &node->flags)) { node 469 fs/hfs/bnode.c hfs_bnode_unhash(node); node 471 fs/hfs/bnode.c hfs_bmap_free(node); node 472 fs/hfs/bnode.c hfs_bnode_free(node); node 23 fs/hfs/brec.c dataoff = node->tree->node_size - (rec + 2) * 2; node 24 fs/hfs/brec.c hfs_bnode_read(node, retval, dataoff, 4); node 34 fs/hfs/brec.c if (node->type != HFS_NODE_INDEX && node->type != HFS_NODE_LEAF) node 37 fs/hfs/brec.c if ((node->type == HFS_NODE_INDEX) && node 38 fs/hfs/brec.c !(node->tree->attributes & HFS_TREE_VARIDXKEYS)) { node 39 fs/hfs/brec.c if (node->tree->attributes & HFS_TREE_BIGKEYS) node 40 fs/hfs/brec.c retval = node->tree->max_key_len + 2; node 42 fs/hfs/brec.c retval = node->tree->max_key_len + 1; node 44 fs/hfs/brec.c recoff = hfs_bnode_read_u16(node, node->tree->node_size - (rec + 1) * 2); node 47 fs/hfs/brec.c if (node->tree->attributes & HFS_TREE_BIGKEYS) { node 48 fs/hfs/brec.c retval = hfs_bnode_read_u16(node, recoff) + 2; node 49 fs/hfs/brec.c if (retval > node->tree->max_key_len + 2) { node 55 fs/hfs/brec.c retval = (hfs_bnode_read_u8(node, recoff) | 1) + 1; node 56 fs/hfs/brec.c if (retval > node->tree->max_key_len + 1) { node 69 fs/hfs/brec.c struct hfs_bnode *node, *new_node; node 91 fs/hfs/brec.c node = fd->bnode; node 92 fs/hfs/brec.c hfs_bnode_dump(node); node 94 fs/hfs/brec.c end_rec_off = tree->node_size - (node->num_recs + 1) * 2; node 95 fs/hfs/brec.c end_off = hfs_bnode_read_u16(node, end_rec_off); node 106 fs/hfs/brec.c if (node->type == HFS_NODE_LEAF) { node 110 fs/hfs/brec.c node->num_recs++; node 112 fs/hfs/brec.c hfs_bnode_write_u16(node, offsetof(struct hfs_bnode_desc, num_recs), node->num_recs); node 113 fs/hfs/brec.c hfs_bnode_write_u16(node, end_rec_off, end_off + size); node 121 fs/hfs/brec.c data_off = hfs_bnode_read_u16(node, data_rec_off + 2); node 122 fs/hfs/brec.c hfs_bnode_write_u16(node, data_rec_off, data_off + size); node 127 fs/hfs/brec.c hfs_bnode_move(node, data_off + size, data_off, node 131 fs/hfs/brec.c hfs_bnode_write(node, fd->search_key, data_off, key_len); node 132 fs/hfs/brec.c hfs_bnode_write(node, entry, data_off + key_len, entry_len); node 133 fs/hfs/brec.c hfs_bnode_dump(node); node 139 fs/hfs/brec.c if (!rec && new_node != node) node 179 fs/hfs/brec.c struct hfs_bnode *node, *parent; node 183 fs/hfs/brec.c node = fd->bnode; node 186 fs/hfs/brec.c end_off = tree->node_size - (node->num_recs + 1) * 2; node 188 fs/hfs/brec.c if (node->type == HFS_NODE_LEAF) { node 192 fs/hfs/brec.c hfs_bnode_dump(node); node 194 fs/hfs/brec.c if (!--node->num_recs) { node 195 fs/hfs/brec.c hfs_bnode_unlink(node); node 196 fs/hfs/brec.c if (!node->parent) node 198 fs/hfs/brec.c parent = hfs_bnode_find(tree, node->parent); node 201 fs/hfs/brec.c hfs_bnode_put(node); node 202 fs/hfs/brec.c node = fd->bnode = parent; node 204 fs/hfs/brec.c __hfs_brec_find(node, fd); node 207 fs/hfs/brec.c hfs_bnode_write_u16(node, offsetof(struct hfs_bnode_desc, num_recs), node->num_recs); node 214 fs/hfs/brec.c data_off = hfs_bnode_read_u16(node, rec_off); node 215 fs/hfs/brec.c hfs_bnode_write_u16(node, rec_off + 2, data_off - size); node 220 fs/hfs/brec.c hfs_bnode_move(node, fd->keyoffset, fd->keyoffset + size, node 223 fs/hfs/brec.c hfs_bnode_dump(node); node 232 fs/hfs/brec.c struct hfs_bnode *node, *new_node, *next_node; node 238 fs/hfs/brec.c node = fd->bnode; node 242 fs/hfs/brec.c hfs_bnode_get(node); node 244 fs/hfs/brec.c node->this, new_node->this, node->next); node 245 fs/hfs/brec.c new_node->next = node->next; node 246 fs/hfs/brec.c new_node->prev = node->this; node 247 fs/hfs/brec.c new_node->parent = node->parent; node 248 fs/hfs/brec.c new_node->type = node->type; node 249 fs/hfs/brec.c new_node->height = node->height; node 251 fs/hfs/brec.c if (node->next) node 252 fs/hfs/brec.c next_node = hfs_bnode_find(tree, node->next); node 257 fs/hfs/brec.c hfs_bnode_put(node); node 262 fs/hfs/brec.c size = tree->node_size / 2 - node->num_recs * 2 - 14; node 266 fs/hfs/brec.c data_start = hfs_bnode_read_u16(node, old_rec_off); node 270 fs/hfs/brec.c if (++num_recs < node->num_recs) node 273 fs/hfs/brec.c hfs_bnode_put(node); node 286 fs/hfs/brec.c data_start = hfs_bnode_read_u16(node, old_rec_off); node 288 fs/hfs/brec.c hfs_bnode_put(node); node 295 fs/hfs/brec.c new_node->num_recs = node->num_recs - num_recs; node 296 fs/hfs/brec.c node->num_recs = num_recs; node 307 fs/hfs/brec.c data_end = hfs_bnode_read_u16(node, old_rec_off); node 312 fs/hfs/brec.c hfs_bnode_copy(new_node, 14, node, data_start, data_end - data_start); node 324 fs/hfs/brec.c node->next = new_node->this; node 325 fs/hfs/brec.c hfs_bnode_read(node, &node_desc, 0, sizeof(node_desc)); node 326 fs/hfs/brec.c node_desc.next = cpu_to_be32(node->next); node 327 fs/hfs/brec.c node_desc.num_recs = cpu_to_be16(node->num_recs); node 328 fs/hfs/brec.c hfs_bnode_write(node, &node_desc, 0, sizeof(node_desc)); node 337 fs/hfs/brec.c } else if (node->this == tree->leaf_tail) { node 343 fs/hfs/brec.c hfs_bnode_dump(node); node 345 fs/hfs/brec.c hfs_bnode_put(node); node 353 fs/hfs/brec.c struct hfs_bnode *node, *new_node, *parent; node 359 fs/hfs/brec.c node = fd->bnode; node 361 fs/hfs/brec.c if (!node->parent) node 365 fs/hfs/brec.c parent = hfs_bnode_find(tree, node->parent); node 374 fs/hfs/brec.c newkeylen = (hfs_bnode_read_u8(node, 14) | 1) + 1; node 412 fs/hfs/brec.c hfs_bnode_copy(parent, fd->keyoffset, node, 14, newkeylen); node 417 fs/hfs/brec.c hfs_bnode_put(node); node 418 fs/hfs/brec.c node = parent; node 434 fs/hfs/brec.c if (new_node == node) node 437 fs/hfs/brec.c hfs_bnode_read_key(node, fd->search_key, 14); node 441 fs/hfs/brec.c if (!rec && node->parent) node 444 fs/hfs/brec.c fd->bnode = node; node 450 fs/hfs/brec.c struct hfs_bnode *node, *new_node; node 455 fs/hfs/brec.c node = NULL; node 457 fs/hfs/brec.c node = hfs_bnode_find(tree, tree->root); node 458 fs/hfs/brec.c if (IS_ERR(node)) node 459 fs/hfs/brec.c return PTR_ERR(node); node 463 fs/hfs/brec.c hfs_bnode_put(node); node 492 fs/hfs/brec.c if (node) { node 494 fs/hfs/brec.c node->parent = tree->root; node 495 fs/hfs/brec.c if (node->type == HFS_NODE_LEAF || node 497 fs/hfs/brec.c key_size = hfs_bnode_read_u8(node, 14) + 1; node 500 fs/hfs/brec.c hfs_bnode_copy(new_node, 14, node, 14, key_size); node 507 fs/hfs/brec.c cnid = cpu_to_be32(node->this); node 513 fs/hfs/brec.c hfs_bnode_put(node); node 123 fs/hfs/btree.c struct hfs_bnode *node; node 130 fs/hfs/btree.c while ((node = tree->node_hash[i])) { node 131 fs/hfs/btree.c tree->node_hash[i] = node->next_hash; node 132 fs/hfs/btree.c if (atomic_read(&node->refcnt)) node 134 fs/hfs/btree.c node->tree->cnid, node->this, atomic_read(&node->refcnt)); node 135 fs/hfs/btree.c hfs_bnode_free(node); node 146 fs/hfs/btree.c struct hfs_bnode *node; node 149 fs/hfs/btree.c node = hfs_bnode_find(tree, 0); node 150 fs/hfs/btree.c if (IS_ERR(node)) node 154 fs/hfs/btree.c page = node->page[0]; node 168 fs/hfs/btree.c hfs_bnode_put(node); node 174 fs/hfs/btree.c struct hfs_bnode *node; node 178 fs/hfs/btree.c node = hfs_bnode_create(tree, idx); node 179 fs/hfs/btree.c if (IS_ERR(node)) node 180 fs/hfs/btree.c return node; node 189 fs/hfs/btree.c node->type = HFS_NODE_MAP; node 190 fs/hfs/btree.c node->num_recs = 1; node 191 fs/hfs/btree.c hfs_bnode_clear(node, 0, tree->node_size); node 198 fs/hfs/btree.c hfs_bnode_write(node, &desc, 0, sizeof(desc)); node 199 fs/hfs/btree.c hfs_bnode_write_u16(node, 14, 0x8000); node 200 fs/hfs/btree.c hfs_bnode_write_u16(node, tree->node_size - 2, 14); node 201 fs/hfs/btree.c hfs_bnode_write_u16(node, tree->node_size - 4, tree->node_size - 6); node 203 fs/hfs/btree.c return node; node 208 fs/hfs/btree.c struct hfs_bnode *node, *next_node; node 237 fs/hfs/btree.c node = hfs_bnode_find(tree, nidx); node 238 fs/hfs/btree.c if (IS_ERR(node)) node 239 fs/hfs/btree.c return node; node 240 fs/hfs/btree.c len = hfs_brec_lenoff(node, 2, &off16); node 243 fs/hfs/btree.c off += node->page_offset; node 244 fs/hfs/btree.c pagep = node->page + (off >> PAGE_CACHE_SHIFT); node 261 fs/hfs/btree.c hfs_bnode_put(node); node 275 fs/hfs/btree.c nidx = node->next; node 278 fs/hfs/btree.c next_node = hfs_bmap_new_bmap(node, idx); node 281 fs/hfs/btree.c hfs_bnode_put(node); node 284 fs/hfs/btree.c node = next_node; node 286 fs/hfs/btree.c len = hfs_brec_lenoff(node, 0, &off16); node 288 fs/hfs/btree.c off += node->page_offset; node 289 fs/hfs/btree.c pagep = node->page + (off >> PAGE_CACHE_SHIFT); node 303 fs/hfs/btree.c dprint(DBG_BNODE_MOD, "btree_free_node: %u\n", node->this); node 304 fs/hfs/btree.c tree = node->tree; node 305 fs/hfs/btree.c nidx = node->this; node 306 fs/hfs/btree.c node = hfs_bnode_find(tree, 0); node 307 fs/hfs/btree.c if (IS_ERR(node)) node 309 fs/hfs/btree.c len = hfs_brec_lenoff(node, 2, &off); node 314 fs/hfs/btree.c i = node->next; node 315 fs/hfs/btree.c hfs_bnode_put(node); node 318 fs/hfs/btree.c printk(KERN_CRIT "hfs: unable to free bnode %u. bmap not found!\n", node->this); node 321 fs/hfs/btree.c node = hfs_bnode_find(tree, i); node 322 fs/hfs/btree.c if (IS_ERR(node)) node 324 fs/hfs/btree.c if (node->type != HFS_NODE_MAP) { node 326 fs/hfs/btree.c printk(KERN_CRIT "hfs: invalid bmap found! (%u,%d)\n", node->this, node->type); node 327 fs/hfs/btree.c hfs_bnode_put(node); node 330 fs/hfs/btree.c len = hfs_brec_lenoff(node, 0, &off); node 332 fs/hfs/btree.c off += node->page_offset + nidx / 8; node 333 fs/hfs/btree.c page = node->page[off >> PAGE_CACHE_SHIFT]; node 339 fs/hfs/btree.c printk(KERN_CRIT "hfs: trying to free free bnode %u(%d)\n", node->this, node->type); node 341 fs/hfs/btree.c hfs_bnode_put(node); node 347 fs/hfs/btree.c hfs_bnode_put(node); node 58 fs/hfs/inode.c struct hfs_bnode *node; node 76 fs/hfs/inode.c node = hfs_bnode_findhash(tree, nidx); node 77 fs/hfs/inode.c if (!node) node 79 fs/hfs/inode.c else if (atomic_read(&node->refcnt)) node 81 fs/hfs/inode.c if (res && node) { node 82 fs/hfs/inode.c hfs_bnode_unhash(node); node 83 fs/hfs/inode.c hfs_bnode_free(node); node 91 fs/hfs/inode.c node = hfs_bnode_findhash(tree, nidx++); node 92 fs/hfs/inode.c if (!node) node 94 fs/hfs/inode.c if (atomic_read(&node->refcnt)) { node 98 fs/hfs/inode.c hfs_bnode_unhash(node); node 99 fs/hfs/inode.c hfs_bnode_free(node); node 26 fs/hfsplus/bnode.c off += node->page_offset; node 27 fs/hfsplus/bnode.c pagep = node->page + (off >> PAGE_CACHE_SHIFT); node 46 fs/hfsplus/bnode.c hfs_bnode_read(node, &data, off, 2); node 54 fs/hfsplus/bnode.c hfs_bnode_read(node, &data, off, 1); node 63 fs/hfsplus/bnode.c tree = node->tree; node 64 fs/hfsplus/bnode.c if (node->type == HFS_NODE_LEAF || node 66 fs/hfsplus/bnode.c key_len = hfs_bnode_read_u16(node, off) + 2; node 70 fs/hfsplus/bnode.c hfs_bnode_read(node, key, off, key_len); node 78 fs/hfsplus/bnode.c off += node->page_offset; node 79 fs/hfsplus/bnode.c pagep = node->page + (off >> PAGE_CACHE_SHIFT); node 100 fs/hfsplus/bnode.c hfs_bnode_write(node, &v, off, 2); node 108 fs/hfsplus/bnode.c off += node->page_offset; node 109 fs/hfsplus/bnode.c pagep = node->page + (off >> PAGE_CACHE_SHIFT); node 193 fs/hfsplus/bnode.c src += node->page_offset; node 194 fs/hfsplus/bnode.c dst += node->page_offset; node 197 fs/hfsplus/bnode.c src_page = node->page + (src >> PAGE_CACHE_SHIFT); node 200 fs/hfsplus/bnode.c dst_page = node->page + (dst >> PAGE_CACHE_SHIFT); node 246 fs/hfsplus/bnode.c src_page = node->page + (src >> PAGE_CACHE_SHIFT); node 248 fs/hfsplus/bnode.c dst_page = node->page + (dst >> PAGE_CACHE_SHIFT); node 300 fs/hfsplus/bnode.c dprint(DBG_BNODE_MOD, "bnode: %d\n", node->this); node 301 fs/hfsplus/bnode.c hfs_bnode_read(node, &desc, 0, sizeof(desc)); node 306 fs/hfsplus/bnode.c off = node->tree->node_size - 2; node 308 fs/hfsplus/bnode.c key_off = hfs_bnode_read_u16(node, off); node 310 fs/hfsplus/bnode.c if (i && node->type == HFS_NODE_INDEX) { node 313 fs/hfsplus/bnode.c if (node->tree->attributes & HFS_TREE_VARIDXKEYS) node 314 fs/hfsplus/bnode.c tmp = hfs_bnode_read_u16(node, key_off) + 2; node 316 fs/hfsplus/bnode.c tmp = node->tree->max_key_len + 2; node 318 fs/hfsplus/bnode.c hfs_bnode_read(node, &cnid, key_off + tmp, 4); node 320 fs/hfsplus/bnode.c } else if (i && node->type == HFS_NODE_LEAF) { node 323 fs/hfsplus/bnode.c tmp = hfs_bnode_read_u16(node, key_off); node 336 fs/hfsplus/bnode.c tree = node->tree; node 337 fs/hfsplus/bnode.c if (node->prev) { node 338 fs/hfsplus/bnode.c tmp = hfs_bnode_find(tree, node->prev); node 341 fs/hfsplus/bnode.c tmp->next = node->next; node 345 fs/hfsplus/bnode.c } else if (node->type == HFS_NODE_LEAF) node 346 fs/hfsplus/bnode.c tree->leaf_head = node->next; node 348 fs/hfsplus/bnode.c if (node->next) { node 349 fs/hfsplus/bnode.c tmp = hfs_bnode_find(tree, node->next); node 352 fs/hfsplus/bnode.c tmp->prev = node->prev; node 356 fs/hfsplus/bnode.c } else if (node->type == HFS_NODE_LEAF) node 357 fs/hfsplus/bnode.c tree->leaf_tail = node->prev; node 360 fs/hfsplus/bnode.c if (!node->prev && !node->next) { node 363 fs/hfsplus/bnode.c if (!node->parent) { node 367 fs/hfsplus/bnode.c set_bit(HFS_BNODE_DELETED, &node->flags); node 379 fs/hfsplus/bnode.c struct hfs_bnode *node; node 386 fs/hfsplus/bnode.c for (node = tree->node_hash[hfs_bnode_hash(cnid)]; node 387 fs/hfsplus/bnode.c node; node = node->next_hash) { node 388 fs/hfsplus/bnode.c if (node->this == cnid) { node 389 fs/hfsplus/bnode.c return node; node 398 fs/hfsplus/bnode.c struct hfs_bnode *node, *node2; node 412 fs/hfsplus/bnode.c node = kzalloc(size, GFP_KERNEL); node 413 fs/hfsplus/bnode.c if (!node) node 415 fs/hfsplus/bnode.c node->tree = tree; node 416 fs/hfsplus/bnode.c node->this = cnid; node 417 fs/hfsplus/bnode.c set_bit(HFS_BNODE_NEW, &node->flags); node 418 fs/hfsplus/bnode.c atomic_set(&node->refcnt, 1); node 420 fs/hfsplus/bnode.c node->tree->cnid, node->this); node 421 fs/hfsplus/bnode.c init_waitqueue_head(&node->lock_wq); node 426 fs/hfsplus/bnode.c node->next_hash = tree->node_hash[hash]; node 427 fs/hfsplus/bnode.c tree->node_hash[hash] = node; node 431 fs/hfsplus/bnode.c kfree(node); node 440 fs/hfsplus/bnode.c node->page_offset = off & ~PAGE_CACHE_MASK; node 450 fs/hfsplus/bnode.c node->page[i] = page; node 453 fs/hfsplus/bnode.c return node; node 455 fs/hfsplus/bnode.c set_bit(HFS_BNODE_ERROR, &node->flags); node 456 fs/hfsplus/bnode.c return node; node 464 fs/hfsplus/bnode.c node->tree->cnid, node->this, atomic_read(&node->refcnt)); node 465 fs/hfsplus/bnode.c for (p = &node->tree->node_hash[hfs_bnode_hash(node->this)]; node 466 fs/hfsplus/bnode.c *p && *p != node; p = &(*p)->next_hash) node 469 fs/hfsplus/bnode.c *p = node->next_hash; node 470 fs/hfsplus/bnode.c node->tree->node_hash_cnt--; node 476 fs/hfsplus/bnode.c struct hfs_bnode *node; node 482 fs/hfsplus/bnode.c node = hfs_bnode_findhash(tree, num); node 483 fs/hfsplus/bnode.c if (node) { node 484 fs/hfsplus/bnode.c hfs_bnode_get(node); node 486 fs/hfsplus/bnode.c wait_event(node->lock_wq, !test_bit(HFS_BNODE_NEW, &node->flags)); node 487 fs/hfsplus/bnode.c if (test_bit(HFS_BNODE_ERROR, &node->flags)) node 489 fs/hfsplus/bnode.c return node; node 492 fs/hfsplus/bnode.c node = __hfs_bnode_create(tree, num); node 493 fs/hfsplus/bnode.c if (!node) node 495 fs/hfsplus/bnode.c if (test_bit(HFS_BNODE_ERROR, &node->flags)) node 497 fs/hfsplus/bnode.c if (!test_bit(HFS_BNODE_NEW, &node->flags)) node 498 fs/hfsplus/bnode.c return node; node 500 fs/hfsplus/bnode.c desc = (struct hfs_bnode_desc *)(kmap(node->page[0]) + node->page_offset); node 501 fs/hfsplus/bnode.c node->prev = be32_to_cpu(desc->prev); node 502 fs/hfsplus/bnode.c node->next = be32_to_cpu(desc->next); node 503 fs/hfsplus/bnode.c node->num_recs = be16_to_cpu(desc->num_recs); node 504 fs/hfsplus/bnode.c node->type = desc->type; node 505 fs/hfsplus/bnode.c node->height = desc->height; node 506 fs/hfsplus/bnode.c kunmap(node->page[0]); node 508 fs/hfsplus/bnode.c switch (node->type) { node 511 fs/hfsplus/bnode.c if (node->height != 0) node 515 fs/hfsplus/bnode.c if (node->height != 1) node 519 fs/hfsplus/bnode.c if (node->height <= 1 || node->height > tree->depth) node 527 fs/hfsplus/bnode.c off = hfs_bnode_read_u16(node, rec_off); node 530 fs/hfsplus/bnode.c for (i = 1; i <= node->num_recs; off = next_off, i++) { node 532 fs/hfsplus/bnode.c next_off = hfs_bnode_read_u16(node, rec_off); node 538 fs/hfsplus/bnode.c if (node->type != HFS_NODE_INDEX && node 539 fs/hfsplus/bnode.c node->type != HFS_NODE_LEAF) node 541 fs/hfsplus/bnode.c key_size = hfs_bnode_read_u16(node, off) + 2; node 545 fs/hfsplus/bnode.c clear_bit(HFS_BNODE_NEW, &node->flags); node 546 fs/hfsplus/bnode.c wake_up(&node->lock_wq); node 547 fs/hfsplus/bnode.c return node; node 550 fs/hfsplus/bnode.c set_bit(HFS_BNODE_ERROR, &node->flags); node 551 fs/hfsplus/bnode.c clear_bit(HFS_BNODE_NEW, &node->flags); node 552 fs/hfsplus/bnode.c wake_up(&node->lock_wq); node 553 fs/hfsplus/bnode.c hfs_bnode_put(node); node 564 fs/hfsplus/bnode.c kfree(node); node 569 fs/hfsplus/bnode.c struct hfs_bnode *node; node 574 fs/hfsplus/bnode.c node = hfs_bnode_findhash(tree, num); node 576 fs/hfsplus/bnode.c if (node) { node 579 fs/hfsplus/bnode.c return node; node 581 fs/hfsplus/bnode.c node = __hfs_bnode_create(tree, num); node 582 fs/hfsplus/bnode.c if (!node) node 584 fs/hfsplus/bnode.c if (test_bit(HFS_BNODE_ERROR, &node->flags)) { node 585 fs/hfsplus/bnode.c hfs_bnode_put(node); node 589 fs/hfsplus/bnode.c pagep = node->page; node 590 fs/hfsplus/bnode.c memset(kmap(*pagep) + node->page_offset, 0, node 599 fs/hfsplus/bnode.c clear_bit(HFS_BNODE_NEW, &node->flags); node 600 fs/hfsplus/bnode.c wake_up(&node->lock_wq); node 602 fs/hfsplus/bnode.c return node; node 607 fs/hfsplus/bnode.c if (node) { node 608 fs/hfsplus/bnode.c atomic_inc(&node->refcnt); node 610 fs/hfsplus/bnode.c node->tree->cnid, node->this, atomic_read(&node->refcnt)); node 617 fs/hfsplus/bnode.c if (node) { node 618 fs/hfsplus/bnode.c struct hfs_btree *tree = node->tree; node 622 fs/hfsplus/bnode.c node->tree->cnid, node->this, atomic_read(&node->refcnt)); node 623 fs/hfsplus/bnode.c BUG_ON(!atomic_read(&node->refcnt)); node 624 fs/hfsplus/bnode.c if (!atomic_dec_and_lock(&node->refcnt, &tree->hash_lock)) node 627 fs/hfsplus/bnode.c if (!node->page[i]) node 629 fs/hfsplus/bnode.c mark_page_accessed(node->page[i]); node 632 fs/hfsplus/bnode.c if (test_bit(HFS_BNODE_DELETED, &node->flags)) { node 633 fs/hfsplus/bnode.c hfs_bnode_unhash(node); node 635 fs/hfsplus/bnode.c hfs_bmap_free(node); node 636 fs/hfsplus/bnode.c hfs_bnode_free(node); node 24 fs/hfsplus/brec.c dataoff = node->tree->node_size - (rec + 2) * 2; node 25 fs/hfsplus/brec.c hfs_bnode_read(node, retval, dataoff, 4); node 35 fs/hfsplus/brec.c if (node->type != HFS_NODE_INDEX && node->type != HFS_NODE_LEAF) node 38 fs/hfsplus/brec.c if ((node->type == HFS_NODE_INDEX) && node 39 fs/hfsplus/brec.c !(node->tree->attributes & HFS_TREE_VARIDXKEYS)) { node 40 fs/hfsplus/brec.c retval = node->tree->max_key_len + 2; node 42 fs/hfsplus/brec.c recoff = hfs_bnode_read_u16(node, node->tree->node_size - (rec + 1) * 2); node 45 fs/hfsplus/brec.c if (node->tree->attributes & HFS_TREE_BIGKEYS) node 46 fs/hfsplus/brec.c retval = hfs_bnode_read_u16(node, recoff) + 2; node 48 fs/hfsplus/brec.c retval = (hfs_bnode_read_u8(node, recoff) | 1) + 1; node 56 fs/hfsplus/brec.c struct hfs_bnode *node, *new_node; node 78 fs/hfsplus/brec.c node = fd->bnode; node 79 fs/hfsplus/brec.c hfs_bnode_dump(node); node 81 fs/hfsplus/brec.c end_rec_off = tree->node_size - (node->num_recs + 1) * 2; node 82 fs/hfsplus/brec.c end_off = hfs_bnode_read_u16(node, end_rec_off); node 93 fs/hfsplus/brec.c if (node->type == HFS_NODE_LEAF) { node 97 fs/hfsplus/brec.c node->num_recs++; node 99 fs/hfsplus/brec.c hfs_bnode_write_u16(node, offsetof(struct hfs_bnode_desc, num_recs), node->num_recs); node 100 fs/hfsplus/brec.c hfs_bnode_write_u16(node, end_rec_off, end_off + size); node 108 fs/hfsplus/brec.c data_off = hfs_bnode_read_u16(node, data_rec_off + 2); node 109 fs/hfsplus/brec.c hfs_bnode_write_u16(node, data_rec_off, data_off + size); node 114 fs/hfsplus/brec.c hfs_bnode_move(node, data_off + size, data_off, node 118 fs/hfsplus/brec.c hfs_bnode_write(node, fd->search_key, data_off, key_len); node 119 fs/hfsplus/brec.c hfs_bnode_write(node, entry, data_off + key_len, entry_len); node 120 fs/hfsplus/brec.c hfs_bnode_dump(node); node 126 fs/hfsplus/brec.c if (!rec && new_node != node) node 166 fs/hfsplus/brec.c struct hfs_bnode *node, *parent; node 170 fs/hfsplus/brec.c node = fd->bnode; node 173 fs/hfsplus/brec.c end_off = tree->node_size - (node->num_recs + 1) * 2; node 175 fs/hfsplus/brec.c if (node->type == HFS_NODE_LEAF) { node 179 fs/hfsplus/brec.c hfs_bnode_dump(node); node 181 fs/hfsplus/brec.c if (!--node->num_recs) { node 182 fs/hfsplus/brec.c hfs_bnode_unlink(node); node 183 fs/hfsplus/brec.c if (!node->parent) node 185 fs/hfsplus/brec.c parent = hfs_bnode_find(tree, node->parent); node 188 fs/hfsplus/brec.c hfs_bnode_put(node); node 189 fs/hfsplus/brec.c node = fd->bnode = parent; node 191 fs/hfsplus/brec.c __hfs_brec_find(node, fd); node 194 fs/hfsplus/brec.c hfs_bnode_write_u16(node, offsetof(struct hfs_bnode_desc, num_recs), node->num_recs); node 201 fs/hfsplus/brec.c data_off = hfs_bnode_read_u16(node, rec_off); node 202 fs/hfsplus/brec.c hfs_bnode_write_u16(node, rec_off + 2, data_off - size); node 207 fs/hfsplus/brec.c hfs_bnode_move(node, fd->keyoffset, fd->keyoffset + size, node 210 fs/hfsplus/brec.c hfs_bnode_dump(node); node 219 fs/hfsplus/brec.c struct hfs_bnode *node, *new_node; node 225 fs/hfsplus/brec.c node = fd->bnode; node 229 fs/hfsplus/brec.c hfs_bnode_get(node); node 231 fs/hfsplus/brec.c node->this, new_node->this, node->next); node 232 fs/hfsplus/brec.c new_node->next = node->next; node 233 fs/hfsplus/brec.c new_node->prev = node->this; node 234 fs/hfsplus/brec.c new_node->parent = node->parent; node 235 fs/hfsplus/brec.c new_node->type = node->type; node 236 fs/hfsplus/brec.c new_node->height = node->height; node 238 fs/hfsplus/brec.c size = tree->node_size / 2 - node->num_recs * 2 - 14; node 242 fs/hfsplus/brec.c data_start = hfs_bnode_read_u16(node, old_rec_off); node 246 fs/hfsplus/brec.c if (++num_recs < node->num_recs) node 249 fs/hfsplus/brec.c hfs_bnode_put(node); node 260 fs/hfsplus/brec.c data_start = hfs_bnode_read_u16(node, old_rec_off); node 262 fs/hfsplus/brec.c hfs_bnode_put(node); node 269 fs/hfsplus/brec.c new_node->num_recs = node->num_recs - num_recs; node 270 fs/hfsplus/brec.c node->num_recs = num_recs; node 281 fs/hfsplus/brec.c data_end = hfs_bnode_read_u16(node, old_rec_off); node 286 fs/hfsplus/brec.c hfs_bnode_copy(new_node, 14, node, data_start, data_end - data_start); node 298 fs/hfsplus/brec.c node->next = new_node->this; node 299 fs/hfsplus/brec.c hfs_bnode_read(node, &node_desc, 0, sizeof(node_desc)); node 300 fs/hfsplus/brec.c node_desc.next = cpu_to_be32(node->next); node 301 fs/hfsplus/brec.c node_desc.num_recs = cpu_to_be16(node->num_recs); node 302 fs/hfsplus/brec.c hfs_bnode_write(node, &node_desc, 0, sizeof(node_desc)); node 312 fs/hfsplus/brec.c } else if (node->this == tree->leaf_tail) { node 318 fs/hfsplus/brec.c hfs_bnode_dump(node); node 320 fs/hfsplus/brec.c hfs_bnode_put(node); node 328 fs/hfsplus/brec.c struct hfs_bnode *node, *new_node, *parent; node 334 fs/hfsplus/brec.c node = fd->bnode; node 336 fs/hfsplus/brec.c if (!node->parent) node 340 fs/hfsplus/brec.c parent = hfs_bnode_find(tree, node->parent); node 349 fs/hfsplus/brec.c newkeylen = hfs_bnode_read_u16(node, 14) + 2; node 387 fs/hfsplus/brec.c hfs_bnode_copy(parent, fd->keyoffset, node, 14, newkeylen); node 390 fs/hfsplus/brec.c hfs_bnode_put(node); node 391 fs/hfsplus/brec.c node = parent; node 407 fs/hfsplus/brec.c if (new_node == node) node 410 fs/hfsplus/brec.c hfs_bnode_read_key(node, fd->search_key, 14); node 414 fs/hfsplus/brec.c if (!rec && node->parent) node 417 fs/hfsplus/brec.c fd->bnode = node; node 423 fs/hfsplus/brec.c struct hfs_bnode *node, *new_node; node 428 fs/hfsplus/brec.c node = NULL; node 430 fs/hfsplus/brec.c node = hfs_bnode_find(tree, tree->root); node 431 fs/hfsplus/brec.c if (IS_ERR(node)) node 432 fs/hfsplus/brec.c return PTR_ERR(node); node 436 fs/hfsplus/brec.c hfs_bnode_put(node); node 465 fs/hfsplus/brec.c if (node) { node 467 fs/hfsplus/brec.c node->parent = tree->root; node 468 fs/hfsplus/brec.c if (node->type == HFS_NODE_LEAF || node 470 fs/hfsplus/brec.c key_size = hfs_bnode_read_u16(node, 14) + 2; node 473 fs/hfsplus/brec.c hfs_bnode_copy(new_node, 14, node, 14, key_size); node 479 fs/hfsplus/brec.c cnid = cpu_to_be32(node->this); node 485 fs/hfsplus/brec.c hfs_bnode_put(node); node 101 fs/hfsplus/btree.c struct hfs_bnode *node; node 108 fs/hfsplus/btree.c while ((node = tree->node_hash[i])) { node 109 fs/hfsplus/btree.c tree->node_hash[i] = node->next_hash; node 110 fs/hfsplus/btree.c if (atomic_read(&node->refcnt)) node 112 fs/hfsplus/btree.c node->tree->cnid, node->this, atomic_read(&node->refcnt)); node 113 fs/hfsplus/btree.c hfs_bnode_free(node); node 124 fs/hfsplus/btree.c struct hfs_bnode *node; node 127 fs/hfsplus/btree.c node = hfs_bnode_find(tree, 0); node 128 fs/hfsplus/btree.c if (IS_ERR(node)) node 132 fs/hfsplus/btree.c page = node->page[0]; node 146 fs/hfsplus/btree.c hfs_bnode_put(node); node 152 fs/hfsplus/btree.c struct hfs_bnode *node; node 156 fs/hfsplus/btree.c node = hfs_bnode_create(tree, idx); node 157 fs/hfsplus/btree.c if (IS_ERR(node)) node 158 fs/hfsplus/btree.c return node; node 165 fs/hfsplus/btree.c node->type = HFS_NODE_MAP; node 166 fs/hfsplus/btree.c node->num_recs = 1; node 167 fs/hfsplus/btree.c hfs_bnode_clear(node, 0, tree->node_size); node 174 fs/hfsplus/btree.c hfs_bnode_write(node, &desc, 0, sizeof(desc)); node 175 fs/hfsplus/btree.c hfs_bnode_write_u16(node, 14, 0x8000); node 176 fs/hfsplus/btree.c hfs_bnode_write_u16(node, tree->node_size - 2, 14); node 177 fs/hfsplus/btree.c hfs_bnode_write_u16(node, tree->node_size - 4, tree->node_size - 6); node 179 fs/hfsplus/btree.c return node; node 184 fs/hfsplus/btree.c struct hfs_bnode *node, *next_node; node 213 fs/hfsplus/btree.c node = hfs_bnode_find(tree, nidx); node 214 fs/hfsplus/btree.c if (IS_ERR(node)) node 215 fs/hfsplus/btree.c return node; node 216 fs/hfsplus/btree.c len = hfs_brec_lenoff(node, 2, &off16); node 219 fs/hfsplus/btree.c off += node->page_offset; node 220 fs/hfsplus/btree.c pagep = node->page + (off >> PAGE_CACHE_SHIFT); node 237 fs/hfsplus/btree.c hfs_bnode_put(node); node 251 fs/hfsplus/btree.c nidx = node->next; node 254 fs/hfsplus/btree.c next_node = hfs_bmap_new_bmap(node, idx); node 257 fs/hfsplus/btree.c hfs_bnode_put(node); node 260 fs/hfsplus/btree.c node = next_node; node 262 fs/hfsplus/btree.c len = hfs_brec_lenoff(node, 0, &off16); node 264 fs/hfsplus/btree.c off += node->page_offset; node 265 fs/hfsplus/btree.c pagep = node->page + (off >> PAGE_CACHE_SHIFT); node 279 fs/hfsplus/btree.c dprint(DBG_BNODE_MOD, "btree_free_node: %u\n", node->this); node 280 fs/hfsplus/btree.c BUG_ON(!node->this); node 281 fs/hfsplus/btree.c tree = node->tree; node 282 fs/hfsplus/btree.c nidx = node->this; node 283 fs/hfsplus/btree.c node = hfs_bnode_find(tree, 0); node 284 fs/hfsplus/btree.c if (IS_ERR(node)) node 286 fs/hfsplus/btree.c len = hfs_brec_lenoff(node, 2, &off); node 291 fs/hfsplus/btree.c i = node->next; node 292 fs/hfsplus/btree.c hfs_bnode_put(node); node 295 fs/hfsplus/btree.c printk(KERN_CRIT "hfs: unable to free bnode %u. bmap not found!\n", node->this); node 298 fs/hfsplus/btree.c node = hfs_bnode_find(tree, i); node 299 fs/hfsplus/btree.c if (IS_ERR(node)) node 301 fs/hfsplus/btree.c if (node->type != HFS_NODE_MAP) { node 303 fs/hfsplus/btree.c printk(KERN_CRIT "hfs: invalid bmap found! (%u,%d)\n", node->this, node->type); node 304 fs/hfsplus/btree.c hfs_bnode_put(node); node 307 fs/hfsplus/btree.c len = hfs_brec_lenoff(node, 0, &off); node 309 fs/hfsplus/btree.c off += node->page_offset + nidx / 8; node 310 fs/hfsplus/btree.c page = node->page[off >> PAGE_CACHE_SHIFT]; node 316 fs/hfsplus/btree.c printk(KERN_CRIT "hfs: trying to free free bnode %u(%d)\n", node->this, node->type); node 318 fs/hfsplus/btree.c hfs_bnode_put(node); node 324 fs/hfsplus/btree.c hfs_bnode_put(node); node 50 fs/hfsplus/inode.c struct hfs_bnode *node; node 73 fs/hfsplus/inode.c node = hfs_bnode_findhash(tree, nidx); node 74 fs/hfsplus/inode.c if (!node) node 76 fs/hfsplus/inode.c else if (atomic_read(&node->refcnt)) node 78 fs/hfsplus/inode.c if (res && node) { node 79 fs/hfsplus/inode.c hfs_bnode_unhash(node); node 80 fs/hfsplus/inode.c hfs_bnode_free(node); node 88 fs/hfsplus/inode.c node = hfs_bnode_findhash(tree, nidx++); node 89 fs/hfsplus/inode.c if (!node) node 91 fs/hfsplus/inode.c if (atomic_read(&node->refcnt)) { node 95 fs/hfsplus/inode.c hfs_bnode_unhash(node); node 96 fs/hfsplus/inode.c hfs_bnode_free(node); node 72 fs/hpfs/anode.c if (!(fnode = hpfs_map_fnode(s, node, &bh))) return -1; node 75 fs/hpfs/anode.c if (!(anode = hpfs_map_anode(s, node, &bh))) return -1; node 78 fs/hpfs/anode.c a = node; node 100 fs/hpfs/anode.c fnod?'f':'a', node); node 112 fs/hpfs/anode.c hpfs_error(s, "empty file %08x, trying to add sector %08x", node, fsecno); node 116 fs/hpfs/anode.c se = !fnod ? node : (node + 16384) & ~16383; node 124 fs/hpfs/anode.c up = a != node ? anode->up : -1; node 130 fs/hpfs/anode.c if (a == node && fnod) { node 131 fs/hpfs/anode.c anode->up = node; node 162 fs/hpfs/anode.c if ((a == node && fnod) || na == -1) return se; node 168 fs/hpfs/anode.c if (up != node || !fnod) { node 187 fs/hpfs/anode.c anode->btree.fnode_parent = up == node && fnod; node 193 fs/hpfs/anode.c up = up != node ? anode->up : -1; node 217 fs/hpfs/anode.c anode->up = node; node 223 fs/hpfs/anode.c if (!(anode = hpfs_map_anode(s, node, &bh))) { node 229 fs/hpfs/anode.c if (!(fnode = hpfs_map_fnode(s, node, &bh))) { node 235 fs/hpfs/anode.c ranode->up = node; node 399 fs/hpfs/anode.c anode_secno node = f; node 426 fs/hpfs/anode.c hpfs_error(s, "internal btree %08x doesn't end with -1", node); node 439 fs/hpfs/anode.c node = btree->u.internal[i].down; node 442 fs/hpfs/anode.c if (hpfs_stop_cycles(s, node, &c1, &c2, "hpfs_truncate_btree")) node 444 fs/hpfs/anode.c if (!(anode = hpfs_map_anode(s, node, &bh))) return; node 495 fs/inode.c struct hlist_node *node; node 499 fs/inode.c hlist_for_each_entry(inode, node, head, i_hash) { node 510 fs/inode.c return node ? inode : NULL; node 519 fs/inode.c struct hlist_node *node; node 523 fs/inode.c hlist_for_each_entry(inode, node, head, i_hash) { node 534 fs/inode.c return node ? inode : NULL; node 77 fs/jffs2/debug.c struct jffs2_full_dnode *fn = frag->node; node 95 fs/jffs2/debug.c && frag_prev(frag)->size < PAGE_CACHE_SIZE && frag_prev(frag)->node) { node 102 fs/jffs2/debug.c && frag_next(frag)->size < PAGE_CACHE_SIZE && frag_next(frag)->node) { node 703 fs/jffs2/debug.c if (this->node) node 705 fs/jffs2/debug.c this->ofs, this->ofs+this->size, ref_offset(this->node->raw), node 706 fs/jffs2/debug.c ref_flags(this->node->raw), this, frag_left(this), frag_right(this), node 767 fs/jffs2/debug.c union jffs2_node_union node; node 775 fs/jffs2/debug.c ret = jffs2_flash_read(c, ofs, len, &retlen, (unsigned char *)&node); node 782 fs/jffs2/debug.c printk(JFFS2_DBG "magic:\t%#04x\n", je16_to_cpu(node.u.magic)); node 783 fs/jffs2/debug.c printk(JFFS2_DBG "nodetype:\t%#04x\n", je16_to_cpu(node.u.nodetype)); node 784 fs/jffs2/debug.c printk(JFFS2_DBG "totlen:\t%#08x\n", je32_to_cpu(node.u.totlen)); node 785 fs/jffs2/debug.c printk(JFFS2_DBG "hdr_crc:\t%#08x\n", je32_to_cpu(node.u.hdr_crc)); node 787 fs/jffs2/debug.c crc = crc32(0, &node.u, sizeof(node.u) - 4); node 788 fs/jffs2/debug.c if (crc != je32_to_cpu(node.u.hdr_crc)) { node 793 fs/jffs2/debug.c if (je16_to_cpu(node.u.magic) != JFFS2_MAGIC_BITMASK && node 794 fs/jffs2/debug.c je16_to_cpu(node.u.magic) != JFFS2_OLD_MAGIC_BITMASK) node 797 fs/jffs2/debug.c je16_to_cpu(node.u.magic), JFFS2_MAGIC_BITMASK); node 801 fs/jffs2/debug.c switch(je16_to_cpu(node.u.nodetype)) { node 806 fs/jffs2/debug.c printk(JFFS2_DBG "ino:\t%#08x\n", je32_to_cpu(node.i.ino)); node 807 fs/jffs2/debug.c printk(JFFS2_DBG "version:\t%#08x\n", je32_to_cpu(node.i.version)); node 808 fs/jffs2/debug.c printk(JFFS2_DBG "mode:\t%#08x\n", node.i.mode.m); node 809 fs/jffs2/debug.c printk(JFFS2_DBG "uid:\t%#04x\n", je16_to_cpu(node.i.uid)); node 810 fs/jffs2/debug.c printk(JFFS2_DBG "gid:\t%#04x\n", je16_to_cpu(node.i.gid)); node 811 fs/jffs2/debug.c printk(JFFS2_DBG "isize:\t%#08x\n", je32_to_cpu(node.i.isize)); node 812 fs/jffs2/debug.c printk(JFFS2_DBG "atime:\t%#08x\n", je32_to_cpu(node.i.atime)); node 813 fs/jffs2/debug.c printk(JFFS2_DBG "mtime:\t%#08x\n", je32_to_cpu(node.i.mtime)); node 814 fs/jffs2/debug.c printk(JFFS2_DBG "ctime:\t%#08x\n", je32_to_cpu(node.i.ctime)); node 815 fs/jffs2/debug.c printk(JFFS2_DBG "offset:\t%#08x\n", je32_to_cpu(node.i.offset)); node 816 fs/jffs2/debug.c printk(JFFS2_DBG "csize:\t%#08x\n", je32_to_cpu(node.i.csize)); node 817 fs/jffs2/debug.c printk(JFFS2_DBG "dsize:\t%#08x\n", je32_to_cpu(node.i.dsize)); node 818 fs/jffs2/debug.c printk(JFFS2_DBG "compr:\t%#02x\n", node.i.compr); node 819 fs/jffs2/debug.c printk(JFFS2_DBG "usercompr:\t%#02x\n", node.i.usercompr); node 820 fs/jffs2/debug.c printk(JFFS2_DBG "flags:\t%#04x\n", je16_to_cpu(node.i.flags)); node 821 fs/jffs2/debug.c printk(JFFS2_DBG "data_crc:\t%#08x\n", je32_to_cpu(node.i.data_crc)); node 822 fs/jffs2/debug.c printk(JFFS2_DBG "node_crc:\t%#08x\n", je32_to_cpu(node.i.node_crc)); node 824 fs/jffs2/debug.c crc = crc32(0, &node.i, sizeof(node.i) - 8); node 825 fs/jffs2/debug.c if (crc != je32_to_cpu(node.i.node_crc)) { node 834 fs/jffs2/debug.c printk(JFFS2_DBG "pino:\t%#08x\n", je32_to_cpu(node.d.pino)); node 835 fs/jffs2/debug.c printk(JFFS2_DBG "version:\t%#08x\n", je32_to_cpu(node.d.version)); node 836 fs/jffs2/debug.c printk(JFFS2_DBG "ino:\t%#08x\n", je32_to_cpu(node.d.ino)); node 837 fs/jffs2/debug.c printk(JFFS2_DBG "mctime:\t%#08x\n", je32_to_cpu(node.d.mctime)); node 838 fs/jffs2/debug.c printk(JFFS2_DBG "nsize:\t%#02x\n", node.d.nsize); node 839 fs/jffs2/debug.c printk(JFFS2_DBG "type:\t%#02x\n", node.d.type); node 840 fs/jffs2/debug.c printk(JFFS2_DBG "node_crc:\t%#08x\n", je32_to_cpu(node.d.node_crc)); node 841 fs/jffs2/debug.c printk(JFFS2_DBG "name_crc:\t%#08x\n", je32_to_cpu(node.d.name_crc)); node 843 fs/jffs2/debug.c node.d.name[node.d.nsize] = '\0'; node 844 fs/jffs2/debug.c printk(JFFS2_DBG "name:\t\"%s\"\n", node.d.name); node 846 fs/jffs2/debug.c crc = crc32(0, &node.d, sizeof(node.d) - 8); node 847 fs/jffs2/debug.c if (crc != je32_to_cpu(node.d.node_crc)) { node 483 fs/jffs2/gc.c if (frag->node && frag->node->raw == raw) { node 484 fs/jffs2/gc.c fn = frag->node; node 488 fs/jffs2/gc.c if (nrfrags == frag->node->frags) node 497 fs/jffs2/gc.c frag->node->raw = f->inocache->nodes; node 543 fs/jffs2/gc.c union jffs2_node_union *node; node 571 fs/jffs2/gc.c node = kmalloc(rawlen, GFP_KERNEL); node 572 fs/jffs2/gc.c if (!node) node 575 fs/jffs2/gc.c ret = jffs2_flash_read(c, ref_offset(raw), rawlen, &retlen, (char *)node); node 581 fs/jffs2/gc.c crc = crc32(0, node, sizeof(struct jffs2_unknown_node)-4); node 582 fs/jffs2/gc.c if (je32_to_cpu(node->u.hdr_crc) != crc) { node 584 fs/jffs2/gc.c ref_offset(raw), je32_to_cpu(node->u.hdr_crc), crc); node 588 fs/jffs2/gc.c switch(je16_to_cpu(node->u.nodetype)) { node 590 fs/jffs2/gc.c crc = crc32(0, node, sizeof(node->i)-8); node 591 fs/jffs2/gc.c if (je32_to_cpu(node->i.node_crc) != crc) { node 593 fs/jffs2/gc.c ref_offset(raw), je32_to_cpu(node->i.node_crc), crc); node 597 fs/jffs2/gc.c if (je32_to_cpu(node->i.dsize)) { node 598 fs/jffs2/gc.c crc = crc32(0, node->i.data, je32_to_cpu(node->i.csize)); node 599 fs/jffs2/gc.c if (je32_to_cpu(node->i.data_crc) != crc) { node 601 fs/jffs2/gc.c ref_offset(raw), je32_to_cpu(node->i.data_crc), crc); node 608 fs/jffs2/gc.c crc = crc32(0, node, sizeof(node->d)-8); node 609 fs/jffs2/gc.c if (je32_to_cpu(node->d.node_crc) != crc) { node 611 fs/jffs2/gc.c ref_offset(raw), je32_to_cpu(node->d.node_crc), crc); node 615 fs/jffs2/gc.c if (strnlen(node->d.name, node->d.nsize) != node->d.nsize) { node 620 fs/jffs2/gc.c if (node->d.nsize) { node 621 fs/jffs2/gc.c crc = crc32(0, node->d.name, node->d.nsize); node 622 fs/jffs2/gc.c if (je32_to_cpu(node->d.name_crc) != crc) { node 624 fs/jffs2/gc.c ref_offset(raw), je32_to_cpu(node->d.name_crc), crc); node 633 fs/jffs2/gc.c ref_offset(raw), je16_to_cpu(node->u.nodetype)); node 642 fs/jffs2/gc.c ret = jffs2_flash_write(c, phys_ofs, rawlen, &retlen, (char *)node); node 689 fs/jffs2/gc.c kfree(node); node 1069 fs/jffs2/gc.c if (frag->node == fn) { node 1070 fs/jffs2/gc.c frag->node = new_fn; node 1142 fs/jffs2/gc.c if (!frag->node || !frag->node->raw) { node 1152 fs/jffs2/gc.c struct jffs2_raw_node_ref *raw = frag->node->raw; node 1192 fs/jffs2/gc.c if (!frag->node || !frag->node->raw) { node 1202 fs/jffs2/gc.c struct jffs2_raw_node_ref *raw = frag->node->raw; node 292 fs/jffs2/malloc.c xd->node = (void *)xd; node 311 fs/jffs2/malloc.c ref->node = (void *)ref; node 92 fs/jffs2/nodelist.c if (frag->node && (frag->ofs & (PAGE_CACHE_SIZE - 1)) == 0) { node 95 fs/jffs2/nodelist.c frag->node->raw->flash_offset = ref_offset(frag->node->raw) | REF_PRISTINE; node 103 fs/jffs2/nodelist.c if (this->node) { node 104 fs/jffs2/nodelist.c this->node->frags--; node 105 fs/jffs2/nodelist.c if (!this->node->frags) { node 108 fs/jffs2/nodelist.c ref_offset(this->node->raw), this->node->ofs, this->node->ofs+this->node->size); node 109 fs/jffs2/nodelist.c jffs2_mark_node_obsolete(c, this->node->raw); node 110 fs/jffs2/nodelist.c jffs2_free_full_dnode(this->node); node 113 fs/jffs2/nodelist.c ref_offset(this->node->raw), this->node->ofs, this->node->ofs+this->node->size, this->node->frags); node 114 fs/jffs2/nodelist.c mark_ref_normal(this->node->raw); node 156 fs/jffs2/nodelist.c newfrag->node = fn; node 172 fs/jffs2/nodelist.c if (lastend < newfrag->node->ofs) { node 176 fs/jffs2/nodelist.c holefrag= new_fragment(NULL, lastend, newfrag->node->ofs - lastend); node 220 fs/jffs2/nodelist.c this = jffs2_lookup_node_frag(root, newfrag->node->ofs); node 224 fs/jffs2/nodelist.c this->ofs, this->ofs+this->size, this->node?(ref_offset(this->node->raw)):0xffffffff, this); node 240 fs/jffs2/nodelist.c if (this->node) node 241 fs/jffs2/nodelist.c mark_ref_normal(this->node->raw); node 242 fs/jffs2/nodelist.c mark_ref_normal(newfrag->node->raw); node 248 fs/jffs2/nodelist.c if (this->node) node 251 fs/jffs2/nodelist.c ref_offset(this->node->raw), ref_flags(this->node->raw)); node 264 fs/jffs2/nodelist.c mark_ref_normal(newfrag->node->raw); node 265 fs/jffs2/nodelist.c if (this->node) node 266 fs/jffs2/nodelist.c mark_ref_normal(this->node->raw); node 272 fs/jffs2/nodelist.c if (this->node) node 274 fs/jffs2/nodelist.c this->ofs, this->ofs+this->size, ref_offset(this->node->raw)); node 280 fs/jffs2/nodelist.c newfrag2 = new_fragment(this->node, newfrag->ofs + newfrag->size, node 284 fs/jffs2/nodelist.c if (this->node) node 285 fs/jffs2/nodelist.c this->node->frags++; node 351 fs/jffs2/nodelist.c if (this->node) node 352 fs/jffs2/nodelist.c mark_ref_normal(this->node->raw); node 353 fs/jffs2/nodelist.c mark_ref_normal(newfrag->node->raw); node 373 fs/jffs2/nodelist.c newfrag->node->frags = 1; node 389 fs/jffs2/nodelist.c if (prev->node) node 390 fs/jffs2/nodelist.c mark_ref_normal(prev->node->raw); node 398 fs/jffs2/nodelist.c if (next->node) node 399 fs/jffs2/nodelist.c mark_ref_normal(next->node->raw); node 585 fs/jffs2/nodelist.c if (frag->node && !(--frag->node->frags)) { node 589 fs/jffs2/nodelist.c jffs2_mark_node_obsolete(c, frag->node->raw); node 591 fs/jffs2/nodelist.c jffs2_free_full_dnode(frag->node); node 268 fs/jffs2/nodelist.h struct jffs2_full_dnode *node; /* NULL for holes */ node 325 fs/jffs2/nodelist.h struct rb_node *node = rb_first(root); node 327 fs/jffs2/nodelist.h if (!node) node 330 fs/jffs2/nodelist.h return rb_entry(node, struct jffs2_node_frag, rb); node 335 fs/jffs2/nodelist.h struct rb_node *node = rb_last(root); node 337 fs/jffs2/nodelist.h if (!node) node 340 fs/jffs2/nodelist.h return rb_entry(node, struct jffs2_node_frag, rb); node 181 fs/jffs2/read.c } else if (unlikely(!frag->node)) { node 197 fs/jffs2/read.c ref_offset(frag->node->raw), ref_flags(frag->node->raw))); node 198 fs/jffs2/read.c ret = jffs2_read_dnode(c, f, frag->node, buf, fragofs + frag->ofs - frag->node->ofs, readlen); node 384 fs/jffs2/readinode.c struct rb_node *parent = rb_parent(node); node 388 fs/jffs2/readinode.c BUG_ON(node->rb_right); node 392 fs/jffs2/readinode.c else if (node == parent->rb_left) node 397 fs/jffs2/readinode.c *link = node->rb_left; node 399 fs/jffs2/readinode.c if (node->rb_left) node 400 fs/jffs2/readinode.c node->rb_left->rb_parent_color = node->rb_parent_color; node 959 fs/jffs2/readinode.c union jffs2_node_union *node; node 1030 fs/jffs2/readinode.c node = (union jffs2_node_union *)buf; node 1033 fs/jffs2/readinode.c if (je32_to_cpu(node->u.hdr_crc) != crc32(0, node, sizeof(node->u)-4)) { node 1035 fs/jffs2/readinode.c ref_offset(ref), je16_to_cpu(node->u.magic), node 1036 fs/jffs2/readinode.c je16_to_cpu(node->u.nodetype), node 1037 fs/jffs2/readinode.c je32_to_cpu(node->u.totlen), node 1038 fs/jffs2/readinode.c je32_to_cpu(node->u.hdr_crc)); node 1043 fs/jffs2/readinode.c if (je16_to_cpu(node->u.magic) != JFFS2_MAGIC_BITMASK) { node 1046 fs/jffs2/readinode.c je16_to_cpu(node->u.magic), ref_offset(ref)); node 1051 fs/jffs2/readinode.c switch (je16_to_cpu(node->u.nodetype)) { node 1062 fs/jffs2/readinode.c err = read_direntry(c, ref, &node->d, retlen, rii); node 1077 fs/jffs2/readinode.c err = read_dnode(c, ref, &node->i, len, rii); node 1091 fs/jffs2/readinode.c err = read_unknown(c, ref, &node->u); node 1300 fs/jffs2/readinode.c f->metadata = frag_first(&f->fragtree)->node; node 354 fs/jffs2/scan.c raw->next_in_ino = xd->node->next_in_ino; node 355 fs/jffs2/scan.c xd->node->next_in_ino = raw; node 434 fs/jffs2/scan.c struct jffs2_unknown_node *node; node 617 fs/jffs2/scan.c if (jeb->offset + c->sector_size < ofs + sizeof(*node)) { node 619 fs/jffs2/scan.c jeb->offset, c->sector_size, ofs, sizeof(*node))); node 625 fs/jffs2/scan.c if (buf_ofs + buf_len < ofs + sizeof(*node)) { node 635 fs/jffs2/scan.c node = (struct jffs2_unknown_node *)&buf[ofs-buf_ofs]; node 694 fs/jffs2/scan.c if (ofs == jeb->offset && je16_to_cpu(node->magic) == KSAMTIB_CIGAM_2SFFJ) { node 701 fs/jffs2/scan.c if (je16_to_cpu(node->magic) == JFFS2_DIRTY_BITMASK) { node 708 fs/jffs2/scan.c if (je16_to_cpu(node->magic) == JFFS2_OLD_MAGIC_BITMASK) { node 716 fs/jffs2/scan.c if (je16_to_cpu(node->magic) != JFFS2_MAGIC_BITMASK) { node 720 fs/jffs2/scan.c je16_to_cpu(node->magic)); node 727 fs/jffs2/scan.c crcnode.magic = node->magic; node 728 fs/jffs2/scan.c crcnode.nodetype = cpu_to_je16( je16_to_cpu(node->nodetype) | JFFS2_NODE_ACCURATE); node 729 fs/jffs2/scan.c crcnode.totlen = node->totlen; node 732 fs/jffs2/scan.c if (hdr_crc != je32_to_cpu(node->hdr_crc)) { node 734 fs/jffs2/scan.c ofs, je16_to_cpu(node->magic), node 735 fs/jffs2/scan.c je16_to_cpu(node->nodetype), node 736 fs/jffs2/scan.c je32_to_cpu(node->totlen), node 737 fs/jffs2/scan.c je32_to_cpu(node->hdr_crc), node 745 fs/jffs2/scan.c if (ofs + je32_to_cpu(node->totlen) > jeb->offset + c->sector_size) { node 748 fs/jffs2/scan.c ofs, je32_to_cpu(node->totlen)); node 756 fs/jffs2/scan.c if (!(je16_to_cpu(node->nodetype) & JFFS2_NODE_ACCURATE)) { node 759 fs/jffs2/scan.c if ((err = jffs2_scan_dirty_space(c, jeb, PAD(je32_to_cpu(node->totlen))))) node 761 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen)); node 765 fs/jffs2/scan.c switch(je16_to_cpu(node->nodetype)) { node 775 fs/jffs2/scan.c node = (void *)buf; node 777 fs/jffs2/scan.c err = jffs2_scan_inode_node(c, jeb, (void *)node, ofs, s); node 779 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen)); node 783 fs/jffs2/scan.c if (buf_ofs + buf_len < ofs + je32_to_cpu(node->totlen)) { node 786 fs/jffs2/scan.c je32_to_cpu(node->totlen), buf_len, ofs)); node 791 fs/jffs2/scan.c node = (void *)buf; node 793 fs/jffs2/scan.c err = jffs2_scan_dirent_node(c, jeb, (void *)node, ofs, s); node 795 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen)); node 800 fs/jffs2/scan.c if (buf_ofs + buf_len < ofs + je32_to_cpu(node->totlen)) { node 804 fs/jffs2/scan.c je32_to_cpu(node->totlen), buf_len, ofs)); node 809 fs/jffs2/scan.c node = (void *)buf; node 811 fs/jffs2/scan.c err = jffs2_scan_xattr_node(c, jeb, (void *)node, ofs, s); node 814 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen)); node 817 fs/jffs2/scan.c if (buf_ofs + buf_len < ofs + je32_to_cpu(node->totlen)) { node 821 fs/jffs2/scan.c je32_to_cpu(node->totlen), buf_len, ofs)); node 826 fs/jffs2/scan.c node = (void *)buf; node 828 fs/jffs2/scan.c err = jffs2_scan_xref_node(c, jeb, (void *)node, ofs, s); node 831 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen)); node 837 fs/jffs2/scan.c if (je32_to_cpu(node->totlen) != c->cleanmarker_size) { node 839 fs/jffs2/scan.c ofs, je32_to_cpu(node->totlen), c->cleanmarker_size); node 857 fs/jffs2/scan.c jffs2_sum_add_padding_mem(s, je32_to_cpu(node->totlen)); node 858 fs/jffs2/scan.c if ((err = jffs2_scan_dirty_space(c, jeb, PAD(je32_to_cpu(node->totlen))))) node 860 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen)); node 864 fs/jffs2/scan.c switch (je16_to_cpu(node->nodetype) & JFFS2_COMPAT_MASK) { node 866 fs/jffs2/scan.c printk(KERN_NOTICE "Read-only compatible feature node (0x%04x) found at offset 0x%08x\n", je16_to_cpu(node->nodetype), ofs); node 870 fs/jffs2/scan.c if ((err = jffs2_scan_dirty_space(c, jeb, PAD(je32_to_cpu(node->totlen))))) node 872 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen)); node 876 fs/jffs2/scan.c printk(KERN_NOTICE "Incompatible feature node (0x%04x) found at offset 0x%08x\n", je16_to_cpu(node->nodetype), ofs); node 880 fs/jffs2/scan.c D1(printk(KERN_NOTICE "Unknown but compatible feature node (0x%04x) found at offset 0x%08x\n", je16_to_cpu(node->nodetype), ofs)); node 881 fs/jffs2/scan.c if ((err = jffs2_scan_dirty_space(c, jeb, PAD(je32_to_cpu(node->totlen))))) node 883 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen)); node 887 fs/jffs2/scan.c D1(printk(KERN_NOTICE "Unknown but compatible feature node (0x%04x) found at offset 0x%08x\n", je16_to_cpu(node->nodetype), ofs)); node 889 fs/jffs2/scan.c jffs2_link_node_ref(c, jeb, ofs | REF_PRISTINE, PAD(je32_to_cpu(node->totlen)), NULL); node 893 fs/jffs2/scan.c ofs += PAD(je32_to_cpu(node->totlen)); node 249 fs/jffs2/summary.c union jffs2_node_union *node; node 257 fs/jffs2/summary.c node = invecs[0].iov_base; node 261 fs/jffs2/summary.c switch (je16_to_cpu(node->u.nodetype)) { node 269 fs/jffs2/summary.c temp->nodetype = node->i.nodetype; node 270 fs/jffs2/summary.c temp->inode = node->i.ino; node 271 fs/jffs2/summary.c temp->version = node->i.version; node 273 fs/jffs2/summary.c temp->totlen = node->i.totlen; node 281 fs/jffs2/summary.c kmalloc(sizeof(struct jffs2_sum_dirent_mem) + node->d.nsize, GFP_KERNEL); node 286 fs/jffs2/summary.c temp->nodetype = node->d.nodetype; node 287 fs/jffs2/summary.c temp->totlen = node->d.totlen; node 289 fs/jffs2/summary.c temp->pino = node->d.pino; node 290 fs/jffs2/summary.c temp->version = node->d.version; node 291 fs/jffs2/summary.c temp->ino = node->d.ino; node 292 fs/jffs2/summary.c temp->nsize = node->d.nsize; node 293 fs/jffs2/summary.c temp->type = node->d.type; node 298 fs/jffs2/summary.c memcpy(temp->name,node->d.name,node->d.nsize); node 302 fs/jffs2/summary.c memcpy(temp->name,invecs[1].iov_base,node->d.nsize); node 319 fs/jffs2/summary.c temp->nodetype = node->x.nodetype; node 320 fs/jffs2/summary.c temp->xid = node->x.xid; node 321 fs/jffs2/summary.c temp->version = node->x.version; node 322 fs/jffs2/summary.c temp->totlen = node->x.totlen; node 333 fs/jffs2/summary.c temp->nodetype = node->r.nodetype; node 342 fs/jffs2/summary.c c->summary->sum_padded += je32_to_cpu(node->u.totlen); node 505 fs/jffs2/summary.c raw->next_in_ino = xd->node->next_in_ino; node 506 fs/jffs2/summary.c xd->node->next_in_ino = raw; node 538 fs/jffs2/summary.c *pseudo_random += ref->node->flash_offset; node 185 fs/jffs2/wbuf.c node, je16_to_cpu(node->u.magic), je16_to_cpu(node->u.nodetype)); node 187 fs/jffs2/wbuf.c BUG_ON(je16_to_cpu(node->u.magic) != 0x1985 && node 188 fs/jffs2/wbuf.c je16_to_cpu(node->u.magic) != 0); node 190 fs/jffs2/wbuf.c switch (je16_to_cpu(node->u.nodetype)) { node 196 fs/jffs2/wbuf.c frag = jffs2_lookup_node_frag(&f->fragtree, je32_to_cpu(node->i.offset)); node 199 fs/jffs2/wbuf.c while (!frag->node || frag->node->raw != raw) { node 203 fs/jffs2/wbuf.c dbg_noderef("Will replace ->raw in full_dnode at %p\n", frag->node); node 204 fs/jffs2/wbuf.c return &frag->node->raw; node 217 fs/jffs2/wbuf.c je16_to_cpu(node->u.nodetype)); node 470 fs/jffs2/wbuf.c BUG_ON(xd->node != raw); node 471 fs/jffs2/wbuf.c adjust_ref = &xd->node; node 476 fs/jffs2/wbuf.c BUG_ON(xr->node != raw); node 477 fs/jffs2/wbuf.c adjust_ref = &xr->node; node 72 fs/jffs2/xattr.c for (raw=xd->node; raw != (void *)xd; raw=raw->next_in_ino) { node 138 fs/jffs2/xattr.c offset = ref_offset(xd->node); node 139 fs/jffs2/xattr.c if (ref_flags(xd->node) == REF_PRISTINE) node 179 fs/jffs2/xattr.c for (raw=xd->node; raw != (void *)xd; raw=raw->next_in_ino) { node 186 fs/jffs2/xattr.c raw->flash_offset = ref_offset(raw) | ((xd->node==raw) ? REF_PRISTINE : REF_NORMAL); node 207 fs/jffs2/xattr.c BUG_ON(ref_flags(xd->node) != REF_PRISTINE); node 215 fs/jffs2/xattr.c ret = jffs2_flash_read(c, ref_offset(xd->node)+sizeof(struct jffs2_raw_xattr), node 220 fs/jffs2/xattr.c ret, length, readlen, ref_offset(xd->node)); node 230 fs/jffs2/xattr.c ref_offset(xd->node), xd->data_crc, crc); node 405 fs/jffs2/xattr.c if (xd->node == (void *)xd) { node 447 fs/jffs2/xattr.c if (ref_flags(ref->node) != REF_UNCHECKED) node 449 fs/jffs2/xattr.c offset = ref_offset(ref->node); node 483 fs/jffs2/xattr.c for (raw=ref->node; raw != (void *)ref; raw=raw->next_in_ino) { node 490 fs/jffs2/xattr.c raw->flash_offset = ref_offset(raw) | ((ref->node==raw) ? REF_PRISTINE : REF_NORMAL); node 495 fs/jffs2/xattr.c ref->ino, ref->xid, ref_offset(ref->node)); node 783 fs/jffs2/xattr.c if (ref_flags(ref->node) != REF_PRISTINE) { node 785 fs/jffs2/xattr.c BUG_ON(ref->node->next_in_ino != (void *)ref); node 786 fs/jffs2/xattr.c ref->node->next_in_ino = NULL; node 787 fs/jffs2/xattr.c jffs2_mark_node_obsolete(c, ref->node); node 799 fs/jffs2/xattr.c raw = ref->node; node 802 fs/jffs2/xattr.c raw->next_in_ino = tmp->node; node 803 fs/jffs2/xattr.c tmp->node = raw; node 805 fs/jffs2/xattr.c raw->next_in_ino = tmp->node->next_in_ino; node 806 fs/jffs2/xattr.c tmp->node->next_in_ino = raw; node 1206 fs/jffs2/xattr.c if (xd->node != raw) node 1216 fs/jffs2/xattr.c old_ofs = ref_offset(xd->node); node 1227 fs/jffs2/xattr.c xd->xid, xd->version, old_ofs, ref_offset(xd->node)); node 1242 fs/jffs2/xattr.c BUG_ON(!ref->node); node 1244 fs/jffs2/xattr.c if (ref->node != raw) node 1249 fs/jffs2/xattr.c old_ofs = ref_offset(ref->node); node 1250 fs/jffs2/xattr.c totlen = ref_totlen(c, c->gcblock, ref->node); node 1262 fs/jffs2/xattr.c ref->ic->ino, ref->xd->xid, old_ofs, ref_offset(ref->node)); node 1285 fs/jffs2/xattr.c for (raw=xd->node; raw != (void *)xd; raw=raw->next_in_ino) { node 1293 fs/jffs2/xattr.c | ((xd->node == (void *)raw) ? REF_PRISTINE : REF_NORMAL); node 1306 fs/jffs2/xattr.c if (atomic_read(&xd->refcnt) || xd->node != (void *)xd) node 1318 fs/jffs2/xattr.c if (ref->node != (void *)ref) node 26 fs/jffs2/xattr.h struct jffs2_raw_node_ref *node; node 48 fs/jffs2/xattr.h struct jffs2_raw_node_ref *node; node 423 fs/ocfs2/alloc.c struct ocfs2_path_item *node; node 429 fs/ocfs2/alloc.c node = &path->p_node[i]; node 431 fs/ocfs2/alloc.c brelse(node->bh); node 432 fs/ocfs2/alloc.c node->bh = NULL; node 433 fs/ocfs2/alloc.c node->el = NULL; node 450 fs/ocfs2/cluster/heartbeat.c (f->hc_func)(node, idx, f->hc_data); node 509 fs/ocfs2/cluster/heartbeat.c event->hn_node = node; node 522 fs/ocfs2/cluster/heartbeat.c struct o2nm_node *node; node 524 fs/ocfs2/cluster/heartbeat.c node = o2nm_get_node_by_num(slot->ds_node_num); node 525 fs/ocfs2/cluster/heartbeat.c if (!node) node 538 fs/ocfs2/cluster/heartbeat.c o2hb_queue_node_event(&event, O2HB_NODE_DOWN_CB, node, node 546 fs/ocfs2/cluster/heartbeat.c o2nm_node_put(node); node 555 fs/ocfs2/cluster/heartbeat.c struct o2nm_node *node; node 565 fs/ocfs2/cluster/heartbeat.c node = o2nm_get_node_by_num(slot->ds_node_num); node 566 fs/ocfs2/cluster/heartbeat.c if (!node) node 638 fs/ocfs2/cluster/heartbeat.c o2hb_queue_node_event(&event, O2HB_NODE_UP_CB, node, node 684 fs/ocfs2/cluster/heartbeat.c o2hb_queue_node_event(&event, O2HB_NODE_DOWN_CB, node, node 705 fs/ocfs2/cluster/heartbeat.c o2nm_node_put(node); node 714 fs/ocfs2/cluster/heartbeat.c int highest, node; node 717 fs/ocfs2/cluster/heartbeat.c node = -1; node 718 fs/ocfs2/cluster/heartbeat.c while ((node = find_next_bit(nodes, numbits, node + 1)) != -1) { node 719 fs/ocfs2/cluster/heartbeat.c if (node >= numbits) node 722 fs/ocfs2/cluster/heartbeat.c highest = node; node 41 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node = NULL; node 47 fs/ocfs2/cluster/nodemanager.c node = o2nm_single_cluster->cl_nodes[node_num]; node 48 fs/ocfs2/cluster/nodemanager.c if (node) node 49 fs/ocfs2/cluster/nodemanager.c config_item_get(&node->nd_item); node 52 fs/ocfs2/cluster/nodemanager.c return node; node 80 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node, *ret = NULL; node 86 fs/ocfs2/cluster/nodemanager.c node = rb_entry(parent, struct o2nm_node, nd_ip_node); node 88 fs/ocfs2/cluster/nodemanager.c cmp = memcmp(&ip_needle, &node->nd_ipv4_address, node 95 fs/ocfs2/cluster/nodemanager.c ret = node; node 110 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node = NULL; node 117 fs/ocfs2/cluster/nodemanager.c node = o2nm_node_ip_tree_lookup(cluster, addr, NULL, NULL); node 118 fs/ocfs2/cluster/nodemanager.c if (node) node 119 fs/ocfs2/cluster/nodemanager.c config_item_get(&node->nd_item); node 123 fs/ocfs2/cluster/nodemanager.c return node; node 129 fs/ocfs2/cluster/nodemanager.c config_item_put(&node->nd_item); node 135 fs/ocfs2/cluster/nodemanager.c config_item_get(&node->nd_item); node 167 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node = to_o2nm_node(item); node 168 fs/ocfs2/cluster/nodemanager.c kfree(node); node 173 fs/ocfs2/cluster/nodemanager.c return sprintf(page, "%d\n", node->nd_num); node 180 fs/ocfs2/cluster/nodemanager.c return to_o2nm_cluster(node->nd_item.ci_parent->ci_parent); node 193 fs/ocfs2/cluster/nodemanager.c struct o2nm_cluster *cluster = to_o2nm_cluster_from_node(node); node 208 fs/ocfs2/cluster/nodemanager.c if (!test_bit(O2NM_NODE_ATTR_ADDRESS, &node->nd_set_attributes) || node 209 fs/ocfs2/cluster/nodemanager.c !test_bit(O2NM_NODE_ATTR_PORT, &node->nd_set_attributes)) node 216 fs/ocfs2/cluster/nodemanager.c cluster->cl_nodes[tmp] = node; node 217 fs/ocfs2/cluster/nodemanager.c node->nd_num = tmp; node 228 fs/ocfs2/cluster/nodemanager.c return sprintf(page, "%u\n", ntohs(node->nd_ipv4_port)); node 246 fs/ocfs2/cluster/nodemanager.c node->nd_ipv4_port = htons(tmp); node 253 fs/ocfs2/cluster/nodemanager.c return sprintf(page, "%u.%u.%u.%u\n", NIPQUAD(node->nd_ipv4_address)); node 260 fs/ocfs2/cluster/nodemanager.c struct o2nm_cluster *cluster = to_o2nm_cluster_from_node(node); node 282 fs/ocfs2/cluster/nodemanager.c rb_link_node(&node->nd_ip_node, parent, p); node 283 fs/ocfs2/cluster/nodemanager.c rb_insert_color(&node->nd_ip_node, &cluster->cl_node_ip_tree); node 289 fs/ocfs2/cluster/nodemanager.c memcpy(&node->nd_ipv4_address, &ipv4_addr, sizeof(ipv4_addr)); node 296 fs/ocfs2/cluster/nodemanager.c return sprintf(page, "%d\n", node->nd_local); node 302 fs/ocfs2/cluster/nodemanager.c struct o2nm_cluster *cluster = to_o2nm_cluster_from_node(node); node 315 fs/ocfs2/cluster/nodemanager.c if (!test_bit(O2NM_NODE_ATTR_ADDRESS, &node->nd_set_attributes) || node 316 fs/ocfs2/cluster/nodemanager.c !test_bit(O2NM_NODE_ATTR_NUM, &node->nd_set_attributes) || node 317 fs/ocfs2/cluster/nodemanager.c !test_bit(O2NM_NODE_ATTR_PORT, &node->nd_set_attributes)) node 323 fs/ocfs2/cluster/nodemanager.c cluster->cl_local_node != node->nd_num) node 328 fs/ocfs2/cluster/nodemanager.c ret = o2net_start_listening(node); node 334 fs/ocfs2/cluster/nodemanager.c cluster->cl_local_node == node->nd_num) { node 335 fs/ocfs2/cluster/nodemanager.c o2net_stop_listening(node); node 339 fs/ocfs2/cluster/nodemanager.c node->nd_local = tmp; node 340 fs/ocfs2/cluster/nodemanager.c if (node->nd_local) { node 342 fs/ocfs2/cluster/nodemanager.c cluster->cl_local_node = node->nd_num; node 409 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node = to_o2nm_node(item); node 415 fs/ocfs2/cluster/nodemanager.c ret = o2nm_node_attr->show(node, page); node 423 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node = to_o2nm_node(item); node 434 fs/ocfs2/cluster/nodemanager.c if (test_bit(attr_index, &node->nd_set_attributes)) node 437 fs/ocfs2/cluster/nodemanager.c ret = o2nm_node_attr->store(node, page, count); node 441 fs/ocfs2/cluster/nodemanager.c set_bit(attr_index, &node->nd_set_attributes); node 650 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node = NULL; node 655 fs/ocfs2/cluster/nodemanager.c node = kzalloc(sizeof(struct o2nm_node), GFP_KERNEL); node 656 fs/ocfs2/cluster/nodemanager.c if (node == NULL) node 659 fs/ocfs2/cluster/nodemanager.c strcpy(node->nd_name, name); /* use item.ci_namebuf instead? */ node 660 fs/ocfs2/cluster/nodemanager.c config_item_init_type_name(&node->nd_item, name, &o2nm_node_type); node 661 fs/ocfs2/cluster/nodemanager.c spin_lock_init(&node->nd_lock); node 663 fs/ocfs2/cluster/nodemanager.c return &node->nd_item; node 669 fs/ocfs2/cluster/nodemanager.c struct o2nm_node *node = to_o2nm_node(item); node 672 fs/ocfs2/cluster/nodemanager.c o2net_disconnect_node(node); node 675 fs/ocfs2/cluster/nodemanager.c (cluster->cl_local_node == node->nd_num)) { node 678 fs/ocfs2/cluster/nodemanager.c o2net_stop_listening(node); node 686 fs/ocfs2/cluster/nodemanager.c if (node->nd_ipv4_address) node 687 fs/ocfs2/cluster/nodemanager.c rb_erase(&node->nd_ip_node, &cluster->cl_node_ip_tree); node 690 fs/ocfs2/cluster/nodemanager.c if (cluster->cl_nodes[node->nd_num] == node) { node 691 fs/ocfs2/cluster/nodemanager.c cluster->cl_nodes[node->nd_num] = NULL; node 692 fs/ocfs2/cluster/nodemanager.c clear_bit(node->nd_num, cluster->cl_nodes_bitmap); node 161 fs/ocfs2/cluster/quorum.c if (!test_and_set_bit(node, qs->qs_hold_bm)) { node 164 fs/ocfs2/cluster/quorum.c "node %u\n", node); node 165 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_holds); node 173 fs/ocfs2/cluster/quorum.c if (test_and_clear_bit(node, qs->qs_hold_bm)) { node 174 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_holds - 1); node 182 fs/ocfs2/cluster/quorum.c node, qs->qs_holds); node 198 fs/ocfs2/cluster/quorum.c "node %u\n", node); node 199 fs/ocfs2/cluster/quorum.c mlog_bug_on_msg(test_bit(node, qs->qs_hb_bm), "node %u\n", node); node 200 fs/ocfs2/cluster/quorum.c set_bit(node, qs->qs_hb_bm); node 202 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_heartbeating); node 204 fs/ocfs2/cluster/quorum.c if (!test_bit(node, qs->qs_conn_bm)) node 205 fs/ocfs2/cluster/quorum.c o2quo_set_hold(qs, node); node 207 fs/ocfs2/cluster/quorum.c o2quo_clear_hold(qs, node); node 223 fs/ocfs2/cluster/quorum.c node, qs->qs_heartbeating); node 224 fs/ocfs2/cluster/quorum.c mlog_bug_on_msg(!test_bit(node, qs->qs_hb_bm), "node %u\n", node); node 225 fs/ocfs2/cluster/quorum.c clear_bit(node, qs->qs_hb_bm); node 227 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_heartbeating); node 229 fs/ocfs2/cluster/quorum.c o2quo_clear_hold(qs, node); node 245 fs/ocfs2/cluster/quorum.c mlog(0, "node %u\n", node); node 248 fs/ocfs2/cluster/quorum.c o2quo_clear_hold(qs, node); node 266 fs/ocfs2/cluster/quorum.c "node %u\n", node); node 267 fs/ocfs2/cluster/quorum.c mlog_bug_on_msg(test_bit(node, qs->qs_conn_bm), "node %u\n", node); node 268 fs/ocfs2/cluster/quorum.c set_bit(node, qs->qs_conn_bm); node 270 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_connected); node 272 fs/ocfs2/cluster/quorum.c if (!test_bit(node, qs->qs_hb_bm)) node 273 fs/ocfs2/cluster/quorum.c o2quo_set_hold(qs, node); node 275 fs/ocfs2/cluster/quorum.c o2quo_clear_hold(qs, node); node 290 fs/ocfs2/cluster/quorum.c if (test_bit(node, qs->qs_conn_bm)) { node 294 fs/ocfs2/cluster/quorum.c node, qs->qs_connected); node 296 fs/ocfs2/cluster/quorum.c clear_bit(node, qs->qs_conn_bm); node 299 fs/ocfs2/cluster/quorum.c mlog(0, "node %u, %d total\n", node, qs->qs_connected); node 301 fs/ocfs2/cluster/quorum.c if (test_bit(node, qs->qs_hb_bm)) node 302 fs/ocfs2/cluster/quorum.c o2quo_set_hold(qs, node); node 153 fs/ocfs2/cluster/tcp.c nst->st_node = node; node 386 fs/ocfs2/cluster/tcp.c o2nm_node_get(node); node 387 fs/ocfs2/cluster/tcp.c sc->sc_node = node; node 1536 fs/ocfs2/cluster/tcp.c struct o2nm_node *node = NULL, *mynode = NULL; node 1547 fs/ocfs2/cluster/tcp.c node = o2nm_get_node_by_num(o2net_num_from_nn(nn)); node 1548 fs/ocfs2/cluster/tcp.c if (node == NULL) { node 1577 fs/ocfs2/cluster/tcp.c sc = sc_alloc(node); node 1619 fs/ocfs2/cluster/tcp.c remoteaddr.sin_addr.s_addr = node->nd_ipv4_address; node 1620 fs/ocfs2/cluster/tcp.c remoteaddr.sin_port = node->nd_ipv4_port; node 1640 fs/ocfs2/cluster/tcp.c if (node) node 1641 fs/ocfs2/cluster/tcp.c o2nm_node_put(node); node 1678 fs/ocfs2/cluster/tcp.c struct o2net_node *nn = o2net_nn_from_num(node->nd_num); node 1700 fs/ocfs2/cluster/tcp.c o2net_disconnect_node(node); node 1761 fs/ocfs2/cluster/tcp.c struct o2nm_node *node = NULL; node 1791 fs/ocfs2/cluster/tcp.c node = o2nm_get_node_by_ip(sin.sin_addr.s_addr); node 1792 fs/ocfs2/cluster/tcp.c if (node == NULL) { node 1800 fs/ocfs2/cluster/tcp.c if (o2nm_this_node() > node->nd_num) { node 1803 fs/ocfs2/cluster/tcp.c node->nd_name, NIPQUAD(sin.sin_addr.s_addr), node 1804 fs/ocfs2/cluster/tcp.c ntohs(sin.sin_port), node->nd_num); node 1811 fs/ocfs2/cluster/tcp.c if (!o2hb_check_node_heartbeating_from_callback(node->nd_num)) { node 1814 fs/ocfs2/cluster/tcp.c node->nd_name, NIPQUAD(sin.sin_addr.s_addr), node 1820 fs/ocfs2/cluster/tcp.c nn = o2net_nn_from_num(node->nd_num); node 1831 fs/ocfs2/cluster/tcp.c node->nd_name, NIPQUAD(sin.sin_addr.s_addr), node 1836 fs/ocfs2/cluster/tcp.c sc = sc_alloc(node); node 1859 fs/ocfs2/cluster/tcp.c if (node) node 1860 fs/ocfs2/cluster/tcp.c o2nm_node_put(node); node 1968 fs/ocfs2/cluster/tcp.c ret = o2net_open_listening_sock(node->nd_ipv4_address, node 1969 fs/ocfs2/cluster/tcp.c node->nd_ipv4_port); node 1974 fs/ocfs2/cluster/tcp.c o2quo_conn_up(node->nd_num); node 1996 fs/ocfs2/cluster/tcp.c struct o2nm_node *node = o2nm_get_node_by_num(i); node 1997 fs/ocfs2/cluster/tcp.c if (node) { node 1998 fs/ocfs2/cluster/tcp.c o2net_disconnect_node(node); node 1999 fs/ocfs2/cluster/tcp.c o2nm_node_put(node); node 2011 fs/ocfs2/cluster/tcp.c o2quo_conn_err(node->nd_num); node 196 fs/ocfs2/dlm/dlmast.c lock->ml.node == dlm->node_num ? "master" : node 224 fs/ocfs2/dlm/dlmast.c BUG_ON(lock->ml.node != dlm->node_num); node 241 fs/ocfs2/dlm/dlmast.c BUG_ON(lock->ml.node == dlm->node_num); node 258 fs/ocfs2/dlm/dlmast.c BUG_ON(lock->ml.node != dlm->node_num); node 432 fs/ocfs2/dlm/dlmast.c res->lockname.len, res->lockname.name, lock->ml.node, node 454 fs/ocfs2/dlm/dlmast.c lock->ml.node, &status); node 460 fs/ocfs2/dlm/dlmast.c "node is dead!\n", lock->ml.node); node 464 fs/ocfs2/dlm/dlmast.c "DLM_MIGRATING!\n", lock->ml.node); node 468 fs/ocfs2/dlm/dlmast.c lock->ml.node, status); node 267 fs/ocfs2/dlm/dlmcommon.h dlm->joining_node = node; node 343 fs/ocfs2/dlm/dlmcommon.h u8 node; node 214 fs/ocfs2/dlm/dlmconvert.c if (lock->ml.node == dlm->node_num) node 477 fs/ocfs2/dlm/dlmconvert.c lock->ml.node == cnv->node_idx) { node 81 fs/ocfs2/dlm/dlmdebug.c lock->ml.type, lock->ml.convert_type, lock->ml.node, node 562 fs/ocfs2/dlm/dlmdebug.c lock->ml.node, node 762 fs/ocfs2/dlm/dlmdebug.c struct dlm_reco_node_data *node; node 857 fs/ocfs2/dlm/dlmdebug.c list_for_each_entry(node, &dlm->reco.node_data, list) { node 858 fs/ocfs2/dlm/dlmdebug.c switch (node->state) { node 885 fs/ocfs2/dlm/dlmdebug.c node->node_num, state); node 508 fs/ocfs2/dlm/dlmdomain.c int node = -1; node 514 fs/ocfs2/dlm/dlmdomain.c while ((node = find_next_bit(dlm->domain_map, O2NM_MAX_NODES, node 515 fs/ocfs2/dlm/dlmdomain.c node + 1)) < O2NM_MAX_NODES) { node 516 fs/ocfs2/dlm/dlmdomain.c printk("%d ", node); node 525 fs/ocfs2/dlm/dlmdomain.c unsigned int node; node 533 fs/ocfs2/dlm/dlmdomain.c node = exit_msg->node_idx; node 535 fs/ocfs2/dlm/dlmdomain.c printk(KERN_INFO "ocfs2_dlm: Node %u leaves domain %s\n", node, dlm->name); node 538 fs/ocfs2/dlm/dlmdomain.c clear_bit(node, dlm->domain_map); node 542 fs/ocfs2/dlm/dlmdomain.c dlm_hb_event_notify_attached(dlm, node, 0); node 558 fs/ocfs2/dlm/dlmdomain.c node, dlm->name, dlm->node_num); node 564 fs/ocfs2/dlm/dlmdomain.c &leave_msg, sizeof(leave_msg), node, node 575 fs/ocfs2/dlm/dlmdomain.c int node, clear_node, status; node 585 fs/ocfs2/dlm/dlmdomain.c while ((node = find_next_bit(dlm->domain_map, O2NM_MAX_NODES, node 597 fs/ocfs2/dlm/dlmdomain.c status = dlm_send_one_domain_exit(dlm, node); node 602 fs/ocfs2/dlm/dlmdomain.c "to node %d\n", status, node); node 615 fs/ocfs2/dlm/dlmdomain.c clear_bit(node, dlm->domain_map); node 709 fs/ocfs2/dlm/dlmdomain.c node, proto_type, node 719 fs/ocfs2/dlm/dlmdomain.c node, proto_type, node 960 fs/ocfs2/dlm/dlmdomain.c &cancel_msg, sizeof(cancel_msg), node, node 977 fs/ocfs2/dlm/dlmdomain.c unsigned int node; node 988 fs/ocfs2/dlm/dlmdomain.c node = -1; node 989 fs/ocfs2/dlm/dlmdomain.c while ((node = find_next_bit(node_map, O2NM_MAX_NODES, node 990 fs/ocfs2/dlm/dlmdomain.c node + 1)) < O2NM_MAX_NODES) { node 991 fs/ocfs2/dlm/dlmdomain.c if (node == dlm->node_num) node 994 fs/ocfs2/dlm/dlmdomain.c tmpstat = dlm_send_one_join_cancel(dlm, node); node 997 fs/ocfs2/dlm/dlmdomain.c "node %d\n", tmpstat, node); node 1017 fs/ocfs2/dlm/dlmdomain.c mlog(0, "querying node %d\n", node); node 1030 fs/ocfs2/dlm/dlmdomain.c sizeof(join_msg), node, node 1058 fs/ocfs2/dlm/dlmdomain.c node); node 1069 fs/ocfs2/dlm/dlmdomain.c node, node 1077 fs/ocfs2/dlm/dlmdomain.c packet.code, node); node 1080 fs/ocfs2/dlm/dlmdomain.c mlog(0, "status %d, node %d response is %d\n", status, node, node 1093 fs/ocfs2/dlm/dlmdomain.c mlog(0, "Sending join assert to node %u\n", node); node 1101 fs/ocfs2/dlm/dlmdomain.c &assert_msg, sizeof(assert_msg), node, node 1112 fs/ocfs2/dlm/dlmdomain.c int status, node, live; node 1115 fs/ocfs2/dlm/dlmdomain.c node = -1; node 1116 fs/ocfs2/dlm/dlmdomain.c while ((node = find_next_bit(node_map, O2NM_MAX_NODES, node 1117 fs/ocfs2/dlm/dlmdomain.c node + 1)) < O2NM_MAX_NODES) { node 1118 fs/ocfs2/dlm/dlmdomain.c if (node == dlm->node_num) node 1125 fs/ocfs2/dlm/dlmdomain.c status = dlm_send_one_join_assert(dlm, node); node 1128 fs/ocfs2/dlm/dlmdomain.c live = test_bit(node, dlm->live_nodes_map); node 1133 fs/ocfs2/dlm/dlmdomain.c "join on node %d\n", status, node); node 1174 fs/ocfs2/dlm/dlmdomain.c int status = 0, tmpstat, node; node 1199 fs/ocfs2/dlm/dlmdomain.c node = -1; node 1200 fs/ocfs2/dlm/dlmdomain.c while ((node = find_next_bit(ctxt->live_map, O2NM_MAX_NODES, node 1201 fs/ocfs2/dlm/dlmdomain.c node + 1)) < O2NM_MAX_NODES) { node 1202 fs/ocfs2/dlm/dlmdomain.c if (node == dlm->node_num) node 1205 fs/ocfs2/dlm/dlmdomain.c status = dlm_request_join(dlm, node, &response); node 1214 fs/ocfs2/dlm/dlmdomain.c set_bit(node, ctxt->yes_resp_map); node 136 fs/ocfs2/dlm/dlmlock.c lock->ml.node != dlm->node_num) { node 165 fs/ocfs2/dlm/dlmlock.c lock->ml.node); node 176 fs/ocfs2/dlm/dlmlock.c lock->ml.node); node 186 fs/ocfs2/dlm/dlmlock.c if (lock->ml.node == dlm->node_num) node 408 fs/ocfs2/dlm/dlmlock.c newlock->ml.node = node; node 447 fs/ocfs2/dlm/dlmlock.c dlm_init_lock(lock, type, node, cookie); node 1196 fs/ocfs2/dlm/dlmmaster.c int node; node 1205 fs/ocfs2/dlm/dlmmaster.c node = dlm_bitmap_diff_iter_next(&bdi, &sc); node 1206 fs/ocfs2/dlm/dlmmaster.c while (node >= 0) { node 1211 fs/ocfs2/dlm/dlmmaster.c mlog(ML_NOTICE, "node %d up while restarting\n", node); node 1215 fs/ocfs2/dlm/dlmmaster.c clear_bit(node, mle->response_map); node 1216 fs/ocfs2/dlm/dlmmaster.c set_bit(node, mle->vote_map); node 1218 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "node down! %d\n", node); node 1224 fs/ocfs2/dlm/dlmmaster.c clear_bit(node, mle->maybe_map); node 1226 fs/ocfs2/dlm/dlmmaster.c if (node == lowest) { node 1229 fs/ocfs2/dlm/dlmmaster.c "waiting on it!\n", node); node 1273 fs/ocfs2/dlm/dlmmaster.c node = dlm_bitmap_diff_iter_next(&bdi, &sc); node 2233 fs/ocfs2/dlm/dlmmaster.c u8 node; node 2244 fs/ocfs2/dlm/dlmmaster.c node = deref->node_idx; node 2251 fs/ocfs2/dlm/dlmmaster.c mlog(ML_ERROR, "Invalid node number: %u\n", node); node 2272 fs/ocfs2/dlm/dlmmaster.c if (test_bit(node, res->refmap)) { node 2273 fs/ocfs2/dlm/dlmmaster.c dlm_lockres_clear_refmap_bit(node, res); node 2285 fs/ocfs2/dlm/dlmmaster.c res->lockname.len, res->lockname.name, node); node 2301 fs/ocfs2/dlm/dlmmaster.c item->u.dl.deref_node = node; node 2322 fs/ocfs2/dlm/dlmmaster.c u8 node; node 2327 fs/ocfs2/dlm/dlmmaster.c node = item->u.dl.deref_node; node 2331 fs/ocfs2/dlm/dlmmaster.c if (test_bit(node, res->refmap)) { node 2333 fs/ocfs2/dlm/dlmmaster.c dlm_lockres_clear_refmap_bit(node, res); node 2340 fs/ocfs2/dlm/dlmmaster.c dlm->name, res->lockname.len, res->lockname.name, node); node 2345 fs/ocfs2/dlm/dlmmaster.c res->lockname.len, res->lockname.name, node); node 2383 fs/ocfs2/dlm/dlmmaster.c if (lock->ml.node == dlm->node_num) { node 2860 fs/ocfs2/dlm/dlmmaster.c if (lock->ml.node != dlm->node_num) { node 2862 fs/ocfs2/dlm/dlmmaster.c lock->ml.node); node 2868 fs/ocfs2/dlm/dlmmaster.c dlm_lockres_clear_refmap_bit(lock->ml.node, res); node 2914 fs/ocfs2/dlm/dlmmaster.c if (lock->ml.node != dlm->node_num) { node 2916 fs/ocfs2/dlm/dlmmaster.c return lock->ml.node; node 349 fs/ocfs2/dlm/dlmrecovery.c dead = !test_bit(node, dlm->domain_map); node 360 fs/ocfs2/dlm/dlmrecovery.c recovered = !test_bit(node, dlm->recovery_map); node 370 fs/ocfs2/dlm/dlmrecovery.c "death of node %u\n", dlm->name, timeout, node); node 372 fs/ocfs2/dlm/dlmrecovery.c dlm_is_node_dead(dlm, node), node 376 fs/ocfs2/dlm/dlmrecovery.c "of death of node %u\n", dlm->name, node); node 378 fs/ocfs2/dlm/dlmrecovery.c dlm_is_node_dead(dlm, node)); node 388 fs/ocfs2/dlm/dlmrecovery.c "recovery of node %u\n", dlm->name, timeout, node); node 390 fs/ocfs2/dlm/dlmrecovery.c dlm_is_node_recovered(dlm, node), node 394 fs/ocfs2/dlm/dlmrecovery.c "of recovery of node %u\n", dlm->name, node); node 396 fs/ocfs2/dlm/dlmrecovery.c dlm_is_node_recovered(dlm, node)); node 1051 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == dead_node) { node 1200 fs/ocfs2/dlm/dlmrecovery.c ml->node = lock->ml.node; node 1218 fs/ocfs2/dlm/dlmrecovery.c dummy.ml.node = dlm->node_num; node 1231 fs/ocfs2/dlm/dlmrecovery.c *nodenum = ml->node; node 1761 fs/ocfs2/dlm/dlmrecovery.c if (ml->node == dlm->node_num) { node 1789 fs/ocfs2/dlm/dlmrecovery.c BUG_ON(lock->ml.node != ml->node); node 1812 fs/ocfs2/dlm/dlmrecovery.c newlock = dlm_new_lock(ml->type, ml->node, node 1899 fs/ocfs2/dlm/dlmrecovery.c ml->type, ml->convert_type, ml->node, node 1914 fs/ocfs2/dlm/dlmrecovery.c res->lockname.len, res->lockname.name, ml->node); node 1915 fs/ocfs2/dlm/dlmrecovery.c dlm_lockres_set_refmap_bit(ml->node, res); node 2131 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == search_node) { node 2166 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == dead_node) { node 2175 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == dead_node) { node 2184 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == dead_node) { node 2252 fs/ocfs2/dlm/dlmrecovery.c if (lock->ml.node == dead_node) { node 59 fs/ocfs2/dlm/dlmthread.c #define dlm_lock_is_remote(dlm, lock) ((lock)->ml.node != (dlm)->node_num) node 356 fs/ocfs2/dlm/dlmthread.c target->ml.convert_type, target->ml.node); node 418 fs/ocfs2/dlm/dlmthread.c target->ml.type, target->ml.node); node 533 fs/ocfs2/dlm/dlmthread.c if (lock->ml.node != dlm->node_num) { node 581 fs/ocfs2/dlm/dlmthread.c if (lock->ml.node != dlm->node_num) { node 465 fs/ocfs2/dlm/dlmunlock.c lock->ml.node == unlock->node_idx) { node 346 fs/ocfs2/stack_o2cb.c *node = node_num; node 439 fs/ocfs2/stackglue.c return active_stack->sp_ops->this_node(node); node 150 fs/ocfs2/stackglue.h int (*this_node)(unsigned int *node); node 92 fs/ocfs2/uptodate.c struct rb_node *node; node 95 fs/ocfs2/uptodate.c while ((node = rb_last(root)) != NULL) { node 96 fs/ocfs2/uptodate.c item = rb_entry(node, struct ocfs2_meta_cache_item, c_node); node 31 fs/openpromfs/inode.c struct device_node *node; node 192 fs/openpromfs/inode.c dp = oi->u.node; node 206 fs/openpromfs/inode.c ent_data.node = child; node 269 fs/openpromfs/inode.c struct device_node *dp = oi->u.node; node 405 fs/openpromfs/inode.c oi->u.node = of_find_node_by_path("/"); node 255 fs/ubifs/debug.c const struct ubifs_ch *ch = node; node 264 fs/ubifs/debug.c (void *)node, UBIFS_CH_SZ, 1); node 269 fs/ubifs/debug.c dump_ch(node); node 274 fs/ubifs/debug.c const struct ubifs_pad_node *pad = node; node 282 fs/ubifs/debug.c const struct ubifs_sb_node *sup = node; node 337 fs/ubifs/debug.c const struct ubifs_mst_node *mst = node; node 399 fs/ubifs/debug.c const struct ubifs_ref_node *ref = node; node 411 fs/ubifs/debug.c const struct ubifs_ino_node *ino = node; node 453 fs/ubifs/debug.c const struct ubifs_dent_node *dent = node; node 477 fs/ubifs/debug.c const struct ubifs_data_node *dn = node; node 495 fs/ubifs/debug.c const struct ubifs_trun_node *trun = node; node 507 fs/ubifs/debug.c const struct ubifs_idx_node *idx = node; node 530 fs/ubifs/debug.c const struct ubifs_orph_node *orph = node; node 673 fs/ubifs/debug.c dbg_dump_node(c, snod->node); node 1682 fs/ubifs/debug.c void *node; node 1693 fs/ubifs/debug.c node = kmalloc(zbr->len, GFP_NOFS); node 1694 fs/ubifs/debug.c if (!node) node 1697 fs/ubifs/debug.c err = ubifs_tnc_read_node(c, zbr, node); node 1706 fs/ubifs/debug.c fscki = add_inode(c, priv, node); node 1723 fs/ubifs/debug.c ch = node; node 1733 fs/ubifs/debug.c struct ubifs_data_node *dn = node; node 1761 fs/ubifs/debug.c struct ubifs_dent_node *dent = node; node 1808 fs/ubifs/debug.c kfree(node); node 1813 fs/ubifs/debug.c dbg_dump_node(c, node); node 1815 fs/ubifs/debug.c kfree(node); node 197 fs/ubifs/gc.c err = ubifs_wbuf_write_nolock(wbuf, snod->node, node 299 fs/ubifs/gc.c struct ubifs_idx_node *idx = snod->node; node 228 fs/ubifs/io.c struct ubifs_ch *ch = node; node 238 fs/ubifs/io.c crc = crc32(UBIFS_CRC32_INIT, node + 8, len - 8); node 244 fs/ubifs/io.c ubifs_pad(c, node + len, pad); node 261 fs/ubifs/io.c struct ubifs_ch *ch = node; node 274 fs/ubifs/io.c crc = crc32(UBIFS_CRC32_INIT, node + 8, len - 8); node 260 fs/ubifs/journal.c ubifs_prepare_node(c, node, len, 0); node 262 fs/ubifs/journal.c return ubifs_wbuf_write_nolock(wbuf, node, len); node 655 fs/ubifs/log.c struct ubifs_ch *ch = node; node 668 fs/ubifs/log.c memcpy(buf + *offs, node, len); node 707 fs/ubifs/log.c struct ubifs_ref_node *ref = snod->node; node 715 fs/ubifs/log.c &offs, snod->node); node 725 fs/ubifs/log.c snod->node); node 1166 fs/ubifs/lprops.c struct ubifs_idx_node *idx = snod->node; node 52 fs/ubifs/master.c memcpy(c->mst_node, snod->node, snod->len); node 72 fs/ubifs/master.c (void *)snod->node + UBIFS_CH_SZ, node 310 fs/ubifs/misc.h return ubifs_tnc_locate(c, key, node, NULL, NULL); node 570 fs/ubifs/orphan.c dbg_dump_node(c, snod->node); node 574 fs/ubifs/orphan.c orph = snod->node; node 598 fs/ubifs/orphan.c dbg_dump_node(c, snod->node); node 735 fs/ubifs/orphan.c struct ubifs_ino_node *node; node 847 fs/ubifs/orphan.c err = ubifs_tnc_read_node(c, zbr, ci->node); node 852 fs/ubifs/orphan.c if (ci->node->nlink == 0) node 875 fs/ubifs/orphan.c orph = snod->node; node 926 fs/ubifs/orphan.c ci.node = kmalloc(UBIFS_MAX_INO_NODE_SZ, GFP_NOFS); node 927 fs/ubifs/orphan.c if (!ci.node) { node 954 fs/ubifs/orphan.c kfree(ci.node); node 570 fs/ubifs/recovery.c ch = snod->node; node 551 fs/ubifs/replay.c struct ubifs_ino_node *ino = snod->node; node 563 fs/ubifs/replay.c struct ubifs_data_node *dn = snod->node; node 576 fs/ubifs/replay.c struct ubifs_dent_node *dent = snod->node; node 590 fs/ubifs/replay.c struct ubifs_trun_node *trun = snod->node; node 644 fs/ubifs/replay.c dbg_dump_node(c, snod->node); node 836 fs/ubifs/replay.c const struct ubifs_cs_node *node; node 852 fs/ubifs/replay.c node = sleb->buf; node 868 fs/ubifs/replay.c if (le64_to_cpu(node->cmt_no) != c->cmt_no) { node 872 fs/ubifs/replay.c (unsigned long long)le64_to_cpu(node->cmt_no), node 877 fs/ubifs/replay.c c->cs_sqnum = le64_to_cpu(node->ch.sqnum); node 919 fs/ubifs/replay.c const struct ubifs_ref_node *ref = snod->node; node 962 fs/ubifs/replay.c dbg_dump_node(c, snod->node); node 209 fs/ubifs/scan.c snod->node = buf; node 352 fs/ubifs/scan.c struct ubifs_scan_node *node; node 357 fs/ubifs/scan.c node = list_entry(head->next, struct ubifs_scan_node, list); node 358 fs/ubifs/scan.c list_del(&node->list); node 359 fs/ubifs/scan.c kfree(node); node 333 fs/ubifs/tnc.c const struct ubifs_dent_node *dent = node; node 351 fs/ubifs/tnc.c memcpy(lnc_node, node, zbr->len); node 373 fs/ubifs/tnc.c err = ubifs_validate_entry(c, node); node 376 fs/ubifs/tnc.c dbg_dump_node(c, node); node 380 fs/ubifs/tnc.c zbr->leaf = node; node 418 fs/ubifs/tnc.c memcpy(node, zbr->leaf, zbr->len); node 422 fs/ubifs/tnc.c err = ubifs_tnc_read_node(c, zbr, node); node 427 fs/ubifs/tnc.c err = lnc_add(c, zbr, node); node 498 fs/ubifs/tnc.c ret = try_read_node(c, node, key_type(c, key), zbr->len, zbr->lnum, node 502 fs/ubifs/tnc.c struct ubifs_dent_node *dent = node; node 1460 fs/ubifs/tnc.c err = tnc_read_node_nm(c, zt, node); node 1464 fs/ubifs/tnc.c err = ubifs_tnc_read_node(c, zt, node); node 1474 fs/ubifs/tnc.c err = ubifs_tnc_read_node(c, &zbr, node); node 1478 fs/ubifs/tnc.c err = fallible_read_node(c, key, &zbr, node); node 1535 fs/ubifs/tnc.c err = tnc_read_node_nm(c, &znode->zbranch[n], node); node 1559 fs/ubifs/tnc.c const struct ubifs_dent_node *dent = node; node 1565 fs/ubifs/tnc.c err = ubifs_tnc_lookup(c, key, node); node 1577 fs/ubifs/tnc.c return do_lookup_nm(c, key, node, nm); node 258 fs/ubifs/tnc_commit.c idx = snod->node; node 471 fs/ubifs/tnc_misc.c err = ubifs_read_node_wbuf(wbuf, node, type, zbr->len, node 474 fs/ubifs/tnc_misc.c err = ubifs_read_node(c, node, type, zbr->len, zbr->lnum, node 484 fs/ubifs/tnc_misc.c if (memcmp(node + UBIFS_KEY_OFFSET, &key1, c->key_len)) { node 489 fs/ubifs/tnc_misc.c dbg_dump_node(c, node); node 282 fs/ubifs/ubifs.h void *node; node 1828 fs/xfs/xfs_attr.c xfs_da_intnode_t *node; node 1848 fs/xfs/xfs_attr.c node = bp->data; node 1849 fs/xfs/xfs_attr.c switch (be16_to_cpu(node->hdr.info.magic)) { node 1851 fs/xfs/xfs_attr.c xfs_attr_trace_l_cn("wrong blk", context, node); node 1898 fs/xfs/xfs_attr.c node = bp->data; node 1899 fs/xfs/xfs_attr.c if (be16_to_cpu(node->hdr.info.magic) node 1902 fs/xfs/xfs_attr.c if (unlikely(be16_to_cpu(node->hdr.info.magic) node 1907 fs/xfs/xfs_attr.c node); node 1911 fs/xfs/xfs_attr.c btree = node->btree; node 1912 fs/xfs/xfs_attr.c for (i = 0; i < be16_to_cpu(node->hdr.count); node 1922 fs/xfs/xfs_attr.c if (i == be16_to_cpu(node->hdr.count)) { node 2288 fs/xfs/xfs_attr.c (__psunsigned_t)be16_to_cpu(node->hdr.count), node 2289 fs/xfs/xfs_attr.c (__psunsigned_t)be32_to_cpu(node->btree[0].hashval), node 2290 fs/xfs/xfs_attr.c (__psunsigned_t)be32_to_cpu(node->btree[ node 2291 fs/xfs/xfs_attr.c be16_to_cpu(node->hdr.count)-1].hashval)); node 849 fs/xfs/xfs_attr_leaf.c xfs_da_intnode_t *node; node 882 fs/xfs/xfs_attr_leaf.c node = bp1->data; node 886 fs/xfs/xfs_attr_leaf.c node->btree[0].hashval = node 888 fs/xfs/xfs_attr_leaf.c node->btree[0].before = cpu_to_be32(blkno); node 889 fs/xfs/xfs_attr_leaf.c node->hdr.count = cpu_to_be16(1); node 2736 fs/xfs/xfs_attr_leaf.c xfs_da_intnode_t *node; node 2750 fs/xfs/xfs_attr_leaf.c node = bp->data; node 2751 fs/xfs/xfs_attr_leaf.c ASSERT(be16_to_cpu(node->hdr.info.magic) == XFS_DA_NODE_MAGIC); node 2753 fs/xfs/xfs_attr_leaf.c count = be16_to_cpu(node->hdr.count); node 2758 fs/xfs/xfs_attr_leaf.c child_fsb = be32_to_cpu(node->btree[0].before); node 2818 fs/xfs/xfs_attr_leaf.c child_fsb = be32_to_cpu(node->btree[i+1].before); node 113 fs/xfs/xfs_da_btree.c xfs_da_intnode_t *node; node 123 fs/xfs/xfs_da_btree.c node = bp->data; node 124 fs/xfs/xfs_da_btree.c node->hdr.info.forw = 0; node 125 fs/xfs/xfs_da_btree.c node->hdr.info.back = 0; node 126 fs/xfs/xfs_da_btree.c node->hdr.info.magic = cpu_to_be16(XFS_DA_NODE_MAGIC); node 127 fs/xfs/xfs_da_btree.c node->hdr.info.pad = 0; node 128 fs/xfs/xfs_da_btree.c node->hdr.count = 0; node 129 fs/xfs/xfs_da_btree.c node->hdr.level = cpu_to_be16(level); node 132 fs/xfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, &node->hdr, sizeof(node->hdr))); node 146 fs/xfs/xfs_da_btree.c xfs_da_intnode_t *node; node 257 fs/xfs/xfs_da_btree.c node = oldblk->bp->data; node 258 fs/xfs/xfs_da_btree.c if (node->hdr.info.forw) { node 259 fs/xfs/xfs_da_btree.c if (be32_to_cpu(node->hdr.info.forw) == addblk->blkno) { node 265 fs/xfs/xfs_da_btree.c node = bp->data; node 266 fs/xfs/xfs_da_btree.c node->hdr.info.back = cpu_to_be32(oldblk->blkno); node 268 fs/xfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, &node->hdr.info, node 269 fs/xfs/xfs_da_btree.c sizeof(node->hdr.info))); node 271 fs/xfs/xfs_da_btree.c node = oldblk->bp->data; node 272 fs/xfs/xfs_da_btree.c if (node->hdr.info.back) { node 273 fs/xfs/xfs_da_btree.c if (be32_to_cpu(node->hdr.info.back) == addblk->blkno) { node 279 fs/xfs/xfs_da_btree.c node = bp->data; node 280 fs/xfs/xfs_da_btree.c node->hdr.info.forw = cpu_to_be32(oldblk->blkno); node 282 fs/xfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, &node->hdr.info, node 283 fs/xfs/xfs_da_btree.c sizeof(node->hdr.info))); node 300 fs/xfs/xfs_da_btree.c xfs_da_intnode_t *node, *oldroot; node 326 fs/xfs/xfs_da_btree.c node = bp->data; node 337 fs/xfs/xfs_da_btree.c memcpy(node, oldroot, size); node 348 fs/xfs/xfs_da_btree.c be16_to_cpu(node->hdr.level) + 1, &bp, args->whichfork); node 351 fs/xfs/xfs_da_btree.c node = bp->data; node 352 fs/xfs/xfs_da_btree.c node->btree[0].hashval = cpu_to_be32(blk1->hashval); node 353 fs/xfs/xfs_da_btree.c node->btree[0].before = cpu_to_be32(blk1->blkno); node 354 fs/xfs/xfs_da_btree.c node->btree[1].hashval = cpu_to_be32(blk2->hashval); node 355 fs/xfs/xfs_da_btree.c node->btree[1].before = cpu_to_be32(blk2->blkno); node 356 fs/xfs/xfs_da_btree.c node->hdr.count = cpu_to_be16(2); node 369 fs/xfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, node->btree, node 385 fs/xfs/xfs_da_btree.c xfs_da_intnode_t *node; node 390 fs/xfs/xfs_da_btree.c node = oldblk->bp->data; node 391 fs/xfs/xfs_da_btree.c ASSERT(be16_to_cpu(node->hdr.info.magic) == XFS_DA_NODE_MAGIC); node 401 fs/xfs/xfs_da_btree.c if ((be16_to_cpu(node->hdr.count) + newcount) > state->node_ents) { node 437 fs/xfs/xfs_da_btree.c node = oldblk->bp->data; node 438 fs/xfs/xfs_da_btree.c if (oldblk->index <= be16_to_cpu(node->hdr.count)) { node 580 fs/xfs/xfs_da_btree.c xfs_da_intnode_t *node; node 585 fs/xfs/xfs_da_btree.c node = oldblk->bp->data; node 587 fs/xfs/xfs_da_btree.c ASSERT(be16_to_cpu(node->hdr.info.magic) == XFS_DA_NODE_MAGIC); node 588 fs/xfs/xfs_da_btree.c ASSERT((oldblk->index >= 0) && (oldblk->index <= be16_to_cpu(node->hdr.count))); node 598 fs/xfs/xfs_da_btree.c btree = &node->btree[ oldblk->index ]; node 599 fs/xfs/xfs_da_btree.c if (oldblk->index < be16_to_cpu(node->hdr.count)) { node 600 fs/xfs/xfs_da_btree.c tmp = (be16_to_cpu(node->hdr.count) - oldblk->index) * (uint)sizeof(*btree); node 606 fs/xfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, btree, tmp + sizeof(*btree))); node 607 fs/xfs/xfs_da_btree.c be16_add_cpu(&node->hdr.count, 1); node 609 fs/xfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, &node->hdr, sizeof(node->hdr))); node 614 fs/xfs/xfs_da_btree.c oldblk->hashval = be32_to_cpu(node->btree[be16_to_cpu(node->hdr.count)-1 ].hashval); node 771 fs/xfs/xfs_da_btree.c xfs_da_intnode_t *node; node 786 fs/xfs/xfs_da_btree.c node = (xfs_da_intnode_t *)info; node 787 fs/xfs/xfs_da_btree.c count = be16_to_cpu(node->hdr.count); node 840 fs/xfs/xfs_da_btree.c node = (xfs_da_intnode_t *)info; node 843 fs/xfs/xfs_da_btree.c count -= be16_to_cpu(node->hdr.count); node 844 fs/xfs/xfs_da_btree.c node = bp->data; node 845 fs/xfs/xfs_da_btree.c ASSERT(be16_to_cpu(node->hdr.info.magic) == XFS_DA_NODE_MAGIC); node 846 fs/xfs/xfs_da_btree.c count -= be16_to_cpu(node->hdr.count); node 894 fs/xfs/xfs_da_btree.c xfs_da_intnode_t *node; node 919 fs/xfs/xfs_da_btree.c node = blk->bp->data; node 920 fs/xfs/xfs_da_btree.c ASSERT(be16_to_cpu(node->hdr.info.magic) == XFS_DA_NODE_MAGIC); node 921 fs/xfs/xfs_da_btree.c btree = &node->btree[ blk->index ]; node 927 fs/xfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, btree, sizeof(*btree))); node 929 fs/xfs/xfs_da_btree.c lasthash = be32_to_cpu(node->btree[be16_to_cpu(node->hdr.count)-1].hashval); node 939 fs/xfs/xfs_da_btree.c xfs_da_intnode_t *node; node 943 fs/xfs/xfs_da_btree.c node = drop_blk->bp->data; node 944 fs/xfs/xfs_da_btree.c ASSERT(drop_blk->index < be16_to_cpu(node->hdr.count)); node 950 fs/xfs/xfs_da_btree.c btree = &node->btree[drop_blk->index]; node 951 fs/xfs/xfs_da_btree.c if (drop_blk->index < (be16_to_cpu(node->hdr.count)-1)) { node 952 fs/xfs/xfs_da_btree.c tmp = be16_to_cpu(node->hdr.count) - drop_blk->index - 1; node 956 fs/xfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, btree, tmp)); node 957 fs/xfs/xfs_da_btree.c btree = &node->btree[be16_to_cpu(node->hdr.count)-1]; node 961 fs/xfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, btree, sizeof(*btree))); node 962 fs/xfs/xfs_da_btree.c be16_add_cpu(&node->hdr.count, -1); node 964 fs/xfs/xfs_da_btree.c XFS_DA_LOGRANGE(node, &node->hdr, sizeof(node->hdr))); node 1053 fs/xfs/xfs_da_btree.c xfs_da_intnode_t *node; node 1091 fs/xfs/xfs_da_btree.c node = blk->bp->data; node 1092 fs/xfs/xfs_da_btree.c max = be16_to_cpu(node->hdr.count); node 1093 fs/xfs/xfs_da_btree.c blk->hashval = be32_to_cpu(node->btree[max-1].hashval); node 1100 fs/xfs/xfs_da_btree.c for (btree = &node->btree[probe]; span > 4; node 1101 fs/xfs/xfs_da_btree.c btree = &node->btree[probe]) { node 1132 fs/xfs/xfs_da_btree.c blkno = be32_to_cpu(node->btree[max-1].before); node 1305 fs/xfs/xfs_da_btree.c xfs_da_intnode_t *node; node 1307 fs/xfs/xfs_da_btree.c node = bp->data; node 1308 fs/xfs/xfs_da_btree.c ASSERT(be16_to_cpu(node->hdr.info.magic) == XFS_DA_NODE_MAGIC); node 1310 fs/xfs/xfs_da_btree.c *count = be16_to_cpu(node->hdr.count); node 1311 fs/xfs/xfs_da_btree.c if (!node->hdr.count) node 1313 fs/xfs/xfs_da_btree.c return be32_to_cpu(node->btree[be16_to_cpu(node->hdr.count)-1].hashval); node 1403 fs/xfs/xfs_da_btree.c xfs_da_intnode_t *node; node 1420 fs/xfs/xfs_da_btree.c node = blk->bp->data; node 1421 fs/xfs/xfs_da_btree.c ASSERT(be16_to_cpu(node->hdr.info.magic) == XFS_DA_NODE_MAGIC); node 1422 fs/xfs/xfs_da_btree.c if (forward && (blk->index < be16_to_cpu(node->hdr.count)-1)) { node 1424 fs/xfs/xfs_da_btree.c blkno = be32_to_cpu(node->btree[blk->index].before); node 1428 fs/xfs/xfs_da_btree.c blkno = be32_to_cpu(node->btree[blk->index].before); node 1465 fs/xfs/xfs_da_btree.c node = (xfs_da_intnode_t *)info; node 1466 fs/xfs/xfs_da_btree.c blk->hashval = be32_to_cpu(node->btree[be16_to_cpu(node->hdr.count)-1].hashval); node 1470 fs/xfs/xfs_da_btree.c blk->index = be16_to_cpu(node->hdr.count)-1; node 1471 fs/xfs/xfs_da_btree.c blkno = be32_to_cpu(node->btree[blk->index].before); node 265 include/acpi/aclocal.h struct acpi_namespace_node *node; node 388 include/acpi/aclocal.h struct acpi_namespace_node *node; node 498 include/acpi/aclocal.h ACPI_STATE_COMMON struct acpi_namespace_node *node; node 540 include/acpi/aclocal.h ACPI_STATE_COMMON struct acpi_namespace_node *node; node 604 include/acpi/aclocal.h struct acpi_namespace_node *node; /* For use by interpreter */\ node 134 include/acpi/acobject.h struct acpi_namespace_node *node; /* Link back to parent node */ node 138 include/acpi/acobject.h ACPI_OBJECT_COMMON_HEADER struct acpi_namespace_node *node; /* Link back to parent node */ node 163 include/acpi/acobject.h struct acpi_namespace_node *node; /* Containing namespace node */ node 169 include/acpi/acobject.h struct acpi_namespace_node *node; /* Containing namespace node */ node 241 include/acpi/acobject.h struct acpi_namespace_node *node; /* Link back to parent node */\ node 284 include/acpi/acobject.h ACPI_OBJECT_COMMON_HEADER struct acpi_namespace_node *node; /* Parent device */ node 293 include/acpi/acobject.h struct acpi_namespace_node *node; /* Parent device */ node 318 include/acpi/acobject.h struct acpi_namespace_node *node; node 417 include/acpi/acobject.h struct acpi_namespace_node node; node 285 include/acpi/acpi_bus.h struct list_head node; node 316 include/acpi/acpi_bus.h struct list_head node; node 38 include/asm-generic/topology.h #define parent_node(node) ((void)(node),0) node 41 include/asm-generic/topology.h #define node_to_cpumask(node) ((void)node, cpu_online_map) node 44 include/asm-generic/topology.h #define node_to_first_cpu(node) ((void)(node),0) node 63 include/asm-generic/topology.h cpumask_t _##v = node_to_cpumask(node); \ node 67 include/asm-generic/topology.h _##v = node_to_cpumask(node) node 44 include/asm-m32r/mmzone.h int node; node 46 include/asm-m32r/mmzone.h for (node = 0 ; node < MAX_NUMNODES ; node++) node 47 include/asm-m32r/mmzone.h if (pfn >= node_start_pfn(node) && pfn <= node_end_pfn(node)) node 50 include/asm-m32r/mmzone.h return node; node 115 include/asm-m68k/dvma.h int node; /* Prom node for this DMA device */ node 277 include/asm-m68k/openprom.h int (*no_nextnode)(int node); node 278 include/asm-m68k/openprom.h int (*no_child)(int node); node 279 include/asm-m68k/openprom.h int (*no_proplen)(int node, char *name); node 280 include/asm-m68k/openprom.h int (*no_getprop)(int node, char *name, char *val); node 281 include/asm-m68k/openprom.h int (*no_setprop)(int node, char *name, char *val, int len); node 282 include/asm-m68k/openprom.h char * (*no_nextprop)(int node, char *name); node 93 include/asm-x86/numaq/apic.h int node = apicid_to_node(logical_apicid); node 96 include/asm-x86/numaq/apic.h return physid_mask_of_physid(cpu + 4*node); node 15 include/asm-x86/pci.h int node; /* NUMA node */ node 105 include/asm-x86/pci.h return sd->node; node 68 include/asm-x86/topology.h return node_to_cpumask_map[node]; node 108 include/asm-x86/topology.h return &node_to_cpumask_map[node]; node 114 include/asm-x86/topology.h return node_to_cpumask_map[node]; node 121 include/asm-x86/topology.h const cpumask_t *v = _node_to_cpumask_ptr(node) node 124 include/asm-x86/topology.h v = _node_to_cpumask_ptr(node) node 132 include/asm-x86/topology.h #define parent_node(node) (node) node 204 include/asm-x86/topology.h const cpumask_t *v = _node_to_cpumask_ptr(node) node 207 include/asm-x86/topology.h v = _node_to_cpumask_ptr(node) node 216 include/asm-x86/topology.h node_to_cpumask_ptr(mask, node); node 309 include/asm-x86/uv/uv_bau.h return constant_test_bit(node, &dstp->bits[0]); node 313 include/asm-x86/uv/uv_bau.h __set_bit(node, &dstp->bits[0]); node 143 include/linux/acpi.h struct list_head node; node 35 include/linux/async_tx.h struct list_head node; node 59 include/linux/atmel_tc.h struct list_head node; node 17 include/linux/attribute_container.h struct list_head node; node 413 include/linux/blkdev.h int node; node 13 include/linux/dca.h struct list_head node; node 27 include/linux/debugobjects.h struct hlist_node node; node 250 include/linux/device.h struct list_head node; node 438 include/linux/device.h dev->numa_node = node; node 86 include/linux/dio.h struct list_head node; node 174 include/linux/elevator.h #define rb_entry_rq(node) rb_entry((node), struct request, rb_node) node 97 include/linux/enclosure.h struct list_head node; node 809 include/linux/fb.h int node; node 56 include/linux/ftrace.h struct hlist_node node; node 50 include/linux/gameport.h struct list_head node; node 67 include/linux/hidraw.h struct list_head node; node 123 include/linux/hrtimer.h struct rb_node node; node 75 include/linux/init_task.h { .first = &init_task.pids[PIDTYPE_PID].node }, \ node 76 include/linux/init_task.h { .first = &init_task.pids[PIDTYPE_PGID].node }, \ node 77 include/linux/init_task.h { .first = &init_task.pids[PIDTYPE_SID].node }, \ node 90 include/linux/init_task.h .node = { \ node 1091 include/linux/input.h struct list_head node; node 1193 include/linux/input.h struct list_head node; node 50 include/linux/leds.h struct list_head node; /* LED Device list */ node 511 include/linux/mm.h return zone->node; node 547 include/linux/mm.h page->flags |= (node & NODES_MASK) << NODES_PGSHIFT; node 560 include/linux/mm.h set_page_node(page, node); node 135 include/linux/mmc/card.h #define mmc_list_to_card(l) container_of(l, struct mmc_card, node) node 221 include/linux/mmzone.h int node; node 739 include/linux/mmzone.h return zoneref->zone->node; node 29 include/linux/node.h extern struct node node_devices[]; node 57 include/linux/node.h #define to_node(sys_device) container_of(sys_device, struct node, sysdev) node 95 include/linux/nodemask.h #define node_set(node, dst) __node_set((node), &(dst)) node 98 include/linux/nodemask.h set_bit(node, dstp->bits); node 101 include/linux/nodemask.h #define node_clear(node, dst) __node_clear((node), &(dst)) node 104 include/linux/nodemask.h clear_bit(node, dstp->bits); node 120 include/linux/nodemask.h #define node_isset(node, nodemask) test_bit((node), (nodemask).bits) node 123 include/linux/nodemask.h __node_test_and_set((node), &(nodemask)) node 126 include/linux/nodemask.h return test_and_set_bit(node, addr->bits); node 246 include/linux/nodemask.h m.bits[0] = 1UL<<(node); \ node 249 include/linux/nodemask.h node_set((node), m); \ node 351 include/linux/nodemask.h for ((node) = first_node(mask); \ node 352 include/linux/nodemask.h (node) < MAX_NUMNODES; \ node 353 include/linux/nodemask.h (node) = next_node((node), (mask))) node 357 include/linux/nodemask.h for ((node) = 0; (node) < 1; (node)++) node 386 include/linux/nodemask.h return node_isset(node, node_states[state]); node 391 include/linux/nodemask.h __node_set(node, &node_states[state]); node 396 include/linux/nodemask.h __node_clear(node, &node_states[state]); node 415 include/linux/nodemask.h return node == 0; node 432 include/linux/nodemask.h for ( (node) = 0; (node) == 0; (node) = 1) node 445 include/linux/nodemask.h int node; \ node 446 include/linux/nodemask.h for_each_node_mask(node, (mask)) \ node 447 include/linux/nodemask.h if (node_online(node)) \ node 449 include/linux/nodemask.h node; \ node 454 include/linux/nodemask.h #define node_online(node) node_state((node), N_ONLINE) node 455 include/linux/nodemask.h #define node_possible(node) node_state((node), N_POSSIBLE) node 457 include/linux/nodemask.h #define node_set_online(node) node_set_state((node), N_ONLINE) node 458 include/linux/nodemask.h #define node_set_offline(node) node_clear_state((node), N_ONLINE) node 460 include/linux/nodemask.h #define for_each_node(node) for_each_node_state(node, N_POSSIBLE) node 461 include/linux/nodemask.h #define for_each_online_node(node) for_each_node_state(node, N_ONLINE) node 283 include/linux/pci.h struct list_head node; /* node in list of buses */ node 312 include/linux/pci.h #define pci_bus_b(n) list_entry(n, struct pci_bus, node) node 402 include/linux/pci.h struct list_head node; node 71 include/linux/pid.h struct hlist_node node; node 152 include/linux/pid.h &pid->tasks[type], pids[type].node) { node 119 include/linux/plist.h .plist = PLIST_HEAD_INIT((node).plist, NULL), \ node 144 include/linux/plist.h node->prio = prio; node 145 include/linux/plist.h plist_head_init(&node->plist, NULL); node 206 include/linux/plist.h return plist_head_empty(&node->plist); node 60 include/linux/posix-timers.h unsigned int node; node 91 include/linux/prio_tree.h return node->parent == node; node 96 include/linux/prio_tree.h return node->left == node; node 101 include/linux/prio_tree.h return node->right == node; node 114 include/linux/prio_tree.h (struct prio_tree_node *) (node)) node 116 include/linux/prio_tree.h prio_tree_insert(root, (struct prio_tree_node *) (node)) node 118 include/linux/prio_tree.h prio_tree_remove(root, (struct prio_tree_node *) (node)) node 136 include/linux/rbtree.h #define RB_EMPTY_NODE(node) (rb_parent(node) == node) node 137 include/linux/rbtree.h #define RB_CLEAR_NODE(node) (rb_set_parent(node, node)) node 155 include/linux/rbtree.h node->rb_parent_color = (unsigned long )parent; node 156 include/linux/rbtree.h node->rb_left = node->rb_right = NULL; node 158 include/linux/rbtree.h *rb_link = node; node 110 include/linux/rfkill.h struct list_head node; node 140 include/linux/rio.h struct list_head node; node 171 include/linux/rio.h struct list_head node; /* node in global list of ports */ node 200 include/linux/rio.h struct list_head node; /* node in list of networks */ node 218 include/linux/rio.h struct list_head node; node 276 include/linux/rio.h struct list_head node; node 52 include/linux/serio.h struct list_head node; node 248 include/linux/slab.h __kmalloc_node_track_caller(size, flags, node, \ node 252 include/linux/slab.h __kmalloc_node(size, flags, node) node 288 include/linux/slab.h return kmalloc_node(size, flags | __GFP_ZERO, node); node 88 include/linux/slab_def.h flags, node); node 91 include/linux/slab_def.h flags, node); node 93 include/linux/slab_def.h return __kmalloc_node(size, flags, node); node 15 include/linux/slob_def.h return __kmalloc_node(size, flags, node); node 38 include/linux/slub_def.h int node; /* The node of the page (or -1 for debug) */ node 103 include/linux/slub_def.h struct kmem_cache_node *node[MAX_NUMNODES]; node 243 include/linux/slub_def.h return kmem_cache_alloc_node(s, flags, node); node 245 include/linux/slub_def.h return __kmalloc_node(size, flags, node); node 81 include/linux/tc.h struct list_head node; /* Node in list of all TC devices. */ node 105 include/linux/tc.h struct list_head node; node 62 include/linux/thermal.h struct list_head node; node 76 include/linux/thermal.h struct list_head node; node 95 include/linux/thermal.h struct list_head node; node 48 include/linux/tipc.h __u32 node; node 66 include/linux/tipc.h return (zone << 24) | (cluster << 12) | node; node 43 include/linux/topology.h node_to_cpumask_ptr(__tmp__, node); \ node 49 include/linux/topology.h for_each_online_node(node) \ node 50 include/linux/topology.h if (nr_cpus_node(node)) node 925 include/linux/usb.h struct list_head node; node 171 include/linux/vmstat.h struct zone *zones = NODE_DATA(node)->node_zones; node 172 include/linux/zorro.h struct list_head node; node 164 include/net/ax25.h hlist_for_each_entry(__ax25, node, list, uid_node) node 250 include/net/ax25.h hlist_for_each_entry(__ax25, node, list, ax25_node) node 15 include/net/datalink.h struct list_head node; node 79 include/net/garp.h struct rb_node node; node 83 include/net/inet_hashtables.h struct hlist_node node; node 88 include/net/inet_hashtables.h hlist_for_each_entry(tb, node, head, node) node 178 include/net/inet_timewait_sock.h hlist_for_each_entry(tw, node, head, tw_node) node 181 include/net/inet_timewait_sock.h hlist_for_each_entry(tw, node, jail, tw_death_node) node 184 include/net/inet_timewait_sock.h hlist_for_each_entry_safe(tw, node, safe, jail, tw_death_node) node 127 include/net/ip6_fib.h struct fib6_node *root, *node; node 19 include/net/ipx.h __u8 node[IPX_NODE_LEN]; node 70 include/net/ipx.h struct list_head node; /* node in ipx_interfaces list */ node 78 include/net/ipx.h struct list_head node; /* node in ipx_routes list */ node 102 include/net/ipx.h unsigned char node[IPX_NODE_LEN]; node 82 include/net/lapb.h struct list_head node; node 55 include/net/llc.h struct list_head node; node 159 include/net/netrom.h hlist_for_each_entry(__nr_neigh, node, list, neigh_node) node 162 include/net/netrom.h hlist_for_each_entry_safe(__nr_neigh, node, node2, list, neigh_node) node 165 include/net/netrom.h hlist_for_each_entry(__nr_node, node, list, node_node) node 168 include/net/netrom.h hlist_for_each_entry_safe(__nr_node, node, node2, list, node_node) node 15 include/net/pkt_cls.h int (*fn)(struct tcf_proto *, unsigned long node, struct tcf_walker *); node 653 include/net/sctp/sctp.h hlist_for_each_entry(epb, node, head, node) node 103 include/net/sctp/structs.h struct hlist_node node; node 1259 include/net/sctp/structs.h struct hlist_node node; node 315 include/net/sock.h node->pprev = NULL; node 387 include/net/sock.h hlist_for_each_entry(__sk, node, list, sk_node) node 389 include/net/sock.h if (__sk && ({ node = &(__sk)->sk_node; 1; })) \ node 390 include/net/sock.h hlist_for_each_entry_from(__sk, node, sk_node) node 392 include/net/sock.h if (__sk && ({ node = &(__sk)->sk_node; 1; })) \ node 393 include/net/sock.h hlist_for_each_entry_continue(__sk, node, sk_node) node 395 include/net/sock.h hlist_for_each_entry_safe(__sk, node, tmp, list, sk_node) node 397 include/net/sock.h hlist_for_each_entry(__sk, node, list, sk_bind_node) node 608 include/net/sock.h struct list_head node; node 124 include/net/x25.h struct list_head node; node 132 include/net/x25.h struct list_head node; node 165 include/net/x25.h struct list_head node; node 61 include/scsi/scsi_device.h struct list_head node; node 53 include/xen/xenbus.h const char *node; node 116 kernel/audit_tree.c size = offsetof(struct audit_chunk, owners) + count * sizeof(struct node); node 323 kernel/audit_tree.c struct node *p; node 433 kernel/audit_tree.c struct node *p; node 436 kernel/audit_tree.c p = list_entry(victim->chunks.next, struct node, list); node 462 kernel/audit_tree.c struct node *node = list_entry(p, struct node, list); node 464 kernel/audit_tree.c if (node->index & (1U<<31)) { node 471 kernel/audit_tree.c struct node *node; node 474 kernel/audit_tree.c node = list_entry(tree->chunks.next, struct node, list); node 477 kernel/audit_tree.c if (!(node->index & (1U<<31))) node 480 kernel/audit_tree.c chunk = find_chunk(node); node 484 kernel/audit_tree.c untag_chunk(chunk, node); node 537 kernel/audit_tree.c struct node *node; node 558 kernel/audit_tree.c list_for_each_entry(node, &tree->chunks, list) { node 559 kernel/audit_tree.c struct audit_chunk *chunk = find_chunk(node); node 562 kernel/audit_tree.c node->index |= 1U<<31; node 565 kernel/audit_tree.c node->index &= ~(1U<<31); node 662 kernel/audit_tree.c struct node *node; node 664 kernel/audit_tree.c list_for_each_entry(node, &tree->chunks, list) node 665 kernel/audit_tree.c node->index &= ~(1U<<31); node 786 kernel/audit_tree.c struct node *node; node 788 kernel/audit_tree.c list_for_each_entry(node, &tree->chunks, list) node 789 kernel/audit_tree.c node->index &= ~(1U<<31); node 328 kernel/cgroup.c struct hlist_node *node; node 347 kernel/cgroup.c hlist_for_each_entry(cg, node, hhead, hlist) { node 645 kernel/cgroup.c struct list_head *node; node 649 kernel/cgroup.c node = dentry->d_subdirs.next; node 650 kernel/cgroup.c while (node != &dentry->d_subdirs) { node 651 kernel/cgroup.c struct dentry *d = list_entry(node, struct dentry, d_u.d_child); node 652 kernel/cgroup.c list_del_init(node); node 664 kernel/cgroup.c node = dentry->d_subdirs.next; node 1041 kernel/cgroup.c struct hlist_node *node; node 1044 kernel/cgroup.c hlist_for_each_entry(cg, node, hhead, hlist) { node 2185 kernel/cpuset.c int node; /* node that zone z is on */ node 2191 kernel/cpuset.c node = zone_to_nid(z); node 2193 kernel/cpuset.c if (node_isset(node, current->mems_allowed)) node 2214 kernel/cpuset.c allowed = node_isset(node, cs->mems_allowed); node 2245 kernel/cpuset.c int node; /* node that zone z is on */ node 2249 kernel/cpuset.c node = zone_to_nid(z); node 2250 kernel/cpuset.c if (node_isset(node, current->mems_allowed)) node 2316 kernel/cpuset.c int node; node 2318 kernel/cpuset.c node = next_node(current->cpuset_mem_spread_rotor, current->mems_allowed); node 2319 kernel/cpuset.c if (node == MAX_NUMNODES) node 2320 kernel/cpuset.c node = first_node(current->mems_allowed); node 2321 kernel/cpuset.c current->cpuset_mem_spread_rotor = node; node 2322 kernel/cpuset.c return node; node 519 kernel/hrtimer.c timer = rb_entry(base->first, struct hrtimer, node); node 846 kernel/hrtimer.c entry = rb_entry(parent, struct hrtimer, node); node 875 kernel/hrtimer.c base->first = &timer->node; node 878 kernel/hrtimer.c rb_link_node(&timer->node, parent, link); node 879 kernel/hrtimer.c rb_insert_color(&timer->node, &base->active); node 909 kernel/hrtimer.c if (base->first == &timer->node) { node 910 kernel/hrtimer.c base->first = rb_next(&timer->node); node 915 kernel/hrtimer.c rb_erase(&timer->node, &base->active); node 1111 kernel/hrtimer.c timer = rb_entry(base->first, struct hrtimer, node); node 1221 kernel/hrtimer.c if (base->first == &timer->node && node 1302 kernel/hrtimer.c struct rb_node *node; node 1308 kernel/hrtimer.c while ((node = base->first)) { node 1311 kernel/hrtimer.c timer = rb_entry(node, struct hrtimer, node); node 1392 kernel/hrtimer.c struct rb_node *node; node 1415 kernel/hrtimer.c while ((node = base->first)) { node 1418 kernel/hrtimer.c timer = rb_entry(node, struct hrtimer, node); node 1600 kernel/hrtimer.c struct rb_node *node; node 1603 kernel/hrtimer.c while ((node = rb_first(&old_base->active))) { node 1604 kernel/hrtimer.c timer = rb_entry(node, struct hrtimer, node); node 293 kernel/kprobes.c struct hlist_node *node; node 297 kernel/kprobes.c hlist_for_each_entry_rcu(p, node, head, hlist) { node 438 kernel/kprobes.c struct hlist_node *node, *tmp; node 448 kernel/kprobes.c hlist_for_each_entry_safe(ri, node, tmp, head, hlist) { node 454 kernel/kprobes.c hlist_for_each_entry_safe(ri, node, tmp, &empty_rp, hlist) { node 1150 kernel/kprobes.c struct hlist_node *node; node 1159 kernel/kprobes.c hlist_for_each_entry_rcu(p, node, head, hlist) { node 1194 kernel/kprobes.c struct hlist_node *node; node 1206 kernel/kprobes.c hlist_for_each_entry_rcu(p, node, head, hlist) node 1221 kernel/kprobes.c struct hlist_node *node; node 1235 kernel/kprobes.c hlist_for_each_entry_rcu(p, node, head, hlist) { node 358 kernel/marker.c struct hlist_node *node; node 363 kernel/marker.c hlist_for_each_entry(e, node, head, hlist) { node 377 kernel/marker.c struct hlist_node *node; node 386 kernel/marker.c hlist_for_each_entry(e, node, head, hlist) { node 432 kernel/marker.c struct hlist_node *node; node 439 kernel/marker.c hlist_for_each_entry(e, node, head, hlist) { node 739 kernel/marker.c struct hlist_node *node; node 743 kernel/marker.c hlist_for_each_entry(entry, node, head, hlist) { node 827 kernel/marker.c struct hlist_node *node; node 834 kernel/marker.c hlist_for_each_entry(e, node, head, hlist) { node 322 kernel/pid.c hlist_add_head_rcu(&link->node, &pid->tasks[type]); node 335 kernel/pid.c hlist_del_rcu(&link->node); node 362 kernel/pid.c hlist_replace_rcu(&old->pids[type].node, &new->pids[type].node); node 372 kernel/pid.c result = hlist_entry(first, struct task_struct, pids[(type)].node); node 142 kernel/pm_qos_params.c struct requirement_list *node; node 148 kernel/pm_qos_params.c list_for_each_entry(node, node 151 kernel/pm_qos_params.c extreme_value, node->value); node 254 kernel/pm_qos_params.c struct requirement_list *node; node 258 kernel/pm_qos_params.c list_for_each_entry(node, node 260 kernel/pm_qos_params.c if (strcmp(node->name, name) == 0) { node 262 kernel/pm_qos_params.c node->value = node 265 kernel/pm_qos_params.c node->value = new_value; node 289 kernel/pm_qos_params.c struct requirement_list *node; node 293 kernel/pm_qos_params.c list_for_each_entry(node, node 295 kernel/pm_qos_params.c if (strcmp(node->name, name) == 0) { node 296 kernel/pm_qos_params.c kfree(node->name); node 297 kernel/pm_qos_params.c list_del(&node->list); node 298 kernel/pm_qos_params.c kfree(node); node 73 kernel/power/swsusp.c struct rb_node node; node 88 kernel/power/swsusp.c ext = container_of(*new, struct swsusp_extent, node); node 116 kernel/power/swsusp.c rb_link_node(&ext->node, parent, new); node 117 kernel/power/swsusp.c rb_insert_color(&ext->node, &swsusp_extents); node 148 kernel/power/swsusp.c struct rb_node *node; node 150 kernel/power/swsusp.c while ((node = swsusp_extents.rb_node)) { node 154 kernel/power/swsusp.c ext = container_of(node, struct swsusp_extent, node); node 155 kernel/power/swsusp.c rb_erase(node, &swsusp_extents); node 336 kernel/profile.c int node, cpu = (unsigned long)__cpu; node 342 kernel/profile.c node = cpu_to_node(cpu); node 345 kernel/profile.c page = alloc_pages_node(node, node 353 kernel/profile.c page = alloc_pages_node(node, node 535 kernel/profile.c int node = cpu_to_node(cpu); node 538 kernel/profile.c page = alloc_pages_node(node, node 545 kernel/profile.c page = alloc_pages_node(node, node 2473 kernel/sched.c struct hlist_node *node; node 2475 kernel/sched.c hlist_for_each_entry(notifier, node, &curr->preempt_notifiers, link) node 2484 kernel/sched.c struct hlist_node *node; node 2486 kernel/sched.c hlist_for_each_entry(notifier, node, &curr->preempt_notifiers, link) node 6989 kernel/sched.c n = (node + i) % nr_node_ids; node 6999 kernel/sched.c val = node_distance(node, n); node 7023 kernel/sched.c node_to_cpumask_ptr(nodemask, node); node 7030 kernel/sched.c node_set(node, used_nodes); node 7033 kernel/sched.c int next_node = find_next_best_node(node, &used_nodes); node 230 kernel/time/clockevents.c struct list_head *node, *tmp; node 241 kernel/time/clockevents.c list_for_each_safe(node, tmp, &clockevents_released) node 242 kernel/time/clockevents.c list_del(node); node 98 kernel/time/timer_list.c timer = rb_entry(curr, struct hrtimer, node); node 240 kernel/trace/ftrace.c hlist_for_each_entry_rcu(rec, t, head, node) { node 273 kernel/trace/ftrace.c hlist_for_each_entry_rcu(p, t, &ftrace_hash[key], node) { node 286 kernel/trace/ftrace.c hlist_add_head_rcu(&node->node, &ftrace_hash[key]); node 292 kernel/trace/ftrace.c hlist_del(&node->node); node 337 kernel/trace/ftrace.c struct dyn_ftrace *node; node 379 kernel/trace/ftrace.c node = ftrace_alloc_dyn_node(ip); node 380 kernel/trace/ftrace.c if (!node) node 383 kernel/trace/ftrace.c node->ip = ip; node 385 kernel/trace/ftrace.c ftrace_add_hash(node, key); node 724 kernel/trace/ftrace.c hlist_for_each_entry_safe(p, t, n, head, node) { node 745 kernel/trace/ftrace.c INIT_HLIST_NODE(&p->node); node 746 kernel/trace/ftrace.c hlist_add_head(&p->node, &temp_list); node 766 kernel/trace/ftrace.c hlist_for_each_entry_safe(p, t, n, &temp_list, node) { node 767 kernel/trace/ftrace.c hlist_del(&p->node); node 768 kernel/trace/ftrace.c INIT_HLIST_NODE(&p->node); node 769 kernel/trace/ftrace.c hlist_add_head(&p->node, head); node 86 lib/debugobjects.c hlist_add_head(&new->node, &obj_pool); node 98 lib/debugobjects.c struct hlist_node *node; node 102 lib/debugobjects.c hlist_for_each_entry(obj, node, &b->list, node) { node 124 lib/debugobjects.c obj = hlist_entry(obj_pool.first, typeof(*obj), node); node 129 lib/debugobjects.c hlist_del(&obj->node); node 131 lib/debugobjects.c hlist_add_head(&obj->node, &b->list); node 156 lib/debugobjects.c hlist_add_head(&obj->node, &obj_pool); node 175 lib/debugobjects.c struct hlist_node *node, *tmp; node 189 lib/debugobjects.c hlist_for_each_entry_safe(obj, node, tmp, &freelist, node) { node 190 lib/debugobjects.c hlist_del(&obj->node); node 506 lib/debugobjects.c hlist_del(&obj->node); node 519 lib/debugobjects.c struct hlist_node *node, *tmp; node 539 lib/debugobjects.c hlist_for_each_entry_safe(obj, node, tmp, &db->list, node) { node 555 lib/debugobjects.c hlist_del(&obj->node); node 556 lib/debugobjects.c hlist_add_head(&obj->node, &freelist); node 563 lib/debugobjects.c hlist_for_each_entry_safe(obj, node, tmp, &freelist, node) { node 564 lib/debugobjects.c hlist_del(&obj->node); node 873 lib/debugobjects.c hlist_add_head(&obj_static_pool[i].node, &obj_pool); node 79 lib/plist.c WARN_ON(!plist_node_empty(node)); node 82 lib/plist.c if (node->prio < iter->prio) node 84 lib/plist.c else if (node->prio == iter->prio) { node 92 lib/plist.c list_add_tail(&node->plist.prio_list, &iter->plist.prio_list); node 94 lib/plist.c list_add_tail(&node->plist.node_list, &iter->plist.node_list); node 109 lib/plist.c if (!list_empty(&node->plist.prio_list)) { node 110 lib/plist.c struct plist_node *next = plist_first(&node->plist); node 112 lib/plist.c list_move_tail(&next->plist.prio_list, &node->plist.prio_list); node 113 lib/plist.c list_del_init(&node->plist.prio_list); node 116 lib/plist.c list_del_init(&node->plist.node_list); node 58 lib/prio_tree.c node, struct vm_area_struct, shared.prio_tree_node); node 64 lib/prio_tree.c *radix = node->start; node 65 lib/prio_tree.c *heap = node->last; node 123 lib/prio_tree.c INIT_PRIO_TREE_NODE(node); node 126 lib/prio_tree.c node->left = first; node 127 lib/prio_tree.c first->parent = node; node 129 lib/prio_tree.c last = node; node 136 lib/prio_tree.c root->prio_tree_node = node; node 137 lib/prio_tree.c return node; node 146 lib/prio_tree.c INIT_PRIO_TREE_NODE(node); node 154 lib/prio_tree.c node->parent = node; node 155 lib/prio_tree.c root->prio_tree_node = node; node 157 lib/prio_tree.c node->parent = old->parent; node 159 lib/prio_tree.c old->parent->left = node; node 161 lib/prio_tree.c old->parent->right = node; node 165 lib/prio_tree.c node->left = old->left; node 166 lib/prio_tree.c old->left->parent = node; node 170 lib/prio_tree.c node->right = old->right; node 171 lib/prio_tree.c old->right->parent = node; node 190 lib/prio_tree.c struct prio_tree_node *cur, *res = node; node 195 lib/prio_tree.c get_index(root, node, &radix_index, &heap_index); node 199 lib/prio_tree.c return prio_tree_expand(root, node, heap_index); node 212 lib/prio_tree.c struct prio_tree_node *tmp = node; node 213 lib/prio_tree.c node = prio_tree_replace(root, cur, node); node 231 lib/prio_tree.c INIT_PRIO_TREE_NODE(node); node 232 lib/prio_tree.c cur->right = node; node 233 lib/prio_tree.c node->parent = cur; node 239 lib/prio_tree.c INIT_PRIO_TREE_NODE(node); node 240 lib/prio_tree.c cur->left = node; node 241 lib/prio_tree.c node->parent = cur; node 269 lib/prio_tree.c cur = node; node 304 lib/prio_tree.c while (cur != node) node 58 lib/radix-tree.c struct radix_tree_node *node; node 94 lib/radix-tree.c __set_bit(offset, node->tags[tag]); node 100 lib/radix-tree.c __clear_bit(offset, node->tags[tag]); node 106 lib/radix-tree.c return test_bit(offset, node->tags[tag]); node 137 lib/radix-tree.c if (node->tags[tag][idx]) node 176 lib/radix-tree.c struct radix_tree_node *node = node 184 lib/radix-tree.c tag_clear(node, 0, 0); node 185 lib/radix-tree.c tag_clear(node, 1, 0); node 186 lib/radix-tree.c node->slots[0] = NULL; node 187 lib/radix-tree.c node->count = 0; node 189 lib/radix-tree.c kmem_cache_free(radix_tree_node_cachep, node); node 195 lib/radix-tree.c call_rcu(&node->rcu_head, radix_tree_node_rcu_free); node 207 lib/radix-tree.c struct radix_tree_node *node; node 214 lib/radix-tree.c node = kmem_cache_alloc(radix_tree_node_cachep, gfp_mask); node 215 lib/radix-tree.c if (node == NULL) node 220 lib/radix-tree.c rtp->nodes[rtp->nr++] = node; node 222 lib/radix-tree.c kmem_cache_free(radix_tree_node_cachep, node); node 244 lib/radix-tree.c struct radix_tree_node *node; node 260 lib/radix-tree.c if (!(node = radix_tree_node_alloc(root))) node 264 lib/radix-tree.c node->slots[0] = radix_tree_indirect_to_ptr(root->rnode); node 269 lib/radix-tree.c tag_set(node, tag, 0); node 273 lib/radix-tree.c node->height = newheight; node 274 lib/radix-tree.c node->count = 1; node 275 lib/radix-tree.c node = radix_tree_ptr_to_indirect(node); node 276 lib/radix-tree.c rcu_assign_pointer(root->rnode, node); node 294 lib/radix-tree.c struct radix_tree_node *node = NULL, *slot; node 320 lib/radix-tree.c if (node) { node 321 lib/radix-tree.c rcu_assign_pointer(node->slots[offset], slot); node 322 lib/radix-tree.c node->count++; node 330 lib/radix-tree.c node = slot; node 331 lib/radix-tree.c slot = node->slots[offset]; node 339 lib/radix-tree.c if (node) { node 340 lib/radix-tree.c node->count++; node 341 lib/radix-tree.c rcu_assign_pointer(node->slots[offset], item); node 342 lib/radix-tree.c BUG_ON(tag_get(node, 0, offset)); node 343 lib/radix-tree.c BUG_ON(tag_get(node, 1, offset)); node 370 lib/radix-tree.c struct radix_tree_node *node, **slot; node 372 lib/radix-tree.c node = rcu_dereference(root->rnode); node 373 lib/radix-tree.c if (node == NULL) node 376 lib/radix-tree.c if (!radix_tree_is_indirect_ptr(node)) { node 381 lib/radix-tree.c node = radix_tree_indirect_to_ptr(node); node 383 lib/radix-tree.c height = node->height; node 391 lib/radix-tree.c (node->slots + ((index>>shift) & RADIX_TREE_MAP_MASK)); node 392 lib/radix-tree.c node = rcu_dereference(*slot); node 393 lib/radix-tree.c if (node == NULL) node 419 lib/radix-tree.c struct radix_tree_node *node, **slot; node 421 lib/radix-tree.c node = rcu_dereference(root->rnode); node 422 lib/radix-tree.c if (node == NULL) node 425 lib/radix-tree.c if (!radix_tree_is_indirect_ptr(node)) { node 428 lib/radix-tree.c return node; node 430 lib/radix-tree.c node = radix_tree_indirect_to_ptr(node); node 432 lib/radix-tree.c height = node->height; node 440 lib/radix-tree.c (node->slots + ((index>>shift) & RADIX_TREE_MAP_MASK)); node 441 lib/radix-tree.c node = rcu_dereference(*slot); node 442 lib/radix-tree.c if (node == NULL) node 449 lib/radix-tree.c return node; node 528 lib/radix-tree.c pathp->node = NULL; node 539 lib/radix-tree.c pathp[1].node = slot; node 549 lib/radix-tree.c while (pathp->node) { node 550 lib/radix-tree.c if (!tag_get(pathp->node, tag, pathp->offset)) node 552 lib/radix-tree.c tag_clear(pathp->node, tag, pathp->offset); node 553 lib/radix-tree.c if (any_tag_set(pathp->node, tag)) node 583 lib/radix-tree.c struct radix_tree_node *node; node 590 lib/radix-tree.c node = rcu_dereference(root->rnode); node 591 lib/radix-tree.c if (node == NULL) node 594 lib/radix-tree.c if (!radix_tree_is_indirect_ptr(node)) node 596 lib/radix-tree.c node = radix_tree_indirect_to_ptr(node); node 598 lib/radix-tree.c height = node->height; node 607 lib/radix-tree.c if (node == NULL) node 616 lib/radix-tree.c if (!tag_get(node, tag, offset)) node 619 lib/radix-tree.c int ret = tag_get(node, tag, offset); node 624 lib/radix-tree.c node = rcu_dereference(node->slots[offset]); node 739 lib/radix-tree.c struct radix_tree_node *node; node 743 lib/radix-tree.c node = rcu_dereference(root->rnode); node 744 lib/radix-tree.c if (!node) node 747 lib/radix-tree.c if (!radix_tree_is_indirect_ptr(node)) { node 750 lib/radix-tree.c results[0] = node; node 753 lib/radix-tree.c node = radix_tree_indirect_to_ptr(node); node 755 lib/radix-tree.c max_index = radix_tree_maxindex(node->height); node 764 lib/radix-tree.c slots_found = __lookup(node, (void ***)results + ret, cur_index, node 807 lib/radix-tree.c struct radix_tree_node *node; node 811 lib/radix-tree.c node = rcu_dereference(root->rnode); node 812 lib/radix-tree.c if (!node) node 815 lib/radix-tree.c if (!radix_tree_is_indirect_ptr(node)) { node 821 lib/radix-tree.c node = radix_tree_indirect_to_ptr(node); node 823 lib/radix-tree.c max_index = radix_tree_maxindex(node->height); node 832 lib/radix-tree.c slots_found = __lookup(node, results + ret, cur_index, node 927 lib/radix-tree.c struct radix_tree_node *node; node 936 lib/radix-tree.c node = rcu_dereference(root->rnode); node 937 lib/radix-tree.c if (!node) node 940 lib/radix-tree.c if (!radix_tree_is_indirect_ptr(node)) { node 943 lib/radix-tree.c results[0] = node; node 946 lib/radix-tree.c node = radix_tree_indirect_to_ptr(node); node 948 lib/radix-tree.c max_index = radix_tree_maxindex(node->height); node 957 lib/radix-tree.c slots_found = __lookup_tag(node, (void ***)results + ret, node 996 lib/radix-tree.c struct radix_tree_node *node; node 1005 lib/radix-tree.c node = rcu_dereference(root->rnode); node 1006 lib/radix-tree.c if (!node) node 1009 lib/radix-tree.c if (!radix_tree_is_indirect_ptr(node)) { node 1015 lib/radix-tree.c node = radix_tree_indirect_to_ptr(node); node 1017 lib/radix-tree.c max_index = radix_tree_maxindex(node->height); node 1026 lib/radix-tree.c slots_found = __lookup_tag(node, results + ret, node 1113 lib/radix-tree.c pathp->node = NULL; node 1122 lib/radix-tree.c pathp->node = slot; node 1135 lib/radix-tree.c if (tag_get(pathp->node, tag, pathp->offset)) node 1141 lib/radix-tree.c while (pathp->node) { node 1142 lib/radix-tree.c pathp->node->slots[pathp->offset] = NULL; node 1143 lib/radix-tree.c pathp->node->count--; node 1151 lib/radix-tree.c if (pathp->node->count) { node 1152 lib/radix-tree.c if (pathp->node == node 1159 lib/radix-tree.c to_free = pathp->node; node 1188 lib/radix-tree.c memset(node, 0, sizeof(struct radix_tree_node)); node 28 lib/rbtree.c struct rb_node *right = node->rb_right; node 29 lib/rbtree.c struct rb_node *parent = rb_parent(node); node 31 lib/rbtree.c if ((node->rb_right = right->rb_left)) node 32 lib/rbtree.c rb_set_parent(right->rb_left, node); node 33 lib/rbtree.c right->rb_left = node; node 39 lib/rbtree.c if (node == parent->rb_left) node 46 lib/rbtree.c rb_set_parent(node, right); node 51 lib/rbtree.c struct rb_node *left = node->rb_left; node 52 lib/rbtree.c struct rb_node *parent = rb_parent(node); node 54 lib/rbtree.c if ((node->rb_left = left->rb_right)) node 55 lib/rbtree.c rb_set_parent(left->rb_right, node); node 56 lib/rbtree.c left->rb_right = node; node 62 lib/rbtree.c if (node == parent->rb_right) node 69 lib/rbtree.c rb_set_parent(node, left); node 76 lib/rbtree.c while ((parent = rb_parent(node)) && rb_is_red(parent)) node 89 lib/rbtree.c node = gparent; node 94 lib/rbtree.c if (parent->rb_right == node) node 99 lib/rbtree.c parent = node; node 100 lib/rbtree.c node = tmp; node 114 lib/rbtree.c node = gparent; node 119 lib/rbtree.c if (parent->rb_left == node) node 124 lib/rbtree.c parent = node; node 125 lib/rbtree.c node = tmp; node 143 lib/rbtree.c while ((!node || rb_is_black(node)) && node != root->rb_node) node 145 lib/rbtree.c if (parent->rb_left == node) node 159 lib/rbtree.c node = parent; node 160 lib/rbtree.c parent = rb_parent(node); node 178 lib/rbtree.c node = root->rb_node; node 196 lib/rbtree.c node = parent; node 197 lib/rbtree.c parent = rb_parent(node); node 215 lib/rbtree.c node = root->rb_node; node 220 lib/rbtree.c if (node) node 221 lib/rbtree.c rb_set_black(node); node 229 lib/rbtree.c if (!node->rb_left) node 230 lib/rbtree.c child = node->rb_right; node 231 lib/rbtree.c else if (!node->rb_right) node 232 lib/rbtree.c child = node->rb_left; node 235 lib/rbtree.c struct rb_node *old = node, *left; node 237 lib/rbtree.c node = node->rb_right; node 238 lib/rbtree.c while ((left = node->rb_left) != NULL) node 239 lib/rbtree.c node = left; node 240 lib/rbtree.c child = node->rb_right; node 241 lib/rbtree.c parent = rb_parent(node); node 242 lib/rbtree.c color = rb_color(node); node 248 lib/rbtree.c parent = node; node 252 lib/rbtree.c node->rb_parent_color = old->rb_parent_color; node 253 lib/rbtree.c node->rb_right = old->rb_right; node 254 lib/rbtree.c node->rb_left = old->rb_left; node 259 lib/rbtree.c rb_parent(old)->rb_left = node; node 261 lib/rbtree.c rb_parent(old)->rb_right = node; node 263 lib/rbtree.c root->rb_node = node; node 265 lib/rbtree.c rb_set_parent(old->rb_left, node); node 267 lib/rbtree.c rb_set_parent(old->rb_right, node); node 271 lib/rbtree.c parent = rb_parent(node); node 272 lib/rbtree.c color = rb_color(node); node 278 lib/rbtree.c if (parent->rb_left == node) node 325 lib/rbtree.c if (rb_parent(node) == node) node 330 lib/rbtree.c if (node->rb_right) { node 331 lib/rbtree.c node = node->rb_right; node 332 lib/rbtree.c while (node->rb_left) node 333 lib/rbtree.c node=node->rb_left; node 334 lib/rbtree.c return node; node 343 lib/rbtree.c while ((parent = rb_parent(node)) && node == parent->rb_right) node 344 lib/rbtree.c node = parent; node 354 lib/rbtree.c if (rb_parent(node) == node) node 359 lib/rbtree.c if (node->rb_left) { node 360 lib/rbtree.c node = node->rb_left; node 361 lib/rbtree.c while (node->rb_right) node 362 lib/rbtree.c node=node->rb_right; node 363 lib/rbtree.c return node; node 368 lib/rbtree.c while ((parent = rb_parent(node)) && node == parent->rb_left) node 369 lib/rbtree.c node = parent; node 548 lib/zlib_deflate/deftree.c int node; /* new node being created */ node 571 lib/zlib_deflate/deftree.c node = s->heap[++(s->heap_len)] = (max_code < 2 ? ++max_code : 0); node 572 lib/zlib_deflate/deftree.c tree[node].Freq = 1; node 573 lib/zlib_deflate/deftree.c s->depth[node] = 0; node 574 lib/zlib_deflate/deftree.c s->opt_len--; if (stree) s->static_len -= stree[node].Len; node 587 lib/zlib_deflate/deftree.c node = elems; /* next internal node of the tree */ node 596 lib/zlib_deflate/deftree.c tree[node].Freq = tree[n].Freq + tree[m].Freq; node 597 lib/zlib_deflate/deftree.c s->depth[node] = (uch) (max(s->depth[n], s->depth[m]) + 1); node 598 lib/zlib_deflate/deftree.c tree[n].Dad = tree[m].Dad = (ush)node; node 602 lib/zlib_deflate/deftree.c node, tree[node].Freq, n, tree[n].Freq, m, tree[m].Freq); node 606 lib/zlib_deflate/deftree.c s->heap[SMALLEST] = node++; node 58 mm/allocpercpu.c int node = cpu_to_node(cpu); node 66 mm/allocpercpu.c if (node_online(node)) node 67 mm/allocpercpu.c pdata->ptrs[cpu] = kmalloc_node(size, gfp|__GFP_ZERO, node); node 1391 mm/hugetlb.c int node; node 1394 mm/hugetlb.c for_each_node_mask(node, cpuset_current_mems_allowed) node 1395 mm/hugetlb.c nr += array[node]; node 911 mm/memcontrol.c int node, zid; node 922 mm/memcontrol.c for_each_node_state(node, N_POSSIBLE) node 925 mm/memcontrol.c mz = mem_cgroup_zoneinfo(mem, node, zid); node 1065 mm/memcontrol.c int zone, tmp = node; node 1074 mm/memcontrol.c if (!node_state(node, N_NORMAL_MEMORY)) node 1080 mm/memcontrol.c mem->info.nodeinfo[node] = pn; node 1094 mm/memcontrol.c kfree(mem->info.nodeinfo[node]); node 1124 mm/memcontrol.c int node; node 1137 mm/memcontrol.c for_each_node_state(node, N_POSSIBLE) node 1138 mm/memcontrol.c if (alloc_mem_cgroup_per_zone_info(mem, node)) node 1143 mm/memcontrol.c for_each_node_state(node, N_POSSIBLE) node 1144 mm/memcontrol.c free_mem_cgroup_per_zone_info(mem, node); node 1160 mm/memcontrol.c int node; node 1163 mm/memcontrol.c for_each_node_state(node, N_POSSIBLE) node 1164 mm/memcontrol.c free_mem_cgroup_per_zone_info(mem, node); node 129 mm/memory_hotplug.c int node = pgdat->node_id; node 137 mm/memory_hotplug.c get_page_bootmem(node, page, NODE_INFO); node 148 mm/memory_hotplug.c get_page_bootmem(node, page, NODE_INFO); node 743 mm/memory_hotplug.c int ret, drain, retry_max, node; node 759 mm/memory_hotplug.c node = zone_to_nid(zone); node 770 mm/memory_hotplug.c if (nr_pages >= node_present_pages(node)) node 771 mm/memory_hotplug.c arg.status_change_nid = node; node 292 mm/mempolicy.c int node = first_node(pol->w.user_nodemask); node 294 mm/mempolicy.c if (node_isset(node, *nodes)) { node 295 mm/mempolicy.c pol->v.preferred_node = node; node 771 mm/mempolicy.c return alloc_pages_node(node, GFP_HIGHUSER_MOVABLE, 0); node 1408 mm/mempolicy.c return zone->node; node 2185 mm/mempolicy.c unsigned long node[MAX_NUMNODES]; node 2212 mm/mempolicy.c md->node[page_to_nid(page)]++; node 2322 mm/mempolicy.c if (md->node[n]) node 2323 mm/mempolicy.c seq_printf(m, " N%d=%lu", n, md->node[n]); node 836 mm/migrate.c int node; node 845 mm/migrate.c while (pm->node != MAX_NUMNODES && pm->page != p) node 848 mm/migrate.c if (pm->node == MAX_NUMNODES) node 853 mm/migrate.c return alloc_pages_node(pm->node, node 875 mm/migrate.c for (pp = pm; pp->node != MAX_NUMNODES; pp++) { node 906 mm/migrate.c if (err == pp->node) node 948 mm/migrate.c for ( ; pm->node != MAX_NUMNODES; pm++) { node 1059 mm/migrate.c int node; node 1061 mm/migrate.c if (get_user(node, nodes + i)) node 1065 mm/migrate.c if (!node_state(node, N_HIGH_MEMORY)) node 1069 mm/migrate.c if (!node_isset(node, task_nodes)) node 1072 mm/migrate.c pm[i].node = node; node 1074 mm/migrate.c pm[i].node = 0; /* anything to not match MAX_NUMNODES */ node 1077 mm/migrate.c pm[nr_pages].node = MAX_NUMNODES; node 56 mm/mm_init.c zone->node, zone->name); node 325 mm/page-writeback.c int node; node 328 mm/page-writeback.c for_each_node_state(node, N_HIGH_MEMORY) { node 330 mm/page-writeback.c &NODE_DATA(node)->node_zones[ZONE_HIGHMEM]; node 2089 mm/page_alloc.c if (!node_isset(node, *used_node_mask)) { node 2090 mm/page_alloc.c node_set(node, *used_node_mask); node 2091 mm/page_alloc.c return node; node 2101 mm/page_alloc.c val = node_distance(node, n); node 2104 mm/page_alloc.c val += (n < node); node 2141 mm/page_alloc.c j = build_zonelists_node(NODE_DATA(node), zonelist, j, node 2171 mm/page_alloc.c int pos, j, node; node 2180 mm/page_alloc.c node = node_order[j]; node 2181 mm/page_alloc.c z = &NODE_DATA(node)->node_zones[zone_type]; node 2257 mm/page_alloc.c int j, node, load; node 2281 mm/page_alloc.c while ((node = find_next_best_node(local_node, &used_mask)) >= 0) { node 2282 mm/page_alloc.c int distance = node_distance(local_node, node); node 2297 mm/page_alloc.c node_load[node] = load; node 2299 mm/page_alloc.c prev_node = node; node 2302 mm/page_alloc.c build_zonelists_in_node_order(pgdat, node); node 2304 mm/page_alloc.c node_order[j++] = node; /* remember order */ node 2339 mm/page_alloc.c int node, local_node; node 2356 mm/page_alloc.c for (node = local_node + 1; node < MAX_NUMNODES; node++) { node 2357 mm/page_alloc.c if (!node_online(node)) node 2359 mm/page_alloc.c j = build_zonelists_node(NODE_DATA(node), zonelist, j, node 2362 mm/page_alloc.c for (node = 0; node < local_node; node++) { node 2363 mm/page_alloc.c if (!node_online(node)) node 2365 mm/page_alloc.c j = build_zonelists_node(NODE_DATA(node), zonelist, j, node 2724 mm/page_alloc.c int node = cpu_to_node(cpu); node 2726 mm/page_alloc.c node_set_state(node, N_CPU); /* this node has a cpu */ node 2734 mm/page_alloc.c GFP_KERNEL, node); node 3454 mm/page_alloc.c zone->node = nid; node 3555 mm/page_alloc.c unsigned int node; node 3558 mm/page_alloc.c for_each_node_mask(node, node_possible_map) node 3559 mm/page_alloc.c highest = node; node 116 mm/prio_tree.c struct vm_area_struct *node, *head, *new_head; node 144 mm/prio_tree.c node = vma->shared.vm_set.head; node 151 mm/prio_tree.c node->shared.vm_set.head = new_head; node 152 mm/prio_tree.c new_head->shared.vm_set.head = node; node 154 mm/prio_tree.c node->shared.vm_set.head = NULL; node 29 mm/quicklist.c int node = numa_node_id(); node 30 mm/quicklist.c struct zone *zones = NODE_DATA(node)->node_zones; node 32 mm/quicklist.c node_to_cpumask_ptr(cpumask_on_node, node); node 902 mm/slab.c int node; node 904 mm/slab.c node = next_node(cpu_to_node(cpu), node_online_map); node 905 mm/slab.c if (node == MAX_NUMNODES) node 906 mm/slab.c node = first_node(node_online_map); node 908 mm/slab.c per_cpu(reap_node, cpu) = node; node 913 mm/slab.c int node = __get_cpu_var(reap_node); node 915 mm/slab.c node = next_node(node, node_online_map); node 916 mm/slab.c if (unlikely(node >= MAX_NUMNODES)) node 917 mm/slab.c node = first_node(node_online_map); node 918 mm/slab.c __get_cpu_var(reap_node) = node; node 956 mm/slab.c nc = kmalloc_node(memsize, GFP_KERNEL, node); node 1035 mm/slab.c ac_ptr = kmalloc_node(memsize, GFP_KERNEL, node); node 1038 mm/slab.c if (i == node || !node_online(i)) { node 1042 mm/slab.c ac_ptr[i] = alloc_arraycache(node, limit, 0xbaadf00d); node 1068 mm/slab.c struct kmem_list3 *rl3 = cachep->nodelists[node]; node 1080 mm/slab.c free_block(cachep, ac->entry, ac->avail, node); node 1091 mm/slab.c int node = __get_cpu_var(reap_node); node 1094 mm/slab.c struct array_cache *ac = l3->alien[node]; node 1097 mm/slab.c __drain_alien_cache(cachep, ac, node); node 1126 mm/slab.c int node; node 1128 mm/slab.c node = numa_node_id(); node 1134 mm/slab.c if (likely(slabp->nodeid == node)) node 1137 mm/slab.c l3 = cachep->nodelists[node]; node 1161 mm/slab.c int node = cpu_to_node(cpu); node 1162 mm/slab.c node_to_cpumask_ptr(mask, node); node 1172 mm/slab.c l3 = cachep->nodelists[node]; node 1182 mm/slab.c free_block(cachep, nc->entry, nc->avail, node); node 1192 mm/slab.c shared->avail, node); node 1215 mm/slab.c l3 = cachep->nodelists[node]; node 1226 mm/slab.c int node = cpu_to_node(cpu); node 1242 mm/slab.c if (!cachep->nodelists[node]) { node 1243 mm/slab.c l3 = kmalloc_node(memsize, GFP_KERNEL, node); node 1255 mm/slab.c cachep->nodelists[node] = l3; node 1258 mm/slab.c spin_lock_irq(&cachep->nodelists[node]->list_lock); node 1259 mm/slab.c cachep->nodelists[node]->free_limit = node 1260 mm/slab.c (1 + nr_cpus_node(node)) * node 1262 mm/slab.c spin_unlock_irq(&cachep->nodelists[node]->list_lock); node 1274 mm/slab.c nc = alloc_arraycache(node, cachep->limit, node 1279 mm/slab.c shared = alloc_arraycache(node, node 1288 mm/slab.c alien = alloc_alien_cache(node, cachep->limit); node 1296 mm/slab.c l3 = cachep->nodelists[node]; node 1413 mm/slab.c int node; node 1415 mm/slab.c for_each_online_node(node) { node 1416 mm/slab.c cachep->nodelists[node] = &initkmem_list3[index + node]; node 1417 mm/slab.c cachep->nodelists[node]->next_reap = jiffies + node 1434 mm/slab.c int node; node 1475 mm/slab.c node = numa_node_id(); node 1482 mm/slab.c cache_cache.nodelists[node] = &initkmem_list3[CACHE_CACHE + node]; node 2088 mm/slab.c int node; node 2089 mm/slab.c for_each_online_node(node) { node 2090 mm/slab.c cachep->nodelists[node] = node 2092 mm/slab.c GFP_KERNEL, node); node 2093 mm/slab.c BUG_ON(!cachep->nodelists[node]); node 2094 mm/slab.c kmem_list3_init(cachep->nodelists[node]); node 2414 mm/slab.c assert_spin_locked(&cachep->nodelists[node]->list_lock); node 2433 mm/slab.c int node = numa_node_id(); node 2437 mm/slab.c spin_lock(&cachep->nodelists[node]->list_lock); node 2438 mm/slab.c free_block(cachep, ac->entry, ac->avail, node); node 2439 mm/slab.c spin_unlock(&cachep->nodelists[node]->list_lock); node 2446 mm/slab.c int node; node 2450 mm/slab.c for_each_online_node(node) { node 2451 mm/slab.c l3 = cachep->nodelists[node]; node 2456 mm/slab.c for_each_online_node(node) { node 2457 mm/slab.c l3 = cachep->nodelists[node]; node 2459 mm/slab.c drain_array(cachep, l3, l3->shared, 1, node); node 2954 mm/slab.c int node; node 2958 mm/slab.c node = numa_node_id(); node 2969 mm/slab.c l3 = cachep->nodelists[node]; node 3007 mm/slab.c node); node 3026 mm/slab.c x = cache_grow(cachep, flags | GFP_THISNODE, node, NULL); node 3489 mm/slab.c l3 = cachep->nodelists[node]; node 3491 mm/slab.c check_spinlock_acquired_node(cachep, node); node 3493 mm/slab.c slab_put_obj(cachep, slabp, objp, node); node 3526 mm/slab.c int node = numa_node_id(); node 3533 mm/slab.c l3 = cachep->nodelists[node]; node 3548 mm/slab.c free_block(cachep, ac->entry, batchcount, node); node 3676 mm/slab.c return kmem_cache_alloc_node(cachep, flags, node); node 3682 mm/slab.c return __do_kmalloc_node(size, flags, node, node 3690 mm/slab.c return __do_kmalloc_node(size, flags, node, caller); node 3696 mm/slab.c return __do_kmalloc_node(size, flags, node, NULL); node 3810 mm/slab.c int node; node 3815 mm/slab.c for_each_online_node(node) { node 3818 mm/slab.c new_alien = alloc_alien_cache(node, cachep->limit); node 3825 mm/slab.c new_shared = alloc_arraycache(node, node 3834 mm/slab.c l3 = cachep->nodelists[node]; node 3842 mm/slab.c shared->avail, node); node 3849 mm/slab.c l3->free_limit = (1 + nr_cpus_node(node)) * node 3856 mm/slab.c l3 = kmalloc_node(sizeof(struct kmem_list3), GFP_KERNEL, node); node 3868 mm/slab.c l3->free_limit = (1 + nr_cpus_node(node)) * node 3870 mm/slab.c cachep->nodelists[node] = l3; node 3877 mm/slab.c node--; node 3878 mm/slab.c while (node >= 0) { node 3879 mm/slab.c if (cachep->nodelists[node]) { node 3880 mm/slab.c l3 = cachep->nodelists[node]; node 3885 mm/slab.c cachep->nodelists[node] = NULL; node 3887 mm/slab.c node--; node 4027 mm/slab.c free_block(cachep, ac->entry, tofree, node); node 4052 mm/slab.c int node = numa_node_id(); node 4068 mm/slab.c l3 = searchp->nodelists[node]; node 4072 mm/slab.c drain_array(searchp, l3, cpu_cache_get(searchp), 0, node); node 4083 mm/slab.c drain_array(searchp, l3, l3->shared, 0, node); node 4161 mm/slab.c int node; node 4166 mm/slab.c for_each_online_node(node) { node 4167 mm/slab.c l3 = cachep->nodelists[node]; node 4393 mm/slab.c int node; node 4405 mm/slab.c for_each_online_node(node) { node 4406 mm/slab.c l3 = cachep->nodelists[node]; node 238 mm/slob.c if (node != -1) node 239 mm/slob.c page = alloc_pages_node(node, gfp, order); node 327 mm/slob.c if (node != -1 && page_to_nid(&sp->page) != node) node 352 mm/slob.c b = slob_new_page(gfp & ~__GFP_ZERO, 0, node); node 471 mm/slob.c m = slob_alloc(size + align, gfp, align, node); node 479 mm/slob.c ret = slob_new_page(gfp | __GFP_COMP, get_order(size), node); node 573 mm/slob.c b = slob_alloc(c->size, flags, c->align, node); node 575 mm/slob.c b = slob_new_page(flags, get_order(c->size), node); node 223 mm/slub.c return s->node[node]; node 828 mm/slub.c struct kmem_cache_node *n = get_node(s, node); node 835 mm/slub.c struct kmem_cache_node *n = get_node(s, node); node 850 mm/slub.c struct kmem_cache_node *n = get_node(s, node); node 1065 mm/slub.c if (node == -1) node 1068 mm/slub.c return alloc_pages_node(node, flags, order); node 1078 mm/slub.c page = alloc_slab_page(flags | __GFP_NOWARN | __GFP_NORETRY, node, node 1086 mm/slub.c page = alloc_slab_page(flags, node, oo); node 1119 mm/slub.c flags & (GFP_RECLAIM_MASK | GFP_CONSTRAINT_MASK), node); node 1348 mm/slub.c int searchnode = (node == -1) ? numa_node_id() : node; node 1477 mm/slub.c if (node != -1 && c->node != node) node 1514 mm/slub.c if (unlikely(!node_match(c, node))) node 1529 mm/slub.c c->node = page_to_nid(c->page); node 1539 mm/slub.c new = get_partial(s, gfpflags, node); node 1549 mm/slub.c new = new_slab(s, gfpflags, node); node 1571 mm/slub.c c->node = -1; node 1596 mm/slub.c if (unlikely(!c->freelist || !node_match(c, node))) node 1598 mm/slub.c object = __slab_alloc(s, gfpflags, node, addr, c); node 1622 mm/slub.c return slab_alloc(s, gfpflags, node, __builtin_return_address(0)); node 1717 mm/slub.c if (likely(page == c->page && c->node >= 0)) { node 1908 mm/slub.c c->node = 0; node 2084 mm/slub.c page = new_slab(kmalloc_caches, gfpflags, node); node 2087 mm/slub.c if (page_to_nid(page) != node) { node 2089 mm/slub.c "node %d\n", node); node 2098 mm/slub.c kmalloc_caches->node[node] = n; node 2104 mm/slub.c inc_slabs_node(kmalloc_caches, node, page->objects); node 2119 mm/slub.c int node; node 2121 mm/slub.c for_each_node_state(node, N_NORMAL_MEMORY) { node 2122 mm/slub.c struct kmem_cache_node *n = s->node[node]; node 2125 mm/slub.c s->node[node] = NULL; node 2131 mm/slub.c int node; node 2139 mm/slub.c for_each_node_state(node, N_NORMAL_MEMORY) { node 2142 mm/slub.c if (local_node == node) node 2147 mm/slub.c node); node 2151 mm/slub.c gfpflags, node); node 2159 mm/slub.c s->node[node] = n; node 2427 mm/slub.c int node; node 2433 mm/slub.c for_each_node_state(node, N_NORMAL_MEMORY) { node 2434 mm/slub.c struct kmem_cache_node *n = get_node(s, node); node 2437 mm/slub.c if (n->nr_partial || slabs_node(s, node)) node 2667 mm/slub.c struct page *page = alloc_pages_node(node, flags | __GFP_COMP, node 2682 mm/slub.c return kmalloc_large_node(size, flags, node); node 2689 mm/slub.c return slab_alloc(s, flags, node, __builtin_return_address(0)); node 2762 mm/slub.c int node; node 2776 mm/slub.c for_each_node_state(node, N_NORMAL_MEMORY) { node 2777 mm/slub.c n = get_node(s, node); node 2866 mm/slub.c s->node[offline_node] = NULL; node 2906 mm/slub.c s->node[nid] = n; node 3223 mm/slub.c return kmalloc_large_node(size, gfpflags, node); node 3230 mm/slub.c return slab_alloc(s, gfpflags, node, caller); node 3346 mm/slub.c int node; node 3355 mm/slub.c for_each_node_state(node, N_NORMAL_MEMORY) { node 3356 mm/slub.c struct kmem_cache_node *n = get_node(s, node); node 3569 mm/slub.c int node; node 3578 mm/slub.c for_each_node_state(node, N_NORMAL_MEMORY) { node 3579 mm/slub.c struct kmem_cache_node *n = get_node(s, node); node 3663 mm/slub.c int node; node 3679 mm/slub.c if (!c || c->node < 0) node 3691 mm/slub.c nodes[c->node] += x; node 3693 mm/slub.c per_cpu[c->node]++; node 3698 mm/slub.c for_each_node_state(node, N_NORMAL_MEMORY) { node 3699 mm/slub.c struct kmem_cache_node *n = get_node(s, node); node 3710 mm/slub.c nodes[node] += x; node 3714 mm/slub.c for_each_node_state(node, N_NORMAL_MEMORY) { node 3715 mm/slub.c struct kmem_cache_node *n = get_node(s, node); node 3724 mm/slub.c nodes[node] += x; node 3729 mm/slub.c for_each_node_state(node, N_NORMAL_MEMORY) node 3730 mm/slub.c if (nodes[node]) node 3732 mm/slub.c node, nodes[node]); node 3740 mm/slub.c int node; node 3742 mm/slub.c for_each_online_node(node) { node 3743 mm/slub.c struct kmem_cache_node *n = get_node(s, node); node 4467 mm/slub.c int node; node 4471 mm/slub.c for_each_online_node(node) { node 4472 mm/slub.c struct kmem_cache_node *n = get_node(s, node); node 43 mm/sparse-vmemmap.c return __alloc_bootmem_node(NODE_DATA(node), size, align, goal); node 51 mm/sparse-vmemmap.c struct page *page = alloc_pages_node(node, node 57 mm/sparse-vmemmap.c return __earlyonly_bootmem_alloc(node, size, size, node 67 mm/sparse-vmemmap.c if (actual_node != node) node 77 mm/sparse-vmemmap.c void *p = vmemmap_alloc_block(PAGE_SIZE, node); node 90 mm/sparse-vmemmap.c void *p = vmemmap_alloc_block(PAGE_SIZE, node); node 102 mm/sparse-vmemmap.c void *p = vmemmap_alloc_block(PAGE_SIZE, node); node 114 mm/sparse-vmemmap.c void *p = vmemmap_alloc_block(PAGE_SIZE, node); node 133 mm/sparse-vmemmap.c pgd = vmemmap_pgd_populate(addr, node); node 136 mm/sparse-vmemmap.c pud = vmemmap_pud_populate(pgd, addr, node); node 139 mm/sparse-vmemmap.c pmd = vmemmap_pmd_populate(pud, addr, node); node 142 mm/sparse-vmemmap.c pte = vmemmap_pte_populate(pmd, addr, node); node 145 mm/sparse-vmemmap.c vmemmap_verify(pte, node, addr, addr + PAGE_SIZE); node 240 mm/vmalloc.c area = kmalloc_node(sizeof(*area), gfp_mask & GFP_RECLAIM_MASK, node); node 327 mm/vmalloc.c return __get_vm_area_node(size, flags, VMALLOC_START, VMALLOC_END, node, node 503 mm/vmalloc.c PAGE_KERNEL, node, caller); node 508 mm/vmalloc.c node); node 521 mm/vmalloc.c if (node < 0) node 524 mm/vmalloc.c page = alloc_pages_node(node, gfp_mask, 0); node 571 mm/vmalloc.c node, gfp_mask, caller); node 576 mm/vmalloc.c return __vmalloc_area_node(area, gfp_mask, prot, node, caller); node 639 mm/vmalloc.c node, __builtin_return_address(0)); node 379 mm/vmstat.c if (z->node == numa_node_id()) node 401 mm/vmstat.c loff_t node = *pos; node 403 mm/vmstat.c pgdat && node; node 405 mm/vmstat.c --node; node 156 net/802/garp.c attr = rb_entry(parent, struct garp_attr, node); node 176 net/802/garp.c attr = rb_entry(parent, struct garp_attr, node); node 183 net/802/garp.c rb_link_node(&new->node, parent, p); node 184 net/802/garp.c rb_insert_color(&new->node, &app->gid); node 205 net/802/garp.c rb_erase(&attr->node, &app->gid); node 385 net/802/garp.c struct rb_node *node, *next; node 388 net/802/garp.c for (node = rb_first(&app->gid); node 389 net/802/garp.c next = node ? rb_next(node) : NULL, node != NULL; node 390 net/802/garp.c node = next) { node 391 net/802/garp.c attr = rb_entry(node, struct garp_attr, node); node 36 net/802/psnap.c list_for_each_entry_rcu(p, &snap_list, node) { node 143 net/802/psnap.c list_add_rcu(&proto->node, &snap_list); node 158 net/802/psnap.c list_del_rcu(&proto->node); node 150 net/appletalk/atalk_proc.c struct hlist_node *node; node 152 net/appletalk/atalk_proc.c sk_for_each(s, node, &atalk_sockets) node 94 net/appletalk/ddp.c struct hlist_node *node; node 97 net/appletalk/ddp.c sk_for_each(s, node, &atalk_sockets) { node 142 net/appletalk/ddp.c struct hlist_node *node; node 146 net/appletalk/ddp.c sk_for_each(s, node, &atalk_sockets) { node 390 net/appletalk/ddp.c if (node != ATADDR_BCAST && node 391 net/appletalk/ddp.c iface->address.s_node != node && node 392 net/appletalk/ddp.c node != ATADDR_ANYNODE) node 408 net/appletalk/ddp.c if ((node == ATADDR_BCAST || node 409 net/appletalk/ddp.c node == ATADDR_ANYNODE || node 410 net/appletalk/ddp.c iface->address.s_node == node) && node 416 net/appletalk/ddp.c if (node == ATADDR_ANYNODE && net != ATADDR_ANYNET && node 1092 net/appletalk/ddp.c struct hlist_node *node; node 1094 net/appletalk/ddp.c sk_for_each(s, node, &atalk_sockets) { node 223 net/atm/common.c struct hlist_node *node, *tmp; node 227 net/atm/common.c sk_for_each_safe(s, node, tmp, head) { node 269 net/atm/common.c struct hlist_node *node; node 273 net/atm/common.c sk_for_each(s, node, head) { node 1010 net/atm/lec.c struct hlist_node *node; node 1020 net/atm/lec.c struct hlist_node *e = state->node; node 1034 net/atm/lec.c state->node = e; node 1127 net/atm/lec.c state->node = SEQ_START_TOKEN; node 1163 net/atm/lec.c struct lec_arp_table *entry = hlist_entry(state->node, struct lec_arp_table, next); node 1500 net/atm/lec.c struct hlist_node *node; node 1517 net/atm/lec.c hlist_for_each_entry(entry, node, &priv->lec_arp_tables[i], next) { node 1558 net/atm/lec.c struct hlist_node *node; node 1565 net/atm/lec.c hlist_for_each_entry(rulla, node, &priv->lec_arp_tables[i], next) { node 1600 net/atm/lec.c hlist_for_each_entry(rulla, node, &priv->lec_no_forward, next) { node 1629 net/atm/lec.c hlist_for_each_entry(rulla, node, &priv->lec_arp_empty_ones, next) { node 1658 net/atm/lec.c hlist_for_each_entry(rulla, node, &priv->mcast_fwds, next) { node 1696 net/atm/lec.c struct hlist_node *node, *next; node 1708 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_tables[i], next) { node 1715 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_empty_ones, next) { node 1723 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->lec_no_forward, next) { node 1731 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->mcast_fwds, next) { node 1748 net/atm/lec.c struct hlist_node *node; node 1757 net/atm/lec.c hlist_for_each_entry(entry, node, head, next) { node 1851 net/atm/lec.c struct hlist_node *node, *next; node 1862 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_tables[i], next) { node 2022 net/atm/lec.c struct hlist_node *node, *next; node 2029 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_tables[i], next) { node 2053 net/atm/lec.c struct hlist_node *node, *next; node 2070 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_empty_ones, next) { node 2114 net/atm/lec.c hlist_for_each_entry(tmp, node, &priv->lec_arp_tables[i], next) { node 2154 net/atm/lec.c struct hlist_node *node; node 2229 net/atm/lec.c hlist_for_each_entry(entry, node, &priv->lec_arp_tables[i], next) { node 2305 net/atm/lec.c struct hlist_node *node; node 2313 net/atm/lec.c hlist_for_each_entry(entry, node, &priv->lec_arp_tables[i], next) { node 2340 net/atm/lec.c struct hlist_node *node; node 2346 net/atm/lec.c hlist_for_each_entry(entry, node, &priv->lec_arp_tables[i], next) { node 2396 net/atm/lec.c struct hlist_node *node, *next; node 2406 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_tables[i], next) { node 2417 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_empty_ones, next) { node 2426 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->lec_no_forward, next) { node 2435 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->mcast_fwds, next) { node 2453 net/atm/lec.c struct hlist_node *node, *next; node 2467 net/atm/lec.c hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_empty_ones, next) { node 218 net/atm/signaling.c struct hlist_node *node; node 232 net/atm/signaling.c sk_for_each(s, node, head) { node 84 net/ax25/af_ax25.c struct hlist_node *node; node 91 net/ax25/af_ax25.c ax25_for_each(s, node, &ax25_list) { node 161 net/ax25/af_ax25.c struct hlist_node *node; node 164 net/ax25/af_ax25.c ax25_for_each(s, node, &ax25_list) { node 190 net/ax25/af_ax25.c struct hlist_node *node; node 193 net/ax25/af_ax25.c ax25_for_each(s, node, &ax25_list) { node 216 net/ax25/af_ax25.c struct hlist_node *node; node 219 net/ax25/af_ax25.c ax25_for_each(s, node, &ax25_list) { node 251 net/ax25/af_ax25.c struct hlist_node *node; node 254 net/ax25/af_ax25.c ax25_for_each(s, node, &ax25_list) { node 1861 net/ax25/af_ax25.c struct hlist_node *node; node 1865 net/ax25/af_ax25.c ax25_for_each(ax25, node, &ax25_list) { node 42 net/ax25/ax25_ds_subr.c struct hlist_node *node; node 83 net/ax25/ax25_ds_subr.c ax25_for_each(ax25o, node, &ax25_list) { node 162 net/ax25/ax25_ds_subr.c struct hlist_node *node; node 165 net/ax25/ax25_ds_subr.c ax25_for_each(ax25, node, &ax25_list) node 74 net/ax25/ax25_ds_timer.c struct hlist_node *node; node 85 net/ax25/ax25_ds_timer.c ax25_for_each(ax25, node, &ax25_list) { node 200 net/ax25/ax25_iface.c struct hlist_node *node; node 203 net/ax25/ax25_iface.c hlist_for_each_entry(lf, node, &ax25_linkfail_list, lf_node) node 56 net/ax25/ax25_uid.c struct hlist_node *node; node 59 net/ax25/ax25_uid.c ax25_uid_for_each(ax25_uid, node, &ax25_uid_list) { node 76 net/ax25/ax25_uid.c struct hlist_node *node; node 84 net/ax25/ax25_uid.c ax25_uid_for_each(ax25_uid, node, &ax25_uid_list) { node 123 net/ax25/ax25_uid.c ax25_uid_for_each(ax25_uid, node, &ax25_uid_list) { node 150 net/ax25/ax25_uid.c struct hlist_node *node; node 154 net/ax25/ax25_uid.c ax25_uid_for_each(pt, node, &ax25_uid_list) { node 218 net/ax25/ax25_uid.c struct hlist_node *node; node 222 net/ax25/ax25_uid.c ax25_uid_for_each(ax25_uid, node, &ax25_uid_list) { node 94 net/bluetooth/hci_sock.c struct hlist_node *node; node 99 net/bluetooth/hci_sock.c sk_for_each(sk, node, &hci_sk_list.head) { node 671 net/bluetooth/hci_sock.c struct hlist_node *node; node 675 net/bluetooth/hci_sock.c sk_for_each(sk, node, &hci_sk_list.head) { node 521 net/bluetooth/l2cap.c struct hlist_node *node; node 522 net/bluetooth/l2cap.c sk_for_each(sk, node, &l2cap_sk_list.head) node 536 net/bluetooth/l2cap.c struct hlist_node *node; node 538 net/bluetooth/l2cap.c sk_for_each(sk, node, &l2cap_sk_list.head) { node 552 net/bluetooth/l2cap.c return node ? sk : sk1; node 2137 net/bluetooth/l2cap.c struct hlist_node *node; node 2146 net/bluetooth/l2cap.c sk_for_each(sk, node, &l2cap_sk_list.head) { node 2424 net/bluetooth/l2cap.c struct hlist_node *node; node 2429 net/bluetooth/l2cap.c sk_for_each(sk, node, &l2cap_sk_list.head) { node 128 net/bluetooth/rfcomm/sock.c struct hlist_node *node; node 130 net/bluetooth/rfcomm/sock.c sk_for_each(sk, node, &rfcomm_sk_list.head) { node 136 net/bluetooth/rfcomm/sock.c return node ? sk : NULL; node 145 net/bluetooth/rfcomm/sock.c struct hlist_node *node; node 147 net/bluetooth/rfcomm/sock.c sk_for_each(sk, node, &rfcomm_sk_list.head) { node 161 net/bluetooth/rfcomm/sock.c return node ? sk : sk1; node 902 net/bluetooth/rfcomm/sock.c struct hlist_node *node; node 907 net/bluetooth/rfcomm/sock.c sk_for_each(sk, node, &rfcomm_sk_list.head) { node 289 net/bluetooth/sco.c struct hlist_node *node; node 291 net/bluetooth/sco.c sk_for_each(sk, node, &sco_sk_list.head) node 305 net/bluetooth/sco.c struct hlist_node *node; node 309 net/bluetooth/sco.c sk_for_each(sk, node, &sco_sk_list.head) { node 324 net/bluetooth/sco.c return node ? sk : sk1; node 899 net/bluetooth/sco.c struct hlist_node *node; node 904 net/bluetooth/sco.c sk_for_each(sk, node, &sco_sk_list.head) { node 190 net/core/skbuff.c skb = kmem_cache_alloc_node(cache, gfp_mask & ~__GFP_DMA, node); node 196 net/core/skbuff.c gfp_mask, node); node 255 net/core/skbuff.c int node = dev->dev.parent ? dev_to_node(dev->dev.parent) : -1; node 258 net/core/skbuff.c skb = __alloc_skb(length + NET_SKB_PAD, gfp_mask, 0, node); node 268 net/core/skbuff.c int node = dev->dev.parent ? dev_to_node(dev->dev.parent) : -1; node 271 net/core/skbuff.c page = alloc_pages_node(node, gfp_mask, 0); node 2092 net/core/sock.c list_add(&prot->node, &proto_list); node 2119 net/core/sock.c list_del(&prot->node); node 2217 net/core/sock.c proto_seq_printf(seq, list_entry(v, struct proto, node)); node 179 net/decnet/af_decnet.c struct hlist_node *node; node 184 net/decnet/af_decnet.c sk_for_each(sk, node, &dn_sk_hash[dn_ntohs(port) & DN_SK_HASH_MASK]) { node 378 net/decnet/af_decnet.c struct hlist_node *node; node 382 net/decnet/af_decnet.c sk_for_each(sk, node, list) { node 418 net/decnet/af_decnet.c struct hlist_node *node; node 422 net/decnet/af_decnet.c sk_for_each(sk, node, &dn_sk_hash[dn_ntohs(cb->dst_port) & DN_SK_HASH_MASK]) { node 667 net/decnet/af_decnet.c unsigned short node, area; node 669 net/decnet/af_decnet.c node = addr & 0x03ff; node 671 net/decnet/af_decnet.c sprintf(buf, "%hd.%hd", area, node); node 470 net/decnet/dn_table.c struct hlist_node *node; node 485 net/decnet/dn_table.c hlist_for_each_entry(tb, node, &dn_fib_table_hash[h], hlist) { node 814 net/decnet/dn_table.c struct hlist_node *node; node 825 net/decnet/dn_table.c hlist_for_each_entry_rcu(t, node, &dn_fib_table_hash[h], hlist) { node 871 net/decnet/dn_table.c struct hlist_node *node; node 875 net/decnet/dn_table.c hlist_for_each_entry(tb, node, &dn_fib_table_hash[h], hlist) node 894 net/decnet/dn_table.c struct hlist_node *node, *next; node 899 net/decnet/dn_table.c hlist_for_each_entry_safe(t, node, next, &dn_fib_table_hash[h], node 90 net/decnet/sysctl_net_decnet.c __u16 area, node; node 109 net/decnet/sysctl_net_decnet.c node = *str++ - '0'; node 111 net/decnet/sysctl_net_decnet.c node *= 10; node 112 net/decnet/sysctl_net_decnet.c node += (*str++ - '0'); node 115 net/decnet/sysctl_net_decnet.c node *= 10; node 116 net/decnet/sysctl_net_decnet.c node += (*str++ - '0'); node 119 net/decnet/sysctl_net_decnet.c node *= 10; node 120 net/decnet/sysctl_net_decnet.c node += (*str++ - '0'); node 123 net/decnet/sysctl_net_decnet.c if ((node > 1023) || (area > 63)) node 129 net/decnet/sysctl_net_decnet.c *addr = dn_htons((area << 10) | node); node 782 net/econet/af_econet.c struct hlist_node *node; node 784 net/econet/af_econet.c sk_for_each(sk, node, &econet_sklist) { node 96 net/ipv4/fib_frontend.c struct hlist_node *node; node 106 net/ipv4/fib_frontend.c hlist_for_each_entry_rcu(tb, node, head, tb_hlist) { node 136 net/ipv4/fib_frontend.c struct hlist_node *node; node 142 net/ipv4/fib_frontend.c hlist_for_each_entry(tb, node, head, tb_hlist) node 631 net/ipv4/fib_frontend.c struct hlist_node *node; node 645 net/ipv4/fib_frontend.c hlist_for_each_entry(tb, node, head, tb_hlist) { node 1008 net/ipv4/fib_frontend.c struct hlist_node *node, *tmp; node 1011 net/ipv4/fib_frontend.c hlist_for_each_entry_safe(tb, node, tmp, head, tb_hlist) { node 1012 net/ipv4/fib_frontend.c hlist_del(node); node 119 net/ipv4/fib_hash.c struct hlist_node *node, *n; node 122 net/ipv4/fib_hash.c hlist_for_each_entry_safe(f, node, n, &old_ht[i], fn_hash) { node 255 net/ipv4/fib_hash.c struct hlist_node *node; node 260 net/ipv4/fib_hash.c hlist_for_each_entry(f, node, head, fn_hash) { node 282 net/ipv4/fib_hash.c struct hlist_node *node; node 297 net/ipv4/fib_hash.c hlist_for_each_entry(f, node, &fz->fz_hash[0], fn_hash) { node 359 net/ipv4/fib_hash.c struct hlist_node *node; node 362 net/ipv4/fib_hash.c hlist_for_each_entry(f, node, head, fn_hash) { node 632 net/ipv4/fib_hash.c struct hlist_node *node, *n; node 636 net/ipv4/fib_hash.c hlist_for_each_entry_safe(f, node, n, head, fn_hash) { node 688 net/ipv4/fib_hash.c struct hlist_node *node; node 694 net/ipv4/fib_hash.c hlist_for_each_entry(f, node, head, fn_hash) { node 845 net/ipv4/fib_hash.c struct hlist_node *node; node 848 net/ipv4/fib_hash.c hlist_for_each_entry(fn,node,iter->hash_head,fn_hash) { node 884 net/ipv4/fib_hash.c struct hlist_node *node = &fn->fn_hash; node 885 net/ipv4/fib_hash.c hlist_for_each_entry_continue(fn, node, fn_hash) { node 902 net/ipv4/fib_hash.c struct hlist_node *node; node 910 net/ipv4/fib_hash.c hlist_for_each_entry(fn, node, iter->hash_head, fn_hash) { node 927 net/ipv4/fib_hash.c hlist_for_each_entry(fn, node, iter->hash_head, fn_hash) { node 223 net/ipv4/fib_semantics.c struct hlist_node *node; node 230 net/ipv4/fib_semantics.c hlist_for_each_entry(fi, node, head, fib_hash) { node 255 net/ipv4/fib_semantics.c struct hlist_node *node; node 263 net/ipv4/fib_semantics.c hlist_for_each_entry(nh, node, head, nh_hash) { node 644 net/ipv4/fib_semantics.c struct hlist_node *node, *n; node 647 net/ipv4/fib_semantics.c hlist_for_each_entry_safe(fi, node, n, head, fib_hash) { node 662 net/ipv4/fib_semantics.c struct hlist_node *node, *n; node 665 net/ipv4/fib_semantics.c hlist_for_each_entry_safe(fi, node, n, lhead, fib_lhash) { node 1043 net/ipv4/fib_semantics.c struct hlist_node *node; node 1049 net/ipv4/fib_semantics.c hlist_for_each_entry(fi, node, head, fib_lhash) { node 1067 net/ipv4/fib_semantics.c struct hlist_node *node; node 1073 net/ipv4/fib_semantics.c hlist_for_each_entry(nh, node, head, nh_hash) { node 1124 net/ipv4/fib_semantics.c struct hlist_node *node; node 1136 net/ipv4/fib_semantics.c hlist_for_each_entry(nh, node, head, nh_hash) { node 92 net/ipv4/fib_trie.c #define NODE_TYPE(node) ((node)->parent & NODE_TYPE_MASK) node 127 net/ipv4/fib_trie.c struct node *child[0]; node 152 net/ipv4/fib_trie.c struct node *trie; node 161 net/ipv4/fib_trie.c static struct node *resize(struct trie *t, struct tnode *tn); node 170 net/ipv4/fib_trie.c return (struct tnode *)(node->parent & ~NODE_TYPE_MASK); node 175 net/ipv4/fib_trie.c struct tnode *ret = node_parent(node); node 186 net/ipv4/fib_trie.c node->parent = (unsigned long)ptr | NODE_TYPE(node); node 189 net/ipv4/fib_trie.c static inline struct node *tnode_get_child(struct tnode *tn, unsigned int i) node 196 net/ipv4/fib_trie.c static inline struct node *tnode_get_child_rcu(struct tnode *tn, unsigned int i) node 198 net/ipv4/fib_trie.c struct node *ret = tnode_get_child(tn, i); node 370 net/ipv4/fib_trie.c (sizeof(struct node *) << tn->bits); node 410 net/ipv4/fib_trie.c size_t sz = sizeof(struct tnode) + (sizeof(struct node *) << bits); node 423 net/ipv4/fib_trie.c (unsigned long) (sizeof(struct node) << bits)); node 454 net/ipv4/fib_trie.c struct node *chi = tn->child[i]; node 481 net/ipv4/fib_trie.c static struct node *resize(struct trie *t, struct tnode *tn) node 504 net/ipv4/fib_trie.c struct node *n; node 664 net/ipv4/fib_trie.c struct node *n; node 677 net/ipv4/fib_trie.c return (struct node *) tn; node 724 net/ipv4/fib_trie.c put_child(t, tn, 2*i, (struct node *) left); node 725 net/ipv4/fib_trie.c put_child(t, tn, 2*i+1, (struct node *) right); node 731 net/ipv4/fib_trie.c struct node *node = tnode_get_child(oldtnode, i); node 736 net/ipv4/fib_trie.c if (node == NULL) node 741 net/ipv4/fib_trie.c if (IS_LEAF(node) || ((struct tnode *) node)->pos > node 743 net/ipv4/fib_trie.c if (tkey_extract_bits(node->key, node 746 net/ipv4/fib_trie.c put_child(t, tn, 2*i, node); node 748 net/ipv4/fib_trie.c put_child(t, tn, 2*i+1, node); node 753 net/ipv4/fib_trie.c inode = (struct tnode *) node; node 826 net/ipv4/fib_trie.c struct node *left, *right; node 857 net/ipv4/fib_trie.c put_child(t, tn, i/2, (struct node *)newn); node 911 net/ipv4/fib_trie.c struct hlist_node *node; node 914 net/ipv4/fib_trie.c hlist_for_each_entry_rcu(li, node, head, hlist) node 934 net/ipv4/fib_trie.c struct hlist_node *node; node 939 net/ipv4/fib_trie.c hlist_for_each_entry(li, node, head, hlist) { node 959 net/ipv4/fib_trie.c struct node *n; node 986 net/ipv4/fib_trie.c static struct node *trie_rebalance(struct trie *t, struct tnode *tn) node 992 net/ipv4/fib_trie.c while (tn != NULL && (tp = node_parent((struct node *)tn)) != NULL) { node 998 net/ipv4/fib_trie.c (struct node *)tn, wasfull); node 1000 net/ipv4/fib_trie.c tp = node_parent((struct node *) tn); node 1010 net/ipv4/fib_trie.c return (struct node *)tn; node 1019 net/ipv4/fib_trie.c struct node *n; node 1105 net/ipv4/fib_trie.c node_set_parent((struct node *)l, tp); node 1108 net/ipv4/fib_trie.c put_child(t, (struct tnode *)tp, cindex, (struct node *)l); node 1135 net/ipv4/fib_trie.c node_set_parent((struct node *)tn, tp); node 1138 net/ipv4/fib_trie.c put_child(t, tn, missbit, (struct node *)l); node 1144 net/ipv4/fib_trie.c (struct node *)tn); node 1146 net/ipv4/fib_trie.c rcu_assign_pointer(t->trie, (struct node *)tn); node 1340 net/ipv4/fib_trie.c struct hlist_node *node; node 1342 net/ipv4/fib_trie.c hlist_for_each_entry_rcu(li, node, hhead, hlist) { node 1371 net/ipv4/fib_trie.c struct node *n; node 1544 net/ipv4/fib_trie.c struct tnode *parent = node_parent((struct node *) pn); node 1571 net/ipv4/fib_trie.c struct tnode *tp = node_parent((struct node *) l); node 1691 net/ipv4/fib_trie.c struct hlist_node *node, *tmp; node 1694 net/ipv4/fib_trie.c hlist_for_each_entry_safe(li, node, tmp, lih, hlist) { node 1735 net/ipv4/fib_trie.c c = (struct node *) p; node 1756 net/ipv4/fib_trie.c struct node *c = (struct node *) l; node 1915 net/ipv4/fib_trie.c struct hlist_node *node; node 1922 net/ipv4/fib_trie.c hlist_for_each_entry_rcu(li, node, &l->list, hlist) { node 2039 net/ipv4/fib_trie.c static struct node *fib_trie_get_next(struct fib_trie_iter *iter) node 2053 net/ipv4/fib_trie.c struct node *n = tnode_get_child_rcu(tn, cindex); node 2072 net/ipv4/fib_trie.c p = node_parent_rcu((struct node *)tn); node 2084 net/ipv4/fib_trie.c static struct node *fib_trie_get_first(struct fib_trie_iter *iter, node 2087 net/ipv4/fib_trie.c struct node *n; node 2111 net/ipv4/fib_trie.c struct node *n; node 2184 net/ipv4/fib_trie.c bytes += sizeof(struct node *) * pointers; node 2229 net/ipv4/fib_trie.c struct hlist_node *node; node 2232 net/ipv4/fib_trie.c hlist_for_each_entry_rcu(tb, node, head, tb_hlist) { node 2265 net/ipv4/fib_trie.c static struct node *fib_trie_get_idx(struct seq_file *seq, loff_t pos) node 2274 net/ipv4/fib_trie.c struct hlist_node *node; node 2277 net/ipv4/fib_trie.c hlist_for_each_entry_rcu(tb, node, head, tb_hlist) { node 2278 net/ipv4/fib_trie.c struct node *n; node 2307 net/ipv4/fib_trie.c struct node *n; node 2392 net/ipv4/fib_trie.c struct node *n = v; node 2409 net/ipv4/fib_trie.c struct hlist_node *node; node 2415 net/ipv4/fib_trie.c hlist_for_each_entry_rcu(li, node, &l->list, hlist) { node 2561 net/ipv4/fib_trie.c struct hlist_node *node; node 2570 net/ipv4/fib_trie.c hlist_for_each_entry_rcu(li, node, &l->list, hlist) { node 57 net/ipv4/inet_connection_sock.c struct hlist_node *node; node 67 net/ipv4/inet_connection_sock.c sk_for_each_bound(sk2, node, &tb->owners) { node 82 net/ipv4/inet_connection_sock.c return node != NULL; node 94 net/ipv4/inet_connection_sock.c struct hlist_node *node; node 111 net/ipv4/inet_connection_sock.c inet_bind_bucket_for_each(tb, node, &head->chain) node 139 net/ipv4/inet_connection_sock.c inet_bind_bucket_for_each(tb, node, &head->chain) node 726 net/ipv4/inet_diag.c struct hlist_node *node; node 729 net/ipv4/inet_diag.c sk_for_each(sk, node, &hashinfo->listening_hash[i]) { node 783 net/ipv4/inet_diag.c struct hlist_node *node; node 794 net/ipv4/inet_diag.c sk_for_each(sk, node, &head->chain) { node 818 net/ipv4/inet_diag.c inet_twsk_for_each(tw, node, node 42 net/ipv4/inet_hashtables.c hlist_add_head(&tb->node, &head->chain); node 53 net/ipv4/inet_hashtables.c __hlist_del(&tb->node); node 155 net/ipv4/inet_hashtables.c const struct hlist_node *node; node 158 net/ipv4/inet_hashtables.c sk_for_each(sk, node, head) { node 226 net/ipv4/inet_hashtables.c const struct hlist_node *node; node 236 net/ipv4/inet_hashtables.c sk_for_each(sk, node, &head->chain) { node 243 net/ipv4/inet_hashtables.c sk_for_each(sk, node, &head->twchain) { node 275 net/ipv4/inet_hashtables.c const struct hlist_node *node; node 282 net/ipv4/inet_hashtables.c sk_for_each(sk2, node, &head->twchain) { node 296 net/ipv4/inet_hashtables.c sk_for_each(sk2, node, &head->chain) { node 434 net/ipv4/inet_hashtables.c struct hlist_node *node; node 451 net/ipv4/inet_hashtables.c inet_bind_bucket_for_each(tb, node, &head->chain) { node 147 net/ipv4/inet_timewait_sock.c struct hlist_node *node; node 160 net/ipv4/inet_timewait_sock.c inet_twsk_for_each_inmate(tw, node, &twdr->cells[slot]) { node 373 net/ipv4/inet_timewait_sock.c struct hlist_node *node, *safe; node 376 net/ipv4/inet_timewait_sock.c inet_twsk_for_each_inmate_safe(tw, node, safe, node 419 net/ipv4/inet_timewait_sock.c struct hlist_node *node; node 429 net/ipv4/inet_timewait_sock.c sk_for_each(sk, node, &head->twchain) { node 188 net/ipv4/inetpeer.c struct inet_peer **nodep, *node, *l, *r; node 193 net/ipv4/inetpeer.c node = *nodep; node 194 net/ipv4/inetpeer.c l = node->avl_left; node 195 net/ipv4/inetpeer.c r = node->avl_right; node 205 net/ipv4/inetpeer.c node->avl_left = lr; /* lr: RH or RH+1 */ node 206 net/ipv4/inetpeer.c node->avl_right = r; /* r: RH */ node 207 net/ipv4/inetpeer.c node->avl_height = lrh + 1; /* RH+1 or RH+2 */ node 209 net/ipv4/inetpeer.c l->avl_right = node; /* node: RH+1 or RH+2 */ node 210 net/ipv4/inetpeer.c l->avl_height = node->avl_height + 1; node 215 net/ipv4/inetpeer.c node->avl_left = lrr; /* lrr: RH or RH-1 */ node 216 net/ipv4/inetpeer.c node->avl_right = r; /* r: RH */ node 217 net/ipv4/inetpeer.c node->avl_height = rh + 1; /* node: RH+1 */ node 222 net/ipv4/inetpeer.c lr->avl_right = node; /* node: RH+1 */ node 233 net/ipv4/inetpeer.c node->avl_right = rl; /* rl: LH or LH+1 */ node 234 net/ipv4/inetpeer.c node->avl_left = l; /* l: LH */ node 235 net/ipv4/inetpeer.c node->avl_height = rlh + 1; /* LH+1 or LH+2 */ node 237 net/ipv4/inetpeer.c r->avl_left = node; /* node: LH+1 or LH+2 */ node 238 net/ipv4/inetpeer.c r->avl_height = node->avl_height + 1; node 243 net/ipv4/inetpeer.c node->avl_right = rll; /* rll: LH or LH-1 */ node 244 net/ipv4/inetpeer.c node->avl_left = l; /* l: LH */ node 245 net/ipv4/inetpeer.c node->avl_height = lh + 1; /* node: LH+1 */ node 250 net/ipv4/inetpeer.c rl->avl_left = node; /* node: LH+1 */ node 255 net/ipv4/inetpeer.c node->avl_height = (lh > rh ? lh : rh) + 1; node 113 net/ipv4/raw.c struct hlist_node *node; node 115 net/ipv4/raw.c sk_for_each_from(sk, node) { node 858 net/ipv4/raw.c struct hlist_node *node; node 860 net/ipv4/raw.c sk_for_each(sk, node, &state->h->ht[state->bucket]) node 1878 net/ipv4/tcp_ipv4.c struct hlist_node *node; node 1921 net/ipv4/tcp_ipv4.c sk_for_each_from(sk, node) { node 1972 net/ipv4/tcp_ipv4.c struct hlist_node *node; node 1981 net/ipv4/tcp_ipv4.c sk_for_each(sk, node, &tcp_hashinfo.ehash[st->bucket].chain) { node 1990 net/ipv4/tcp_ipv4.c inet_twsk_for_each(tw, node, node 2010 net/ipv4/tcp_ipv4.c struct hlist_node *node; node 2042 net/ipv4/tcp_ipv4.c sk_for_each_from(sk, node) { node 132 net/ipv4/udp.c struct hlist_node *node; node 134 net/ipv4/udp.c sk_for_each(sk2, node, &udptable[udp_hashfn(net, num)]) node 219 net/ipv4/udp.c struct hlist_node *node; node 224 net/ipv4/udp.c sk_for_each(sk, node, &udptable[udp_hashfn(net, hnum)]) { node 292 net/ipv4/udp.c struct hlist_node *node; node 296 net/ipv4/udp.c sk_for_each_from(s, node) { node 1508 net/ipv4/udp.c struct hlist_node *node; node 1509 net/ipv4/udp.c sk_for_each(sk, node, state->hashtable + state->bucket) { node 33 net/ipv6/inet6_connection_sock.c const struct hlist_node *node; node 40 net/ipv6/inet6_connection_sock.c sk_for_each_bound(sk2, node, &tb->owners) { node 51 net/ipv6/inet6_connection_sock.c return node != NULL; node 66 net/ipv6/inet6_hashtables.c const struct hlist_node *node; node 77 net/ipv6/inet6_hashtables.c sk_for_each(sk, node, &head->chain) { node 83 net/ipv6/inet6_hashtables.c sk_for_each(sk, node, &head->twchain) { node 102 net/ipv6/inet6_hashtables.c const struct hlist_node *node; node 107 net/ipv6/inet6_hashtables.c sk_for_each(sk, node, node 175 net/ipv6/inet6_hashtables.c const struct hlist_node *node; node 182 net/ipv6/inet6_hashtables.c sk_for_each(sk2, node, &head->twchain) { node 195 net/ipv6/inet6_hashtables.c sk_for_each(sk2, node, &head->chain) { node 229 net/ipv6/ip6_fib.c struct hlist_node *node; node 237 net/ipv6/ip6_fib.c hlist_for_each_entry_rcu(tb, node, head, tb6_hlist) { node 353 net/ipv6/ip6_fib.c struct hlist_node *node; node 387 net/ipv6/ip6_fib.c hlist_for_each_entry(tb, node, head, tb6_hlist) { node 1052 net/ipv6/ip6_fib.c w->root = w->node = NULL; node 1054 net/ipv6/ip6_fib.c } else if (w->node == fn) { node 1056 net/ipv6/ip6_fib.c w->node = pn; node 1064 net/ipv6/ip6_fib.c if (w->node == fn) { node 1065 net/ipv6/ip6_fib.c w->node = child; node 1223 net/ipv6/ip6_fib.c fn = w->node; node 1236 net/ipv6/ip6_fib.c w->node = FIB6_SUBTREE(fn); node 1243 net/ipv6/ip6_fib.c w->node = fn->left; node 1250 net/ipv6/ip6_fib.c w->node = fn->right; node 1268 net/ipv6/ip6_fib.c w->node = pn; node 1282 net/ipv6/ip6_fib.c w->leaf = w->node->leaf; node 1297 net/ipv6/ip6_fib.c w->node = w->root; node 1365 net/ipv6/ip6_fib.c struct hlist_node *node; node 1372 net/ipv6/ip6_fib.c hlist_for_each_entry_rcu(table, node, head, tb6_hlist) { node 71 net/ipv6/raw.c struct hlist_node *node; node 74 net/ipv6/raw.c sk_for_each_from(sk, node) node 63 net/ipv6/udp.c struct hlist_node *node; node 68 net/ipv6/udp.c sk_for_each(sk, node, &udptable[udp_hashfn(net, hnum)]) { node 336 net/ipv6/udp.c struct hlist_node *node; node 340 net/ipv6/udp.c sk_for_each_from(s, node) { node 101 net/ipx/af_ipx.c struct ipx_interface, node); node 172 net/ipx/af_ipx.c list_for_each_entry(i, &ipx_interfaces, node) node 199 net/ipx/af_ipx.c list_for_each_entry(i, &ipx_interfaces, node) node 231 net/ipx/af_ipx.c struct hlist_node *node; node 233 net/ipx/af_ipx.c sk_for_each(s, node, &intrfc->if_sklist) node 262 net/ipx/af_ipx.c struct hlist_node *node; node 267 net/ipx/af_ipx.c sk_for_each(s, node, &intrfc->if_sklist) { node 271 net/ipx/af_ipx.c !memcmp(ipx_node, ipxs->node, IPX_NODE_LEN)) node 285 net/ipx/af_ipx.c struct hlist_node *node, *t; node 292 net/ipx/af_ipx.c sk_for_each_safe(s, node, t, &intrfc->if_sklist) { node 306 net/ipx/af_ipx.c list_del(&intrfc->node); node 345 net/ipx/af_ipx.c list_for_each_entry_safe(i, tmp, &ipx_interfaces, node) node 363 net/ipx/af_ipx.c list_for_each_entry_safe(i, tmp, &ipx_interfaces, node) node 385 net/ipx/af_ipx.c int is_broadcast = !memcmp(ipx->ipx_dest.node, ipx_broadcast_node, node 388 net/ipx/af_ipx.c struct hlist_node *node; node 393 net/ipx/af_ipx.c sk_for_each(s, node, &intrfc->if_sklist) { node 397 net/ipx/af_ipx.c (is_broadcast || !memcmp(ipx->ipx_dest.node, node 398 net/ipx/af_ipx.c ipxs->node, IPX_NODE_LEN))) { node 449 net/ipx/af_ipx.c struct hlist_node *node; node 454 net/ipx/af_ipx.c sk_for_each(sk, node, &intrfc->if_sklist) node 625 net/ipx/af_ipx.c !memcmp(intrfc->if_node, node, IPX_NODE_LEN)) { node 634 net/ipx/af_ipx.c if (!memcmp(ipx_broadcast_node, node, IPX_NODE_LEN)) { node 667 net/ipx/af_ipx.c if (!memcmp(ipx_broadcast_node, node, IPX_NODE_LEN)) node 670 net/ipx/af_ipx.c memcpy(dest_node, &(node[IPX_NODE_LEN-addr_len]), addr_len); node 736 net/ipx/af_ipx.c if (!memcmp(ipx_broadcast_node, ipx->ipx_dest.node, IPX_NODE_LEN) || node 737 net/ipx/af_ipx.c !memcmp(intrfc->if_node, ipx->ipx_dest.node, IPX_NODE_LEN)) { node 843 net/ipx/af_ipx.c list_for_each_entry(ifcs, &ipx_interfaces, node) { node 874 net/ipx/af_ipx.c list_add_tail(&intrfc->node, &ipx_interfaces); node 1468 net/ipx/af_ipx.c memcpy(ipxs->node, intrfc->if_node, IPX_NODE_LEN); node 1470 net/ipx/af_ipx.c memcpy(ipxs->node, addr->sipx_node, IPX_NODE_LEN); node 1473 net/ipx/af_ipx.c if (ipxitf_find_internal_socket(intrfc, ipxs->node, node 1486 net/ipx/af_ipx.c memcpy(ipxs->node, intrfc->if_node, IPX_NODE_LEN); node 1567 net/ipx/af_ipx.c memcpy(ipxs->dest_addr.node, addr->sipx_node, IPX_NODE_LEN); node 1602 net/ipx/af_ipx.c memcpy(sipx.sipx_node, addr->node, IPX_NODE_LEN); node 1607 net/ipx/af_ipx.c memcpy(sipx.sipx_node, ipxs->node, IPX_NODE_LEN); node 1745 net/ipx/af_ipx.c memcpy(usipx->sipx_node, ipxs->dest_addr.node, IPX_NODE_LEN); node 1815 net/ipx/af_ipx.c memcpy(sipx->sipx_node, ipx->ipx_source.node, IPX_NODE_LEN); node 20 net/ipx/ipx_proc.c list_for_each_entry(i, &ipx_interfaces, node) node 32 net/ipx/ipx_proc.c if (i->node.next != &ipx_interfaces) node 33 net/ipx/ipx_proc.c rc = list_entry(i->node.next, struct ipx_interface, node); node 97 net/ipx/ipx_proc.c rc = list_entry(ipx_routes.next, struct ipx_route, node); node 105 net/ipx/ipx_proc.c if (r->node.next != &ipx_routes) node 106 net/ipx/ipx_proc.c rc = list_entry(r->node.next, struct ipx_route, node); node 114 net/ipx/ipx_proc.c list_for_each_entry(r, &ipx_routes, node) node 171 net/ipx/ipx_proc.c struct hlist_node *node; node 174 net/ipx/ipx_proc.c list_for_each_entry(i, &ipx_interfaces, node) { node 176 net/ipx/ipx_proc.c sk_for_each(s, node, &i->if_sklist) { node 183 net/ipx/ipx_proc.c if (node) node 265 net/ipx/ipx_proc.c ipxs->node[0], ipxs->node[1], ipxs->node[2], ipxs->node[3], node 266 net/ipx/ipx_proc.c ipxs->node[4], ipxs->node[5], ntohs(ipxs->port)); node 276 net/ipx/ipx_proc.c ipxs->dest_addr.node[0], ipxs->dest_addr.node[1], node 277 net/ipx/ipx_proc.c ipxs->dest_addr.node[2], ipxs->dest_addr.node[3], node 278 net/ipx/ipx_proc.c ipxs->dest_addr.node[4], ipxs->dest_addr.node[5], node 37 net/ipx/ipx_route.c list_for_each_entry(r, &ipx_routes, node) node 68 net/ipx/ipx_route.c list_add(&rt->node, &ipx_routes); node 78 net/ipx/ipx_route.c if (!node) { node 82 net/ipx/ipx_route.c memcpy(rt->ir_router_node, node, IPX_NODE_LEN); node 98 net/ipx/ipx_route.c list_for_each_entry_safe(r, tmp, &ipx_routes, node) node 100 net/ipx/ipx_route.c list_del(&r->node); node 127 net/ipx/ipx_route.c list_for_each_entry_safe(r, tmp, &ipx_routes, node) node 133 net/ipx/ipx_route.c list_del(&r->node); node 160 net/ipx/ipx_route.c r->ir_router_node : ipx->ipx_dest.node); node 217 net/ipx/ipx_route.c memcpy(ipx->ipx_source.node, ipxs->node, IPX_NODE_LEN); node 223 net/ipx/ipx_route.c memcpy(ipx->ipx_source.node, intrfc->if_node, IPX_NODE_LEN); node 226 net/ipx/ipx_route.c memcpy(ipx->ipx_source.node, ipxs->intrfc->if_node, node 232 net/ipx/ipx_route.c memcpy(ipx->ipx_dest.node, usipx->sipx_node, IPX_NODE_LEN); node 248 net/ipx/ipx_route.c rt->ir_router_node : ipx->ipx_dest.node); node 62 net/irda/discovery.c discovery_t *discovery, *node; node 78 net/irda/discovery.c node = discovery; node 83 net/irda/discovery.c if ((node->data.saddr == new->data.saddr) && node 84 net/irda/discovery.c ((node->data.daddr == new->data.daddr) || node 85 net/irda/discovery.c (strcmp(node->data.info, new->data.info) == 0))) node 90 net/irda/discovery.c hashbin_remove_this(cachelog, (irda_queue_t *) node); node 92 net/irda/discovery.c if (get_unaligned((__u16 *)node->data.hints) == get_unaligned((__u16 *)new->data.hints)) node 94 net/irda/discovery.c new->firststamp = node->firststamp; node 95 net/irda/discovery.c kfree(node); node 128 net/irda/irias_object.c struct ias_object *node; node 134 net/irda/irias_object.c node = hashbin_remove_this(irias_objects, (irda_queue_t *) obj); node 135 net/irda/irias_object.c if (!node) node 156 net/irda/irias_object.c struct ias_attrib *node; node 163 net/irda/irias_object.c node = hashbin_remove_this(obj->attribs, (irda_queue_t *) attrib); node 164 net/irda/irias_object.c if (!node) node 168 net/irda/irias_object.c __irias_delete_attrib(node); node 175 net/irda/irias_object.c node = (struct ias_attrib *) hashbin_get_first(obj->attribs); node 176 net/irda/irias_object.c if (cleanobject && !node) node 100 net/iucv/af_iucv.c struct hlist_node *node; node 102 net/iucv/af_iucv.c sk_for_each(sk, node, &iucv_sk_list.head) node 978 net/iucv/af_iucv.c struct hlist_node *node; node 989 net/iucv/af_iucv.c sk_for_each(sk, node, &iucv_sk_list.head) node 261 net/key/af_key.c struct hlist_node *node; node 272 net/key/af_key.c sk_for_each(sk, node, &pfkey_table) { node 3649 net/key/af_key.c struct hlist_node *node; node 3656 net/key/af_key.c sk_for_each(s, node, &pfkey_table) node 69 net/lapb/lapb_iface.c if (lapb->node.next) { node 70 net/lapb/lapb_iface.c list_del(&lapb->node); node 80 net/lapb/lapb_iface.c list_add(&lapb->node, &lapb_list); node 90 net/lapb/lapb_iface.c lapb = list_entry(entry, struct lapb_cb, node); node 492 net/llc/llc_conn.c struct hlist_node *node; node 495 net/llc/llc_conn.c sk_for_each(rc, node, &sap->sk_list.list) { node 538 net/llc/llc_conn.c struct hlist_node *node; node 541 net/llc/llc_conn.c sk_for_each(rc, node, &sap->sk_list.list) { node 54 net/llc/llc_core.c list_add_tail(&sap->node, &llc_sap_list); node 66 net/llc/llc_core.c list_del(&sap->node); node 74 net/llc/llc_core.c list_for_each_entry(sap, &llc_sap_list, node) node 38 net/llc/llc_proc.c struct hlist_node *node; node 42 net/llc/llc_proc.c sap = list_entry(sap_entry, struct llc_sap, node); node 45 net/llc/llc_proc.c sk_for_each(sk, node, &sap->sk_list.list) { node 87 net/llc/llc_proc.c if (sap->node.next == &llc_sap_list) node 89 net/llc/llc_proc.c sap = list_entry(sap->node.next, struct llc_sap, node); node 312 net/llc/llc_sap.c struct hlist_node *node; node 315 net/llc/llc_sap.c sk_for_each(rc, node, &sap->sk_list.list) { node 344 net/llc/llc_sap.c struct hlist_node *node; node 347 net/llc/llc_sap.c sk_for_each(sk, node, &sap->sk_list.list) { node 80 net/llc/llc_station.c struct list_head node; /* node in station->ev_q.list */ node 298 net/mac80211/mesh.h hlist_for_each_entry_rcu(node, p, &x->hash_buckets[i], list) node 77 net/mac80211/mesh_pathtbl.c struct mpath_node *node; node 82 net/mac80211/mesh_pathtbl.c hlist_for_each_entry_rcu(node, n, bucket, list) { node 83 net/mac80211/mesh_pathtbl.c mpath = node->mpath; node 104 net/mac80211/mesh_pathtbl.c struct mpath_node *node; node 109 net/mac80211/mesh_pathtbl.c hlist_for_each_entry_rcu(node, n, bucket, list) { node 110 net/mac80211/mesh_pathtbl.c mpath = node->mpath; node 137 net/mac80211/mesh_pathtbl.c struct mpath_node *node; node 142 net/mac80211/mesh_pathtbl.c for_each_mesh_entry(mesh_paths, p, node, i) { node 143 net/mac80211/mesh_pathtbl.c if (sdata && node->mpath->sdata != sdata) node 146 net/mac80211/mesh_pathtbl.c if (MPATH_EXPIRED(node->mpath)) { node 147 net/mac80211/mesh_pathtbl.c spin_lock_bh(&node->mpath->state_lock); node 148 net/mac80211/mesh_pathtbl.c if (MPATH_EXPIRED(node->mpath)) node 149 net/mac80211/mesh_pathtbl.c node->mpath->flags &= ~MESH_PATH_ACTIVE; node 150 net/mac80211/mesh_pathtbl.c spin_unlock_bh(&node->mpath->state_lock); node 152 net/mac80211/mesh_pathtbl.c return node->mpath; node 171 net/mac80211/mesh_pathtbl.c struct mpath_node *node, *new_node; node 215 net/mac80211/mesh_pathtbl.c hlist_for_each_entry(node, n, bucket, list) { node 216 net/mac80211/mesh_pathtbl.c mpath = node->mpath; node 261 net/mac80211/mesh_pathtbl.c struct mpath_node *node, *new_node; node 301 net/mac80211/mesh_pathtbl.c hlist_for_each_entry(node, n, bucket, list) { node 302 net/mac80211/mesh_pathtbl.c mpath = node->mpath; node 354 net/mac80211/mesh_pathtbl.c struct mpath_node *node; node 360 net/mac80211/mesh_pathtbl.c for_each_mesh_entry(mesh_paths, p, node, i) { node 361 net/mac80211/mesh_pathtbl.c mpath = node->mpath; node 392 net/mac80211/mesh_pathtbl.c struct mpath_node *node; node 396 net/mac80211/mesh_pathtbl.c for_each_mesh_entry(mesh_paths, p, node, i) { node 397 net/mac80211/mesh_pathtbl.c mpath = node->mpath; node 406 net/mac80211/mesh_pathtbl.c struct mpath_node *node; node 410 net/mac80211/mesh_pathtbl.c for_each_mesh_entry(mesh_paths, p, node, i) { node 411 net/mac80211/mesh_pathtbl.c mpath = node->mpath; node 419 net/mac80211/mesh_pathtbl.c struct mpath_node *node = container_of(rp, struct mpath_node, rcu); node 420 net/mac80211/mesh_pathtbl.c struct ieee80211_sub_if_data *sdata = node->mpath->sdata; node 422 net/mac80211/mesh_pathtbl.c del_timer_sync(&node->mpath->timer); node 424 net/mac80211/mesh_pathtbl.c kfree(node->mpath); node 425 net/mac80211/mesh_pathtbl.c kfree(node); node 439 net/mac80211/mesh_pathtbl.c struct mpath_node *node; node 450 net/mac80211/mesh_pathtbl.c hlist_for_each_entry(node, n, bucket, list) { node 451 net/mac80211/mesh_pathtbl.c mpath = node->mpath; node 456 net/mac80211/mesh_pathtbl.c hlist_del_rcu(&node->list); node 457 net/mac80211/mesh_pathtbl.c call_rcu(&node->rcu, mesh_path_node_reclaim); node 562 net/mac80211/mesh_pathtbl.c struct mpath_node *node = hlist_entry(p, struct mpath_node, list); node 563 net/mac80211/mesh_pathtbl.c mpath = node->mpath; node 567 net/mac80211/mesh_pathtbl.c kfree(node); node 573 net/mac80211/mesh_pathtbl.c struct mpath_node *node, *new_node; node 580 net/mac80211/mesh_pathtbl.c node = hlist_entry(p, struct mpath_node, list); node 581 net/mac80211/mesh_pathtbl.c mpath = node->mpath; node 613 net/mac80211/mesh_pathtbl.c struct mpath_node *node; node 618 net/mac80211/mesh_pathtbl.c for_each_mesh_entry(mesh_paths, p, node, i) { node 619 net/mac80211/mesh_pathtbl.c if (node->mpath->sdata != sdata) node 621 net/mac80211/mesh_pathtbl.c mpath = node->mpath; node 68 net/netfilter/xt_hashlimit.c struct hlist_node node; node 81 net/netfilter/xt_hashlimit.c struct hlist_node node; /* global list of all htables */ node 135 net/netfilter/xt_hashlimit.c hlist_for_each_entry(ent, pos, &ht->hash[hash], node) node 174 net/netfilter/xt_hashlimit.c hlist_add_head(&ent->node, &ht->hash[hash_dst(ht, dst)]); node 182 net/netfilter/xt_hashlimit.c hlist_del(&ent->node); node 254 net/netfilter/xt_hashlimit.c hlist_add_head(&hinfo->node, &hashlimit_htables); node 316 net/netfilter/xt_hashlimit.c hlist_add_head(&hinfo->node, &hashlimit_htables); node 345 net/netfilter/xt_hashlimit.c hlist_for_each_entry_safe(dh, pos, n, &ht->hash[i], node) { node 384 net/netfilter/xt_hashlimit.c hlist_for_each_entry(hinfo, pos, &hashlimit_htables, node) { node 400 net/netfilter/xt_hashlimit.c hlist_del(&hinfo->node); node 932 net/netfilter/xt_hashlimit.c hlist_for_each_entry(ent, pos, &htable->hash[*bucket], node) node 226 net/netlink/af_netlink.c struct hlist_node *node; node 230 net/netlink/af_netlink.c sk_for_each(sk, node, head) { node 290 net/netlink/af_netlink.c struct hlist_node *node, *tmp; node 292 net/netlink/af_netlink.c sk_for_each_safe(sk, node, tmp, &otable[i]) node 322 net/netlink/af_netlink.c struct hlist_node *node; node 328 net/netlink/af_netlink.c sk_for_each_bound(sk, node, &tbl->mc_list) { node 344 net/netlink/af_netlink.c struct hlist_node *node; node 350 net/netlink/af_netlink.c sk_for_each(osk, node, head) { node 355 net/netlink/af_netlink.c if (node) node 525 net/netlink/af_netlink.c struct hlist_node *node; node 534 net/netlink/af_netlink.c sk_for_each(osk, node, head) { node 1011 net/netlink/af_netlink.c struct hlist_node *node; node 1031 net/netlink/af_netlink.c sk_for_each_bound(sk, node, &nl_table[ssk->sk_protocol].mc_list) node 1083 net/netlink/af_netlink.c struct hlist_node *node; node 1093 net/netlink/af_netlink.c sk_for_each_bound(sk, node, &nl_table[ssk->sk_protocol].mc_list) node 1492 net/netlink/af_netlink.c struct hlist_node *node; node 1497 net/netlink/af_netlink.c sk_for_each_bound(sk, node, &tbl->mc_list) node 1754 net/netlink/af_netlink.c struct hlist_node *node; node 1761 net/netlink/af_netlink.c sk_for_each(s, node, &hash->table[j]) { node 107 net/netrom/af_netrom.c struct hlist_node *node; node 110 net/netrom/af_netrom.c sk_for_each(s, node, &nr_list) node 152 net/netrom/af_netrom.c struct hlist_node *node; node 155 net/netrom/af_netrom.c sk_for_each(s, node, &nr_list) node 173 net/netrom/af_netrom.c struct hlist_node *node; node 176 net/netrom/af_netrom.c sk_for_each(s, node, &nr_list) { node 197 net/netrom/af_netrom.c struct hlist_node *node; node 200 net/netrom/af_netrom.c sk_for_each(s, node, &nr_list) { node 1264 net/netrom/af_netrom.c struct hlist_node *node; node 1271 net/netrom/af_netrom.c sk_for_each(s, node, &nr_list) { node 51 net/netrom/nr_route.c struct hlist_node *node; node 54 net/netrom/nr_route.c nr_node_for_each(nr_node, node, &nr_node_list) node 69 net/netrom/nr_route.c struct hlist_node *node; node 72 net/netrom/nr_route.c nr_neigh_for_each(nr_neigh, node, &nr_neigh_list) node 116 net/netrom/nr_route.c struct hlist_node *node; node 119 net/netrom/nr_route.c nr_node_for_each(nr_nodet, node, &nr_node_list) { node 484 net/netrom/nr_route.c struct hlist_node *node, *nodet; node 488 net/netrom/nr_route.c nr_node_for_each_safe(s, node, nodet, &nr_node_list) { node 538 net/netrom/nr_route.c struct hlist_node *node, *nodet, *node2, *node2t; node 543 net/netrom/nr_route.c nr_neigh_for_each_safe(s, node, nodet, &nr_neigh_list) { node 728 net/netrom/nr_route.c struct hlist_node *node; node 732 net/netrom/nr_route.c nr_neigh_for_each(s, node, &nr_neigh_list) { node 752 net/netrom/nr_route.c nr_node_for_each(nr_node, node, &nr_node_list) { node 862 net/netrom/nr_route.c struct hlist_node *node; node 869 net/netrom/nr_route.c nr_node_for_each(nr_node, node, &nr_node_list) { node 880 net/netrom/nr_route.c struct hlist_node *node; node 883 net/netrom/nr_route.c node = (v == SEQ_START_TOKEN) node 887 net/netrom/nr_route.c return hlist_entry(node, struct nr_node, node_node); node 948 net/netrom/nr_route.c struct hlist_node *node; node 955 net/netrom/nr_route.c nr_neigh_for_each(nr_neigh, node, &nr_neigh_list) { node 964 net/netrom/nr_route.c struct hlist_node *node; node 967 net/netrom/nr_route.c node = (v == SEQ_START_TOKEN) node 971 net/netrom/nr_route.c return hlist_entry(node, struct nr_neigh, neigh_node); node 1038 net/netrom/nr_route.c struct hlist_node *node, *nodet; node 1042 net/netrom/nr_route.c nr_node_for_each_safe(t, node, nodet, &nr_node_list) { node 1047 net/netrom/nr_route.c nr_neigh_for_each_safe(s, node, nodet, &nr_neigh_list) { node 1588 net/packet/af_packet.c struct hlist_node *node; node 1593 net/packet/af_packet.c sk_for_each(sk, node, &net->packet.sklist) { node 2004 net/packet/af_packet.c struct hlist_node *node; node 2006 net/packet/af_packet.c sk_for_each(s, node, &net->packet.sklist) { node 527 net/phonet/pep.c struct hlist_node *node; node 531 net/phonet/pep.c sk_for_each(sknode, node, hlist) { node 62 net/phonet/socket.c struct hlist_node *node; node 70 net/phonet/socket.c sk_for_each(sknode, node, &pnsocks.hlist) { node 251 net/rfkill/rfkill.c list_for_each_entry(rfkill, &rfkill_list, node) { node 292 net/rfkill/rfkill.c list_for_each_entry(rfkill, &rfkill_list, node) { node 597 net/rfkill/rfkill.c list_for_each_entry(p, &rfkill_list, node) { node 630 net/rfkill/rfkill.c list_add_tail(&rfkill->node, &rfkill_list); node 642 net/rfkill/rfkill.c list_del_init(&rfkill->node); node 680 net/rfkill/rfkill.c INIT_LIST_HEAD(&rfkill->node); node 170 net/rose/af_rose.c struct hlist_node *node; node 173 net/rose/af_rose.c sk_for_each(s, node, &rose_list) { node 191 net/rose/af_rose.c struct hlist_node *node; node 194 net/rose/af_rose.c sk_for_each(s, node, &rose_list) { node 251 net/rose/af_rose.c struct hlist_node *node; node 254 net/rose/af_rose.c sk_for_each(s, node, &rose_list) { node 263 net/rose/af_rose.c sk_for_each(s, node, &rose_list) { node 283 net/rose/af_rose.c struct hlist_node *node; node 286 net/rose/af_rose.c sk_for_each(s, node, &rose_list) { node 1405 net/rose/af_rose.c struct hlist_node *node; node 1412 net/rose/af_rose.c sk_for_each(s, node, &rose_list) { node 671 net/rose/rose_route.c struct rose_node *node; node 676 net/rose/rose_route.c for (node = rose_node_list; node != NULL; node = node->next) { node 677 net/rose/rose_route.c if (rosecmpm(addr, &node->address, node->mask) == 0) { node 678 net/rose/rose_route.c for (i = 0; i < node->count; i++) { node 680 net/rose/rose_route.c if (node->neighbour[i]->restarted) { node 681 net/rose/rose_route.c res = node->neighbour[i]; node 686 net/rose/rose_route.c if (!rose_ftimer_running(node->neighbour[i])) { node 687 net/rose/rose_route.c res = node->neighbour[i]; node 87 net/rxrpc/ar-connection.c bundle = rb_entry(p, struct rxrpc_conn_bundle, node); node 116 net/rxrpc/ar-connection.c bundle = rb_entry(parent, struct rxrpc_conn_bundle, node); node 130 net/rxrpc/ar-connection.c rb_link_node(&bundle->node, parent, pp); node 131 net/rxrpc/ar-connection.c rb_insert_color(&bundle->node, &trans->bundles); node 177 net/rxrpc/ar-connection.c rb_erase(&bundle->node, &trans->bundles); node 247 net/rxrpc/ar-connection.c xconn = rb_entry(parent, struct rxrpc_connection, node); node 263 net/rxrpc/ar-connection.c rb_link_node(&conn->node, parent, p); node 264 net/rxrpc/ar-connection.c rb_insert_color(&conn->node, &conn->trans->client_conns); node 287 net/rxrpc/ar-connection.c xconn = rb_entry(parent, struct rxrpc_connection, node); node 627 net/rxrpc/ar-connection.c conn = rb_entry(p, struct rxrpc_connection, node); node 670 net/rxrpc/ar-connection.c conn = rb_entry(p, struct rxrpc_connection, node); node 687 net/rxrpc/ar-connection.c rb_link_node(&conn->node, p, pp); node 688 net/rxrpc/ar-connection.c rb_insert_color(&conn->node, &trans->server_conns); node 757 net/rxrpc/ar-connection.c conn = rb_entry(p, struct rxrpc_connection, node); node 859 net/rxrpc/ar-connection.c rb_erase(&conn->node, node 862 net/rxrpc/ar-connection.c rb_erase(&conn->node, node 222 net/rxrpc/ar-internal.h struct rb_node node; /* node in transport's lookup tree */ node 244 net/rxrpc/ar-internal.h struct rb_node node; /* node in transport's lookup tree */ node 106 net/sched/sch_htb.c struct rb_node node[TC_HTB_NUMPRIO]; /* node for self or feed tree */ node 262 net/sched/sch_htb.c c = rb_entry(parent, struct htb_class, node[prio]); node 269 net/sched/sch_htb.c rb_link_node(&cl->node[prio], parent, p); node 270 net/sched/sch_htb.c rb_insert_color(&cl->node[prio], root); node 361 net/sched/sch_htb.c if (q->ptr[cl->level][prio] == cl->node + prio) node 364 net/sched/sch_htb.c htb_safe_rb_erase(cl->node + prio, q->row[cl->level] + prio); node 424 net/sched/sch_htb.c if (p->un.inner.ptr[prio] == cl->node + prio) { node 432 net/sched/sch_htb.c htb_safe_rb_erase(cl->node + prio, p->un.inner.feed + prio); node 735 net/sched/sch_htb.c rb_entry(n, struct htb_class, node[prio]); node 791 net/sched/sch_htb.c cl = rb_entry(*sp->pptr, struct htb_class, node[prio]); node 1383 net/sched/sch_htb.c RB_CLEAR_NODE(&cl->node[prio]); node 331 net/sctp/endpointola.c struct hlist_node *node; node 341 net/sctp/endpointola.c sctp_for_each_hentry(epb, node, &head->chain) { node 693 net/sctp/input.c hlist_add_head(&epb->node, &head->chain); node 713 net/sctp/input.c if (hlist_unhashed(&epb->node)) node 721 net/sctp/input.c __hlist_del(&epb->node); node 739 net/sctp/input.c struct hlist_node *node; node 745 net/sctp/input.c sctp_for_each_hentry(epb, node, &head->chain) { node 773 net/sctp/input.c hlist_add_head(&epb->node, &head->chain); node 802 net/sctp/input.c __hlist_del(&epb->node); node 827 net/sctp/input.c struct hlist_node *node; node 836 net/sctp/input.c sctp_for_each_hentry(epb, node, &head->chain) { node 204 net/sctp/proc.c struct hlist_node *node; node 213 net/sctp/proc.c sctp_for_each_hentry(epb, node, &head->chain) { node 308 net/sctp/proc.c struct hlist_node *node; node 317 net/sctp/proc.c sctp_for_each_hentry(epb, node, &head->chain) { node 420 net/sctp/proc.c struct hlist_node *node; node 430 net/sctp/proc.c sctp_for_each_hentry(epb, node, &head->chain) { node 5551 net/sctp/socket.c struct hlist_node *node; node 5576 net/sctp/socket.c sctp_for_each_hentry(pp, node, &head->chain) node 5603 net/sctp/socket.c sctp_for_each_hentry(pp, node, &head->chain) { node 5619 net/sctp/socket.c struct hlist_node *node; node 5636 net/sctp/socket.c sk_for_each_bound(sk2, node, &pp->owner) { node 5954 net/sctp/socket.c hlist_add_head(&pp->node, &head->chain); node 5963 net/sctp/socket.c __hlist_del(&pp->node); node 125 net/sunrpc/svc.c unsigned int node; node 135 net/sunrpc/svc.c node = any_online_node(node_online_map); node 136 net/sunrpc/svc.c if (nr_cpus_node(node) > 2) { node 209 net/sunrpc/svc.c unsigned int node; node 216 net/sunrpc/svc.c for_each_node_with_cpus(node) { node 219 net/sunrpc/svc.c m->to_pool[node] = pidx; node 220 net/sunrpc/svc.c m->pool_to[pidx] = node; node 304 net/sunrpc/svc.c unsigned int node = m->pool_to[pidx]; node 315 net/sunrpc/svc.c set_cpus_allowed_ptr(task, &cpumask_of_cpu(node)); node 320 net/sunrpc/svc.c node_to_cpumask_ptr(nodecpumask, node); node 113 net/tipc/bcast.c struct tipc_node node; node 421 net/tipc/bcast.c struct tipc_node* node = tipc_node_find(msg_prevnode(msg)); node 428 net/tipc/bcast.c if (unlikely(!node || !tipc_node_is_up(node) || !node->bclink.supported || node 437 net/tipc/bcast.c tipc_node_lock(node); node 438 net/tipc/bcast.c tipc_bclink_acknowledge(node, msg_bcast_ack(msg)); node 439 net/tipc/bcast.c tipc_node_unlock(node); node 442 net/tipc/bcast.c bcl->owner->next = node; /* remember requestor */ node 465 net/tipc/bcast.c tipc_node_lock(node); node 467 net/tipc/bcast.c deferred = node->bclink.deferred_head; node 468 net/tipc/bcast.c next_in = mod(node->bclink.last_in + 1); node 473 net/tipc/bcast.c node->bclink.last_in++; node 474 net/tipc/bcast.c bclink_set_gap(node); node 476 net/tipc/bcast.c bclink_send_ack(node); node 480 net/tipc/bcast.c tipc_node_unlock(node); node 485 net/tipc/bcast.c tipc_node_unlock(node); node 489 net/tipc/bcast.c if (tipc_link_recv_fragment(&node->bclink.defragm, node 492 net/tipc/bcast.c tipc_node_unlock(node); node 495 net/tipc/bcast.c tipc_node_unlock(node); node 499 net/tipc/bcast.c tipc_node_lock(node); node 502 net/tipc/bcast.c node->bclink.deferred_head = deferred->next; node 507 net/tipc/bcast.c u32 gap_after = node->bclink.gap_after; node 508 net/tipc/bcast.c u32 gap_to = node->bclink.gap_to; node 510 net/tipc/bcast.c if (tipc_link_defer_pkt(&node->bclink.deferred_head, node 511 net/tipc/bcast.c &node->bclink.deferred_tail, node 513 net/tipc/bcast.c node->bclink.nack_sync++; node 516 net/tipc/bcast.c node->bclink.gap_after = seqno; node 518 net/tipc/bcast.c node->bclink.gap_to = seqno; node 520 net/tipc/bcast.c if (bclink_ack_allowed(node->bclink.nack_sync)) { node 522 net/tipc/bcast.c bclink_send_nack(node); node 523 net/tipc/bcast.c bclink_set_gap(node); node 529 net/tipc/bcast.c tipc_node_unlock(node); node 797 net/tipc/bcast.c spin_lock_init(&bclink->node.lock); node 798 net/tipc/bcast.c bcl->owner = &bclink->node; node 82 net/tipc/bcast.h int n = tipc_node(node); node 98 net/tipc/bcast.h int n = tipc_node(node); node 645 net/tipc/config.c rep_buf = tipc_cfg_do_cmd(orig->node, node 1273 net/tipc/link.c struct tipc_node *node; node 1287 net/tipc/link.c node = tipc_node_select(destaddr, selector); node 1288 net/tipc/link.c if (likely(node)) { node 1289 net/tipc/link.c tipc_node_lock(node); node 1290 net/tipc/link.c l_ptr = node->active_links[selector]; node 1298 net/tipc/link.c tipc_node_unlock(node); node 1323 net/tipc/link.c tipc_node_unlock(node); node 1333 net/tipc/link.c tipc_node_unlock(node); node 1367 net/tipc/link.c struct tipc_node *node; node 1477 net/tipc/link.c node = tipc_node_select(destaddr, sender->publ.ref & 1); node 1478 net/tipc/link.c if (likely(node)) { node 1479 net/tipc/link.c tipc_node_lock(node); node 1480 net/tipc/link.c l_ptr = node->active_links[sender->publ.ref & 1]; node 1482 net/tipc/link.c tipc_node_unlock(node); node 1487 net/tipc/link.c tipc_node_unlock(node); node 1525 net/tipc/link.c tipc_node_unlock(node); node 2951 net/tipc/link.c *node = tipc_node_find(link_name_parts.addr_peer); node 2952 net/tipc/link.c if (!*node) node 2955 net/tipc/link.c l_ptr = (*node)->links[b_ptr->identity]; node 2968 net/tipc/link.c struct tipc_node *node; node 2986 net/tipc/link.c l_ptr = link_find_link(args->name, &node); node 2992 net/tipc/link.c tipc_node_lock(node); node 3021 net/tipc/link.c tipc_node_unlock(node); node 3046 net/tipc/link.c struct tipc_node *node; node 3059 net/tipc/link.c l_ptr = link_find_link(link_name, &node); node 3065 net/tipc/link.c tipc_node_lock(node); node 3067 net/tipc/link.c tipc_node_unlock(node); node 3094 net/tipc/link.c struct tipc_node *node; node 3104 net/tipc/link.c l_ptr = link_find_link(name, &node); node 3109 net/tipc/link.c tipc_node_lock(node); node 3169 net/tipc/link.c tipc_node_unlock(node); node 3210 net/tipc/link.c struct tipc_node * node; node 3215 net/tipc/link.c node = tipc_node_find(a); node 3216 net/tipc/link.c if (node) { node 3217 net/tipc/link.c tipc_node_lock(node); node 3218 net/tipc/link.c l_ptr = node->links[bearer_id]; node 3235 net/tipc/link.c tipc_node_unlock(node); node 182 net/tipc/name_distr.c buf = named_prepare_buf(PUBLICATION, left, node); node 193 net/tipc/name_distr.c msg_set_link_selector(buf_msg(buf), node); node 195 net/tipc/name_distr.c "<%u.%u.%u>\n", tipc_zone(node), node 196 net/tipc/name_distr.c tipc_cluster(node), tipc_node(node)); node 197 net/tipc/name_distr.c tipc_link_send(buf, node, node); node 225 net/tipc/name_distr.c publ->node, publ->ref, publ->key); node 231 net/tipc/name_distr.c publ->type, publ->lower, publ->node, publ->ref, publ->key); node 313 net/tipc/name_distr.c if (publ->node == tipc_own_addr) node 315 net/tipc/name_distr.c publ->node = tipc_own_addr; node 140 net/tipc/name_table.c publ->node = node; node 319 net/tipc/name_table.c type, lower, upper, node, port, sseq, node 324 net/tipc/name_table.c publ = publ_create(type, lower, upper, scope, node, port, key); node 328 net/tipc/name_table.c publ, node, publ->node, publ->subscr.node); node 338 net/tipc/name_table.c if (in_own_cluster(node)) { node 349 net/tipc/name_table.c if (node == tipc_own_addr) { node 369 net/tipc/name_table.c publ->node, node 409 net/tipc/name_table.c (publ->node && (publ->node != node))) { node 431 net/tipc/name_table.c if (in_own_cluster(node)) { node 443 net/tipc/name_table.c publ->type, publ->lower, publ->node, node 462 net/tipc/name_table.c if (node == tipc_own_addr) { node 474 net/tipc/name_table.c publ->type, publ->lower, publ->node, node 507 net/tipc/name_table.c publ->node, node 541 net/tipc/name_table.c crs->node, node 583 net/tipc/name_table.c dbg("Publishing {%u,%u,%u} from 0x%x\n", type, lower, upper, node); node 592 net/tipc/name_table.c scope, node, port, key); node 604 net/tipc/name_table.c dbg("Withdrawing {%u,%u} from 0x%x\n", type, lower, node); node 605 net/tipc/name_table.c publ = tipc_nameseq_remove_publ(seq, lower, node, ref, key); node 649 net/tipc/name_table.c *destnode = publ->node; node 892 net/tipc/name_table.c tipc_zone(publ->node), tipc_cluster(publ->node), node 893 net/tipc/name_table.c tipc_node(publ->node), publ->ref); node 896 net/tipc/name_table.c if (publ->node != tipc_own_addr) node 76 net/tipc/name_table.h u32 node; node 456 net/tipc/node.c ns->node = NULL; node 51 net/tipc/node_subscr.c node_sub->node = NULL; node 55 net/tipc/node_subscr.c node_sub->node = tipc_node_find(addr); node 56 net/tipc/node_subscr.c if (!node_sub->node) { node 63 net/tipc/node_subscr.c tipc_node_lock(node_sub->node); node 64 net/tipc/node_subscr.c list_add_tail(&node_sub->nodesub_list, &node_sub->node->nsub); node 65 net/tipc/node_subscr.c tipc_node_unlock(node_sub->node); node 74 net/tipc/node_subscr.c if (!node_sub->node) node 77 net/tipc/node_subscr.c tipc_node_lock(node_sub->node); node 79 net/tipc/node_subscr.c tipc_node_unlock(node_sub->node); node 53 net/tipc/node_subscr.h struct tipc_node *node; node 821 net/tipc/port.c orig.node = msg_orignode(msg); node 1077 net/tipc/port.c id->node = tipc_own_addr; node 1202 net/tipc/port.c msg_set_destnode(msg, peer->node); node 1208 net/tipc/port.c if (!may_route(peer->node)) node 1218 net/tipc/port.c tipc_nodesub_subscribe(&p_ptr->subscription,peer->node, node 1224 net/tipc/port.c p_ptr->publ.max_pkt = tipc_link_get_max_pkt(peer->node, ref); node 1321 net/tipc/port.c peer->node = port_peernode(p_ptr); node 1465 net/tipc/port.c msg_set_orignode(msg, orig->node); node 1508 net/tipc/port.c orig.node = tipc_own_addr; node 1539 net/tipc/port.c msg_set_orignode(msg, orig->node); node 1582 net/tipc/port.c orig.node = tipc_own_addr; node 1608 net/tipc/port.c msg_set_orignode(msg, orig->node); node 1610 net/tipc/port.c msg_set_destnode(msg, dest->node); node 1616 net/tipc/port.c if (dest->node == tipc_own_addr) node 1618 net/tipc/port.c res = tipc_link_send_sections_fast(p_ptr, msg_sect, num_sect, dest->node); node 1640 net/tipc/port.c orig.node = tipc_own_addr; node 1665 net/tipc/port.c msg_set_orignode(msg, orig->node); node 1667 net/tipc/port.c msg_set_destnode(msg, dest->node); node 1680 net/tipc/port.c if (dest->node == tipc_own_addr) node 1682 net/tipc/port.c res = tipc_send_buf_fast(buf, dest->node); node 1702 net/tipc/port.c orig.node = tipc_own_addr; node 401 net/tipc/socket.c addr->addr.id.node = tsock->peer_name.node; node 780 net/tipc/socket.c tsock->peer_name.node = msg_orignode(msg); node 803 net/tipc/socket.c addr->addr.id.node = msg_orignode(msg); node 1551 net/tipc/socket.c new_tsock->peer_name.node = msg_orignode(msg); node 114 net/tipc/subscr.c sub->evt.port.node = htohl(node, sub->swap); node 157 net/tipc/subscr.c sub->event_cb(sub, found_lower, found_upper, event, port_ref, node); node 253 net/unix/af_unix.c struct hlist_node *node; node 255 net/unix/af_unix.c sk_for_each(s, node, &unix_socket_table[hash ^ type]) { node 288 net/unix/af_unix.c struct hlist_node *node; node 291 net/unix/af_unix.c sk_for_each(s, node, node 174 net/x25/af_x25.c struct hlist_node *node; node 178 net/x25/af_x25.c sk_for_each(s, node, &x25_list) node 246 net/x25/af_x25.c struct hlist_node *node; node 251 net/x25/af_x25.c sk_for_each(s, node, &x25_list) node 288 net/x25/af_x25.c struct hlist_node *node; node 290 net/x25/af_x25.c sk_for_each(s, node, &x25_list) node 1623 net/x25/af_x25.c struct hlist_node *node; node 1627 net/x25/af_x25.c sk_for_each(s, node, &x25_list) node 51 net/x25/x25_forward.c x25_frwd = list_entry(entry, struct x25_forward, node); node 70 net/x25/x25_forward.c list_add(&new_frwd->node, &x25_forward_list); node 104 net/x25/x25_forward.c frwd = list_entry(entry, struct x25_forward, node); node 141 net/x25/x25_forward.c fwd = list_entry(entry, struct x25_forward, node); node 143 net/x25/x25_forward.c list_del(&fwd->node); node 159 net/x25/x25_forward.c fwd = list_entry(entry, struct x25_forward, node); node 161 net/x25/x25_forward.c list_del(&fwd->node); node 267 net/x25/x25_link.c list_add(&nb->node, &x25_neigh_list); node 283 net/x25/x25_link.c if (nb->node.next) { node 284 net/x25/x25_link.c list_del(&nb->node); node 300 net/x25/x25_link.c nb = list_entry(entry, struct x25_neigh, node); node 321 net/x25/x25_link.c nb = list_entry(entry, struct x25_neigh, node); node 394 net/x25/x25_link.c nb = list_entry(entry, struct x25_neigh, node); node 34 net/x25/x25_proc.c rt = list_entry(route_entry, struct x25_route, node); node 61 net/x25/x25_proc.c struct x25_route, node); node 65 net/x25/x25_proc.c if (rt->node.next != &x25_route_list) node 66 net/x25/x25_proc.c rt = list_entry(rt->node.next, struct x25_route, node); node 99 net/x25/x25_proc.c struct hlist_node *node; node 101 net/x25/x25_proc.c sk_for_each(s, node, &x25_list) node 179 net/x25/x25_proc.c f = list_entry(entry, struct x25_forward, node); node 207 net/x25/x25_proc.c struct x25_forward, node); node 211 net/x25/x25_proc.c if (f->node.next != &x25_forward_list) node 212 net/x25/x25_proc.c f = list_entry(f->node.next, struct x25_forward, node); node 40 net/x25/x25_route.c rt = list_entry(entry, struct x25_route, node); node 59 net/x25/x25_route.c list_add(&rt->node, &x25_route_list); node 75 net/x25/x25_route.c if (rt->node.next) { node 76 net/x25/x25_route.c list_del(&rt->node); node 91 net/x25/x25_route.c rt = list_entry(entry, struct x25_route, node); node 116 net/x25/x25_route.c rt = list_entry(entry, struct x25_route, node); node 159 net/x25/x25_route.c rt = list_entry(entry, struct x25_route, node); node 219 net/x25/x25_route.c rt = list_entry(entry, struct x25_route, node); node 203 scripts/genksyms/genksyms.c free(node->string); node 204 scripts/genksyms/genksyms.c free(node); node 221 scripts/genksyms/genksyms.c newnode->string = xstrdup(node->string); node 222 scripts/genksyms/genksyms.c newnode->tag = node->tag; node 40 scripts/genksyms/parse.y struct string_list *node = *p; node 41 scripts/genksyms/parse.y *p = node->next; node 42 scripts/genksyms/parse.y free_node(node); node 1295 scripts/kconfig/gconf.c gtk_tree_store_set(tree, node, node 1321 scripts/kconfig/gconf.c GtkTreeIter *node = parents[indent]; node 1323 scripts/kconfig/gconf.c gtk_tree_store_append(tree, node, parent); node 1324 scripts/kconfig/gconf.c set_node(node, menu, row); node 174 scripts/kconfig/lxdialog/dialog.h struct dialog_item node; node 561 scripts/kconfig/lxdialog/util.c vsnprintf(item_cur->node.str, sizeof(item_cur->node.str), fmt, ap); node 570 scripts/kconfig/lxdialog/util.c avail = sizeof(item_cur->node.str) - strlen(item_cur->node.str); node 573 scripts/kconfig/lxdialog/util.c vsnprintf(item_cur->node.str + strlen(item_cur->node.str), node 575 scripts/kconfig/lxdialog/util.c item_cur->node.str[sizeof(item_cur->node.str) - 1] = '\0'; node 581 scripts/kconfig/lxdialog/util.c item_cur->node.tag = tag; node 585 scripts/kconfig/lxdialog/util.c item_cur->node.data = ptr; node 590 scripts/kconfig/lxdialog/util.c item_cur->node.selected = val; node 603 scripts/kconfig/lxdialog/util.c return item_cur->node.data; node 608 scripts/kconfig/lxdialog/util.c return item_cur->node.tag; node 644 scripts/kconfig/lxdialog/util.c return item_cur->node.str; node 649 scripts/kconfig/lxdialog/util.c return (item_cur->node.selected != 0); node 654 scripts/kconfig/lxdialog/util.c return (item_cur->node.tag == tag); node 49 security/keys/internal.h struct rb_node node; node 76 security/keys/key.c user = rb_entry(parent, struct key_user, node); node 114 security/keys/key.c rb_link_node(&candidate->node, parent, p); node 115 security/keys/key.c rb_insert_color(&candidate->node, &key_user_tree); node 137 security/keys/key.c rb_erase(&user->node, &key_user_tree); node 999 security/keys/key.c rb_link_node(&root_key_user.node, node 1003 security/keys/key.c rb_insert_color(&root_key_user.node, node 244 security/keys/proc.c struct key_user *user = rb_entry(_p, struct key_user, node); node 250 security/selinux/avc.c struct avc_node *node; node 260 security/selinux/avc.c list_for_each_entry_rcu(node, &avc_cache.slots[i], list) node 277 security/selinux/avc.c struct avc_node *node = container_of(rhead, struct avc_node, rhead); node 278 security/selinux/avc.c kmem_cache_free(avc_node_cachep, node); node 284 security/selinux/avc.c list_del_rcu(&node->list); node 285 security/selinux/avc.c call_rcu(&node->rhead, avc_node_free); node 291 security/selinux/avc.c kmem_cache_free(avc_node_cachep, node); node 305 security/selinux/avc.c struct avc_node *node; node 316 security/selinux/avc.c list_for_each_entry(node, &avc_cache.slots[hvalue], list) { node 317 security/selinux/avc.c if (atomic_dec_and_test(&node->ae.used)) { node 319 security/selinux/avc.c avc_node_delete(node); node 338 security/selinux/avc.c struct avc_node *node; node 340 security/selinux/avc.c node = kmem_cache_zalloc(avc_node_cachep, GFP_ATOMIC); node 341 security/selinux/avc.c if (!node) node 344 security/selinux/avc.c INIT_RCU_HEAD(&node->rhead); node 345 security/selinux/avc.c INIT_LIST_HEAD(&node->list); node 346 security/selinux/avc.c atomic_set(&node->ae.used, 1); node 353 security/selinux/avc.c return node; node 358 security/selinux/avc.c node->ae.ssid = ssid; node 359 security/selinux/avc.c node->ae.tsid = tsid; node 360 security/selinux/avc.c node->ae.tclass = tclass; node 361 security/selinux/avc.c memcpy(&node->ae.avd, &ae->avd, sizeof(node->ae.avd)); node 366 security/selinux/avc.c struct avc_node *node, *ret = NULL; node 370 security/selinux/avc.c list_for_each_entry_rcu(node, &avc_cache.slots[hvalue], list) { node 371 security/selinux/avc.c if (ssid == node->ae.ssid && node 372 security/selinux/avc.c tclass == node->ae.tclass && node 373 security/selinux/avc.c tsid == node->ae.tsid) { node 374 security/selinux/avc.c ret = node; node 407 security/selinux/avc.c struct avc_node *node; node 410 security/selinux/avc.c node = avc_search_node(ssid, tsid, tclass); node 412 security/selinux/avc.c if (node && ((node->ae.avd.decided & requested) == requested)) { node 417 security/selinux/avc.c node = NULL; node 420 security/selinux/avc.c return node; node 464 security/selinux/avc.c struct avc_node *pos, *node = NULL; node 471 security/selinux/avc.c node = avc_alloc_node(); node 472 security/selinux/avc.c if (node) { node 474 security/selinux/avc.c avc_node_populate(node, ssid, tsid, tclass, ae); node 481 security/selinux/avc.c avc_node_replace(node, pos); node 485 security/selinux/avc.c list_add_rcu(&node->list, &avc_cache.slots[hvalue]); node 490 security/selinux/avc.c return node; node 753 security/selinux/avc.c struct avc_node *pos, *node, *orig = NULL; node 755 security/selinux/avc.c node = avc_alloc_node(); node 756 security/selinux/avc.c if (!node) { node 776 security/selinux/avc.c avc_node_kill(node); node 784 security/selinux/avc.c avc_node_populate(node, ssid, tsid, tclass, &orig->ae); node 788 security/selinux/avc.c node->ae.avd.allowed |= perms; node 792 security/selinux/avc.c node->ae.avd.allowed &= ~perms; node 795 security/selinux/avc.c node->ae.avd.auditallow |= perms; node 798 security/selinux/avc.c node->ae.avd.auditallow &= ~perms; node 801 security/selinux/avc.c node->ae.avd.auditdeny |= perms; node 804 security/selinux/avc.c node->ae.avd.auditdeny &= ~perms; node 807 security/selinux/avc.c avc_node_replace(node, orig); node 823 security/selinux/avc.c struct avc_node *node; node 832 security/selinux/avc.c list_for_each_entry(node, &avc_cache.slots[i], list) node 833 security/selinux/avc.c avc_node_delete(node); node 878 security/selinux/avc.c struct avc_node *node; node 887 security/selinux/avc.c node = avc_lookup(ssid, tsid, tclass, requested); node 888 security/selinux/avc.c if (!node) { node 894 security/selinux/avc.c node = avc_insert(ssid, tsid, tclass, &entry); node 897 security/selinux/avc.c p_ae = node ? &node->ae : &entry; node 82 security/selinux/netnode.c struct sel_netnode *node = container_of(p, struct sel_netnode, rcu); node 83 security/selinux/netnode.c kfree(node); node 132 security/selinux/netnode.c struct sel_netnode *node; node 145 security/selinux/netnode.c list_for_each_entry_rcu(node, &sel_netnode_hash[idx].list, list) node 146 security/selinux/netnode.c if (node->nsec.family == family) node 149 security/selinux/netnode.c if (node->nsec.addr.ipv4 == *(__be32 *)addr) node 150 security/selinux/netnode.c return node; node 153 security/selinux/netnode.c if (ipv6_addr_equal(&node->nsec.addr.ipv6, node 155 security/selinux/netnode.c return node; node 174 security/selinux/netnode.c switch (node->nsec.family) { node 176 security/selinux/netnode.c idx = sel_netnode_hashfn_ipv4(node->nsec.addr.ipv4); node 179 security/selinux/netnode.c idx = sel_netnode_hashfn_ipv6(&node->nsec.addr.ipv6); node 185 security/selinux/netnode.c INIT_RCU_HEAD(&node->rcu); node 189 security/selinux/netnode.c list_add_rcu(&node->list, &sel_netnode_hash[idx].list); node 217 security/selinux/netnode.c struct sel_netnode *node; node 221 security/selinux/netnode.c node = sel_netnode_find(addr, family); node 222 security/selinux/netnode.c if (node != NULL) { node 223 security/selinux/netnode.c *sid = node->nsec.sid; node 278 security/selinux/netnode.c struct sel_netnode *node; node 281 security/selinux/netnode.c node = sel_netnode_find(addr, family); node 282 security/selinux/netnode.c if (node != NULL) { node 283 security/selinux/netnode.c *sid = node->nsec.sid; node 302 security/selinux/netnode.c struct sel_netnode *node, *node_tmp; node 306 security/selinux/netnode.c list_for_each_entry_safe(node, node_tmp, node 308 security/selinux/netnode.c list_del_rcu(&node->list); node 309 security/selinux/netnode.c call_rcu(&node->rcu, sel_netnode_free); node 1012 security/selinux/selinuxfs.c struct list_head *node; node 1015 security/selinux/selinuxfs.c node = de->d_subdirs.next; node 1016 security/selinux/selinuxfs.c while (node != &de->d_subdirs) { node 1017 security/selinux/selinuxfs.c struct dentry *d = list_entry(node, struct dentry, d_u.d_child); node 1018 security/selinux/selinuxfs.c list_del_init(node); node 1028 security/selinux/selinuxfs.c node = de->d_subdirs.next; node 198 security/selinux/ss/avtab.c if (!node) node 202 security/selinux/ss/avtab.c for (cur = node->next; cur; cur = cur->next) { node 203 security/selinux/ss/avtab.c if (node->key.source_type == cur->key.source_type && node 204 security/selinux/ss/avtab.c node->key.target_type == cur->key.target_type && node 205 security/selinux/ss/avtab.c node->key.target_class == cur->key.target_class && node 209 security/selinux/ss/avtab.c if (node->key.source_type < cur->key.source_type) node 211 security/selinux/ss/avtab.c if (node->key.source_type == cur->key.source_type && node 212 security/selinux/ss/avtab.c node->key.target_type < cur->key.target_type) node 214 security/selinux/ss/avtab.c if (node->key.source_type == cur->key.source_type && node 215 security/selinux/ss/avtab.c node->key.target_type == cur->key.target_type && node 216 security/selinux/ss/avtab.c node->key.target_class < cur->key.target_class) node 94 security/selinux/ss/conditional.c new_state = cond_evaluate_expr(p, node->expr); node 95 security/selinux/ss/conditional.c if (new_state != node->cur_state) { node 96 security/selinux/ss/conditional.c node->cur_state = new_state; node 100 security/selinux/ss/conditional.c for (cur = node->true_list; cur; cur = cur->next) { node 102 security/selinux/ss/conditional.c cur->node->key.specified &= ~AVTAB_ENABLED; node 104 security/selinux/ss/conditional.c cur->node->key.specified |= AVTAB_ENABLED; node 107 security/selinux/ss/conditional.c for (cur = node->false_list; cur; cur = cur->next) { node 110 security/selinux/ss/conditional.c cur->node->key.specified &= ~AVTAB_ENABLED; node 112 security/selinux/ss/conditional.c cur->node->key.specified |= AVTAB_ENABLED; node 142 security/selinux/ss/conditional.c for (cur_expr = node->expr; cur_expr; cur_expr = next_expr) { node 146 security/selinux/ss/conditional.c cond_av_list_destroy(node->true_list); node 147 security/selinux/ss/conditional.c cond_av_list_destroy(node->false_list); node 148 security/selinux/ss/conditional.c kfree(node); node 295 security/selinux/ss/conditional.c if (cur->node == node_ptr) { node 323 security/selinux/ss/conditional.c list->node = node_ptr; node 396 security/selinux/ss/conditional.c node->cur_state = le32_to_cpu(buf[0]); node 424 security/selinux/ss/conditional.c node->expr = expr; node 430 security/selinux/ss/conditional.c if (cond_read_av_list(p, fp, &node->true_list, NULL) != 0) node 432 security/selinux/ss/conditional.c if (cond_read_av_list(p, fp, &node->false_list, node->true_list) != 0) node 436 security/selinux/ss/conditional.c cond_node_destroy(node); node 442 security/selinux/ss/conditional.c struct cond_node *node, *last = NULL; node 458 security/selinux/ss/conditional.c node = kzalloc(sizeof(struct cond_node), GFP_KERNEL); node 459 security/selinux/ss/conditional.c if (!node) node 462 security/selinux/ss/conditional.c if (cond_read_node(p, node, fp) != 0) node 466 security/selinux/ss/conditional.c p->cond_list = node; node 468 security/selinux/ss/conditional.c last->next = node; node 469 security/selinux/ss/conditional.c last = node; node 483 security/selinux/ss/conditional.c struct avtab_node *node; node 488 security/selinux/ss/conditional.c for (node = avtab_search_node(ctab, key); node; node 489 security/selinux/ss/conditional.c node = avtab_search_node_next(node, key->specified)) { node 491 security/selinux/ss/conditional.c (node->key.specified & (AVTAB_ALLOWED|AVTAB_ENABLED))) node 492 security/selinux/ss/conditional.c avd->allowed |= node->datum.data; node 494 security/selinux/ss/conditional.c (node->key.specified & (AVTAB_AUDITDENY|AVTAB_ENABLED))) node 500 security/selinux/ss/conditional.c avd->auditdeny &= node->datum.data; node 502 security/selinux/ss/conditional.c (node->key.specified & (AVTAB_AUDITALLOW|AVTAB_ENABLED))) node 503 security/selinux/ss/conditional.c avd->auditallow |= node->datum.data; node 43 security/selinux/ss/conditional.h struct avtab_node *node; node 32 security/selinux/ss/ebitmap.c n1 = e1->node; node 33 security/selinux/ss/ebitmap.c n2 = e2->node; node 52 security/selinux/ss/ebitmap.c n = src->node; node 66 security/selinux/ss/ebitmap.c dst->node = new; node 89 security/selinux/ss/ebitmap.c struct ebitmap_node *e_iter = ebmap->node; node 186 security/selinux/ss/ebitmap.c ebmap->node = e_iter; node 222 security/selinux/ss/ebitmap.c n1 = e1->node; node 223 security/selinux/ss/ebitmap.c n2 = e2->node; node 251 security/selinux/ss/ebitmap.c n = e->node; node 266 security/selinux/ss/ebitmap.c n = e->node; node 295 security/selinux/ss/ebitmap.c e->node = n->next; node 322 security/selinux/ss/ebitmap.c new->next = e->node; node 323 security/selinux/ss/ebitmap.c e->node = new; node 336 security/selinux/ss/ebitmap.c n = e->node; node 344 security/selinux/ss/ebitmap.c e->node = NULL; node 378 security/selinux/ss/ebitmap.c e->node = NULL; node 417 security/selinux/ss/ebitmap.c e->node = tmp; node 34 security/selinux/ss/ebitmap.h struct ebitmap_node *node; /* first node in the bitmap */ node 39 security/selinux/ss/ebitmap.h #define ebitmap_startbit(e) ((e)->node ? (e)->node->startbit : 0) node 46 security/selinux/ss/ebitmap.h for (*n = e->node; *n; *n = (*n)->next) { node 78 security/selinux/ss/ebitmap.h (((bit) - (node)->startbit) / EBITMAP_UNIT_SIZE) node 80 security/selinux/ss/ebitmap.h (((bit) - (node)->startbit) % EBITMAP_UNIT_SIZE) node 40 security/selinux/ss/mls.c struct ebitmap_node *node; node 54 security/selinux/ss/mls.c ebitmap_for_each_positive_bit(e, node, i) { node 94 security/selinux/ss/mls.c struct ebitmap_node *node; node 113 security/selinux/ss/mls.c ebitmap_for_each_positive_bit(e, node, i) { node 163 security/selinux/ss/mls.c struct ebitmap_node *node; node 173 security/selinux/ss/mls.c ebitmap_for_each_positive_bit(&l->cat, node, i) { node 477 security/selinux/ss/mls.c struct ebitmap_node *node; node 492 security/selinux/ss/mls.c ebitmap_for_each_positive_bit(&c->range.level[l].cat, node, i) { node 1522 security/selinux/ss/policydb.c struct ebitmap_node *node; node 1533 security/selinux/ss/policydb.c ebitmap_for_each_positive_bit(&user->roles, node, bit) { node 1559 security/selinux/ss/policydb.c struct ebitmap_node *node; node 1570 security/selinux/ss/policydb.c ebitmap_for_each_positive_bit(&role->types, node, bit) { node 1929 security/selinux/ss/policydb.c c->u.node.addr = nodebuf[0]; /* network order */ node 1930 security/selinux/ss/policydb.c c->u.node.mask = nodebuf[1]; /* network order */ node 146 security/selinux/ss/policydb.h } node; /* node information */ node 388 security/selinux/ss/services.c struct avtab_node *node; node 457 security/selinux/ss/services.c for (node = avtab_search_node(&policydb.te_avtab, &avkey); node 458 security/selinux/ss/services.c node; node 459 security/selinux/ss/services.c node = avtab_search_node_next(node, avkey.specified)) { node 460 security/selinux/ss/services.c if (node->key.specified == AVTAB_ALLOWED) node 461 security/selinux/ss/services.c avd->allowed |= node->datum.data; node 462 security/selinux/ss/services.c else if (node->key.specified == AVTAB_AUDITALLOW) node 463 security/selinux/ss/services.c avd->auditallow |= node->datum.data; node 464 security/selinux/ss/services.c else if (node->key.specified == AVTAB_AUDITDENY) node 465 security/selinux/ss/services.c avd->auditdeny &= node->datum.data; node 1139 security/selinux/ss/services.c struct avtab_node *node; node 1209 security/selinux/ss/services.c node = avtab_search_node(&policydb.te_cond_avtab, &avkey); node 1210 security/selinux/ss/services.c for (; node; node = avtab_search_node_next(node, specified)) { node 1211 security/selinux/ss/services.c if (node->key.specified & AVTAB_ENABLED) { node 1212 security/selinux/ss/services.c avdatum = &node->datum; node 1872 security/selinux/ss/services.c if (c->u.node.addr == (addr & c->u.node.mask)) node 71 sound/aoa/aoa-gpio.h struct device_node *node; node 36 sound/aoa/aoa.h struct device_node *node; node 1036 sound/aoa/codecs/snd-aoa-codec-onyx.c onyx->codec.node = of_node_get(node); node 1091 sound/aoa/codecs/snd-aoa-codec-onyx.c of_node_put(onyx->codec.node); node 915 sound/aoa/codecs/snd-aoa-codec-tas.c tas->codec.node = of_node_get(node); node 922 sound/aoa/codecs/snd-aoa-codec-tas.c addr, node->full_name); node 982 sound/aoa/codecs/snd-aoa-codec-tas.c of_node_put(tas->codec.node); node 20 sound/aoa/core/snd-aoa-gpio-pmf.c rc = pmf_call_function(rt->node, #name "-mute", &args); \ node 43 sound/aoa/core/snd-aoa-gpio-pmf.c rc = pmf_call_function(rt->node, "hw-reset", &args); node 188 sound/aoa/core/snd-aoa-gpio-pmf.c err = pmf_register_irq_client(rt->node, node 229 sound/aoa/core/snd-aoa-gpio-pmf.c err = pmf_call_function(rt->node, name, &args); node 724 sound/aoa/fabrics/snd-aoa-fabric-layout.c if (codec->node && (strcmp(codec->node->name, "codec") == 0)) { node 733 sound/aoa/fabrics/snd-aoa-fabric-layout.c if (*ref != codec->node->linux_phandle) { node 951 sound/aoa/fabrics/snd-aoa-fabric-layout.c while ((sound = of_get_next_child(sdev->ofdev.node, sound))) { node 977 sound/aoa/fabrics/snd-aoa-fabric-layout.c ldev->gpio.node = sound->parent; node 77 sound/aoa/soundbus/core.c retval = add_uevent_var(env, "OF_NAME=%s", of->node->name); node 81 sound/aoa/soundbus/core.c retval = add_uevent_var(env, "OF_TYPE=%s", of->node->type); node 89 sound/aoa/soundbus/core.c compat = of_get_property(of->node, "compatible", &cplen); node 172 sound/aoa/soundbus/core.c !dev->ofdev.node || node 44 sound/aoa/soundbus/i2sbus/i2sbus-control.c np = i2sdev->sound.ofdev.node; node 211 sound/aoa/soundbus/i2sbus/i2sbus-core.c dev->sound.ofdev.node = np; node 336 sound/aoa/soundbus/i2sbus/i2sbus-core.c while ((np = of_get_next_child(dev->ofdev.node, np))) { node 12 sound/aoa/soundbus/sysfs.c return sprintf (buf, format_string, mdev->ofdev.node->field); \ node 28 sound/aoa/soundbus/sysfs.c of->node->name, of->node->type); node 50 sound/pci/hda/hda_generic.c struct hda_gnode *node; /* Node for PCM volume */ node 84 sound/pci/hda/hda_generic.c #define defcfg_type(node) (((node)->def_cfg & AC_DEFCFG_DEVICE) >> \ node 86 sound/pci/hda/hda_generic.c #define defcfg_location(node) (((node)->def_cfg & AC_DEFCFG_LOCATION) >> \ node 88 sound/pci/hda/hda_generic.c #define defcfg_port_conn(node) (((node)->def_cfg & AC_DEFCFG_PORT_CONN) >> \ node 97 sound/pci/hda/hda_generic.c struct hda_gnode *node, *n; node 102 sound/pci/hda/hda_generic.c list_for_each_entry_safe(node, n, &spec->nid_list, list) { node 103 sound/pci/hda/hda_generic.c if (node->conn_list != node->slist) node 104 sound/pci/hda/hda_generic.c kfree(node->conn_list); node 105 sound/pci/hda/hda_generic.c kfree(node); node 116 sound/pci/hda/hda_generic.c struct hda_gnode *node; node 120 sound/pci/hda/hda_generic.c node = kzalloc(sizeof(*node), GFP_KERNEL); node 121 sound/pci/hda/hda_generic.c if (node == NULL) node 123 sound/pci/hda/hda_generic.c node->nid = nid; node 127 sound/pci/hda/hda_generic.c kfree(node); node 130 sound/pci/hda/hda_generic.c if (nconns <= ARRAY_SIZE(node->slist)) node 131 sound/pci/hda/hda_generic.c node->conn_list = node->slist; node 133 sound/pci/hda/hda_generic.c node->conn_list = kmalloc(sizeof(hda_nid_t) * nconns, node 135 sound/pci/hda/hda_generic.c if (! node->conn_list) { node 137 sound/pci/hda/hda_generic.c kfree(node); node 141 sound/pci/hda/hda_generic.c memcpy(node->conn_list, conn_list, nconns * sizeof(hda_nid_t)); node 142 sound/pci/hda/hda_generic.c node->nconns = nconns; node 143 sound/pci/hda/hda_generic.c node->wid_caps = get_wcaps(codec, nid); node 144 sound/pci/hda/hda_generic.c node->type = (node->wid_caps & AC_WCAP_TYPE) >> AC_WCAP_TYPE_SHIFT; node 146 sound/pci/hda/hda_generic.c if (node->type == AC_WID_PIN) { node 147 sound/pci/hda/hda_generic.c node->pin_caps = snd_hda_param_read(codec, node->nid, AC_PAR_PIN_CAP); node 148 sound/pci/hda/hda_generic.c node->pin_ctl = snd_hda_codec_read(codec, node->nid, 0, AC_VERB_GET_PIN_WIDGET_CONTROL, 0); node 149 sound/pci/hda/hda_generic.c node->def_cfg = snd_hda_codec_read(codec, node->nid, 0, AC_VERB_GET_CONFIG_DEFAULT, 0); node 152 sound/pci/hda/hda_generic.c if (node->wid_caps & AC_WCAP_OUT_AMP) { node 153 sound/pci/hda/hda_generic.c if (node->wid_caps & AC_WCAP_AMP_OVRD) node 154 sound/pci/hda/hda_generic.c node->amp_out_caps = snd_hda_param_read(codec, node->nid, AC_PAR_AMP_OUT_CAP); node 155 sound/pci/hda/hda_generic.c if (! node->amp_out_caps) node 156 sound/pci/hda/hda_generic.c node->amp_out_caps = spec->def_amp_out_caps; node 158 sound/pci/hda/hda_generic.c if (node->wid_caps & AC_WCAP_IN_AMP) { node 159 sound/pci/hda/hda_generic.c if (node->wid_caps & AC_WCAP_AMP_OVRD) node 160 sound/pci/hda/hda_generic.c node->amp_in_caps = snd_hda_param_read(codec, node->nid, AC_PAR_AMP_IN_CAP); node 161 sound/pci/hda/hda_generic.c if (! node->amp_in_caps) node 162 sound/pci/hda/hda_generic.c node->amp_in_caps = spec->def_amp_in_caps; node 164 sound/pci/hda/hda_generic.c list_add_tail(&node->list, &spec->nid_list); node 205 sound/pci/hda/hda_generic.c struct hda_gnode *node; node 207 sound/pci/hda/hda_generic.c list_for_each_entry(node, &spec->nid_list, list) { node 208 sound/pci/hda/hda_generic.c if (node->nid == nid) node 209 sound/pci/hda/hda_generic.c return node; node 220 sound/pci/hda/hda_generic.c snd_printdd("UNMUTE OUT: NID=0x%x\n", node->nid); node 221 sound/pci/hda/hda_generic.c val = (node->amp_out_caps & AC_AMPCAP_NUM_STEPS) >> AC_AMPCAP_NUM_STEPS_SHIFT; node 222 sound/pci/hda/hda_generic.c ofs = (node->amp_out_caps & AC_AMPCAP_OFFSET) >> AC_AMPCAP_OFFSET_SHIFT; node 225 sound/pci/hda/hda_generic.c snd_hda_codec_amp_stereo(codec, node->nid, HDA_OUTPUT, 0, 0xff, val); node 235 sound/pci/hda/hda_generic.c snd_printdd("UNMUTE IN: NID=0x%x IDX=0x%x\n", node->nid, index); node 236 sound/pci/hda/hda_generic.c val = (node->amp_in_caps & AC_AMPCAP_NUM_STEPS) >> AC_AMPCAP_NUM_STEPS_SHIFT; node 237 sound/pci/hda/hda_generic.c ofs = (node->amp_in_caps & AC_AMPCAP_OFFSET) >> AC_AMPCAP_OFFSET_SHIFT; node 240 sound/pci/hda/hda_generic.c snd_hda_codec_amp_stereo(codec, node->nid, HDA_INPUT, index, 0xff, val); node 250 sound/pci/hda/hda_generic.c snd_printdd("CONNECT: NID=0x%x IDX=0x%x\n", node->nid, index); node 251 sound/pci/hda/hda_generic.c return snd_hda_codec_write_cache(codec, node->nid, 0, node 260 sound/pci/hda/hda_generic.c struct hda_gnode *node; node 262 sound/pci/hda/hda_generic.c list_for_each_entry(node, &spec->nid_list, list) { node 263 sound/pci/hda/hda_generic.c node->checked = 0; node 278 sound/pci/hda/hda_generic.c if (node->checked) node 281 sound/pci/hda/hda_generic.c node->checked = 1; node 282 sound/pci/hda/hda_generic.c if (node->type == AC_WID_AUD_OUT) { node 283 sound/pci/hda/hda_generic.c if (node->wid_caps & AC_WCAP_DIGITAL) { node 284 sound/pci/hda/hda_generic.c snd_printdd("Skip Digital OUT node %x\n", node->nid); node 287 sound/pci/hda/hda_generic.c snd_printdd("AUD_OUT found %x\n", node->nid); node 290 sound/pci/hda/hda_generic.c return node == spec->dac_node[dac_idx]; node 292 sound/pci/hda/hda_generic.c spec->dac_node[dac_idx] = node; node 293 sound/pci/hda/hda_generic.c if ((node->wid_caps & AC_WCAP_OUT_AMP) && node 295 sound/pci/hda/hda_generic.c spec->pcm_vol[spec->pcm_vol_nodes].node = node; node 302 sound/pci/hda/hda_generic.c for (i = 0; i < node->nconns; i++) { node 303 sound/pci/hda/hda_generic.c child = hda_get_node(spec, node->conn_list[i]); node 313 sound/pci/hda/hda_generic.c if (node->nconns > 1) node 314 sound/pci/hda/hda_generic.c select_input_connection(codec, node, i); node 315 sound/pci/hda/hda_generic.c unmute_input(codec, node, i); node 316 sound/pci/hda/hda_generic.c unmute_output(codec, node); node 321 sound/pci/hda/hda_generic.c if ((node->wid_caps & AC_WCAP_IN_AMP) || node 322 sound/pci/hda/hda_generic.c (node->wid_caps & AC_WCAP_OUT_AMP)) { node 324 sound/pci/hda/hda_generic.c spec->pcm_vol[n].node = node; node 345 sound/pci/hda/hda_generic.c struct hda_gnode *node; node 348 sound/pci/hda/hda_generic.c list_for_each_entry(node, &spec->nid_list, list) { node 349 sound/pci/hda/hda_generic.c if (node->type != AC_WID_PIN) node 352 sound/pci/hda/hda_generic.c if (! (node->pin_caps & AC_PINCAP_OUT)) node 354 sound/pci/hda/hda_generic.c if (defcfg_port_conn(node) == AC_JACK_PORT_NONE) node 357 sound/pci/hda/hda_generic.c if (jack_type != defcfg_type(node)) node 359 sound/pci/hda/hda_generic.c if (node->wid_caps & AC_WCAP_DIGITAL) node 363 sound/pci/hda/hda_generic.c if (! (node->pin_ctl & AC_PINCTL_OUT_EN)) node 367 sound/pci/hda/hda_generic.c err = parse_output_path(codec, spec, node, 0); node 371 sound/pci/hda/hda_generic.c err = parse_output_path(codec, spec, node, 1); node 377 sound/pci/hda/hda_generic.c unmute_output(codec, node); node 379 sound/pci/hda/hda_generic.c snd_hda_codec_write_cache(codec, node->nid, 0, node 382 sound/pci/hda/hda_generic.c ((node->pin_caps & AC_PINCAP_HP_DRV) ? node 384 sound/pci/hda/hda_generic.c return node; node 397 sound/pci/hda/hda_generic.c struct hda_gnode *node; node 403 sound/pci/hda/hda_generic.c node = parse_output_jack(codec, spec, AC_JACK_LINE_OUT); node 404 sound/pci/hda/hda_generic.c if (node) /* found, remember the PIN node */ node 405 sound/pci/hda/hda_generic.c spec->out_pin_node[0] = node; node 408 sound/pci/hda/hda_generic.c node = parse_output_jack(codec, spec, AC_JACK_SPEAKER); node 409 sound/pci/hda/hda_generic.c if (node) node 410 sound/pci/hda/hda_generic.c spec->out_pin_node[0] = node; node 413 sound/pci/hda/hda_generic.c node = parse_output_jack(codec, spec, AC_JACK_HP_OUT); node 414 sound/pci/hda/hda_generic.c if (node) { node 416 sound/pci/hda/hda_generic.c spec->out_pin_node[0] = node; node 418 sound/pci/hda/hda_generic.c spec->out_pin_node[1] = node; node 467 sound/pci/hda/hda_generic.c unsigned int location = defcfg_location(node); node 468 sound/pci/hda/hda_generic.c switch (defcfg_type(node)) { node 485 sound/pci/hda/hda_generic.c (node->pin_caps & node 512 sound/pci/hda/hda_generic.c if (node->checked) node 515 sound/pci/hda/hda_generic.c node->checked = 1; node 516 sound/pci/hda/hda_generic.c if (node->type != AC_WID_PIN) { node 517 sound/pci/hda/hda_generic.c for (i = 0; i < node->nconns; i++) { node 519 sound/pci/hda/hda_generic.c child = hda_get_node(spec, node->conn_list[i]); node 529 sound/pci/hda/hda_generic.c if (node->nconns > 1) node 530 sound/pci/hda/hda_generic.c select_input_connection(codec, node, i); node 531 sound/pci/hda/hda_generic.c unmute_input(codec, node, i); node 532 sound/pci/hda/hda_generic.c unmute_output(codec, node); node 540 sound/pci/hda/hda_generic.c if (! (node->pin_caps & AC_PINCAP_IN)) node 543 sound/pci/hda/hda_generic.c if (defcfg_port_conn(node) == AC_JACK_PORT_NONE) node 546 sound/pci/hda/hda_generic.c if (node->wid_caps & AC_WCAP_DIGITAL) node 556 sound/pci/hda/hda_generic.c type = get_input_type(node, &pinctl); node 559 sound/pci/hda/hda_generic.c if (! (node->pin_ctl & AC_PINCTL_IN_EN)) node 568 sound/pci/hda/hda_generic.c unmute_input(codec, node, 0); /* index = 0? */ node 570 sound/pci/hda/hda_generic.c snd_hda_codec_write_cache(codec, node->nid, 0, node 605 sound/pci/hda/hda_generic.c struct hda_gnode *node; node 621 sound/pci/hda/hda_generic.c node = hda_get_node(spec, adc_node->conn_list[i]); node 622 sound/pci/hda/hda_generic.c if (node && node->type == AC_WID_PIN) { node 623 sound/pci/hda/hda_generic.c err = parse_adc_sub_nodes(codec, spec, node); node 632 sound/pci/hda/hda_generic.c node = hda_get_node(spec, adc_node->conn_list[i]); node 633 sound/pci/hda/hda_generic.c if (node && node->type != AC_WID_PIN) { node 634 sound/pci/hda/hda_generic.c err = parse_adc_sub_nodes(codec, spec, node); node 660 sound/pci/hda/hda_generic.c struct hda_gnode *node; node 668 sound/pci/hda/hda_generic.c list_for_each_entry(node, &spec->nid_list, list) { node 669 sound/pci/hda/hda_generic.c if (node->wid_caps & AC_WCAP_DIGITAL) node 671 sound/pci/hda/hda_generic.c if (node->type == AC_WID_AUD_IN) { node 672 sound/pci/hda/hda_generic.c err = parse_input_path(codec, node); node 720 sound/pci/hda/hda_generic.c if ((node->wid_caps & AC_WCAP_IN_AMP) && node 721 sound/pci/hda/hda_generic.c (node->amp_in_caps & AC_AMPCAP_MUTE)) { node 722 sound/pci/hda/hda_generic.c knew = (struct snd_kcontrol_new)HDA_CODEC_MUTE(name, node->nid, index, HDA_INPUT); node 724 sound/pci/hda/hda_generic.c add_input_loopback(codec, node->nid, HDA_INPUT, index); node 725 sound/pci/hda/hda_generic.c snd_printdd("[%s] NID=0x%x, DIR=IN, IDX=0x%x\n", name, node->nid, index); node 729 sound/pci/hda/hda_generic.c } else if ((node->wid_caps & AC_WCAP_OUT_AMP) && node 730 sound/pci/hda/hda_generic.c (node->amp_out_caps & AC_AMPCAP_MUTE)) { node 731 sound/pci/hda/hda_generic.c knew = (struct snd_kcontrol_new)HDA_CODEC_MUTE(name, node->nid, 0, HDA_OUTPUT); node 733 sound/pci/hda/hda_generic.c add_input_loopback(codec, node->nid, HDA_OUTPUT, 0); node 734 sound/pci/hda/hda_generic.c snd_printdd("[%s] NID=0x%x, DIR=OUT\n", name, node->nid); node 744 sound/pci/hda/hda_generic.c if ((node->wid_caps & AC_WCAP_IN_AMP) && node 745 sound/pci/hda/hda_generic.c (node->amp_in_caps & AC_AMPCAP_NUM_STEPS)) { node 746 sound/pci/hda/hda_generic.c knew = (struct snd_kcontrol_new)HDA_CODEC_VOLUME(name, node->nid, index, HDA_INPUT); node 747 sound/pci/hda/hda_generic.c snd_printdd("[%s] NID=0x%x, DIR=IN, IDX=0x%x\n", name, node->nid, index); node 751 sound/pci/hda/hda_generic.c } else if ((node->wid_caps & AC_WCAP_OUT_AMP) && node 752 sound/pci/hda/hda_generic.c (node->amp_out_caps & AC_AMPCAP_NUM_STEPS)) { node 753 sound/pci/hda/hda_generic.c knew = (struct snd_kcontrol_new)HDA_CODEC_VOLUME(name, node->nid, 0, HDA_OUTPUT); node 754 sound/pci/hda/hda_generic.c snd_printdd("[%s] NID=0x%x, DIR=OUT\n", name, node->nid); node 790 sound/pci/hda/hda_generic.c err = create_mixer(codec, spec->pcm_vol[i].node, node 807 sound/pci/hda/hda_generic.c return create_mixer(codec, spec->pcm_vol[0].node, node 893 sound/pci/hda/hda_generic.c if (node->checked) node 896 sound/pci/hda/hda_generic.c node->checked = 1; node 897 sound/pci/hda/hda_generic.c if (node == dest_node) { node 902 sound/pci/hda/hda_generic.c for (i = 0; i < node->nconns; i++) { node 903 sound/pci/hda/hda_generic.c struct hda_gnode *child = hda_get_node(spec, node->conn_list[i]); node 911 sound/pci/hda/hda_generic.c err = create_mixer(codec, node, i, type, node 921 sound/pci/hda/hda_generic.c if (node->nconns > 1) node 922 sound/pci/hda/hda_generic.c select_input_connection(codec, node, i); node 923 sound/pci/hda/hda_generic.c unmute_input(codec, node, i); node 924 sound/pci/hda/hda_generic.c unmute_output(codec, node); node 937 sound/pci/hda/hda_generic.c struct hda_gnode *node; node 944 sound/pci/hda/hda_generic.c list_for_each_entry(node, &spec->nid_list, list) { node 945 sound/pci/hda/hda_generic.c if (node->type != AC_WID_PIN) node 948 sound/pci/hda/hda_generic.c if (! (node->pin_caps & AC_PINCAP_IN)) node 950 sound/pci/hda/hda_generic.c type = get_input_type(node, NULL); node 957 sound/pci/hda/hda_generic.c node, type); node 235 sound/pcmcia/pdaudiocf/pdaudiocf.c link->dev_node = &pdacf->node; node 120 sound/pcmcia/pdaudiocf/pdaudiocf.h dev_node_t node; node 244 sound/pcmcia/vx/vxpocket.c link->dev_node = &vxp->node; node 46 sound/pcmcia/vx/vxpocket.h dev_node_t node; node 832 sound/ppc/pmac.c ppc_md.feature_call(PMAC_FTR_SOUND_CHIP_ENABLE, chip->node, 0, enable); node 848 sound/ppc/pmac.c if (chip->node) node 879 sound/ppc/pmac.c if (chip->node) { node 891 sound/ppc/pmac.c of_node_put(chip->node); node 916 sound/ppc/pmac.c for (mio = chip->node->parent; mio; mio = mio->parent) { node 968 sound/ppc/pmac.c chip->node = of_find_node_by_name(NULL, "awacs"); node 969 sound/ppc/pmac.c sound = of_node_get(chip->node); node 975 sound/ppc/pmac.c if (!chip->node) node 976 sound/ppc/pmac.c chip->node = of_find_node_by_name(NULL, "davbus"); node 981 sound/ppc/pmac.c if (! chip->node) { node 982 sound/ppc/pmac.c chip->node = of_find_node_by_name(NULL, "i2s-a"); node 983 sound/ppc/pmac.c if (chip->node && chip->node->parent && node 984 sound/ppc/pmac.c chip->node->parent->parent) { node 985 sound/ppc/pmac.c if (of_device_is_compatible(chip->node->parent->parent, node 990 sound/ppc/pmac.c if (! chip->node) node 995 sound/ppc/pmac.c while (sound && sound->parent != chip->node) node 999 sound/ppc/pmac.c of_node_put(chip->node); node 1000 sound/ppc/pmac.c chip->node = NULL; node 1014 sound/ppc/pmac.c of_node_put(chip->node); node 1015 sound/ppc/pmac.c chip->node = NULL; node 1060 sound/ppc/pmac.c macio = macio_find(chip->node, macio_unknown); node 1212 sound/ppc/pmac.c np = chip->node; node 1322 sound/ppc/pmac.c for (mio = chip->node->parent; mio; mio = mio->parent) { node 95 sound/ppc/pmac.h struct device_node *node; node 1096 sound/ppc/tumbler.c struct device_node *node; node 1102 sound/ppc/tumbler.c node = find_compatible_audio_device(device); node 1104 sound/ppc/tumbler.c node = find_audio_device(device); node 1105 sound/ppc/tumbler.c if (! node) { node 1111 sound/ppc/tumbler.c base = of_get_property(node, "AAPL,address", NULL); node 1113 sound/ppc/tumbler.c base = of_get_property(node, "reg", NULL); node 1117 sound/ppc/tumbler.c of_node_put(node); node 1128 sound/ppc/tumbler.c base = of_get_property(node, "audio-gpio-active-state", NULL); node 1143 sound/ppc/tumbler.c prop = of_get_property(node, platform, NULL); node 1159 sound/ppc/tumbler.c ret = irq_of_parse_and_map(node, 0); node 1160 sound/ppc/tumbler.c of_node_put(node); node 1360 sound/ppc/tumbler.c for (np = chip->node->child; np; np = np->sibling) { node 724 sound/soc/fsl/mpc5200_psc_i2s.c prop = of_get_property(op->node, "cell-index", &size); node 730 sound/soc/fsl/mpc5200_psc_i2s.c irq = irq_of_parse_and_map(op->node, 0); node 731 sound/soc/fsl/mpc5200_psc_i2s.c if (of_address_to_resource(op->node, 0, &res)) { node 787 sound/soc/fsl/mpc5200_psc_i2s.c if (of_get_property(op->node, "fsl,cellslave", NULL)) node 795 sound/soc/fsl/mpc5200_psc_i2s.c if (!of_get_property(op->node, "codec-handle", NULL)) node 828 sound/soc/fsl/mpc5200_psc_i2s.c of_snd_soc_register_platform(&psc_i2s_pcm_soc_platform, op->node, node 215 sound/soc/fsl/mpc8610_hpcd.c struct device_node *np = ofdev->node; node 108 sound/soc/fsl/soc-of-simple.c pr_info("registering ASoC codec driver: %s\n", node->full_name); node 111 sound/soc/fsl/soc-of-simple.c of_soc = of_snd_soc_get_device(node); node 120 sound/soc/fsl/soc-of-simple.c of_soc->dai_link.name = (char *)node->name; node 121 sound/soc/fsl/soc-of-simple.c of_soc->dai_link.stream_name = (char *)node->name; node 142 sound/soc/fsl/soc-of-simple.c pr_info("registering ASoC platform driver: %s\n", node->full_name); node 144 sound/soc/fsl/soc-of-simple.c handle = of_get_property(node, "codec-handle", &len); node 159 sound/soc/fsl/soc-of-simple.c of_soc->platform_node = node; node 2078 sound/sparc/cs4231.c if (!strcmp(op->node->parent->name, "ebus")) node 2082 sound/sparc/cs4231.c if (!strcmp(op->node->parent->name, "sbus") || node 2083 sound/sparc/cs4231.c !strcmp(op->node->parent->name, "sbi")) node 2650 sound/sparc/dbri.c dbri->irq, op->node->name[9], dbri->mm.version);