node              369 arch/x86/kernel/aperture_64.c 	int i, node;
node              381 arch/x86/kernel/aperture_64.c 	node = 0;
node              403 arch/x86/kernel/aperture_64.c 					node, aper_base, aper_size >> 20);
node              404 arch/x86/kernel/aperture_64.c 			node++;
node              186 arch/x86/kernel/cpu/amd.c 	int i, node;
node              189 arch/x86/kernel/cpu/amd.c 		node = apicid_to_node[i];
node              190 arch/x86/kernel/cpu/amd.c 		if (node != NUMA_NO_NODE && node_online(node))
node              191 arch/x86/kernel/cpu/amd.c 			return node;
node              194 arch/x86/kernel/cpu/amd.c 		node = apicid_to_node[i];
node              195 arch/x86/kernel/cpu/amd.c 		if (node != NUMA_NO_NODE && node_online(node))
node              196 arch/x86/kernel/cpu/amd.c 			return node;
node              224 arch/x86/kernel/cpu/amd.c 	int node;
node              227 arch/x86/kernel/cpu/amd.c 	node = c->phys_proc_id;
node              229 arch/x86/kernel/cpu/amd.c 		node = apicid_to_node[apicid];
node              230 arch/x86/kernel/cpu/amd.c 	if (!node_online(node)) {
node              245 arch/x86/kernel/cpu/amd.c 			node = apicid_to_node[ht_nodeid];
node              247 arch/x86/kernel/cpu/amd.c 		if (!node_online(node))
node              248 arch/x86/kernel/cpu/amd.c 			node = nearby_node(apicid);
node              250 arch/x86/kernel/cpu/amd.c 	numa_set_node(cpu, node);
node              252 arch/x86/kernel/cpu/amd.c 	printk(KERN_INFO "CPU %d/%x -> Node %d\n", cpu, apicid, node);
node              166 arch/x86/kernel/cpu/intel.c 	unsigned node;
node              172 arch/x86/kernel/cpu/intel.c 	node = apicid_to_node[apicid];
node              173 arch/x86/kernel/cpu/intel.c 	if (node == NUMA_NO_NODE || !node_online(node))
node              174 arch/x86/kernel/cpu/intel.c 		node = first_node(node_online_map);
node              175 arch/x86/kernel/cpu/intel.c 	numa_set_node(cpu, node);
node              177 arch/x86/kernel/cpu/intel.c 	printk(KERN_INFO "CPU %d/%x -> Node %d\n", cpu, apicid, node);
node              673 arch/x86/kernel/cpu/intel_cacheinfo.c 	for (i = 0; i <= node; i++) {
node              693 arch/x86/kernel/cpu/intel_cacheinfo.c 	int node = cpu_to_node(first_cpu(this_leaf->shared_cpu_map));
node              701 arch/x86/kernel/cpu/intel_cacheinfo.c 	dev = get_k8_northbridge(node);
node              727 arch/x86/kernel/cpu/intel_cacheinfo.c 	int node = cpu_to_node(first_cpu(this_leaf->shared_cpu_map));
node              744 arch/x86/kernel/cpu/intel_cacheinfo.c 	dev = get_k8_northbridge(node);
node               33 arch/x86/kernel/kdebugfs.c 	struct setup_data_node *node = file->private_data;
node               42 arch/x86/kernel/kdebugfs.c 	if (pos >= node->len)
node               45 arch/x86/kernel/kdebugfs.c 	if (count > node->len - pos)
node               46 arch/x86/kernel/kdebugfs.c 		count = node->len - pos;
node               47 arch/x86/kernel/kdebugfs.c 	pa = node->paddr + sizeof(struct setup_data) + pos;
node               95 arch/x86/kernel/kdebugfs.c 	type = debugfs_create_x32("type", S_IRUGO, d, &node->type);
node              100 arch/x86/kernel/kdebugfs.c 	data = debugfs_create_file("data", S_IRUGO, d, node, &fops_setup_data);
node              117 arch/x86/kernel/kdebugfs.c 	struct setup_data_node *node;
node              133 arch/x86/kernel/kdebugfs.c 		node = kmalloc(sizeof(*node), GFP_KERNEL);
node              134 arch/x86/kernel/kdebugfs.c 		if (!node) {
node              142 arch/x86/kernel/kdebugfs.c 				kfree(node);
node              150 arch/x86/kernel/kdebugfs.c 		node->paddr = pa_data;
node              151 arch/x86/kernel/kdebugfs.c 		node->type = data->type;
node              152 arch/x86/kernel/kdebugfs.c 		node->len = data->len;
node              153 arch/x86/kernel/kdebugfs.c 		error = create_setup_data_node(d, no, node);
node              679 arch/x86/kernel/kprobes.c 	struct hlist_node *node, *tmp;
node              707 arch/x86/kernel/kprobes.c 	hlist_for_each_entry_safe(ri, node, tmp, head, hlist) {
node              735 arch/x86/kernel/kprobes.c 	hlist_for_each_entry_safe(ri, node, tmp, &empty_rp, hlist) {
node               48 arch/x86/kernel/numaq_32.c 	int node;
node               54 arch/x86/kernel/numaq_32.c 	for_each_node(node) {
node               55 arch/x86/kernel/numaq_32.c 		if (scd->quads_present31_0 & (1 << node)) {
node               56 arch/x86/kernel/numaq_32.c 			node_set_online(node);
node               57 arch/x86/kernel/numaq_32.c 			eq = &scd->eq[node];
node               59 arch/x86/kernel/numaq_32.c 			node_start_pfn[node] = MB_TO_PAGES(
node               61 arch/x86/kernel/numaq_32.c 			node_end_pfn[node] = MB_TO_PAGES(
node               64 arch/x86/kernel/numaq_32.c 			e820_register_active_regions(node, node_start_pfn[node],
node               65 arch/x86/kernel/numaq_32.c 							node_end_pfn[node]);
node               66 arch/x86/kernel/numaq_32.c 			memory_present(node,
node               67 arch/x86/kernel/numaq_32.c 				node_start_pfn[node], node_end_pfn[node]);
node               68 arch/x86/kernel/numaq_32.c 			node_remap_size[node] = node_memmap_size_bytes(node,
node               69 arch/x86/kernel/numaq_32.c 							node_start_pfn[node],
node               70 arch/x86/kernel/numaq_32.c 							node_end_pfn[node]);
node              159 arch/x86/kernel/setup_percpu.c 		int node = early_cpu_to_node(cpu);
node              160 arch/x86/kernel/setup_percpu.c 		if (!node_online(node) || !NODE_DATA(node)) {
node              164 arch/x86/kernel/setup_percpu.c 				cpu, node);
node              170 arch/x86/kernel/setup_percpu.c 			ptr = alloc_bootmem_pages_node(NODE_DATA(node), size);
node              173 arch/x86/kernel/setup_percpu.c 					 cpu, node, __pa(ptr));
node              203 arch/x86/kernel/setup_percpu.c 	unsigned int node, num = 0;
node              208 arch/x86/kernel/setup_percpu.c 		for_each_node_mask(node, node_possible_map)
node              209 arch/x86/kernel/setup_percpu.c 			num = node;
node              227 arch/x86/kernel/setup_percpu.c 	if (cpu_pda(cpu) && node != NUMA_NO_NODE)
node              228 arch/x86/kernel/setup_percpu.c 		cpu_pda(cpu)->nodenumber = node;
node              231 arch/x86/kernel/setup_percpu.c 		cpu_to_node_map[cpu] = node;
node              234 arch/x86/kernel/setup_percpu.c 		per_cpu(x86_cpu_to_node_map, cpu) = node;
node              264 arch/x86/kernel/setup_percpu.c 	int node = cpu_to_node(cpu);
node              274 arch/x86/kernel/setup_percpu.c 	mask = &node_to_cpumask_map[node];
node              282 arch/x86/kernel/setup_percpu.c 		enable? "numa_add_cpu":"numa_remove_cpu", cpu, node, buf);
node              337 arch/x86/kernel/setup_percpu.c 			node);
node              341 arch/x86/kernel/setup_percpu.c 	if (node >= nr_node_ids) {
node              344 arch/x86/kernel/setup_percpu.c 			node, nr_node_ids);
node              348 arch/x86/kernel/setup_percpu.c 	return &node_to_cpumask_map[node];
node              363 arch/x86/kernel/setup_percpu.c 			"node_to_cpumask(%d): no node_to_cpumask_map!\n", node);
node              367 arch/x86/kernel/setup_percpu.c 	if (node >= nr_node_ids) {
node              370 arch/x86/kernel/setup_percpu.c 			node, nr_node_ids);
node              374 arch/x86/kernel/setup_percpu.c 	return node_to_cpumask_map[node];
node              147 arch/x86/kernel/smpboot.c 	printk(KERN_INFO "Mapping cpu %d to node %d\n", cpu, node);
node              148 arch/x86/kernel/smpboot.c 	cpu_set(cpu, node_to_cpumask_map[node]);
node              149 arch/x86/kernel/smpboot.c 	cpu_to_node_map[cpu] = node;
node              155 arch/x86/kernel/smpboot.c 	int node;
node              158 arch/x86/kernel/smpboot.c 	for (node = 0; node < MAX_NUMNODES; node++)
node              159 arch/x86/kernel/smpboot.c 		cpu_clear(cpu, node_to_cpumask_map[node]);
node              177 arch/x86/kernel/smpboot.c 	int node = apicid_to_node(apicid);
node              179 arch/x86/kernel/smpboot.c 	if (!node_online(node))
node              180 arch/x86/kernel/smpboot.c 		node = first_online_node;
node              183 arch/x86/kernel/smpboot.c 	map_cpu_to_node(cpu, node);
node              776 arch/x86/kernel/smpboot.c 	int node = cpu_to_node(cpu);
node              782 arch/x86/kernel/smpboot.c 	newpda = kmalloc_node(size, GFP_ATOMIC, node);
node              785 arch/x86/kernel/smpboot.c 			"for CPU %d on node %d\n", cpu, node);
node               45 arch/x86/kernel/summit_32.c 	int twister = 0, node = 0;
node               61 arch/x86/kernel/summit_32.c 			node = scal_devs[i]->node_id;
node              101 arch/x86/kernel/summit_32.c 		mp_bus_id_to_node[bus] = node;
node              594 arch/x86/kernel/tlb_uv.c 	    kmalloc_node(sizeof(struct bau_control), GFP_KERNEL, node);
node              599 arch/x86/kernel/tlb_uv.c 			 DEST_Q_SIZE, GFP_KERNEL, node);
node              607 arch/x86/kernel/tlb_uv.c 	    kmalloc_node(sizeof(int) * DEST_NUM_RESOURCES, GFP_KERNEL, node);
node              656 arch/x86/kernel/tlb_uv.c 	    kmalloc_node(16384, GFP_KERNEL, node);
node              696 arch/x86/kernel/tlb_uv.c 		GFP_KERNEL, node);
node              730 arch/x86/kernel/tlb_uv.c 	bau_tablesp = uv_table_bases_init(blade, node);
node              732 arch/x86/kernel/tlb_uv.c 	adp = uv_activation_descriptor_init(node, pnode);
node              733 arch/x86/kernel/tlb_uv.c 	pqp = uv_payload_queue_init(node, pnode, bau_tablesp);
node              734 arch/x86/kernel/tlb_uv.c 	uv_table_bases_finish(blade, node, cur_cpu, bau_tablesp, adp);
node              754 arch/x86/kernel/tlb_uv.c 	int node;
node              767 arch/x86/kernel/tlb_uv.c 	for_each_online_node(node) {
node              768 arch/x86/kernel/tlb_uv.c 		blade = uv_node_to_blade_id(node);
node              779 arch/x86/kernel/tlb_uv.c 	for_each_online_node(node) {
node              780 arch/x86/kernel/tlb_uv.c 		blade = uv_node_to_blade_id(node);
node              784 arch/x86/kernel/tlb_uv.c 		uv_init_blade(blade, node, cur_cpu);
node              215 arch/x86/kernel/vsyscall_64.c 	if (node)
node              216 arch/x86/kernel/vsyscall_64.c 		*node = p >> 12;
node              254 arch/x86/kernel/vsyscall_64.c 	unsigned long node = 0;
node              256 arch/x86/kernel/vsyscall_64.c 	node = cpu_to_node(cpu);
node              259 arch/x86/kernel/vsyscall_64.c 		write_rdtscp_aux((node << 12) | cpu);
node              266 arch/x86/kernel/vsyscall_64.c 	d |= (node & 0xf) << 12;
node              267 arch/x86/kernel/vsyscall_64.c 	d |= (node >> 4) << 48;
node              791 arch/x86/kvm/mmu.c 	struct hlist_node *node;
node              809 arch/x86/kvm/mmu.c 	hlist_for_each_entry(pte_chain, node, &sp->parent_ptes, link) {
node              828 arch/x86/kvm/mmu.c 	struct hlist_node *node;
node              836 arch/x86/kvm/mmu.c 	hlist_for_each_entry(pte_chain, node, &sp->parent_ptes, link)
node              876 arch/x86/kvm/mmu.c 	struct hlist_node *node;
node              881 arch/x86/kvm/mmu.c 	hlist_for_each_entry(sp, node, bucket, hash_link)
node              904 arch/x86/kvm/mmu.c 	struct hlist_node *node;
node              920 arch/x86/kvm/mmu.c 	hlist_for_each_entry(sp, node, bucket, hash_link)
node             1069 arch/x86/kvm/mmu.c 	struct hlist_node *node, *n;
node             1076 arch/x86/kvm/mmu.c 	hlist_for_each_entry_safe(sp, node, n, bucket, hash_link)
node             1805 arch/x86/kvm/mmu.c 	struct hlist_node *node, *n;
node             1839 arch/x86/kvm/mmu.c 	hlist_for_each_entry_safe(sp, node, n, bucket, hash_link) {
node             2075 arch/x86/kvm/mmu.c 	struct kvm_mmu_page *sp, *node;
node             2078 arch/x86/kvm/mmu.c 	list_for_each_entry_safe(sp, node, &kvm->arch.active_mmu_pages, link)
node             1207 arch/x86/kvm/vmx.c 	int node = cpu_to_node(cpu);
node             1211 arch/x86/kvm/vmx.c 	pages = alloc_pages_node(node, GFP_KERNEL, vmcs_config.order);
node             1155 arch/x86/mm/init_64.c 		pgd = vmemmap_pgd_populate(addr, node);
node             1159 arch/x86/mm/init_64.c 		pud = vmemmap_pud_populate(pgd, addr, node);
node             1165 arch/x86/mm/init_64.c 			pmd = vmemmap_pmd_populate(pud, addr, node);
node             1170 arch/x86/mm/init_64.c 			p = vmemmap_pte_populate(pmd, addr, node);
node             1184 arch/x86/mm/init_64.c 				p = vmemmap_alloc_block(PMD_SIZE, node);
node             1193 arch/x86/mm/init_64.c 				if (p_end != p || node_start != node) {
node             1198 arch/x86/mm/init_64.c 					node_start = node;
node             1205 arch/x86/mm/init_64.c 				vmemmap_verify((pte_t *)pmd, node, addr, next);
node              209 arch/x86/mm/numa_32.c 	int node;
node              211 arch/x86/mm/numa_32.c 	for_each_online_node(node) {
node              212 arch/x86/mm/numa_32.c 		printk(KERN_DEBUG "remap_numa_kva: node %d\n", node);
node              213 arch/x86/mm/numa_32.c 		for (pfn=0; pfn < node_remap_size[node]; pfn += PTRS_PER_PTE) {
node              214 arch/x86/mm/numa_32.c 			vaddr = node_remap_start_vaddr[node]+(pfn<<PAGE_SHIFT);
node              217 arch/x86/mm/numa_32.c 				node_remap_start_pfn[node] + pfn);
node              219 arch/x86/mm/numa_32.c 				node_remap_start_pfn[node] + pfn, 
node              627 arch/x86/mm/numa_64.c 		int node;
node              632 arch/x86/mm/numa_64.c 		node = apicid_to_node[apicid];
node              633 arch/x86/mm/numa_64.c 		if (node == NUMA_NO_NODE)
node              635 arch/x86/mm/numa_64.c 		if (!node_online(node))
node              637 arch/x86/mm/numa_64.c 		numa_set_node(cpu, node);
node              122 arch/x86/mm/srat_64.c 	int pxm, node;
node              134 arch/x86/mm/srat_64.c 	node = setup_node(pxm);
node              135 arch/x86/mm/srat_64.c 	if (node < 0) {
node              145 arch/x86/mm/srat_64.c 	apicid_to_node[apic_id] = node;
node              148 arch/x86/mm/srat_64.c 	       pxm, apic_id, node);
node              169 arch/x86/mm/srat_64.c 	struct bootnode *nd = &nodes_add[node];
node              190 arch/x86/mm/srat_64.c 	if (!hotadd_enough_memory(&nodes_add[node]))  {
node              227 arch/x86/mm/srat_64.c 	int node, pxm;
node              244 arch/x86/mm/srat_64.c 	node = setup_node(pxm);
node              245 arch/x86/mm/srat_64.c 	if (node < 0) {
node              251 arch/x86/mm/srat_64.c 	if (i == node) {
node              263 arch/x86/mm/srat_64.c 	nd = &nodes[node];
node              265 arch/x86/mm/srat_64.c 	if (!node_test_and_set(node, nodes_parsed)) {
node              275 arch/x86/mm/srat_64.c 	printk(KERN_INFO "SRAT: Node %u PXM %u %lx-%lx\n", node, pxm,
node              277 arch/x86/mm/srat_64.c 	e820_register_active_regions(node, start >> PAGE_SHIFT,
node              279 arch/x86/mm/srat_64.c 	push_node_boundaries(node, nd->start >> PAGE_SHIFT,
node              283 arch/x86/mm/srat_64.c 	    (reserve_hotadd(node, start, end) < 0)) {
node              288 arch/x86/mm/srat_64.c 			node_clear(node, nodes_parsed);
node              293 arch/x86/mm/srat_64.c 	memblk_nodeid[num_node_memblks] = node;
node              329 arch/x86/mm/srat_64.c 	node_clear(node, nodes_parsed);
node              331 arch/x86/mm/srat_64.c 		if (apicid_to_node[i] == node)
node              386 arch/x86/mm/srat_64.c 		int node = early_cpu_to_node(i);
node              388 arch/x86/mm/srat_64.c 		if (node == NUMA_NO_NODE)
node              390 arch/x86/mm/srat_64.c 		if (!node_isset(node, node_possible_map))
node              155 arch/x86/pci/acpi.c 	int node;
node              166 arch/x86/pci/acpi.c 	node = -1;
node              170 arch/x86/pci/acpi.c 		node = pxm_to_node(pxm);
node              171 arch/x86/pci/acpi.c 	if (node != -1)
node              172 arch/x86/pci/acpi.c 		set_mp_bus_to_node(busnum, node);
node              175 arch/x86/pci/acpi.c 		node = get_mp_bus_to_node(busnum);
node              177 arch/x86/pci/acpi.c 	if (node != -1 && !node_online(node))
node              178 arch/x86/pci/acpi.c 		node = -1;
node              191 arch/x86/pci/acpi.c 	sd->node = node;
node              210 arch/x86/pci/acpi.c 	if (bus && node != -1) {
node              214 arch/x86/pci/acpi.c 				busnum, pxm, node);
node              217 arch/x86/pci/acpi.c 			busnum, node);
node               29 arch/x86/pci/amd_bus.c 		mp_bus_to_node[busnum] = node;
node               34 arch/x86/pci/amd_bus.c 	int node = -1;
node               37 arch/x86/pci/amd_bus.c 		return node;
node               39 arch/x86/pci/amd_bus.c 	node = mp_bus_to_node[busnum];
node               45 arch/x86/pci/amd_bus.c 	if (node != -1 && !node_online(node))
node               46 arch/x86/pci/amd_bus.c 		node = -1;
node               48 arch/x86/pci/amd_bus.c 	return node;
node               58 arch/x86/pci/amd_bus.c 	mp_bus_to_node[busnum] = (unsigned char) node;
node               63 arch/x86/pci/amd_bus.c 	int node;
node               67 arch/x86/pci/amd_bus.c 	node = mp_bus_to_node[busnum];
node               68 arch/x86/pci/amd_bus.c 	return node;
node               88 arch/x86/pci/amd_bus.c 	int node;
node              283 arch/x86/pci/amd_bus.c 	int node;
node              338 arch/x86/pci/amd_bus.c 		node = (reg >> 4) & 0x07;
node              341 arch/x86/pci/amd_bus.c 			mp_bus_to_node[j] = (unsigned char) node;
node              348 arch/x86/pci/amd_bus.c 		info->node = node;
node              370 arch/x86/pci/amd_bus.c 		node = reg & 0x07;
node              377 arch/x86/pci/amd_bus.c 			if (info->node == node && info->link == link)
node              385 arch/x86/pci/amd_bus.c 		       node, link, (u64)start, (u64)end);
node              397 arch/x86/pci/amd_bus.c 		if (info->node == def_node && info->link == def_link)
node              440 arch/x86/pci/amd_bus.c 		node = reg & 0x07;
node              449 arch/x86/pci/amd_bus.c 			if (info->node == node && info->link == link)
node              458 arch/x86/pci/amd_bus.c 		       node, link, (u64)start, (u64)end);
node              523 arch/x86/pci/amd_bus.c 		if (info->node == def_node && info->link == def_link)
node              546 arch/x86/pci/amd_bus.c 		       info->bus_min, info->bus_max, info->node, info->link);
node              395 arch/x86/pci/common.c 	sd->node = get_mp_bus_to_node(busnum);
node              564 arch/x86/pci/common.c 	sd->node = node;
node              122 arch/x86/pci/i386.c 	list_for_each_entry(bus, bus_list, node) {
node              141 arch/x86/pci/irq.c 		int node;
node              144 arch/x86/pci/irq.c 		node = get_mp_bus_to_node(i);
node              145 arch/x86/pci/irq.c 		if (pci_scan_bus_on_node(i, &pci_root_ops, node))
node               15 arch/x86/pci/legacy.c 	long node;
node               25 arch/x86/pci/legacy.c 		node = get_mp_bus_to_node(n);
node               31 arch/x86/pci/legacy.c 				pci_scan_bus_on_node(n, &pci_root_ops, node);
node               30 arch/x86/vdso/vgetcpu.c 	if (node)
node               31 arch/x86/vdso/vgetcpu.c 		*node = p >> 12;
node              229 block/as-iosched.c 	struct io_context *ioc = get_io_context(GFP_ATOMIC, node);
node             1196 block/as-iosched.c 	rq->elevator_private = as_get_io_context(q->node);
node             1333 block/as-iosched.c 		ioc = as_get_io_context(q->node);
node             1363 block/as-iosched.c 	ad = kmalloc_node(sizeof(*ad), GFP_KERNEL | __GFP_ZERO, q->node);
node              469 block/blk-core.c 				mempool_free_slab, request_cachep, q->node);
node              561 block/blk-core.c 	q->node = node_id;
node              734 block/blk-core.c 			ioc = current_io_context(GFP_ATOMIC, q->node);
node              849 block/blk-core.c 		ioc = current_io_context(GFP_NOIO, q->node);
node               91 block/blk-ioc.c 	ret = kmem_cache_alloc_node(iocontext_cachep, gfp_flags, node);
node              126 block/blk-ioc.c 	ret = alloc_io_context(gfp_flags, node);
node              151 block/blk-ioc.c 		ret = current_io_context(gfp_flags, node);
node             1341 block/cfq-iosched.c 							cfqd->queue->node);
node             1457 block/cfq-iosched.c 					cfqd->queue->node);
node             1463 block/cfq-iosched.c 					cfqd->queue->node);
node             1659 block/cfq-iosched.c 	ioc = get_io_context(gfp_mask, cfqd->queue->node);
node             2214 block/cfq-iosched.c 	cfqd = kmalloc_node(sizeof(*cfqd), GFP_KERNEL | __GFP_ZERO, q->node);
node               68 block/deadline-iosched.c 	struct rb_node *node = rb_next(&rq->rb_node);
node               70 block/deadline-iosched.c 	if (node)
node               71 block/deadline-iosched.c 		return rb_entry_rq(node);
node              354 block/deadline-iosched.c 	dd = kmalloc_node(sizeof(*dd), GFP_KERNEL | __GFP_ZERO, q->node);
node              211 block/elevator.c 	eq = kmalloc_node(sizeof(elevator_t), GFP_KERNEL | __GFP_ZERO, q->node);
node              221 block/elevator.c 					GFP_KERNEL, q->node);
node               72 block/noop-iosched.c 	nd = kmalloc_node(sizeof(*nd), GFP_KERNEL, q->node);
node               69 crypto/async_tx/async_tx.c 	list_for_each_entry_rcu(ref, &async_tx_master_list, node)
node              163 crypto/async_tx/async_tx.c 	INIT_LIST_HEAD(&ref->node);
node              183 crypto/async_tx/async_tx.c 	list_for_each_entry_rcu(ref, &async_tx_master_list, node)
node              260 crypto/async_tx/async_tx.c 		list_for_each_entry_rcu(ref, &async_tx_master_list, node)
node              282 crypto/async_tx/async_tx.c 			list_add_tail_rcu(&master_ref->node,
node              299 crypto/async_tx/async_tx.c 		list_for_each_entry(ref, &async_tx_master_list, node)
node              303 crypto/async_tx/async_tx.c 				list_del_rcu(&ref->node);
node              480 fs/afs/cmservice.c 			r->node[loop] = ntohl(b[loop + 5]);
node              547 fs/afs/cmservice.c 		reply.ia.uuid[loop + 5] = htonl((s8) afs_uuid.node[loop]);
node              426 fs/afs/internal.h 	u8		node[6];			/* spatially unique node ID (MAC addr) */
node               58 fs/afs/main.c  	ret = afs_get_MAC_address(afs_uuid.node, sizeof(afs_uuid.node));
node               83 fs/afs/main.c  	       afs_uuid.node[0], afs_uuid.node[1], afs_uuid.node[2],
node               84 fs/afs/main.c  	       afs_uuid.node[3], afs_uuid.node[4], afs_uuid.node[5]);
node               27 fs/afs/vnode.c 	if (!node)
node               30 fs/afs/vnode.c 	if (node->rb_left)
node               31 fs/afs/vnode.c 		bad = dump_tree_aux(node->rb_left, node, depth + 2, '/');
node               33 fs/afs/vnode.c 	vnode = rb_entry(node, struct afs_vnode, cb_promise);
node               35 fs/afs/vnode.c 	       rb_is_red(node) ? 'R' : 'B',
node               38 fs/afs/vnode.c 	if (rb_parent(node) != parent) {
node               39 fs/afs/vnode.c 		printk("BAD: %p != %p\n", rb_parent(node), parent);
node               43 fs/afs/vnode.c 	if (node->rb_right)
node               44 fs/afs/vnode.c 		bad |= dump_tree_aux(node->rb_right, node, depth + 2, '\\');
node              200 fs/befs/btree.c 	if (node->bh)
node              201 fs/befs/btree.c 		brelse(node->bh);
node              203 fs/befs/btree.c 	node->bh = befs_read_datastream(sb, ds, node_off, &off);
node              204 fs/befs/btree.c 	if (!node->bh) {
node              211 fs/befs/btree.c 	node->od_node =
node              212 fs/befs/btree.c 	    (befs_btree_nodehead *) ((void *) node->bh->b_data + off);
node              214 fs/befs/btree.c 	befs_dump_index_node(sb, node->od_node);
node              216 fs/befs/btree.c 	node->head.left = fs64_to_cpu(sb, node->od_node->left);
node              217 fs/befs/btree.c 	node->head.right = fs64_to_cpu(sb, node->od_node->right);
node              218 fs/befs/btree.c 	node->head.overflow = fs64_to_cpu(sb, node->od_node->overflow);
node              219 fs/befs/btree.c 	node->head.all_key_count =
node              220 fs/befs/btree.c 	    fs16_to_cpu(sb, node->od_node->all_key_count);
node              221 fs/befs/btree.c 	node->head.all_key_length =
node              222 fs/befs/btree.c 	    fs16_to_cpu(sb, node->od_node->all_key_length);
node              353 fs/befs/btree.c 	last = node->head.all_key_count - 1;
node              354 fs/befs/btree.c 	thiskey = befs_bt_get_key(sb, node, last, &keylen);
node              362 fs/befs/btree.c 	valarray = befs_bt_valarray(node);
node              371 fs/befs/btree.c 		thiskey = befs_bt_get_key(sb, node, mid, &keylen);
node              605 fs/befs/btree.c 	if (node->head.overflow == befs_bt_inval)
node              629 fs/befs/btree.c 	    (sizeof (befs_btree_nodehead) + node->head.all_key_length);
node              635 fs/befs/btree.c 	return (fs16 *) ((void *) node->od_node + off);
node              648 fs/befs/btree.c 	void *keylen_index_start = (void *) befs_bt_keylen_index(node);
node              649 fs/befs/btree.c 	size_t keylen_index_size = node->head.all_key_count * sizeof (fs16);
node              664 fs/befs/btree.c 	return (char *) ((void *) node->od_node + sizeof (befs_btree_nodehead));
node              685 fs/befs/btree.c 	if (index < 0 || index > node->head.all_key_count) {
node              690 fs/befs/btree.c 	keystart = befs_bt_keydata(node);
node              691 fs/befs/btree.c 	keylen_index = befs_bt_keylen_index(node);
node              272 fs/befs/debug.c 	befs_debug(sb, "  left %016LX", fs64_to_cpu(sb, node->left));
node              273 fs/befs/debug.c 	befs_debug(sb, "  right %016LX", fs64_to_cpu(sb, node->right));
node              274 fs/befs/debug.c 	befs_debug(sb, "  overflow %016LX", fs64_to_cpu(sb, node->overflow));
node              276 fs/befs/debug.c 		   fs16_to_cpu(sb, node->all_key_count));
node              278 fs/befs/debug.c 		   fs16_to_cpu(sb, node->all_key_length));
node             3981 fs/cifs/cifssmb.c 		struct dfs_info3_param *node = (*target_nodes)+i;
node             3983 fs/cifs/cifssmb.c 		node->flags = le16_to_cpu(pSMBr->DFSFlags);
node             3984 fs/cifs/cifssmb.c 		node->path_consumed = le16_to_cpu(pSMBr->PathConsumed);
node             3985 fs/cifs/cifssmb.c 		node->server_type = le16_to_cpu(ref->ServerType);
node             3986 fs/cifs/cifssmb.c 		node->ref_flag = le16_to_cpu(ref->ReferralEntryFlags);
node             3991 fs/cifs/cifssmb.c 		rc = cifs_strncpy_to_host(&(node->path_name), temp,
node             3999 fs/cifs/cifssmb.c 		rc = cifs_strncpy_to_host(&(node->node_name), temp,
node             1375 fs/dcache.c    	struct hlist_node *node;
node             1380 fs/dcache.c    	hlist_for_each_entry_rcu(dentry, node, head, d_hash) {
node               83 fs/dlm/dir.c   	uint32_t node, n = 0;
node               92 fs/dlm/dir.c   		node = (hash >> 16) % ls->ls_total_weight;
node               93 fs/dlm/dir.c   		nodeid = ls->ls_node_array[node];
node               99 fs/dlm/dir.c   	node = (hash >> 16) % ls->ls_num_nodes;
node              102 fs/dlm/dir.c   		if (n++ != node)
node              109 fs/dlm/dir.c   				 ls->ls_num_nodes, n, node););
node              243 fs/dquot.c     	struct hlist_node *node;
node              246 fs/dquot.c     	hlist_for_each (node, dquot_hash+hashent) {
node              247 fs/dquot.c     		dquot = hlist_entry(node, struct dquot, dq_hash);
node              398 fs/ecryptfs/ecryptfs_kernel.h 	struct list_head node;
node               65 fs/ecryptfs/messaging.c 		*msg_ctx = list_entry(p, struct ecryptfs_msg_ctx, node);
node               85 fs/ecryptfs/messaging.c 	list_move(&msg_ctx->node, &ecryptfs_msg_ctx_alloc_list);
node               98 fs/ecryptfs/messaging.c 	list_move(&(msg_ctx->node), &ecryptfs_msg_ctx_free_list);
node              628 fs/ecryptfs/messaging.c 		INIT_LIST_HEAD(&ecryptfs_msg_ctx_arr[i].node);
node              637 fs/ecryptfs/messaging.c 		list_add_tail(&ecryptfs_msg_ctx_arr[i].node,
node              343 fs/ext2/balloc.c 	struct rb_node *node = &rsv->rsv_node;
node              365 fs/ext2/balloc.c 	rb_link_node(node, parent, p);
node              366 fs/ext2/balloc.c 	rb_insert_color(node, root);
node              330 fs/ext3/balloc.c 	struct rb_node *node = &rsv->rsv_node;
node              352 fs/ext3/balloc.c 	rb_link_node(node, parent, p);
node              353 fs/ext3/balloc.c 	rb_insert_color(node, root);
node               21 fs/hfs/bnode.c 	off += node->page_offset;
node               22 fs/hfs/bnode.c 	page = node->page[0];
node               32 fs/hfs/bnode.c 	hfs_bnode_read(node, &data, off, 2);
node               40 fs/hfs/bnode.c 	hfs_bnode_read(node, &data, off, 1);
node               49 fs/hfs/bnode.c 	tree = node->tree;
node               50 fs/hfs/bnode.c 	if (node->type == HFS_NODE_LEAF ||
node               52 fs/hfs/bnode.c 		key_len = hfs_bnode_read_u8(node, off) + 1;
node               56 fs/hfs/bnode.c 	hfs_bnode_read(node, key, off, key_len);
node               63 fs/hfs/bnode.c 	off += node->page_offset;
node               64 fs/hfs/bnode.c 	page = node->page[0];
node               75 fs/hfs/bnode.c 	hfs_bnode_write(node, &v, off, 2);
node               81 fs/hfs/bnode.c 	hfs_bnode_write(node, &data, off, 1);
node               88 fs/hfs/bnode.c 	off += node->page_offset;
node               89 fs/hfs/bnode.c 	page = node->page[0];
node              125 fs/hfs/bnode.c 	src += node->page_offset;
node              126 fs/hfs/bnode.c 	dst += node->page_offset;
node              127 fs/hfs/bnode.c 	page = node->page[0];
node              140 fs/hfs/bnode.c 	dprint(DBG_BNODE_MOD, "bnode: %d\n", node->this);
node              141 fs/hfs/bnode.c 	hfs_bnode_read(node, &desc, 0, sizeof(desc));
node              146 fs/hfs/bnode.c 	off = node->tree->node_size - 2;
node              148 fs/hfs/bnode.c 		key_off = hfs_bnode_read_u16(node, off);
node              150 fs/hfs/bnode.c 		if (i && node->type == HFS_NODE_INDEX) {
node              153 fs/hfs/bnode.c 			if (node->tree->attributes & HFS_TREE_VARIDXKEYS)
node              154 fs/hfs/bnode.c 				tmp = (hfs_bnode_read_u8(node, key_off) | 1) + 1;
node              156 fs/hfs/bnode.c 				tmp = node->tree->max_key_len + 1;
node              157 fs/hfs/bnode.c 			dprint(DBG_BNODE_MOD, " (%d,%d", tmp, hfs_bnode_read_u8(node, key_off));
node              158 fs/hfs/bnode.c 			hfs_bnode_read(node, &cnid, key_off + tmp, 4);
node              160 fs/hfs/bnode.c 		} else if (i && node->type == HFS_NODE_LEAF) {
node              163 fs/hfs/bnode.c 			tmp = hfs_bnode_read_u8(node, key_off);
node              176 fs/hfs/bnode.c 	tree = node->tree;
node              177 fs/hfs/bnode.c 	if (node->prev) {
node              178 fs/hfs/bnode.c 		tmp = hfs_bnode_find(tree, node->prev);
node              181 fs/hfs/bnode.c 		tmp->next = node->next;
node              185 fs/hfs/bnode.c 	} else if (node->type == HFS_NODE_LEAF)
node              186 fs/hfs/bnode.c 		tree->leaf_head = node->next;
node              188 fs/hfs/bnode.c 	if (node->next) {
node              189 fs/hfs/bnode.c 		tmp = hfs_bnode_find(tree, node->next);
node              192 fs/hfs/bnode.c 		tmp->prev = node->prev;
node              196 fs/hfs/bnode.c 	} else if (node->type == HFS_NODE_LEAF)
node              197 fs/hfs/bnode.c 		tree->leaf_tail = node->prev;
node              200 fs/hfs/bnode.c 	if (!node->prev && !node->next) {
node              203 fs/hfs/bnode.c 	if (!node->parent) {
node              207 fs/hfs/bnode.c 	set_bit(HFS_BNODE_DELETED, &node->flags);
node              219 fs/hfs/bnode.c 	struct hfs_bnode *node;
node              226 fs/hfs/bnode.c 	for (node = tree->node_hash[hfs_bnode_hash(cnid)];
node              227 fs/hfs/bnode.c 	     node; node = node->next_hash) {
node              228 fs/hfs/bnode.c 		if (node->this == cnid) {
node              229 fs/hfs/bnode.c 			return node;
node              238 fs/hfs/bnode.c 	struct hfs_bnode *node, *node2;
node              252 fs/hfs/bnode.c 	node = kzalloc(size, GFP_KERNEL);
node              253 fs/hfs/bnode.c 	if (!node)
node              255 fs/hfs/bnode.c 	node->tree = tree;
node              256 fs/hfs/bnode.c 	node->this = cnid;
node              257 fs/hfs/bnode.c 	set_bit(HFS_BNODE_NEW, &node->flags);
node              258 fs/hfs/bnode.c 	atomic_set(&node->refcnt, 1);
node              260 fs/hfs/bnode.c 	       node->tree->cnid, node->this);
node              261 fs/hfs/bnode.c 	init_waitqueue_head(&node->lock_wq);
node              266 fs/hfs/bnode.c 		node->next_hash = tree->node_hash[hash];
node              267 fs/hfs/bnode.c 		tree->node_hash[hash] = node;
node              271 fs/hfs/bnode.c 		kfree(node);
node              280 fs/hfs/bnode.c 	node->page_offset = off & ~PAGE_CACHE_MASK;
node              290 fs/hfs/bnode.c 		node->page[i] = page;
node              293 fs/hfs/bnode.c 	return node;
node              295 fs/hfs/bnode.c 	set_bit(HFS_BNODE_ERROR, &node->flags);
node              296 fs/hfs/bnode.c 	return node;
node              304 fs/hfs/bnode.c 		node->tree->cnid, node->this, atomic_read(&node->refcnt));
node              305 fs/hfs/bnode.c 	for (p = &node->tree->node_hash[hfs_bnode_hash(node->this)];
node              306 fs/hfs/bnode.c 	     *p && *p != node; p = &(*p)->next_hash)
node              309 fs/hfs/bnode.c 	*p = node->next_hash;
node              310 fs/hfs/bnode.c 	node->tree->node_hash_cnt--;
node              316 fs/hfs/bnode.c 	struct hfs_bnode *node;
node              322 fs/hfs/bnode.c 	node = hfs_bnode_findhash(tree, num);
node              323 fs/hfs/bnode.c 	if (node) {
node              324 fs/hfs/bnode.c 		hfs_bnode_get(node);
node              326 fs/hfs/bnode.c 		wait_event(node->lock_wq, !test_bit(HFS_BNODE_NEW, &node->flags));
node              327 fs/hfs/bnode.c 		if (test_bit(HFS_BNODE_ERROR, &node->flags))
node              329 fs/hfs/bnode.c 		return node;
node              332 fs/hfs/bnode.c 	node = __hfs_bnode_create(tree, num);
node              333 fs/hfs/bnode.c 	if (!node)
node              335 fs/hfs/bnode.c 	if (test_bit(HFS_BNODE_ERROR, &node->flags))
node              337 fs/hfs/bnode.c 	if (!test_bit(HFS_BNODE_NEW, &node->flags))
node              338 fs/hfs/bnode.c 		return node;
node              340 fs/hfs/bnode.c 	desc = (struct hfs_bnode_desc *)(kmap(node->page[0]) + node->page_offset);
node              341 fs/hfs/bnode.c 	node->prev = be32_to_cpu(desc->prev);
node              342 fs/hfs/bnode.c 	node->next = be32_to_cpu(desc->next);
node              343 fs/hfs/bnode.c 	node->num_recs = be16_to_cpu(desc->num_recs);
node              344 fs/hfs/bnode.c 	node->type = desc->type;
node              345 fs/hfs/bnode.c 	node->height = desc->height;
node              346 fs/hfs/bnode.c 	kunmap(node->page[0]);
node              348 fs/hfs/bnode.c 	switch (node->type) {
node              351 fs/hfs/bnode.c 		if (node->height != 0)
node              355 fs/hfs/bnode.c 		if (node->height != 1)
node              359 fs/hfs/bnode.c 		if (node->height <= 1 || node->height > tree->depth)
node              367 fs/hfs/bnode.c 	off = hfs_bnode_read_u16(node, rec_off);
node              370 fs/hfs/bnode.c 	for (i = 1; i <= node->num_recs; off = next_off, i++) {
node              372 fs/hfs/bnode.c 		next_off = hfs_bnode_read_u16(node, rec_off);
node              378 fs/hfs/bnode.c 		if (node->type != HFS_NODE_INDEX &&
node              379 fs/hfs/bnode.c 		    node->type != HFS_NODE_LEAF)
node              381 fs/hfs/bnode.c 		key_size = hfs_bnode_read_u8(node, off) + 1;
node              385 fs/hfs/bnode.c 	clear_bit(HFS_BNODE_NEW, &node->flags);
node              386 fs/hfs/bnode.c 	wake_up(&node->lock_wq);
node              387 fs/hfs/bnode.c 	return node;
node              390 fs/hfs/bnode.c 	set_bit(HFS_BNODE_ERROR, &node->flags);
node              391 fs/hfs/bnode.c 	clear_bit(HFS_BNODE_NEW, &node->flags);
node              392 fs/hfs/bnode.c 	wake_up(&node->lock_wq);
node              393 fs/hfs/bnode.c 	hfs_bnode_put(node);
node              404 fs/hfs/bnode.c 	kfree(node);
node              409 fs/hfs/bnode.c 	struct hfs_bnode *node;
node              414 fs/hfs/bnode.c 	node = hfs_bnode_findhash(tree, num);
node              416 fs/hfs/bnode.c 	BUG_ON(node);
node              417 fs/hfs/bnode.c 	node = __hfs_bnode_create(tree, num);
node              418 fs/hfs/bnode.c 	if (!node)
node              420 fs/hfs/bnode.c 	if (test_bit(HFS_BNODE_ERROR, &node->flags)) {
node              421 fs/hfs/bnode.c 		hfs_bnode_put(node);
node              425 fs/hfs/bnode.c 	pagep = node->page;
node              426 fs/hfs/bnode.c 	memset(kmap(*pagep) + node->page_offset, 0,
node              435 fs/hfs/bnode.c 	clear_bit(HFS_BNODE_NEW, &node->flags);
node              436 fs/hfs/bnode.c 	wake_up(&node->lock_wq);
node              438 fs/hfs/bnode.c 	return node;
node              443 fs/hfs/bnode.c 	if (node) {
node              444 fs/hfs/bnode.c 		atomic_inc(&node->refcnt);
node              446 fs/hfs/bnode.c 		       node->tree->cnid, node->this, atomic_read(&node->refcnt));
node              453 fs/hfs/bnode.c 	if (node) {
node              454 fs/hfs/bnode.c 		struct hfs_btree *tree = node->tree;
node              458 fs/hfs/bnode.c 		       node->tree->cnid, node->this, atomic_read(&node->refcnt));
node              459 fs/hfs/bnode.c 		BUG_ON(!atomic_read(&node->refcnt));
node              460 fs/hfs/bnode.c 		if (!atomic_dec_and_lock(&node->refcnt, &tree->hash_lock))
node              463 fs/hfs/bnode.c 			if (!node->page[i])
node              465 fs/hfs/bnode.c 			mark_page_accessed(node->page[i]);
node              468 fs/hfs/bnode.c 		if (test_bit(HFS_BNODE_DELETED, &node->flags)) {
node              469 fs/hfs/bnode.c 			hfs_bnode_unhash(node);
node              471 fs/hfs/bnode.c 			hfs_bmap_free(node);
node              472 fs/hfs/bnode.c 			hfs_bnode_free(node);
node               23 fs/hfs/brec.c  	dataoff = node->tree->node_size - (rec + 2) * 2;
node               24 fs/hfs/brec.c  	hfs_bnode_read(node, retval, dataoff, 4);
node               34 fs/hfs/brec.c  	if (node->type != HFS_NODE_INDEX && node->type != HFS_NODE_LEAF)
node               37 fs/hfs/brec.c  	if ((node->type == HFS_NODE_INDEX) &&
node               38 fs/hfs/brec.c  	   !(node->tree->attributes & HFS_TREE_VARIDXKEYS)) {
node               39 fs/hfs/brec.c  		if (node->tree->attributes & HFS_TREE_BIGKEYS)
node               40 fs/hfs/brec.c  			retval = node->tree->max_key_len + 2;
node               42 fs/hfs/brec.c  			retval = node->tree->max_key_len + 1;
node               44 fs/hfs/brec.c  		recoff = hfs_bnode_read_u16(node, node->tree->node_size - (rec + 1) * 2);
node               47 fs/hfs/brec.c  		if (node->tree->attributes & HFS_TREE_BIGKEYS) {
node               48 fs/hfs/brec.c  			retval = hfs_bnode_read_u16(node, recoff) + 2;
node               49 fs/hfs/brec.c  			if (retval > node->tree->max_key_len + 2) {
node               55 fs/hfs/brec.c  			retval = (hfs_bnode_read_u8(node, recoff) | 1) + 1;
node               56 fs/hfs/brec.c  			if (retval > node->tree->max_key_len + 1) {
node               69 fs/hfs/brec.c  	struct hfs_bnode *node, *new_node;
node               91 fs/hfs/brec.c  	node = fd->bnode;
node               92 fs/hfs/brec.c  	hfs_bnode_dump(node);
node               94 fs/hfs/brec.c  	end_rec_off = tree->node_size - (node->num_recs + 1) * 2;
node               95 fs/hfs/brec.c  	end_off = hfs_bnode_read_u16(node, end_rec_off);
node              106 fs/hfs/brec.c  	if (node->type == HFS_NODE_LEAF) {
node              110 fs/hfs/brec.c  	node->num_recs++;
node              112 fs/hfs/brec.c  	hfs_bnode_write_u16(node, offsetof(struct hfs_bnode_desc, num_recs), node->num_recs);
node              113 fs/hfs/brec.c  	hfs_bnode_write_u16(node, end_rec_off, end_off + size);
node              121 fs/hfs/brec.c  		data_off = hfs_bnode_read_u16(node, data_rec_off + 2);
node              122 fs/hfs/brec.c  		hfs_bnode_write_u16(node, data_rec_off, data_off + size);
node              127 fs/hfs/brec.c  	hfs_bnode_move(node, data_off + size, data_off,
node              131 fs/hfs/brec.c  	hfs_bnode_write(node, fd->search_key, data_off, key_len);
node              132 fs/hfs/brec.c  	hfs_bnode_write(node, entry, data_off + key_len, entry_len);
node              133 fs/hfs/brec.c  	hfs_bnode_dump(node);
node              139 fs/hfs/brec.c  		if (!rec && new_node != node)
node              179 fs/hfs/brec.c  	struct hfs_bnode *node, *parent;
node              183 fs/hfs/brec.c  	node = fd->bnode;
node              186 fs/hfs/brec.c  	end_off = tree->node_size - (node->num_recs + 1) * 2;
node              188 fs/hfs/brec.c  	if (node->type == HFS_NODE_LEAF) {
node              192 fs/hfs/brec.c  	hfs_bnode_dump(node);
node              194 fs/hfs/brec.c  	if (!--node->num_recs) {
node              195 fs/hfs/brec.c  		hfs_bnode_unlink(node);
node              196 fs/hfs/brec.c  		if (!node->parent)
node              198 fs/hfs/brec.c  		parent = hfs_bnode_find(tree, node->parent);
node              201 fs/hfs/brec.c  		hfs_bnode_put(node);
node              202 fs/hfs/brec.c  		node = fd->bnode = parent;
node              204 fs/hfs/brec.c  		__hfs_brec_find(node, fd);
node              207 fs/hfs/brec.c  	hfs_bnode_write_u16(node, offsetof(struct hfs_bnode_desc, num_recs), node->num_recs);
node              214 fs/hfs/brec.c  		data_off = hfs_bnode_read_u16(node, rec_off);
node              215 fs/hfs/brec.c  		hfs_bnode_write_u16(node, rec_off + 2, data_off - size);
node              220 fs/hfs/brec.c  	hfs_bnode_move(node, fd->keyoffset, fd->keyoffset + size,
node              223 fs/hfs/brec.c  	hfs_bnode_dump(node);
node              232 fs/hfs/brec.c  	struct hfs_bnode *node, *new_node, *next_node;
node              238 fs/hfs/brec.c  	node = fd->bnode;
node              242 fs/hfs/brec.c  	hfs_bnode_get(node);
node              244 fs/hfs/brec.c  		node->this, new_node->this, node->next);
node              245 fs/hfs/brec.c  	new_node->next = node->next;
node              246 fs/hfs/brec.c  	new_node->prev = node->this;
node              247 fs/hfs/brec.c  	new_node->parent = node->parent;
node              248 fs/hfs/brec.c  	new_node->type = node->type;
node              249 fs/hfs/brec.c  	new_node->height = node->height;
node              251 fs/hfs/brec.c  	if (node->next)
node              252 fs/hfs/brec.c  		next_node = hfs_bnode_find(tree, node->next);
node              257 fs/hfs/brec.c  		hfs_bnode_put(node);
node              262 fs/hfs/brec.c  	size = tree->node_size / 2 - node->num_recs * 2 - 14;
node              266 fs/hfs/brec.c  		data_start = hfs_bnode_read_u16(node, old_rec_off);
node              270 fs/hfs/brec.c  		if (++num_recs < node->num_recs)
node              273 fs/hfs/brec.c  		hfs_bnode_put(node);
node              286 fs/hfs/brec.c  		data_start = hfs_bnode_read_u16(node, old_rec_off);
node              288 fs/hfs/brec.c  		hfs_bnode_put(node);
node              295 fs/hfs/brec.c  	new_node->num_recs = node->num_recs - num_recs;
node              296 fs/hfs/brec.c  	node->num_recs = num_recs;
node              307 fs/hfs/brec.c  		data_end = hfs_bnode_read_u16(node, old_rec_off);
node              312 fs/hfs/brec.c  	hfs_bnode_copy(new_node, 14, node, data_start, data_end - data_start);
node              324 fs/hfs/brec.c  	node->next = new_node->this;
node              325 fs/hfs/brec.c  	hfs_bnode_read(node, &node_desc, 0, sizeof(node_desc));
node              326 fs/hfs/brec.c  	node_desc.next = cpu_to_be32(node->next);
node              327 fs/hfs/brec.c  	node_desc.num_recs = cpu_to_be16(node->num_recs);
node              328 fs/hfs/brec.c  	hfs_bnode_write(node, &node_desc, 0, sizeof(node_desc));
node              337 fs/hfs/brec.c  	} else if (node->this == tree->leaf_tail) {
node              343 fs/hfs/brec.c  	hfs_bnode_dump(node);
node              345 fs/hfs/brec.c  	hfs_bnode_put(node);
node              353 fs/hfs/brec.c  	struct hfs_bnode *node, *new_node, *parent;
node              359 fs/hfs/brec.c  	node = fd->bnode;
node              361 fs/hfs/brec.c  	if (!node->parent)
node              365 fs/hfs/brec.c  	parent = hfs_bnode_find(tree, node->parent);
node              374 fs/hfs/brec.c  		newkeylen = (hfs_bnode_read_u8(node, 14) | 1) + 1;
node              412 fs/hfs/brec.c  	hfs_bnode_copy(parent, fd->keyoffset, node, 14, newkeylen);
node              417 fs/hfs/brec.c  	hfs_bnode_put(node);
node              418 fs/hfs/brec.c  	node = parent;
node              434 fs/hfs/brec.c  			if (new_node == node)
node              437 fs/hfs/brec.c  			hfs_bnode_read_key(node, fd->search_key, 14);
node              441 fs/hfs/brec.c  	if (!rec && node->parent)
node              444 fs/hfs/brec.c  	fd->bnode = node;
node              450 fs/hfs/brec.c  	struct hfs_bnode *node, *new_node;
node              455 fs/hfs/brec.c  	node = NULL;
node              457 fs/hfs/brec.c  		node = hfs_bnode_find(tree, tree->root);
node              458 fs/hfs/brec.c  		if (IS_ERR(node))
node              459 fs/hfs/brec.c  			return PTR_ERR(node);
node              463 fs/hfs/brec.c  		hfs_bnode_put(node);
node              492 fs/hfs/brec.c  	if (node) {
node              494 fs/hfs/brec.c  		node->parent = tree->root;
node              495 fs/hfs/brec.c  		if (node->type == HFS_NODE_LEAF ||
node              497 fs/hfs/brec.c  			key_size = hfs_bnode_read_u8(node, 14) + 1;
node              500 fs/hfs/brec.c  		hfs_bnode_copy(new_node, 14, node, 14, key_size);
node              507 fs/hfs/brec.c  		cnid = cpu_to_be32(node->this);
node              513 fs/hfs/brec.c  		hfs_bnode_put(node);
node              123 fs/hfs/btree.c 	struct hfs_bnode *node;
node              130 fs/hfs/btree.c 		while ((node = tree->node_hash[i])) {
node              131 fs/hfs/btree.c 			tree->node_hash[i] = node->next_hash;
node              132 fs/hfs/btree.c 			if (atomic_read(&node->refcnt))
node              134 fs/hfs/btree.c 					node->tree->cnid, node->this, atomic_read(&node->refcnt));
node              135 fs/hfs/btree.c 			hfs_bnode_free(node);
node              146 fs/hfs/btree.c 	struct hfs_bnode *node;
node              149 fs/hfs/btree.c 	node = hfs_bnode_find(tree, 0);
node              150 fs/hfs/btree.c 	if (IS_ERR(node))
node              154 fs/hfs/btree.c 	page = node->page[0];
node              168 fs/hfs/btree.c 	hfs_bnode_put(node);
node              174 fs/hfs/btree.c 	struct hfs_bnode *node;
node              178 fs/hfs/btree.c 	node = hfs_bnode_create(tree, idx);
node              179 fs/hfs/btree.c 	if (IS_ERR(node))
node              180 fs/hfs/btree.c 		return node;
node              189 fs/hfs/btree.c 	node->type = HFS_NODE_MAP;
node              190 fs/hfs/btree.c 	node->num_recs = 1;
node              191 fs/hfs/btree.c 	hfs_bnode_clear(node, 0, tree->node_size);
node              198 fs/hfs/btree.c 	hfs_bnode_write(node, &desc, 0, sizeof(desc));
node              199 fs/hfs/btree.c 	hfs_bnode_write_u16(node, 14, 0x8000);
node              200 fs/hfs/btree.c 	hfs_bnode_write_u16(node, tree->node_size - 2, 14);
node              201 fs/hfs/btree.c 	hfs_bnode_write_u16(node, tree->node_size - 4, tree->node_size - 6);
node              203 fs/hfs/btree.c 	return node;
node              208 fs/hfs/btree.c 	struct hfs_bnode *node, *next_node;
node              237 fs/hfs/btree.c 	node = hfs_bnode_find(tree, nidx);
node              238 fs/hfs/btree.c 	if (IS_ERR(node))
node              239 fs/hfs/btree.c 		return node;
node              240 fs/hfs/btree.c 	len = hfs_brec_lenoff(node, 2, &off16);
node              243 fs/hfs/btree.c 	off += node->page_offset;
node              244 fs/hfs/btree.c 	pagep = node->page + (off >> PAGE_CACHE_SHIFT);
node              261 fs/hfs/btree.c 						hfs_bnode_put(node);
node              275 fs/hfs/btree.c 		nidx = node->next;
node              278 fs/hfs/btree.c 			next_node = hfs_bmap_new_bmap(node, idx);
node              281 fs/hfs/btree.c 		hfs_bnode_put(node);
node              284 fs/hfs/btree.c 		node = next_node;
node              286 fs/hfs/btree.c 		len = hfs_brec_lenoff(node, 0, &off16);
node              288 fs/hfs/btree.c 		off += node->page_offset;
node              289 fs/hfs/btree.c 		pagep = node->page + (off >> PAGE_CACHE_SHIFT);
node              303 fs/hfs/btree.c 	dprint(DBG_BNODE_MOD, "btree_free_node: %u\n", node->this);
node              304 fs/hfs/btree.c 	tree = node->tree;
node              305 fs/hfs/btree.c 	nidx = node->this;
node              306 fs/hfs/btree.c 	node = hfs_bnode_find(tree, 0);
node              307 fs/hfs/btree.c 	if (IS_ERR(node))
node              309 fs/hfs/btree.c 	len = hfs_brec_lenoff(node, 2, &off);
node              314 fs/hfs/btree.c 		i = node->next;
node              315 fs/hfs/btree.c 		hfs_bnode_put(node);
node              318 fs/hfs/btree.c 			printk(KERN_CRIT "hfs: unable to free bnode %u. bmap not found!\n", node->this);
node              321 fs/hfs/btree.c 		node = hfs_bnode_find(tree, i);
node              322 fs/hfs/btree.c 		if (IS_ERR(node))
node              324 fs/hfs/btree.c 		if (node->type != HFS_NODE_MAP) {
node              326 fs/hfs/btree.c 			printk(KERN_CRIT "hfs: invalid bmap found! (%u,%d)\n", node->this, node->type);
node              327 fs/hfs/btree.c 			hfs_bnode_put(node);
node              330 fs/hfs/btree.c 		len = hfs_brec_lenoff(node, 0, &off);
node              332 fs/hfs/btree.c 	off += node->page_offset + nidx / 8;
node              333 fs/hfs/btree.c 	page = node->page[off >> PAGE_CACHE_SHIFT];
node              339 fs/hfs/btree.c 		printk(KERN_CRIT "hfs: trying to free free bnode %u(%d)\n", node->this, node->type);
node              341 fs/hfs/btree.c 		hfs_bnode_put(node);
node              347 fs/hfs/btree.c 	hfs_bnode_put(node);
node               58 fs/hfs/inode.c 	struct hfs_bnode *node;
node               76 fs/hfs/inode.c 		node = hfs_bnode_findhash(tree, nidx);
node               77 fs/hfs/inode.c 		if (!node)
node               79 fs/hfs/inode.c 		else if (atomic_read(&node->refcnt))
node               81 fs/hfs/inode.c 		if (res && node) {
node               82 fs/hfs/inode.c 			hfs_bnode_unhash(node);
node               83 fs/hfs/inode.c 			hfs_bnode_free(node);
node               91 fs/hfs/inode.c 			node = hfs_bnode_findhash(tree, nidx++);
node               92 fs/hfs/inode.c 			if (!node)
node               94 fs/hfs/inode.c 			if (atomic_read(&node->refcnt)) {
node               98 fs/hfs/inode.c 			hfs_bnode_unhash(node);
node               99 fs/hfs/inode.c 			hfs_bnode_free(node);
node               26 fs/hfsplus/bnode.c 	off += node->page_offset;
node               27 fs/hfsplus/bnode.c 	pagep = node->page + (off >> PAGE_CACHE_SHIFT);
node               46 fs/hfsplus/bnode.c 	hfs_bnode_read(node, &data, off, 2);
node               54 fs/hfsplus/bnode.c 	hfs_bnode_read(node, &data, off, 1);
node               63 fs/hfsplus/bnode.c 	tree = node->tree;
node               64 fs/hfsplus/bnode.c 	if (node->type == HFS_NODE_LEAF ||
node               66 fs/hfsplus/bnode.c 		key_len = hfs_bnode_read_u16(node, off) + 2;
node               70 fs/hfsplus/bnode.c 	hfs_bnode_read(node, key, off, key_len);
node               78 fs/hfsplus/bnode.c 	off += node->page_offset;
node               79 fs/hfsplus/bnode.c 	pagep = node->page + (off >> PAGE_CACHE_SHIFT);
node              100 fs/hfsplus/bnode.c 	hfs_bnode_write(node, &v, off, 2);
node              108 fs/hfsplus/bnode.c 	off += node->page_offset;
node              109 fs/hfsplus/bnode.c 	pagep = node->page + (off >> PAGE_CACHE_SHIFT);
node              193 fs/hfsplus/bnode.c 	src += node->page_offset;
node              194 fs/hfsplus/bnode.c 	dst += node->page_offset;
node              197 fs/hfsplus/bnode.c 		src_page = node->page + (src >> PAGE_CACHE_SHIFT);
node              200 fs/hfsplus/bnode.c 		dst_page = node->page + (dst >> PAGE_CACHE_SHIFT);
node              246 fs/hfsplus/bnode.c 		src_page = node->page + (src >> PAGE_CACHE_SHIFT);
node              248 fs/hfsplus/bnode.c 		dst_page = node->page + (dst >> PAGE_CACHE_SHIFT);
node              300 fs/hfsplus/bnode.c 	dprint(DBG_BNODE_MOD, "bnode: %d\n", node->this);
node              301 fs/hfsplus/bnode.c 	hfs_bnode_read(node, &desc, 0, sizeof(desc));
node              306 fs/hfsplus/bnode.c 	off = node->tree->node_size - 2;
node              308 fs/hfsplus/bnode.c 		key_off = hfs_bnode_read_u16(node, off);
node              310 fs/hfsplus/bnode.c 		if (i && node->type == HFS_NODE_INDEX) {
node              313 fs/hfsplus/bnode.c 			if (node->tree->attributes & HFS_TREE_VARIDXKEYS)
node              314 fs/hfsplus/bnode.c 				tmp = hfs_bnode_read_u16(node, key_off) + 2;
node              316 fs/hfsplus/bnode.c 				tmp = node->tree->max_key_len + 2;
node              318 fs/hfsplus/bnode.c 			hfs_bnode_read(node, &cnid, key_off + tmp, 4);
node              320 fs/hfsplus/bnode.c 		} else if (i && node->type == HFS_NODE_LEAF) {
node              323 fs/hfsplus/bnode.c 			tmp = hfs_bnode_read_u16(node, key_off);
node              336 fs/hfsplus/bnode.c 	tree = node->tree;
node              337 fs/hfsplus/bnode.c 	if (node->prev) {
node              338 fs/hfsplus/bnode.c 		tmp = hfs_bnode_find(tree, node->prev);
node              341 fs/hfsplus/bnode.c 		tmp->next = node->next;
node              345 fs/hfsplus/bnode.c 	} else if (node->type == HFS_NODE_LEAF)
node              346 fs/hfsplus/bnode.c 		tree->leaf_head = node->next;
node              348 fs/hfsplus/bnode.c 	if (node->next) {
node              349 fs/hfsplus/bnode.c 		tmp = hfs_bnode_find(tree, node->next);
node              352 fs/hfsplus/bnode.c 		tmp->prev = node->prev;
node              356 fs/hfsplus/bnode.c 	} else if (node->type == HFS_NODE_LEAF)
node              357 fs/hfsplus/bnode.c 		tree->leaf_tail = node->prev;
node              360 fs/hfsplus/bnode.c 	if (!node->prev && !node->next) {
node              363 fs/hfsplus/bnode.c 	if (!node->parent) {
node              367 fs/hfsplus/bnode.c 	set_bit(HFS_BNODE_DELETED, &node->flags);
node              379 fs/hfsplus/bnode.c 	struct hfs_bnode *node;
node              386 fs/hfsplus/bnode.c 	for (node = tree->node_hash[hfs_bnode_hash(cnid)];
node              387 fs/hfsplus/bnode.c 	     node; node = node->next_hash) {
node              388 fs/hfsplus/bnode.c 		if (node->this == cnid) {
node              389 fs/hfsplus/bnode.c 			return node;
node              398 fs/hfsplus/bnode.c 	struct hfs_bnode *node, *node2;
node              412 fs/hfsplus/bnode.c 	node = kzalloc(size, GFP_KERNEL);
node              413 fs/hfsplus/bnode.c 	if (!node)
node              415 fs/hfsplus/bnode.c 	node->tree = tree;
node              416 fs/hfsplus/bnode.c 	node->this = cnid;
node              417 fs/hfsplus/bnode.c 	set_bit(HFS_BNODE_NEW, &node->flags);
node              418 fs/hfsplus/bnode.c 	atomic_set(&node->refcnt, 1);
node              420 fs/hfsplus/bnode.c 	       node->tree->cnid, node->this);
node              421 fs/hfsplus/bnode.c 	init_waitqueue_head(&node->lock_wq);
node              426 fs/hfsplus/bnode.c 		node->next_hash = tree->node_hash[hash];
node              427 fs/hfsplus/bnode.c 		tree->node_hash[hash] = node;
node              431 fs/hfsplus/bnode.c 		kfree(node);
node              440 fs/hfsplus/bnode.c 	node->page_offset = off & ~PAGE_CACHE_MASK;
node              450 fs/hfsplus/bnode.c 		node->page[i] = page;
node              453 fs/hfsplus/bnode.c 	return node;
node              455 fs/hfsplus/bnode.c 	set_bit(HFS_BNODE_ERROR, &node->flags);
node              456 fs/hfsplus/bnode.c 	return node;
node              464 fs/hfsplus/bnode.c 		node->tree->cnid, node->this, atomic_read(&node->refcnt));
node              465 fs/hfsplus/bnode.c 	for (p = &node->tree->node_hash[hfs_bnode_hash(node->this)];
node              466 fs/hfsplus/bnode.c 	     *p && *p != node; p = &(*p)->next_hash)
node              469 fs/hfsplus/bnode.c 	*p = node->next_hash;
node              470 fs/hfsplus/bnode.c 	node->tree->node_hash_cnt--;
node              476 fs/hfsplus/bnode.c 	struct hfs_bnode *node;
node              482 fs/hfsplus/bnode.c 	node = hfs_bnode_findhash(tree, num);
node              483 fs/hfsplus/bnode.c 	if (node) {
node              484 fs/hfsplus/bnode.c 		hfs_bnode_get(node);
node              486 fs/hfsplus/bnode.c 		wait_event(node->lock_wq, !test_bit(HFS_BNODE_NEW, &node->flags));
node              487 fs/hfsplus/bnode.c 		if (test_bit(HFS_BNODE_ERROR, &node->flags))
node              489 fs/hfsplus/bnode.c 		return node;
node              492 fs/hfsplus/bnode.c 	node = __hfs_bnode_create(tree, num);
node              493 fs/hfsplus/bnode.c 	if (!node)
node              495 fs/hfsplus/bnode.c 	if (test_bit(HFS_BNODE_ERROR, &node->flags))
node              497 fs/hfsplus/bnode.c 	if (!test_bit(HFS_BNODE_NEW, &node->flags))
node              498 fs/hfsplus/bnode.c 		return node;
node              500 fs/hfsplus/bnode.c 	desc = (struct hfs_bnode_desc *)(kmap(node->page[0]) + node->page_offset);
node              501 fs/hfsplus/bnode.c 	node->prev = be32_to_cpu(desc->prev);
node              502 fs/hfsplus/bnode.c 	node->next = be32_to_cpu(desc->next);
node              503 fs/hfsplus/bnode.c 	node->num_recs = be16_to_cpu(desc->num_recs);
node              504 fs/hfsplus/bnode.c 	node->type = desc->type;
node              505 fs/hfsplus/bnode.c 	node->height = desc->height;
node              506 fs/hfsplus/bnode.c 	kunmap(node->page[0]);
node              508 fs/hfsplus/bnode.c 	switch (node->type) {
node              511 fs/hfsplus/bnode.c 		if (node->height != 0)
node              515 fs/hfsplus/bnode.c 		if (node->height != 1)
node              519 fs/hfsplus/bnode.c 		if (node->height <= 1 || node->height > tree->depth)
node              527 fs/hfsplus/bnode.c 	off = hfs_bnode_read_u16(node, rec_off);
node              530 fs/hfsplus/bnode.c 	for (i = 1; i <= node->num_recs; off = next_off, i++) {
node              532 fs/hfsplus/bnode.c 		next_off = hfs_bnode_read_u16(node, rec_off);
node              538 fs/hfsplus/bnode.c 		if (node->type != HFS_NODE_INDEX &&
node              539 fs/hfsplus/bnode.c 		    node->type != HFS_NODE_LEAF)
node              541 fs/hfsplus/bnode.c 		key_size = hfs_bnode_read_u16(node, off) + 2;
node              545 fs/hfsplus/bnode.c 	clear_bit(HFS_BNODE_NEW, &node->flags);
node              546 fs/hfsplus/bnode.c 	wake_up(&node->lock_wq);
node              547 fs/hfsplus/bnode.c 	return node;
node              550 fs/hfsplus/bnode.c 	set_bit(HFS_BNODE_ERROR, &node->flags);
node              551 fs/hfsplus/bnode.c 	clear_bit(HFS_BNODE_NEW, &node->flags);
node              552 fs/hfsplus/bnode.c 	wake_up(&node->lock_wq);
node              553 fs/hfsplus/bnode.c 	hfs_bnode_put(node);
node              564 fs/hfsplus/bnode.c 	kfree(node);
node              569 fs/hfsplus/bnode.c 	struct hfs_bnode *node;
node              574 fs/hfsplus/bnode.c 	node = hfs_bnode_findhash(tree, num);
node              576 fs/hfsplus/bnode.c 	if (node) {
node              579 fs/hfsplus/bnode.c 		return node;
node              581 fs/hfsplus/bnode.c 	node = __hfs_bnode_create(tree, num);
node              582 fs/hfsplus/bnode.c 	if (!node)
node              584 fs/hfsplus/bnode.c 	if (test_bit(HFS_BNODE_ERROR, &node->flags)) {
node              585 fs/hfsplus/bnode.c 		hfs_bnode_put(node);
node              589 fs/hfsplus/bnode.c 	pagep = node->page;
node              590 fs/hfsplus/bnode.c 	memset(kmap(*pagep) + node->page_offset, 0,
node              599 fs/hfsplus/bnode.c 	clear_bit(HFS_BNODE_NEW, &node->flags);
node              600 fs/hfsplus/bnode.c 	wake_up(&node->lock_wq);
node              602 fs/hfsplus/bnode.c 	return node;
node              607 fs/hfsplus/bnode.c 	if (node) {
node              608 fs/hfsplus/bnode.c 		atomic_inc(&node->refcnt);
node              610 fs/hfsplus/bnode.c 		       node->tree->cnid, node->this, atomic_read(&node->refcnt));
node              617 fs/hfsplus/bnode.c 	if (node) {
node              618 fs/hfsplus/bnode.c 		struct hfs_btree *tree = node->tree;
node              622 fs/hfsplus/bnode.c 		       node->tree->cnid, node->this, atomic_read(&node->refcnt));
node              623 fs/hfsplus/bnode.c 		BUG_ON(!atomic_read(&node->refcnt));
node              624 fs/hfsplus/bnode.c 		if (!atomic_dec_and_lock(&node->refcnt, &tree->hash_lock))
node              627 fs/hfsplus/bnode.c 			if (!node->page[i])
node              629 fs/hfsplus/bnode.c 			mark_page_accessed(node->page[i]);
node              632 fs/hfsplus/bnode.c 		if (test_bit(HFS_BNODE_DELETED, &node->flags)) {
node              633 fs/hfsplus/bnode.c 			hfs_bnode_unhash(node);
node              635 fs/hfsplus/bnode.c 			hfs_bmap_free(node);
node              636 fs/hfsplus/bnode.c 			hfs_bnode_free(node);
node               24 fs/hfsplus/brec.c 	dataoff = node->tree->node_size - (rec + 2) * 2;
node               25 fs/hfsplus/brec.c 	hfs_bnode_read(node, retval, dataoff, 4);
node               35 fs/hfsplus/brec.c 	if (node->type != HFS_NODE_INDEX && node->type != HFS_NODE_LEAF)
node               38 fs/hfsplus/brec.c 	if ((node->type == HFS_NODE_INDEX) &&
node               39 fs/hfsplus/brec.c 	   !(node->tree->attributes & HFS_TREE_VARIDXKEYS)) {
node               40 fs/hfsplus/brec.c 		retval = node->tree->max_key_len + 2;
node               42 fs/hfsplus/brec.c 		recoff = hfs_bnode_read_u16(node, node->tree->node_size - (rec + 1) * 2);
node               45 fs/hfsplus/brec.c 		if (node->tree->attributes & HFS_TREE_BIGKEYS)
node               46 fs/hfsplus/brec.c 			retval = hfs_bnode_read_u16(node, recoff) + 2;
node               48 fs/hfsplus/brec.c 			retval = (hfs_bnode_read_u8(node, recoff) | 1) + 1;
node               56 fs/hfsplus/brec.c 	struct hfs_bnode *node, *new_node;
node               78 fs/hfsplus/brec.c 	node = fd->bnode;
node               79 fs/hfsplus/brec.c 	hfs_bnode_dump(node);
node               81 fs/hfsplus/brec.c 	end_rec_off = tree->node_size - (node->num_recs + 1) * 2;
node               82 fs/hfsplus/brec.c 	end_off = hfs_bnode_read_u16(node, end_rec_off);
node               93 fs/hfsplus/brec.c 	if (node->type == HFS_NODE_LEAF) {
node               97 fs/hfsplus/brec.c 	node->num_recs++;
node               99 fs/hfsplus/brec.c 	hfs_bnode_write_u16(node, offsetof(struct hfs_bnode_desc, num_recs), node->num_recs);
node              100 fs/hfsplus/brec.c 	hfs_bnode_write_u16(node, end_rec_off, end_off + size);
node              108 fs/hfsplus/brec.c 		data_off = hfs_bnode_read_u16(node, data_rec_off + 2);
node              109 fs/hfsplus/brec.c 		hfs_bnode_write_u16(node, data_rec_off, data_off + size);
node              114 fs/hfsplus/brec.c 	hfs_bnode_move(node, data_off + size, data_off,
node              118 fs/hfsplus/brec.c 	hfs_bnode_write(node, fd->search_key, data_off, key_len);
node              119 fs/hfsplus/brec.c 	hfs_bnode_write(node, entry, data_off + key_len, entry_len);
node              120 fs/hfsplus/brec.c 	hfs_bnode_dump(node);
node              126 fs/hfsplus/brec.c 		if (!rec && new_node != node)
node              166 fs/hfsplus/brec.c 	struct hfs_bnode *node, *parent;
node              170 fs/hfsplus/brec.c 	node = fd->bnode;
node              173 fs/hfsplus/brec.c 	end_off = tree->node_size - (node->num_recs + 1) * 2;
node              175 fs/hfsplus/brec.c 	if (node->type == HFS_NODE_LEAF) {
node              179 fs/hfsplus/brec.c 	hfs_bnode_dump(node);
node              181 fs/hfsplus/brec.c 	if (!--node->num_recs) {
node              182 fs/hfsplus/brec.c 		hfs_bnode_unlink(node);
node              183 fs/hfsplus/brec.c 		if (!node->parent)
node              185 fs/hfsplus/brec.c 		parent = hfs_bnode_find(tree, node->parent);
node              188 fs/hfsplus/brec.c 		hfs_bnode_put(node);
node              189 fs/hfsplus/brec.c 		node = fd->bnode = parent;
node              191 fs/hfsplus/brec.c 		__hfs_brec_find(node, fd);
node              194 fs/hfsplus/brec.c 	hfs_bnode_write_u16(node, offsetof(struct hfs_bnode_desc, num_recs), node->num_recs);
node              201 fs/hfsplus/brec.c 		data_off = hfs_bnode_read_u16(node, rec_off);
node              202 fs/hfsplus/brec.c 		hfs_bnode_write_u16(node, rec_off + 2, data_off - size);
node              207 fs/hfsplus/brec.c 	hfs_bnode_move(node, fd->keyoffset, fd->keyoffset + size,
node              210 fs/hfsplus/brec.c 	hfs_bnode_dump(node);
node              219 fs/hfsplus/brec.c 	struct hfs_bnode *node, *new_node;
node              225 fs/hfsplus/brec.c 	node = fd->bnode;
node              229 fs/hfsplus/brec.c 	hfs_bnode_get(node);
node              231 fs/hfsplus/brec.c 		node->this, new_node->this, node->next);
node              232 fs/hfsplus/brec.c 	new_node->next = node->next;
node              233 fs/hfsplus/brec.c 	new_node->prev = node->this;
node              234 fs/hfsplus/brec.c 	new_node->parent = node->parent;
node              235 fs/hfsplus/brec.c 	new_node->type = node->type;
node              236 fs/hfsplus/brec.c 	new_node->height = node->height;
node              238 fs/hfsplus/brec.c 	size = tree->node_size / 2 - node->num_recs * 2 - 14;
node              242 fs/hfsplus/brec.c 		data_start = hfs_bnode_read_u16(node, old_rec_off);
node              246 fs/hfsplus/brec.c 		if (++num_recs < node->num_recs)
node              249 fs/hfsplus/brec.c 		hfs_bnode_put(node);
node              260 fs/hfsplus/brec.c 		data_start = hfs_bnode_read_u16(node, old_rec_off);
node              262 fs/hfsplus/brec.c 		hfs_bnode_put(node);
node              269 fs/hfsplus/brec.c 	new_node->num_recs = node->num_recs - num_recs;
node              270 fs/hfsplus/brec.c 	node->num_recs = num_recs;
node              281 fs/hfsplus/brec.c 		data_end = hfs_bnode_read_u16(node, old_rec_off);
node              286 fs/hfsplus/brec.c 	hfs_bnode_copy(new_node, 14, node, data_start, data_end - data_start);
node              298 fs/hfsplus/brec.c 	node->next = new_node->this;
node              299 fs/hfsplus/brec.c 	hfs_bnode_read(node, &node_desc, 0, sizeof(node_desc));
node              300 fs/hfsplus/brec.c 	node_desc.next = cpu_to_be32(node->next);
node              301 fs/hfsplus/brec.c 	node_desc.num_recs = cpu_to_be16(node->num_recs);
node              302 fs/hfsplus/brec.c 	hfs_bnode_write(node, &node_desc, 0, sizeof(node_desc));
node              312 fs/hfsplus/brec.c 	} else if (node->this == tree->leaf_tail) {
node              318 fs/hfsplus/brec.c 	hfs_bnode_dump(node);
node              320 fs/hfsplus/brec.c 	hfs_bnode_put(node);
node              328 fs/hfsplus/brec.c 	struct hfs_bnode *node, *new_node, *parent;
node              334 fs/hfsplus/brec.c 	node = fd->bnode;
node              336 fs/hfsplus/brec.c 	if (!node->parent)
node              340 fs/hfsplus/brec.c 	parent = hfs_bnode_find(tree, node->parent);
node              349 fs/hfsplus/brec.c 		newkeylen = hfs_bnode_read_u16(node, 14) + 2;
node              387 fs/hfsplus/brec.c 	hfs_bnode_copy(parent, fd->keyoffset, node, 14, newkeylen);
node              390 fs/hfsplus/brec.c 	hfs_bnode_put(node);
node              391 fs/hfsplus/brec.c 	node = parent;
node              407 fs/hfsplus/brec.c 			if (new_node == node)
node              410 fs/hfsplus/brec.c 			hfs_bnode_read_key(node, fd->search_key, 14);
node              414 fs/hfsplus/brec.c 	if (!rec && node->parent)
node              417 fs/hfsplus/brec.c 	fd->bnode = node;
node              423 fs/hfsplus/brec.c 	struct hfs_bnode *node, *new_node;
node              428 fs/hfsplus/brec.c 	node = NULL;
node              430 fs/hfsplus/brec.c 		node = hfs_bnode_find(tree, tree->root);
node              431 fs/hfsplus/brec.c 		if (IS_ERR(node))
node              432 fs/hfsplus/brec.c 			return PTR_ERR(node);
node              436 fs/hfsplus/brec.c 		hfs_bnode_put(node);
node              465 fs/hfsplus/brec.c 	if (node) {
node              467 fs/hfsplus/brec.c 		node->parent = tree->root;
node              468 fs/hfsplus/brec.c 		if (node->type == HFS_NODE_LEAF ||
node              470 fs/hfsplus/brec.c 			key_size = hfs_bnode_read_u16(node, 14) + 2;
node              473 fs/hfsplus/brec.c 		hfs_bnode_copy(new_node, 14, node, 14, key_size);
node              479 fs/hfsplus/brec.c 		cnid = cpu_to_be32(node->this);
node              485 fs/hfsplus/brec.c 		hfs_bnode_put(node);
node              101 fs/hfsplus/btree.c 	struct hfs_bnode *node;
node              108 fs/hfsplus/btree.c 		while ((node = tree->node_hash[i])) {
node              109 fs/hfsplus/btree.c 			tree->node_hash[i] = node->next_hash;
node              110 fs/hfsplus/btree.c 			if (atomic_read(&node->refcnt))
node              112 fs/hfsplus/btree.c 					node->tree->cnid, node->this, atomic_read(&node->refcnt));
node              113 fs/hfsplus/btree.c 			hfs_bnode_free(node);
node              124 fs/hfsplus/btree.c 	struct hfs_bnode *node;
node              127 fs/hfsplus/btree.c 	node = hfs_bnode_find(tree, 0);
node              128 fs/hfsplus/btree.c 	if (IS_ERR(node))
node              132 fs/hfsplus/btree.c 	page = node->page[0];
node              146 fs/hfsplus/btree.c 	hfs_bnode_put(node);
node              152 fs/hfsplus/btree.c 	struct hfs_bnode *node;
node              156 fs/hfsplus/btree.c 	node = hfs_bnode_create(tree, idx);
node              157 fs/hfsplus/btree.c 	if (IS_ERR(node))
node              158 fs/hfsplus/btree.c 		return node;
node              165 fs/hfsplus/btree.c 	node->type = HFS_NODE_MAP;
node              166 fs/hfsplus/btree.c 	node->num_recs = 1;
node              167 fs/hfsplus/btree.c 	hfs_bnode_clear(node, 0, tree->node_size);
node              174 fs/hfsplus/btree.c 	hfs_bnode_write(node, &desc, 0, sizeof(desc));
node              175 fs/hfsplus/btree.c 	hfs_bnode_write_u16(node, 14, 0x8000);
node              176 fs/hfsplus/btree.c 	hfs_bnode_write_u16(node, tree->node_size - 2, 14);
node              177 fs/hfsplus/btree.c 	hfs_bnode_write_u16(node, tree->node_size - 4, tree->node_size - 6);
node              179 fs/hfsplus/btree.c 	return node;
node              184 fs/hfsplus/btree.c 	struct hfs_bnode *node, *next_node;
node              213 fs/hfsplus/btree.c 	node = hfs_bnode_find(tree, nidx);
node              214 fs/hfsplus/btree.c 	if (IS_ERR(node))
node              215 fs/hfsplus/btree.c 		return node;
node              216 fs/hfsplus/btree.c 	len = hfs_brec_lenoff(node, 2, &off16);
node              219 fs/hfsplus/btree.c 	off += node->page_offset;
node              220 fs/hfsplus/btree.c 	pagep = node->page + (off >> PAGE_CACHE_SHIFT);
node              237 fs/hfsplus/btree.c 						hfs_bnode_put(node);
node              251 fs/hfsplus/btree.c 		nidx = node->next;
node              254 fs/hfsplus/btree.c 			next_node = hfs_bmap_new_bmap(node, idx);
node              257 fs/hfsplus/btree.c 		hfs_bnode_put(node);
node              260 fs/hfsplus/btree.c 		node = next_node;
node              262 fs/hfsplus/btree.c 		len = hfs_brec_lenoff(node, 0, &off16);
node              264 fs/hfsplus/btree.c 		off += node->page_offset;
node              265 fs/hfsplus/btree.c 		pagep = node->page + (off >> PAGE_CACHE_SHIFT);
node              279 fs/hfsplus/btree.c 	dprint(DBG_BNODE_MOD, "btree_free_node: %u\n", node->this);
node              280 fs/hfsplus/btree.c 	BUG_ON(!node->this);
node              281 fs/hfsplus/btree.c 	tree = node->tree;
node              282 fs/hfsplus/btree.c 	nidx = node->this;
node              283 fs/hfsplus/btree.c 	node = hfs_bnode_find(tree, 0);
node              284 fs/hfsplus/btree.c 	if (IS_ERR(node))
node              286 fs/hfsplus/btree.c 	len = hfs_brec_lenoff(node, 2, &off);
node              291 fs/hfsplus/btree.c 		i = node->next;
node              292 fs/hfsplus/btree.c 		hfs_bnode_put(node);
node              295 fs/hfsplus/btree.c 			printk(KERN_CRIT "hfs: unable to free bnode %u. bmap not found!\n", node->this);
node              298 fs/hfsplus/btree.c 		node = hfs_bnode_find(tree, i);
node              299 fs/hfsplus/btree.c 		if (IS_ERR(node))
node              301 fs/hfsplus/btree.c 		if (node->type != HFS_NODE_MAP) {
node              303 fs/hfsplus/btree.c 			printk(KERN_CRIT "hfs: invalid bmap found! (%u,%d)\n", node->this, node->type);
node              304 fs/hfsplus/btree.c 			hfs_bnode_put(node);
node              307 fs/hfsplus/btree.c 		len = hfs_brec_lenoff(node, 0, &off);
node              309 fs/hfsplus/btree.c 	off += node->page_offset + nidx / 8;
node              310 fs/hfsplus/btree.c 	page = node->page[off >> PAGE_CACHE_SHIFT];
node              316 fs/hfsplus/btree.c 		printk(KERN_CRIT "hfs: trying to free free bnode %u(%d)\n", node->this, node->type);
node              318 fs/hfsplus/btree.c 		hfs_bnode_put(node);
node              324 fs/hfsplus/btree.c 	hfs_bnode_put(node);
node               50 fs/hfsplus/inode.c 	struct hfs_bnode *node;
node               73 fs/hfsplus/inode.c 		node = hfs_bnode_findhash(tree, nidx);
node               74 fs/hfsplus/inode.c 		if (!node)
node               76 fs/hfsplus/inode.c 		else if (atomic_read(&node->refcnt))
node               78 fs/hfsplus/inode.c 		if (res && node) {
node               79 fs/hfsplus/inode.c 			hfs_bnode_unhash(node);
node               80 fs/hfsplus/inode.c 			hfs_bnode_free(node);
node               88 fs/hfsplus/inode.c 			node = hfs_bnode_findhash(tree, nidx++);
node               89 fs/hfsplus/inode.c 			if (!node)
node               91 fs/hfsplus/inode.c 			if (atomic_read(&node->refcnt)) {
node               95 fs/hfsplus/inode.c 			hfs_bnode_unhash(node);
node               96 fs/hfsplus/inode.c 			hfs_bnode_free(node);
node               72 fs/hpfs/anode.c 		if (!(fnode = hpfs_map_fnode(s, node, &bh))) return -1;
node               75 fs/hpfs/anode.c 		if (!(anode = hpfs_map_anode(s, node, &bh))) return -1;
node               78 fs/hpfs/anode.c 	a = node;
node              100 fs/hpfs/anode.c 				fnod?'f':'a', node);
node              112 fs/hpfs/anode.c 			hpfs_error(s, "empty file %08x, trying to add sector %08x", node, fsecno);
node              116 fs/hpfs/anode.c 		se = !fnod ? node : (node + 16384) & ~16383;
node              124 fs/hpfs/anode.c 		up = a != node ? anode->up : -1;
node              130 fs/hpfs/anode.c 		if (a == node && fnod) {
node              131 fs/hpfs/anode.c 			anode->up = node;
node              162 fs/hpfs/anode.c 	if ((a == node && fnod) || na == -1) return se;
node              168 fs/hpfs/anode.c 		if (up != node || !fnod) {
node              187 fs/hpfs/anode.c 				anode->btree.fnode_parent = up == node && fnod;
node              193 fs/hpfs/anode.c 		up = up != node ? anode->up : -1;
node              217 fs/hpfs/anode.c 		anode->up = node;
node              223 fs/hpfs/anode.c 		if (!(anode = hpfs_map_anode(s, node, &bh))) {
node              229 fs/hpfs/anode.c 		if (!(fnode = hpfs_map_fnode(s, node, &bh))) {
node              235 fs/hpfs/anode.c 	ranode->up = node;
node              399 fs/hpfs/anode.c 	anode_secno node = f;
node              426 fs/hpfs/anode.c 		hpfs_error(s, "internal btree %08x doesn't end with -1", node);
node              439 fs/hpfs/anode.c 		node = btree->u.internal[i].down;
node              442 fs/hpfs/anode.c 			if (hpfs_stop_cycles(s, node, &c1, &c2, "hpfs_truncate_btree"))
node              444 fs/hpfs/anode.c 		if (!(anode = hpfs_map_anode(s, node, &bh))) return;
node              495 fs/inode.c     	struct hlist_node *node;
node              499 fs/inode.c     	hlist_for_each_entry(inode, node, head, i_hash) {
node              510 fs/inode.c     	return node ? inode : NULL;
node              519 fs/inode.c     	struct hlist_node *node;
node              523 fs/inode.c     	hlist_for_each_entry(inode, node, head, i_hash) {
node              534 fs/inode.c     	return node ? inode : NULL;
node               77 fs/jffs2/debug.c 		struct jffs2_full_dnode *fn = frag->node;
node               95 fs/jffs2/debug.c 					&& frag_prev(frag)->size < PAGE_CACHE_SIZE && frag_prev(frag)->node) {
node              102 fs/jffs2/debug.c 					&& frag_next(frag)->size < PAGE_CACHE_SIZE && frag_next(frag)->node) {
node              703 fs/jffs2/debug.c 		if (this->node)
node              705 fs/jffs2/debug.c 				this->ofs, this->ofs+this->size, ref_offset(this->node->raw),
node              706 fs/jffs2/debug.c 				ref_flags(this->node->raw), this, frag_left(this), frag_right(this),
node              767 fs/jffs2/debug.c 	union jffs2_node_union node;
node              775 fs/jffs2/debug.c 	ret = jffs2_flash_read(c, ofs, len, &retlen, (unsigned char *)&node);
node              782 fs/jffs2/debug.c 	printk(JFFS2_DBG "magic:\t%#04x\n", je16_to_cpu(node.u.magic));
node              783 fs/jffs2/debug.c 	printk(JFFS2_DBG "nodetype:\t%#04x\n", je16_to_cpu(node.u.nodetype));
node              784 fs/jffs2/debug.c 	printk(JFFS2_DBG "totlen:\t%#08x\n", je32_to_cpu(node.u.totlen));
node              785 fs/jffs2/debug.c 	printk(JFFS2_DBG "hdr_crc:\t%#08x\n", je32_to_cpu(node.u.hdr_crc));
node              787 fs/jffs2/debug.c 	crc = crc32(0, &node.u, sizeof(node.u) - 4);
node              788 fs/jffs2/debug.c 	if (crc != je32_to_cpu(node.u.hdr_crc)) {
node              793 fs/jffs2/debug.c 	if (je16_to_cpu(node.u.magic) != JFFS2_MAGIC_BITMASK &&
node              794 fs/jffs2/debug.c 		je16_to_cpu(node.u.magic) != JFFS2_OLD_MAGIC_BITMASK)
node              797 fs/jffs2/debug.c 			je16_to_cpu(node.u.magic), JFFS2_MAGIC_BITMASK);
node              801 fs/jffs2/debug.c 	switch(je16_to_cpu(node.u.nodetype)) {
node              806 fs/jffs2/debug.c 		printk(JFFS2_DBG "ino:\t%#08x\n", je32_to_cpu(node.i.ino));
node              807 fs/jffs2/debug.c 		printk(JFFS2_DBG "version:\t%#08x\n", je32_to_cpu(node.i.version));
node              808 fs/jffs2/debug.c 		printk(JFFS2_DBG "mode:\t%#08x\n", node.i.mode.m);
node              809 fs/jffs2/debug.c 		printk(JFFS2_DBG "uid:\t%#04x\n", je16_to_cpu(node.i.uid));
node              810 fs/jffs2/debug.c 		printk(JFFS2_DBG "gid:\t%#04x\n", je16_to_cpu(node.i.gid));
node              811 fs/jffs2/debug.c 		printk(JFFS2_DBG "isize:\t%#08x\n", je32_to_cpu(node.i.isize));
node              812 fs/jffs2/debug.c 		printk(JFFS2_DBG "atime:\t%#08x\n", je32_to_cpu(node.i.atime));
node              813 fs/jffs2/debug.c 		printk(JFFS2_DBG "mtime:\t%#08x\n", je32_to_cpu(node.i.mtime));
node              814 fs/jffs2/debug.c 		printk(JFFS2_DBG "ctime:\t%#08x\n", je32_to_cpu(node.i.ctime));
node              815 fs/jffs2/debug.c 		printk(JFFS2_DBG "offset:\t%#08x\n", je32_to_cpu(node.i.offset));
node              816 fs/jffs2/debug.c 		printk(JFFS2_DBG "csize:\t%#08x\n", je32_to_cpu(node.i.csize));
node              817 fs/jffs2/debug.c 		printk(JFFS2_DBG "dsize:\t%#08x\n", je32_to_cpu(node.i.dsize));
node              818 fs/jffs2/debug.c 		printk(JFFS2_DBG "compr:\t%#02x\n", node.i.compr);
node              819 fs/jffs2/debug.c 		printk(JFFS2_DBG "usercompr:\t%#02x\n", node.i.usercompr);
node              820 fs/jffs2/debug.c 		printk(JFFS2_DBG "flags:\t%#04x\n", je16_to_cpu(node.i.flags));
node              821 fs/jffs2/debug.c 		printk(JFFS2_DBG "data_crc:\t%#08x\n", je32_to_cpu(node.i.data_crc));
node              822 fs/jffs2/debug.c 		printk(JFFS2_DBG "node_crc:\t%#08x\n", je32_to_cpu(node.i.node_crc));
node              824 fs/jffs2/debug.c 		crc = crc32(0, &node.i, sizeof(node.i) - 8);
node              825 fs/jffs2/debug.c 		if (crc != je32_to_cpu(node.i.node_crc)) {
node              834 fs/jffs2/debug.c 		printk(JFFS2_DBG "pino:\t%#08x\n", je32_to_cpu(node.d.pino));
node              835 fs/jffs2/debug.c 		printk(JFFS2_DBG "version:\t%#08x\n", je32_to_cpu(node.d.version));
node              836 fs/jffs2/debug.c 		printk(JFFS2_DBG "ino:\t%#08x\n", je32_to_cpu(node.d.ino));
node              837 fs/jffs2/debug.c 		printk(JFFS2_DBG "mctime:\t%#08x\n", je32_to_cpu(node.d.mctime));
node              838 fs/jffs2/debug.c 		printk(JFFS2_DBG "nsize:\t%#02x\n", node.d.nsize);
node              839 fs/jffs2/debug.c 		printk(JFFS2_DBG "type:\t%#02x\n", node.d.type);
node              840 fs/jffs2/debug.c 		printk(JFFS2_DBG "node_crc:\t%#08x\n", je32_to_cpu(node.d.node_crc));
node              841 fs/jffs2/debug.c 		printk(JFFS2_DBG "name_crc:\t%#08x\n", je32_to_cpu(node.d.name_crc));
node              843 fs/jffs2/debug.c 		node.d.name[node.d.nsize] = '\0';
node              844 fs/jffs2/debug.c 		printk(JFFS2_DBG "name:\t\"%s\"\n", node.d.name);
node              846 fs/jffs2/debug.c 		crc = crc32(0, &node.d, sizeof(node.d) - 8);
node              847 fs/jffs2/debug.c 		if (crc != je32_to_cpu(node.d.node_crc)) {
node              483 fs/jffs2/gc.c  		if (frag->node && frag->node->raw == raw) {
node              484 fs/jffs2/gc.c  			fn = frag->node;
node              488 fs/jffs2/gc.c  			if (nrfrags == frag->node->frags)
node              497 fs/jffs2/gc.c  				frag->node->raw = f->inocache->nodes;
node              543 fs/jffs2/gc.c  	union jffs2_node_union *node;
node              571 fs/jffs2/gc.c  	node = kmalloc(rawlen, GFP_KERNEL);
node              572 fs/jffs2/gc.c  	if (!node)
node              575 fs/jffs2/gc.c  	ret = jffs2_flash_read(c, ref_offset(raw), rawlen, &retlen, (char *)node);
node              581 fs/jffs2/gc.c  	crc = crc32(0, node, sizeof(struct jffs2_unknown_node)-4);
node              582 fs/jffs2/gc.c  	if (je32_to_cpu(node->u.hdr_crc) != crc) {
node              584 fs/jffs2/gc.c  		       ref_offset(raw), je32_to_cpu(node->u.hdr_crc), crc);
node              588 fs/jffs2/gc.c  	switch(je16_to_cpu(node->u.nodetype)) {
node              590 fs/jffs2/gc.c  		crc = crc32(0, node, sizeof(node->i)-8);
node              591 fs/jffs2/gc.c  		if (je32_to_cpu(node->i.node_crc) != crc) {
node              593 fs/jffs2/gc.c  			       ref_offset(raw), je32_to_cpu(node->i.node_crc), crc);
node              597 fs/jffs2/gc.c  		if (je32_to_cpu(node->i.dsize)) {
node              598 fs/jffs2/gc.c  			crc = crc32(0, node->i.data, je32_to_cpu(node->i.csize));
node              599 fs/jffs2/gc.c  			if (je32_to_cpu(node->i.data_crc) != crc) {
node              601 fs/jffs2/gc.c  				       ref_offset(raw), je32_to_cpu(node->i.data_crc), crc);
node              608 fs/jffs2/gc.c  		crc = crc32(0, node, sizeof(node->d)-8);
node              609 fs/jffs2/gc.c  		if (je32_to_cpu(node->d.node_crc) != crc) {
node              611 fs/jffs2/gc.c  			       ref_offset(raw), je32_to_cpu(node->d.node_crc), crc);
node              615 fs/jffs2/gc.c  		if (strnlen(node->d.name, node->d.nsize) != node->d.nsize) {
node              620 fs/jffs2/gc.c  		if (node->d.nsize) {
node              621 fs/jffs2/gc.c  			crc = crc32(0, node->d.name, node->d.nsize);
node              622 fs/jffs2/gc.c  			if (je32_to_cpu(node->d.name_crc) != crc) {
node              624 fs/jffs2/gc.c  				       ref_offset(raw), je32_to_cpu(node->d.name_crc), crc);
node              633 fs/jffs2/gc.c  			       ref_offset(raw), je16_to_cpu(node->u.nodetype));
node              642 fs/jffs2/gc.c  	ret = jffs2_flash_write(c, phys_ofs, rawlen, &retlen, (char *)node);
node              689 fs/jffs2/gc.c  	kfree(node);
node             1069 fs/jffs2/gc.c  		if (frag->node == fn) {
node             1070 fs/jffs2/gc.c  			frag->node = new_fn;
node             1142 fs/jffs2/gc.c  			if (!frag->node || !frag->node->raw) {
node             1152 fs/jffs2/gc.c  				struct jffs2_raw_node_ref *raw = frag->node->raw;
node             1192 fs/jffs2/gc.c  			if (!frag->node || !frag->node->raw) {
node             1202 fs/jffs2/gc.c  				struct jffs2_raw_node_ref *raw = frag->node->raw;
node              292 fs/jffs2/malloc.c 	xd->node = (void *)xd;
node              311 fs/jffs2/malloc.c 	ref->node = (void *)ref;
node               92 fs/jffs2/nodelist.c 	if (frag->node && (frag->ofs & (PAGE_CACHE_SIZE - 1)) == 0) {
node               95 fs/jffs2/nodelist.c 		frag->node->raw->flash_offset = ref_offset(frag->node->raw) | REF_PRISTINE;
node              103 fs/jffs2/nodelist.c 	if (this->node) {
node              104 fs/jffs2/nodelist.c 		this->node->frags--;
node              105 fs/jffs2/nodelist.c 		if (!this->node->frags) {
node              108 fs/jffs2/nodelist.c 				ref_offset(this->node->raw), this->node->ofs, this->node->ofs+this->node->size);
node              109 fs/jffs2/nodelist.c 			jffs2_mark_node_obsolete(c, this->node->raw);
node              110 fs/jffs2/nodelist.c 			jffs2_free_full_dnode(this->node);
node              113 fs/jffs2/nodelist.c 				ref_offset(this->node->raw), this->node->ofs, this->node->ofs+this->node->size, this->node->frags);
node              114 fs/jffs2/nodelist.c 			mark_ref_normal(this->node->raw);
node              156 fs/jffs2/nodelist.c 		newfrag->node = fn;
node              172 fs/jffs2/nodelist.c 	if (lastend < newfrag->node->ofs) {
node              176 fs/jffs2/nodelist.c 		holefrag= new_fragment(NULL, lastend, newfrag->node->ofs - lastend);
node              220 fs/jffs2/nodelist.c 	this = jffs2_lookup_node_frag(root, newfrag->node->ofs);
node              224 fs/jffs2/nodelist.c 			  this->ofs, this->ofs+this->size, this->node?(ref_offset(this->node->raw)):0xffffffff, this);
node              240 fs/jffs2/nodelist.c 			if (this->node)
node              241 fs/jffs2/nodelist.c 				mark_ref_normal(this->node->raw);
node              242 fs/jffs2/nodelist.c 			mark_ref_normal(newfrag->node->raw);
node              248 fs/jffs2/nodelist.c 	if (this->node)
node              251 fs/jffs2/nodelist.c 		ref_offset(this->node->raw), ref_flags(this->node->raw));
node              264 fs/jffs2/nodelist.c 		mark_ref_normal(newfrag->node->raw);
node              265 fs/jffs2/nodelist.c 		if (this->node)
node              266 fs/jffs2/nodelist.c 			mark_ref_normal(this->node->raw);
node              272 fs/jffs2/nodelist.c 			if (this->node)
node              274 fs/jffs2/nodelist.c 					this->ofs, this->ofs+this->size, ref_offset(this->node->raw));
node              280 fs/jffs2/nodelist.c 			newfrag2 = new_fragment(this->node, newfrag->ofs + newfrag->size,
node              284 fs/jffs2/nodelist.c 			if (this->node)
node              285 fs/jffs2/nodelist.c 				this->node->frags++;
node              351 fs/jffs2/nodelist.c 	if (this->node)
node              352 fs/jffs2/nodelist.c 		mark_ref_normal(this->node->raw);
node              353 fs/jffs2/nodelist.c 	mark_ref_normal(newfrag->node->raw);
node              373 fs/jffs2/nodelist.c 	newfrag->node->frags = 1;
node              389 fs/jffs2/nodelist.c 		if (prev->node)
node              390 fs/jffs2/nodelist.c 			mark_ref_normal(prev->node->raw);
node              398 fs/jffs2/nodelist.c 			if (next->node)
node              399 fs/jffs2/nodelist.c 				mark_ref_normal(next->node->raw);
node              585 fs/jffs2/nodelist.c 		if (frag->node && !(--frag->node->frags)) {
node              589 fs/jffs2/nodelist.c 				jffs2_mark_node_obsolete(c, frag->node->raw);
node              591 fs/jffs2/nodelist.c 			jffs2_free_full_dnode(frag->node);
node              268 fs/jffs2/nodelist.h 	struct jffs2_full_dnode *node; /* NULL for holes */
node              325 fs/jffs2/nodelist.h 	struct rb_node *node = rb_first(root);
node              327 fs/jffs2/nodelist.h 	if (!node)
node              330 fs/jffs2/nodelist.h 	return rb_entry(node, struct jffs2_node_frag, rb);
node              335 fs/jffs2/nodelist.h 	struct rb_node *node = rb_last(root);
node              337 fs/jffs2/nodelist.h 	if (!node)
node              340 fs/jffs2/nodelist.h 	return rb_entry(node, struct jffs2_node_frag, rb);
node              181 fs/jffs2/read.c 		} else if (unlikely(!frag->node)) {
node              197 fs/jffs2/read.c 				  ref_offset(frag->node->raw), ref_flags(frag->node->raw)));
node              198 fs/jffs2/read.c 			ret = jffs2_read_dnode(c, f, frag->node, buf, fragofs + frag->ofs - frag->node->ofs, readlen);
node              384 fs/jffs2/readinode.c 	struct rb_node *parent = rb_parent(node);
node              388 fs/jffs2/readinode.c 	BUG_ON(node->rb_right);
node              392 fs/jffs2/readinode.c 	else if (node == parent->rb_left)
node              397 fs/jffs2/readinode.c 	*link = node->rb_left;
node              399 fs/jffs2/readinode.c 	if (node->rb_left)
node              400 fs/jffs2/readinode.c 		node->rb_left->rb_parent_color = node->rb_parent_color;
node              959 fs/jffs2/readinode.c 	union jffs2_node_union *node;
node             1030 fs/jffs2/readinode.c 		node = (union jffs2_node_union *)buf;
node             1033 fs/jffs2/readinode.c 		if (je32_to_cpu(node->u.hdr_crc) != crc32(0, node, sizeof(node->u)-4)) {
node             1035 fs/jffs2/readinode.c 				     ref_offset(ref), je16_to_cpu(node->u.magic),
node             1036 fs/jffs2/readinode.c 				     je16_to_cpu(node->u.nodetype),
node             1037 fs/jffs2/readinode.c 				     je32_to_cpu(node->u.totlen),
node             1038 fs/jffs2/readinode.c 				     je32_to_cpu(node->u.hdr_crc));
node             1043 fs/jffs2/readinode.c 		if (je16_to_cpu(node->u.magic) != JFFS2_MAGIC_BITMASK) {
node             1046 fs/jffs2/readinode.c 				     je16_to_cpu(node->u.magic), ref_offset(ref));
node             1051 fs/jffs2/readinode.c 		switch (je16_to_cpu(node->u.nodetype)) {
node             1062 fs/jffs2/readinode.c 			err = read_direntry(c, ref, &node->d, retlen, rii);
node             1077 fs/jffs2/readinode.c 			err = read_dnode(c, ref, &node->i, len, rii);
node             1091 fs/jffs2/readinode.c 			err = read_unknown(c, ref, &node->u);
node             1300 fs/jffs2/readinode.c 		f->metadata = frag_first(&f->fragtree)->node;
node              354 fs/jffs2/scan.c 		raw->next_in_ino = xd->node->next_in_ino;
node              355 fs/jffs2/scan.c 		xd->node->next_in_ino = raw;
node              434 fs/jffs2/scan.c 	struct jffs2_unknown_node *node;
node              617 fs/jffs2/scan.c 		if (jeb->offset + c->sector_size < ofs + sizeof(*node)) {
node              619 fs/jffs2/scan.c 				  jeb->offset, c->sector_size, ofs, sizeof(*node)));
node              625 fs/jffs2/scan.c 		if (buf_ofs + buf_len < ofs + sizeof(*node)) {
node              635 fs/jffs2/scan.c 		node = (struct jffs2_unknown_node *)&buf[ofs-buf_ofs];
node              694 fs/jffs2/scan.c 		if (ofs == jeb->offset && je16_to_cpu(node->magic) == KSAMTIB_CIGAM_2SFFJ) {
node              701 fs/jffs2/scan.c 		if (je16_to_cpu(node->magic) == JFFS2_DIRTY_BITMASK) {
node              708 fs/jffs2/scan.c 		if (je16_to_cpu(node->magic) == JFFS2_OLD_MAGIC_BITMASK) {
node              716 fs/jffs2/scan.c 		if (je16_to_cpu(node->magic) != JFFS2_MAGIC_BITMASK) {
node              720 fs/jffs2/scan.c 				     je16_to_cpu(node->magic));
node              727 fs/jffs2/scan.c 		crcnode.magic = node->magic;
node              728 fs/jffs2/scan.c 		crcnode.nodetype = cpu_to_je16( je16_to_cpu(node->nodetype) | JFFS2_NODE_ACCURATE);
node              729 fs/jffs2/scan.c 		crcnode.totlen = node->totlen;
node              732 fs/jffs2/scan.c 		if (hdr_crc != je32_to_cpu(node->hdr_crc)) {
node              734 fs/jffs2/scan.c 				     ofs, je16_to_cpu(node->magic),
node              735 fs/jffs2/scan.c 				     je16_to_cpu(node->nodetype),
node              736 fs/jffs2/scan.c 				     je32_to_cpu(node->totlen),
node              737 fs/jffs2/scan.c 				     je32_to_cpu(node->hdr_crc),
node              745 fs/jffs2/scan.c 		if (ofs + je32_to_cpu(node->totlen) > jeb->offset + c->sector_size) {
node              748 fs/jffs2/scan.c 			       ofs, je32_to_cpu(node->totlen));
node              756 fs/jffs2/scan.c 		if (!(je16_to_cpu(node->nodetype) & JFFS2_NODE_ACCURATE)) {
node              759 fs/jffs2/scan.c 			if ((err = jffs2_scan_dirty_space(c, jeb, PAD(je32_to_cpu(node->totlen)))))
node              761 fs/jffs2/scan.c 			ofs += PAD(je32_to_cpu(node->totlen));
node              765 fs/jffs2/scan.c 		switch(je16_to_cpu(node->nodetype)) {
node              775 fs/jffs2/scan.c 				node = (void *)buf;
node              777 fs/jffs2/scan.c 			err = jffs2_scan_inode_node(c, jeb, (void *)node, ofs, s);
node              779 fs/jffs2/scan.c 			ofs += PAD(je32_to_cpu(node->totlen));
node              783 fs/jffs2/scan.c 			if (buf_ofs + buf_len < ofs + je32_to_cpu(node->totlen)) {
node              786 fs/jffs2/scan.c 					  je32_to_cpu(node->totlen), buf_len, ofs));
node              791 fs/jffs2/scan.c 				node = (void *)buf;
node              793 fs/jffs2/scan.c 			err = jffs2_scan_dirent_node(c, jeb, (void *)node, ofs, s);
node              795 fs/jffs2/scan.c 			ofs += PAD(je32_to_cpu(node->totlen));
node              800 fs/jffs2/scan.c 			if (buf_ofs + buf_len < ofs + je32_to_cpu(node->totlen)) {
node              804 fs/jffs2/scan.c 					  je32_to_cpu(node->totlen), buf_len, ofs));
node              809 fs/jffs2/scan.c 				node = (void *)buf;
node              811 fs/jffs2/scan.c 			err = jffs2_scan_xattr_node(c, jeb, (void *)node, ofs, s);
node              814 fs/jffs2/scan.c 			ofs += PAD(je32_to_cpu(node->totlen));
node              817 fs/jffs2/scan.c 			if (buf_ofs + buf_len < ofs + je32_to_cpu(node->totlen)) {
node              821 fs/jffs2/scan.c 					  je32_to_cpu(node->totlen), buf_len, ofs));
node              826 fs/jffs2/scan.c 				node = (void *)buf;
node              828 fs/jffs2/scan.c 			err = jffs2_scan_xref_node(c, jeb, (void *)node, ofs, s);
node              831 fs/jffs2/scan.c 			ofs += PAD(je32_to_cpu(node->totlen));
node              837 fs/jffs2/scan.c 			if (je32_to_cpu(node->totlen) != c->cleanmarker_size) {
node              839 fs/jffs2/scan.c 				       ofs, je32_to_cpu(node->totlen), c->cleanmarker_size);
node              857 fs/jffs2/scan.c 				jffs2_sum_add_padding_mem(s, je32_to_cpu(node->totlen));
node              858 fs/jffs2/scan.c 			if ((err = jffs2_scan_dirty_space(c, jeb, PAD(je32_to_cpu(node->totlen)))))
node              860 fs/jffs2/scan.c 			ofs += PAD(je32_to_cpu(node->totlen));
node              864 fs/jffs2/scan.c 			switch (je16_to_cpu(node->nodetype) & JFFS2_COMPAT_MASK) {
node              866 fs/jffs2/scan.c 				printk(KERN_NOTICE "Read-only compatible feature node (0x%04x) found at offset 0x%08x\n", je16_to_cpu(node->nodetype), ofs);
node              870 fs/jffs2/scan.c 				if ((err = jffs2_scan_dirty_space(c, jeb, PAD(je32_to_cpu(node->totlen)))))
node              872 fs/jffs2/scan.c 				ofs += PAD(je32_to_cpu(node->totlen));
node              876 fs/jffs2/scan.c 				printk(KERN_NOTICE "Incompatible feature node (0x%04x) found at offset 0x%08x\n", je16_to_cpu(node->nodetype), ofs);
node              880 fs/jffs2/scan.c 				D1(printk(KERN_NOTICE "Unknown but compatible feature node (0x%04x) found at offset 0x%08x\n", je16_to_cpu(node->nodetype), ofs));
node              881 fs/jffs2/scan.c 				if ((err = jffs2_scan_dirty_space(c, jeb, PAD(je32_to_cpu(node->totlen)))))
node              883 fs/jffs2/scan.c 				ofs += PAD(je32_to_cpu(node->totlen));
node              887 fs/jffs2/scan.c 				D1(printk(KERN_NOTICE "Unknown but compatible feature node (0x%04x) found at offset 0x%08x\n", je16_to_cpu(node->nodetype), ofs));
node              889 fs/jffs2/scan.c 				jffs2_link_node_ref(c, jeb, ofs | REF_PRISTINE, PAD(je32_to_cpu(node->totlen)), NULL);
node              893 fs/jffs2/scan.c 				ofs += PAD(je32_to_cpu(node->totlen));
node              249 fs/jffs2/summary.c 	union jffs2_node_union *node;
node              257 fs/jffs2/summary.c 	node = invecs[0].iov_base;
node              261 fs/jffs2/summary.c 	switch (je16_to_cpu(node->u.nodetype)) {
node              269 fs/jffs2/summary.c 			temp->nodetype = node->i.nodetype;
node              270 fs/jffs2/summary.c 			temp->inode = node->i.ino;
node              271 fs/jffs2/summary.c 			temp->version = node->i.version;
node              273 fs/jffs2/summary.c 			temp->totlen = node->i.totlen;
node              281 fs/jffs2/summary.c 				kmalloc(sizeof(struct jffs2_sum_dirent_mem) + node->d.nsize, GFP_KERNEL);
node              286 fs/jffs2/summary.c 			temp->nodetype = node->d.nodetype;
node              287 fs/jffs2/summary.c 			temp->totlen = node->d.totlen;
node              289 fs/jffs2/summary.c 			temp->pino = node->d.pino;
node              290 fs/jffs2/summary.c 			temp->version = node->d.version;
node              291 fs/jffs2/summary.c 			temp->ino = node->d.ino;
node              292 fs/jffs2/summary.c 			temp->nsize = node->d.nsize;
node              293 fs/jffs2/summary.c 			temp->type = node->d.type;
node              298 fs/jffs2/summary.c 					memcpy(temp->name,node->d.name,node->d.nsize);
node              302 fs/jffs2/summary.c 					memcpy(temp->name,invecs[1].iov_base,node->d.nsize);
node              319 fs/jffs2/summary.c 			temp->nodetype = node->x.nodetype;
node              320 fs/jffs2/summary.c 			temp->xid = node->x.xid;
node              321 fs/jffs2/summary.c 			temp->version = node->x.version;
node              322 fs/jffs2/summary.c 			temp->totlen = node->x.totlen;
node              333 fs/jffs2/summary.c 			temp->nodetype = node->r.nodetype;
node              342 fs/jffs2/summary.c 			c->summary->sum_padded += je32_to_cpu(node->u.totlen);
node              505 fs/jffs2/summary.c 					raw->next_in_ino = xd->node->next_in_ino;
node              506 fs/jffs2/summary.c 					xd->node->next_in_ino = raw;
node              538 fs/jffs2/summary.c 				*pseudo_random += ref->node->flash_offset;
node              185 fs/jffs2/wbuf.c 		    node, je16_to_cpu(node->u.magic), je16_to_cpu(node->u.nodetype));
node              187 fs/jffs2/wbuf.c 	BUG_ON(je16_to_cpu(node->u.magic) != 0x1985 &&
node              188 fs/jffs2/wbuf.c 	       je16_to_cpu(node->u.magic) != 0);
node              190 fs/jffs2/wbuf.c 	switch (je16_to_cpu(node->u.nodetype)) {
node              196 fs/jffs2/wbuf.c 		frag = jffs2_lookup_node_frag(&f->fragtree, je32_to_cpu(node->i.offset));
node              199 fs/jffs2/wbuf.c 		while (!frag->node || frag->node->raw != raw) {
node              203 fs/jffs2/wbuf.c 		dbg_noderef("Will replace ->raw in full_dnode at %p\n", frag->node);
node              204 fs/jffs2/wbuf.c 		return &frag->node->raw;
node              217 fs/jffs2/wbuf.c 			    je16_to_cpu(node->u.nodetype));
node              470 fs/jffs2/wbuf.c 			BUG_ON(xd->node != raw);
node              471 fs/jffs2/wbuf.c 			adjust_ref = &xd->node;
node              476 fs/jffs2/wbuf.c 			BUG_ON(xr->node != raw);
node              477 fs/jffs2/wbuf.c 			adjust_ref = &xr->node;
node               72 fs/jffs2/xattr.c 	for (raw=xd->node; raw != (void *)xd; raw=raw->next_in_ino) {
node              138 fs/jffs2/xattr.c 	offset = ref_offset(xd->node);
node              139 fs/jffs2/xattr.c 	if (ref_flags(xd->node) == REF_PRISTINE)
node              179 fs/jffs2/xattr.c 	for (raw=xd->node; raw != (void *)xd; raw=raw->next_in_ino) {
node              186 fs/jffs2/xattr.c 		raw->flash_offset = ref_offset(raw) | ((xd->node==raw) ? REF_PRISTINE : REF_NORMAL);
node              207 fs/jffs2/xattr.c 	BUG_ON(ref_flags(xd->node) != REF_PRISTINE);
node              215 fs/jffs2/xattr.c 	ret = jffs2_flash_read(c, ref_offset(xd->node)+sizeof(struct jffs2_raw_xattr),
node              220 fs/jffs2/xattr.c 			      ret, length, readlen, ref_offset(xd->node));
node              230 fs/jffs2/xattr.c 			      ref_offset(xd->node), xd->data_crc, crc);
node              405 fs/jffs2/xattr.c 		if (xd->node == (void *)xd) {
node              447 fs/jffs2/xattr.c 	if (ref_flags(ref->node) != REF_UNCHECKED)
node              449 fs/jffs2/xattr.c 	offset = ref_offset(ref->node);
node              483 fs/jffs2/xattr.c 	for (raw=ref->node; raw != (void *)ref; raw=raw->next_in_ino) {
node              490 fs/jffs2/xattr.c 		raw->flash_offset = ref_offset(raw) | ((ref->node==raw) ? REF_PRISTINE : REF_NORMAL);
node              495 fs/jffs2/xattr.c 		  ref->ino, ref->xid, ref_offset(ref->node));
node              783 fs/jffs2/xattr.c 		if (ref_flags(ref->node) != REF_PRISTINE) {
node              785 fs/jffs2/xattr.c 				BUG_ON(ref->node->next_in_ino != (void *)ref);
node              786 fs/jffs2/xattr.c 				ref->node->next_in_ino = NULL;
node              787 fs/jffs2/xattr.c 				jffs2_mark_node_obsolete(c, ref->node);
node              799 fs/jffs2/xattr.c 			raw = ref->node;
node              802 fs/jffs2/xattr.c 				raw->next_in_ino = tmp->node;
node              803 fs/jffs2/xattr.c 				tmp->node = raw;
node              805 fs/jffs2/xattr.c 				raw->next_in_ino = tmp->node->next_in_ino;
node              806 fs/jffs2/xattr.c 				tmp->node->next_in_ino = raw;
node             1206 fs/jffs2/xattr.c 	if (xd->node != raw)
node             1216 fs/jffs2/xattr.c 	old_ofs = ref_offset(xd->node);
node             1227 fs/jffs2/xattr.c 			  xd->xid, xd->version, old_ofs, ref_offset(xd->node));
node             1242 fs/jffs2/xattr.c 	BUG_ON(!ref->node);
node             1244 fs/jffs2/xattr.c 	if (ref->node != raw)
node             1249 fs/jffs2/xattr.c 	old_ofs = ref_offset(ref->node);
node             1250 fs/jffs2/xattr.c 	totlen = ref_totlen(c, c->gcblock, ref->node);
node             1262 fs/jffs2/xattr.c 			  ref->ic->ino, ref->xd->xid, old_ofs, ref_offset(ref->node));
node             1285 fs/jffs2/xattr.c 		for (raw=xd->node; raw != (void *)xd; raw=raw->next_in_ino) {
node             1293 fs/jffs2/xattr.c 				| ((xd->node == (void *)raw) ? REF_PRISTINE : REF_NORMAL);
node             1306 fs/jffs2/xattr.c 	if (atomic_read(&xd->refcnt) || xd->node != (void *)xd)
node             1318 fs/jffs2/xattr.c 	if (ref->node != (void *)ref)
node               26 fs/jffs2/xattr.h 	struct jffs2_raw_node_ref *node;
node               48 fs/jffs2/xattr.h 	struct jffs2_raw_node_ref *node;
node              423 fs/ocfs2/alloc.c 	struct ocfs2_path_item *node;
node              429 fs/ocfs2/alloc.c 		node = &path->p_node[i];
node              431 fs/ocfs2/alloc.c 		brelse(node->bh);
node              432 fs/ocfs2/alloc.c 		node->bh = NULL;
node              433 fs/ocfs2/alloc.c 		node->el = NULL;
node              450 fs/ocfs2/cluster/heartbeat.c 		(f->hc_func)(node, idx, f->hc_data);
node              509 fs/ocfs2/cluster/heartbeat.c 	event->hn_node = node;
node              522 fs/ocfs2/cluster/heartbeat.c 	struct o2nm_node *node;
node              524 fs/ocfs2/cluster/heartbeat.c 	node = o2nm_get_node_by_num(slot->ds_node_num);
node              525 fs/ocfs2/cluster/heartbeat.c 	if (!node)
node              538 fs/ocfs2/cluster/heartbeat.c 			o2hb_queue_node_event(&event, O2HB_NODE_DOWN_CB, node,
node              546 fs/ocfs2/cluster/heartbeat.c 	o2nm_node_put(node);
node              555 fs/ocfs2/cluster/heartbeat.c 	struct o2nm_node *node;
node              565 fs/ocfs2/cluster/heartbeat.c 	node = o2nm_get_node_by_num(slot->ds_node_num);
node              566 fs/ocfs2/cluster/heartbeat.c 	if (!node)
node              638 fs/ocfs2/cluster/heartbeat.c 			o2hb_queue_node_event(&event, O2HB_NODE_UP_CB, node,
node              684 fs/ocfs2/cluster/heartbeat.c 			o2hb_queue_node_event(&event, O2HB_NODE_DOWN_CB, node,
node              705 fs/ocfs2/cluster/heartbeat.c 	o2nm_node_put(node);
node              714 fs/ocfs2/cluster/heartbeat.c 	int highest, node;
node              717 fs/ocfs2/cluster/heartbeat.c 	node = -1;
node              718 fs/ocfs2/cluster/heartbeat.c 	while ((node = find_next_bit(nodes, numbits, node + 1)) != -1) {
node              719 fs/ocfs2/cluster/heartbeat.c 		if (node >= numbits)
node              722 fs/ocfs2/cluster/heartbeat.c 		highest = node;
node               41 fs/ocfs2/cluster/nodemanager.c 	struct o2nm_node *node = NULL;
node               47 fs/ocfs2/cluster/nodemanager.c 	node = o2nm_single_cluster->cl_nodes[node_num];
node               48 fs/ocfs2/cluster/nodemanager.c 	if (node)
node               49 fs/ocfs2/cluster/nodemanager.c 		config_item_get(&node->nd_item);
node               52 fs/ocfs2/cluster/nodemanager.c 	return node;
node               80 fs/ocfs2/cluster/nodemanager.c 	struct o2nm_node *node, *ret = NULL;
node               86 fs/ocfs2/cluster/nodemanager.c 		node = rb_entry(parent, struct o2nm_node, nd_ip_node);
node               88 fs/ocfs2/cluster/nodemanager.c 		cmp = memcmp(&ip_needle, &node->nd_ipv4_address,
node               95 fs/ocfs2/cluster/nodemanager.c 			ret = node;
node              110 fs/ocfs2/cluster/nodemanager.c 	struct o2nm_node *node = NULL;
node              117 fs/ocfs2/cluster/nodemanager.c 	node = o2nm_node_ip_tree_lookup(cluster, addr, NULL, NULL);
node              118 fs/ocfs2/cluster/nodemanager.c 	if (node)
node              119 fs/ocfs2/cluster/nodemanager.c 		config_item_get(&node->nd_item);
node              123 fs/ocfs2/cluster/nodemanager.c 	return node;
node              129 fs/ocfs2/cluster/nodemanager.c 	config_item_put(&node->nd_item);
node              135 fs/ocfs2/cluster/nodemanager.c 	config_item_get(&node->nd_item);
node              167 fs/ocfs2/cluster/nodemanager.c 	struct o2nm_node *node = to_o2nm_node(item);
node              168 fs/ocfs2/cluster/nodemanager.c 	kfree(node);
node              173 fs/ocfs2/cluster/nodemanager.c 	return sprintf(page, "%d\n", node->nd_num);
node              180 fs/ocfs2/cluster/nodemanager.c 	return to_o2nm_cluster(node->nd_item.ci_parent->ci_parent);
node              193 fs/ocfs2/cluster/nodemanager.c 	struct o2nm_cluster *cluster = to_o2nm_cluster_from_node(node);
node              208 fs/ocfs2/cluster/nodemanager.c 	if (!test_bit(O2NM_NODE_ATTR_ADDRESS, &node->nd_set_attributes) ||
node              209 fs/ocfs2/cluster/nodemanager.c 	    !test_bit(O2NM_NODE_ATTR_PORT, &node->nd_set_attributes))
node              216 fs/ocfs2/cluster/nodemanager.c 		cluster->cl_nodes[tmp] = node;
node              217 fs/ocfs2/cluster/nodemanager.c 		node->nd_num = tmp;
node              228 fs/ocfs2/cluster/nodemanager.c 	return sprintf(page, "%u\n", ntohs(node->nd_ipv4_port));
node              246 fs/ocfs2/cluster/nodemanager.c 	node->nd_ipv4_port = htons(tmp);
node              253 fs/ocfs2/cluster/nodemanager.c 	return sprintf(page, "%u.%u.%u.%u\n", NIPQUAD(node->nd_ipv4_address));
node              260 fs/ocfs2/cluster/nodemanager.c 	struct o2nm_cluster *cluster = to_o2nm_cluster_from_node(node);
node              282 fs/ocfs2/cluster/nodemanager.c 		rb_link_node(&node->nd_ip_node, parent, p);
node              283 fs/ocfs2/cluster/nodemanager.c 		rb_insert_color(&node->nd_ip_node, &cluster->cl_node_ip_tree);
node              289 fs/ocfs2/cluster/nodemanager.c 	memcpy(&node->nd_ipv4_address, &ipv4_addr, sizeof(ipv4_addr));
node              296 fs/ocfs2/cluster/nodemanager.c 	return sprintf(page, "%d\n", node->nd_local);
node              302 fs/ocfs2/cluster/nodemanager.c 	struct o2nm_cluster *cluster = to_o2nm_cluster_from_node(node);
node              315 fs/ocfs2/cluster/nodemanager.c 	if (!test_bit(O2NM_NODE_ATTR_ADDRESS, &node->nd_set_attributes) ||
node              316 fs/ocfs2/cluster/nodemanager.c 	    !test_bit(O2NM_NODE_ATTR_NUM, &node->nd_set_attributes) ||
node              317 fs/ocfs2/cluster/nodemanager.c 	    !test_bit(O2NM_NODE_ATTR_PORT, &node->nd_set_attributes))
node              323 fs/ocfs2/cluster/nodemanager.c 	    cluster->cl_local_node != node->nd_num)
node              328 fs/ocfs2/cluster/nodemanager.c 		ret = o2net_start_listening(node);
node              334 fs/ocfs2/cluster/nodemanager.c 	    cluster->cl_local_node == node->nd_num) {
node              335 fs/ocfs2/cluster/nodemanager.c 		o2net_stop_listening(node);
node              339 fs/ocfs2/cluster/nodemanager.c 	node->nd_local = tmp;
node              340 fs/ocfs2/cluster/nodemanager.c 	if (node->nd_local) {
node              342 fs/ocfs2/cluster/nodemanager.c 		cluster->cl_local_node = node->nd_num;
node              409 fs/ocfs2/cluster/nodemanager.c 	struct o2nm_node *node = to_o2nm_node(item);
node              415 fs/ocfs2/cluster/nodemanager.c 		ret = o2nm_node_attr->show(node, page);
node              423 fs/ocfs2/cluster/nodemanager.c 	struct o2nm_node *node = to_o2nm_node(item);
node              434 fs/ocfs2/cluster/nodemanager.c 	if (test_bit(attr_index, &node->nd_set_attributes))
node              437 fs/ocfs2/cluster/nodemanager.c 	ret = o2nm_node_attr->store(node, page, count);
node              441 fs/ocfs2/cluster/nodemanager.c 	set_bit(attr_index, &node->nd_set_attributes);
node              650 fs/ocfs2/cluster/nodemanager.c 	struct o2nm_node *node = NULL;
node              655 fs/ocfs2/cluster/nodemanager.c 	node = kzalloc(sizeof(struct o2nm_node), GFP_KERNEL);
node              656 fs/ocfs2/cluster/nodemanager.c 	if (node == NULL)
node              659 fs/ocfs2/cluster/nodemanager.c 	strcpy(node->nd_name, name); /* use item.ci_namebuf instead? */
node              660 fs/ocfs2/cluster/nodemanager.c 	config_item_init_type_name(&node->nd_item, name, &o2nm_node_type);
node              661 fs/ocfs2/cluster/nodemanager.c 	spin_lock_init(&node->nd_lock);
node              663 fs/ocfs2/cluster/nodemanager.c 	return &node->nd_item;
node              669 fs/ocfs2/cluster/nodemanager.c 	struct o2nm_node *node = to_o2nm_node(item);
node              672 fs/ocfs2/cluster/nodemanager.c 	o2net_disconnect_node(node);
node              675 fs/ocfs2/cluster/nodemanager.c 	    (cluster->cl_local_node == node->nd_num)) {
node              678 fs/ocfs2/cluster/nodemanager.c 		o2net_stop_listening(node);
node              686 fs/ocfs2/cluster/nodemanager.c 	if (node->nd_ipv4_address)
node              687 fs/ocfs2/cluster/nodemanager.c 		rb_erase(&node->nd_ip_node, &cluster->cl_node_ip_tree);
node              690 fs/ocfs2/cluster/nodemanager.c 	if (cluster->cl_nodes[node->nd_num] == node) {
node              691 fs/ocfs2/cluster/nodemanager.c 		cluster->cl_nodes[node->nd_num] = NULL;
node              692 fs/ocfs2/cluster/nodemanager.c 		clear_bit(node->nd_num, cluster->cl_nodes_bitmap);
node              161 fs/ocfs2/cluster/quorum.c 	if (!test_and_set_bit(node, qs->qs_hold_bm)) {
node              164 fs/ocfs2/cluster/quorum.c 			        "node %u\n", node);
node              165 fs/ocfs2/cluster/quorum.c 		mlog(0, "node %u, %d total\n", node, qs->qs_holds);
node              173 fs/ocfs2/cluster/quorum.c 	if (test_and_clear_bit(node, qs->qs_hold_bm)) {
node              174 fs/ocfs2/cluster/quorum.c 		mlog(0, "node %u, %d total\n", node, qs->qs_holds - 1);
node              182 fs/ocfs2/cluster/quorum.c 				node, qs->qs_holds);
node              198 fs/ocfs2/cluster/quorum.c 		        "node %u\n", node);
node              199 fs/ocfs2/cluster/quorum.c 	mlog_bug_on_msg(test_bit(node, qs->qs_hb_bm), "node %u\n", node);
node              200 fs/ocfs2/cluster/quorum.c 	set_bit(node, qs->qs_hb_bm);
node              202 fs/ocfs2/cluster/quorum.c 	mlog(0, "node %u, %d total\n", node, qs->qs_heartbeating);
node              204 fs/ocfs2/cluster/quorum.c 	if (!test_bit(node, qs->qs_conn_bm))
node              205 fs/ocfs2/cluster/quorum.c 		o2quo_set_hold(qs, node);
node              207 fs/ocfs2/cluster/quorum.c 		o2quo_clear_hold(qs, node);
node              223 fs/ocfs2/cluster/quorum.c 			node, qs->qs_heartbeating);
node              224 fs/ocfs2/cluster/quorum.c 	mlog_bug_on_msg(!test_bit(node, qs->qs_hb_bm), "node %u\n", node);
node              225 fs/ocfs2/cluster/quorum.c 	clear_bit(node, qs->qs_hb_bm);
node              227 fs/ocfs2/cluster/quorum.c 	mlog(0, "node %u, %d total\n", node, qs->qs_heartbeating);
node              229 fs/ocfs2/cluster/quorum.c 	o2quo_clear_hold(qs, node);
node              245 fs/ocfs2/cluster/quorum.c 	mlog(0, "node %u\n", node);
node              248 fs/ocfs2/cluster/quorum.c 	o2quo_clear_hold(qs, node);
node              266 fs/ocfs2/cluster/quorum.c 		        "node %u\n", node);
node              267 fs/ocfs2/cluster/quorum.c 	mlog_bug_on_msg(test_bit(node, qs->qs_conn_bm), "node %u\n", node);
node              268 fs/ocfs2/cluster/quorum.c 	set_bit(node, qs->qs_conn_bm);
node              270 fs/ocfs2/cluster/quorum.c 	mlog(0, "node %u, %d total\n", node, qs->qs_connected);
node              272 fs/ocfs2/cluster/quorum.c 	if (!test_bit(node, qs->qs_hb_bm))
node              273 fs/ocfs2/cluster/quorum.c 		o2quo_set_hold(qs, node);
node              275 fs/ocfs2/cluster/quorum.c 		o2quo_clear_hold(qs, node);
node              290 fs/ocfs2/cluster/quorum.c 	if (test_bit(node, qs->qs_conn_bm)) {
node              294 fs/ocfs2/cluster/quorum.c 				node, qs->qs_connected);
node              296 fs/ocfs2/cluster/quorum.c 		clear_bit(node, qs->qs_conn_bm);
node              299 fs/ocfs2/cluster/quorum.c 	mlog(0, "node %u, %d total\n", node, qs->qs_connected);
node              301 fs/ocfs2/cluster/quorum.c 	if (test_bit(node, qs->qs_hb_bm))
node              302 fs/ocfs2/cluster/quorum.c 		o2quo_set_hold(qs, node);
node              153 fs/ocfs2/cluster/tcp.c 	nst->st_node = node;
node              386 fs/ocfs2/cluster/tcp.c 	o2nm_node_get(node);
node              387 fs/ocfs2/cluster/tcp.c 	sc->sc_node = node;
node             1536 fs/ocfs2/cluster/tcp.c 	struct o2nm_node *node = NULL, *mynode = NULL;
node             1547 fs/ocfs2/cluster/tcp.c 	node = o2nm_get_node_by_num(o2net_num_from_nn(nn));
node             1548 fs/ocfs2/cluster/tcp.c 	if (node == NULL) {
node             1577 fs/ocfs2/cluster/tcp.c 	sc = sc_alloc(node);
node             1619 fs/ocfs2/cluster/tcp.c 	remoteaddr.sin_addr.s_addr = node->nd_ipv4_address;
node             1620 fs/ocfs2/cluster/tcp.c 	remoteaddr.sin_port = node->nd_ipv4_port;
node             1640 fs/ocfs2/cluster/tcp.c 	if (node)
node             1641 fs/ocfs2/cluster/tcp.c 		o2nm_node_put(node);
node             1678 fs/ocfs2/cluster/tcp.c 	struct o2net_node *nn = o2net_nn_from_num(node->nd_num);
node             1700 fs/ocfs2/cluster/tcp.c 		o2net_disconnect_node(node);
node             1761 fs/ocfs2/cluster/tcp.c 	struct o2nm_node *node = NULL;
node             1791 fs/ocfs2/cluster/tcp.c 	node = o2nm_get_node_by_ip(sin.sin_addr.s_addr);
node             1792 fs/ocfs2/cluster/tcp.c 	if (node == NULL) {
node             1800 fs/ocfs2/cluster/tcp.c 	if (o2nm_this_node() > node->nd_num) {
node             1803 fs/ocfs2/cluster/tcp.c 		     node->nd_name, NIPQUAD(sin.sin_addr.s_addr),
node             1804 fs/ocfs2/cluster/tcp.c 		     ntohs(sin.sin_port), node->nd_num);
node             1811 fs/ocfs2/cluster/tcp.c 	if (!o2hb_check_node_heartbeating_from_callback(node->nd_num)) {
node             1814 fs/ocfs2/cluster/tcp.c 		     node->nd_name, NIPQUAD(sin.sin_addr.s_addr),
node             1820 fs/ocfs2/cluster/tcp.c 	nn = o2net_nn_from_num(node->nd_num);
node             1831 fs/ocfs2/cluster/tcp.c 		     node->nd_name, NIPQUAD(sin.sin_addr.s_addr),
node             1836 fs/ocfs2/cluster/tcp.c 	sc = sc_alloc(node);
node             1859 fs/ocfs2/cluster/tcp.c 	if (node)
node             1860 fs/ocfs2/cluster/tcp.c 		o2nm_node_put(node);
node             1968 fs/ocfs2/cluster/tcp.c 	ret = o2net_open_listening_sock(node->nd_ipv4_address,
node             1969 fs/ocfs2/cluster/tcp.c 					node->nd_ipv4_port);
node             1974 fs/ocfs2/cluster/tcp.c 		o2quo_conn_up(node->nd_num);
node             1996 fs/ocfs2/cluster/tcp.c 		struct o2nm_node *node = o2nm_get_node_by_num(i);
node             1997 fs/ocfs2/cluster/tcp.c 		if (node) {
node             1998 fs/ocfs2/cluster/tcp.c 			o2net_disconnect_node(node);
node             1999 fs/ocfs2/cluster/tcp.c 			o2nm_node_put(node);
node             2011 fs/ocfs2/cluster/tcp.c 	o2quo_conn_err(node->nd_num);
node              196 fs/ocfs2/dlm/dlmast.c 				  lock->ml.node == dlm->node_num ? "master" :
node              224 fs/ocfs2/dlm/dlmast.c 	BUG_ON(lock->ml.node != dlm->node_num);
node              241 fs/ocfs2/dlm/dlmast.c 	BUG_ON(lock->ml.node == dlm->node_num);
node              258 fs/ocfs2/dlm/dlmast.c 	BUG_ON(lock->ml.node != dlm->node_num);
node              432 fs/ocfs2/dlm/dlmast.c 		   res->lockname.len, res->lockname.name, lock->ml.node,
node              454 fs/ocfs2/dlm/dlmast.c 				     lock->ml.node, &status);
node              460 fs/ocfs2/dlm/dlmast.c 			     "node is dead!\n", lock->ml.node);
node              464 fs/ocfs2/dlm/dlmast.c 			     "DLM_MIGRATING!\n", lock->ml.node);
node              468 fs/ocfs2/dlm/dlmast.c 			     lock->ml.node, status);
node              267 fs/ocfs2/dlm/dlmcommon.h 	dlm->joining_node = node;
node              343 fs/ocfs2/dlm/dlmcommon.h 	u8 node;
node              214 fs/ocfs2/dlm/dlmconvert.c 	if (lock->ml.node == dlm->node_num)
node              477 fs/ocfs2/dlm/dlmconvert.c 		    lock->ml.node == cnv->node_idx) {
node               81 fs/ocfs2/dlm/dlmdebug.c 	       lock->ml.type, lock->ml.convert_type, lock->ml.node,
node              562 fs/ocfs2/dlm/dlmdebug.c 		       lock->ml.node,
node              762 fs/ocfs2/dlm/dlmdebug.c 	struct dlm_reco_node_data *node;
node              857 fs/ocfs2/dlm/dlmdebug.c 	list_for_each_entry(node, &dlm->reco.node_data, list) {
node              858 fs/ocfs2/dlm/dlmdebug.c 		switch (node->state) {
node              885 fs/ocfs2/dlm/dlmdebug.c 				node->node_num, state);
node              508 fs/ocfs2/dlm/dlmdomain.c 	int node = -1;
node              514 fs/ocfs2/dlm/dlmdomain.c 	while ((node = find_next_bit(dlm->domain_map, O2NM_MAX_NODES,
node              515 fs/ocfs2/dlm/dlmdomain.c 				     node + 1)) < O2NM_MAX_NODES) {
node              516 fs/ocfs2/dlm/dlmdomain.c 		printk("%d ", node);
node              525 fs/ocfs2/dlm/dlmdomain.c 	unsigned int node;
node              533 fs/ocfs2/dlm/dlmdomain.c 	node = exit_msg->node_idx;
node              535 fs/ocfs2/dlm/dlmdomain.c 	printk(KERN_INFO "ocfs2_dlm: Node %u leaves domain %s\n", node, dlm->name);
node              538 fs/ocfs2/dlm/dlmdomain.c 	clear_bit(node, dlm->domain_map);
node              542 fs/ocfs2/dlm/dlmdomain.c 	dlm_hb_event_notify_attached(dlm, node, 0);
node              558 fs/ocfs2/dlm/dlmdomain.c 		  node, dlm->name, dlm->node_num);
node              564 fs/ocfs2/dlm/dlmdomain.c 				    &leave_msg, sizeof(leave_msg), node,
node              575 fs/ocfs2/dlm/dlmdomain.c 	int node, clear_node, status;
node              585 fs/ocfs2/dlm/dlmdomain.c 	while ((node = find_next_bit(dlm->domain_map, O2NM_MAX_NODES,
node              597 fs/ocfs2/dlm/dlmdomain.c 		status = dlm_send_one_domain_exit(dlm, node);
node              602 fs/ocfs2/dlm/dlmdomain.c 			     "to node %d\n", status, node);
node              615 fs/ocfs2/dlm/dlmdomain.c 			clear_bit(node, dlm->domain_map);
node              709 fs/ocfs2/dlm/dlmdomain.c 		     node, proto_type,
node              719 fs/ocfs2/dlm/dlmdomain.c 		     node, proto_type,
node              960 fs/ocfs2/dlm/dlmdomain.c 				    &cancel_msg, sizeof(cancel_msg), node,
node              977 fs/ocfs2/dlm/dlmdomain.c 	unsigned int node;
node              988 fs/ocfs2/dlm/dlmdomain.c 	node = -1;
node              989 fs/ocfs2/dlm/dlmdomain.c 	while ((node = find_next_bit(node_map, O2NM_MAX_NODES,
node              990 fs/ocfs2/dlm/dlmdomain.c 				     node + 1)) < O2NM_MAX_NODES) {
node              991 fs/ocfs2/dlm/dlmdomain.c 		if (node == dlm->node_num)
node              994 fs/ocfs2/dlm/dlmdomain.c 		tmpstat = dlm_send_one_join_cancel(dlm, node);
node              997 fs/ocfs2/dlm/dlmdomain.c 			     "node %d\n", tmpstat, node);
node             1017 fs/ocfs2/dlm/dlmdomain.c 	mlog(0, "querying node %d\n", node);
node             1030 fs/ocfs2/dlm/dlmdomain.c 				    sizeof(join_msg), node,
node             1058 fs/ocfs2/dlm/dlmdomain.c 		     node);
node             1069 fs/ocfs2/dlm/dlmdomain.c 		     node,
node             1077 fs/ocfs2/dlm/dlmdomain.c 		     packet.code, node);
node             1080 fs/ocfs2/dlm/dlmdomain.c 	mlog(0, "status %d, node %d response is %d\n", status, node,
node             1093 fs/ocfs2/dlm/dlmdomain.c 	mlog(0, "Sending join assert to node %u\n", node);
node             1101 fs/ocfs2/dlm/dlmdomain.c 				    &assert_msg, sizeof(assert_msg), node,
node             1112 fs/ocfs2/dlm/dlmdomain.c 	int status, node, live;
node             1115 fs/ocfs2/dlm/dlmdomain.c 	node = -1;
node             1116 fs/ocfs2/dlm/dlmdomain.c 	while ((node = find_next_bit(node_map, O2NM_MAX_NODES,
node             1117 fs/ocfs2/dlm/dlmdomain.c 				     node + 1)) < O2NM_MAX_NODES) {
node             1118 fs/ocfs2/dlm/dlmdomain.c 		if (node == dlm->node_num)
node             1125 fs/ocfs2/dlm/dlmdomain.c 			status = dlm_send_one_join_assert(dlm, node);
node             1128 fs/ocfs2/dlm/dlmdomain.c 			live = test_bit(node, dlm->live_nodes_map);
node             1133 fs/ocfs2/dlm/dlmdomain.c 				     "join on node %d\n", status, node);
node             1174 fs/ocfs2/dlm/dlmdomain.c 	int status = 0, tmpstat, node;
node             1199 fs/ocfs2/dlm/dlmdomain.c 	node = -1;
node             1200 fs/ocfs2/dlm/dlmdomain.c 	while ((node = find_next_bit(ctxt->live_map, O2NM_MAX_NODES,
node             1201 fs/ocfs2/dlm/dlmdomain.c 				     node + 1)) < O2NM_MAX_NODES) {
node             1202 fs/ocfs2/dlm/dlmdomain.c 		if (node == dlm->node_num)
node             1205 fs/ocfs2/dlm/dlmdomain.c 		status = dlm_request_join(dlm, node, &response);
node             1214 fs/ocfs2/dlm/dlmdomain.c 			set_bit(node, ctxt->yes_resp_map);
node              136 fs/ocfs2/dlm/dlmlock.c 	    lock->ml.node != dlm->node_num) {
node              165 fs/ocfs2/dlm/dlmlock.c 			     lock->ml.node);
node              176 fs/ocfs2/dlm/dlmlock.c 				     lock->ml.node);
node              186 fs/ocfs2/dlm/dlmlock.c 	if (lock->ml.node == dlm->node_num)
node              408 fs/ocfs2/dlm/dlmlock.c 	newlock->ml.node = node;
node              447 fs/ocfs2/dlm/dlmlock.c 	dlm_init_lock(lock, type, node, cookie);
node             1196 fs/ocfs2/dlm/dlmmaster.c 	int node;
node             1205 fs/ocfs2/dlm/dlmmaster.c 	node = dlm_bitmap_diff_iter_next(&bdi, &sc);
node             1206 fs/ocfs2/dlm/dlmmaster.c 	while (node >= 0) {
node             1211 fs/ocfs2/dlm/dlmmaster.c 			mlog(ML_NOTICE, "node %d up while restarting\n", node);
node             1215 fs/ocfs2/dlm/dlmmaster.c 			clear_bit(node, mle->response_map);
node             1216 fs/ocfs2/dlm/dlmmaster.c 			set_bit(node, mle->vote_map);
node             1218 fs/ocfs2/dlm/dlmmaster.c 			mlog(ML_ERROR, "node down! %d\n", node);
node             1224 fs/ocfs2/dlm/dlmmaster.c 				clear_bit(node, mle->maybe_map);
node             1226 fs/ocfs2/dlm/dlmmaster.c 			       	if (node == lowest) {
node             1229 fs/ocfs2/dlm/dlmmaster.c 					    "waiting on it!\n", node);
node             1273 fs/ocfs2/dlm/dlmmaster.c 		node = dlm_bitmap_diff_iter_next(&bdi, &sc);
node             2233 fs/ocfs2/dlm/dlmmaster.c 	u8 node;
node             2244 fs/ocfs2/dlm/dlmmaster.c 	node = deref->node_idx;
node             2251 fs/ocfs2/dlm/dlmmaster.c 		mlog(ML_ERROR, "Invalid node number: %u\n", node);
node             2272 fs/ocfs2/dlm/dlmmaster.c 		if (test_bit(node, res->refmap)) {
node             2273 fs/ocfs2/dlm/dlmmaster.c 			dlm_lockres_clear_refmap_bit(node, res);
node             2285 fs/ocfs2/dlm/dlmmaster.c 		     	res->lockname.len, res->lockname.name, node);
node             2301 fs/ocfs2/dlm/dlmmaster.c 	item->u.dl.deref_node = node;
node             2322 fs/ocfs2/dlm/dlmmaster.c 	u8 node;
node             2327 fs/ocfs2/dlm/dlmmaster.c 	node = item->u.dl.deref_node;
node             2331 fs/ocfs2/dlm/dlmmaster.c 	if (test_bit(node, res->refmap)) {
node             2333 fs/ocfs2/dlm/dlmmaster.c 		dlm_lockres_clear_refmap_bit(node, res);
node             2340 fs/ocfs2/dlm/dlmmaster.c 		     dlm->name, res->lockname.len, res->lockname.name, node);
node             2345 fs/ocfs2/dlm/dlmmaster.c 		     res->lockname.len, res->lockname.name, node);
node             2383 fs/ocfs2/dlm/dlmmaster.c 			if (lock->ml.node == dlm->node_num) {
node             2860 fs/ocfs2/dlm/dlmmaster.c 			if (lock->ml.node != dlm->node_num) {
node             2862 fs/ocfs2/dlm/dlmmaster.c 				     lock->ml.node);
node             2868 fs/ocfs2/dlm/dlmmaster.c 				dlm_lockres_clear_refmap_bit(lock->ml.node, res);
node             2914 fs/ocfs2/dlm/dlmmaster.c 			if (lock->ml.node != dlm->node_num) {
node             2916 fs/ocfs2/dlm/dlmmaster.c 				return lock->ml.node;
node              349 fs/ocfs2/dlm/dlmrecovery.c 	dead = !test_bit(node, dlm->domain_map);
node              360 fs/ocfs2/dlm/dlmrecovery.c 	recovered = !test_bit(node, dlm->recovery_map);
node              370 fs/ocfs2/dlm/dlmrecovery.c 		     "death of node %u\n", dlm->name, timeout, node);
node              372 fs/ocfs2/dlm/dlmrecovery.c 			   dlm_is_node_dead(dlm, node),
node              376 fs/ocfs2/dlm/dlmrecovery.c 		     "of death of node %u\n", dlm->name, node);
node              378 fs/ocfs2/dlm/dlmrecovery.c 			   dlm_is_node_dead(dlm, node));
node              388 fs/ocfs2/dlm/dlmrecovery.c 		     "recovery of node %u\n", dlm->name, timeout, node);
node              390 fs/ocfs2/dlm/dlmrecovery.c 			   dlm_is_node_recovered(dlm, node),
node              394 fs/ocfs2/dlm/dlmrecovery.c 		     "of recovery of node %u\n", dlm->name, node);
node              396 fs/ocfs2/dlm/dlmrecovery.c 			   dlm_is_node_recovered(dlm, node));
node             1051 fs/ocfs2/dlm/dlmrecovery.c 				if (lock->ml.node == dead_node) {
node             1200 fs/ocfs2/dlm/dlmrecovery.c 	ml->node = lock->ml.node;
node             1218 fs/ocfs2/dlm/dlmrecovery.c 	dummy.ml.node = dlm->node_num;
node             1231 fs/ocfs2/dlm/dlmrecovery.c 		*nodenum = ml->node;
node             1761 fs/ocfs2/dlm/dlmrecovery.c 		if (ml->node == dlm->node_num) {
node             1789 fs/ocfs2/dlm/dlmrecovery.c 			BUG_ON(lock->ml.node != ml->node);
node             1812 fs/ocfs2/dlm/dlmrecovery.c 		newlock = dlm_new_lock(ml->type, ml->node,
node             1899 fs/ocfs2/dlm/dlmrecovery.c 	      			     ml->type, ml->convert_type, ml->node,
node             1914 fs/ocfs2/dlm/dlmrecovery.c 			     res->lockname.len, res->lockname.name, ml->node);
node             1915 fs/ocfs2/dlm/dlmrecovery.c 			dlm_lockres_set_refmap_bit(ml->node, res);
node             2131 fs/ocfs2/dlm/dlmrecovery.c 			if (lock->ml.node == search_node) {
node             2166 fs/ocfs2/dlm/dlmrecovery.c 		if (lock->ml.node == dead_node) {
node             2175 fs/ocfs2/dlm/dlmrecovery.c 		if (lock->ml.node == dead_node) {
node             2184 fs/ocfs2/dlm/dlmrecovery.c 		if (lock->ml.node == dead_node) {
node             2252 fs/ocfs2/dlm/dlmrecovery.c 					if (lock->ml.node == dead_node) {
node               59 fs/ocfs2/dlm/dlmthread.c #define dlm_lock_is_remote(dlm, lock)     ((lock)->ml.node != (dlm)->node_num)
node              356 fs/ocfs2/dlm/dlmthread.c 		     target->ml.convert_type, target->ml.node);
node              418 fs/ocfs2/dlm/dlmthread.c 		     target->ml.type, target->ml.node);
node              533 fs/ocfs2/dlm/dlmthread.c 		if (lock->ml.node != dlm->node_num) {
node              581 fs/ocfs2/dlm/dlmthread.c 		if (lock->ml.node != dlm->node_num) {
node              465 fs/ocfs2/dlm/dlmunlock.c 		    	    lock->ml.node == unlock->node_idx) {
node              346 fs/ocfs2/stack_o2cb.c 	*node = node_num;
node              439 fs/ocfs2/stackglue.c 	return active_stack->sp_ops->this_node(node);
node              150 fs/ocfs2/stackglue.h 	int (*this_node)(unsigned int *node);
node               92 fs/ocfs2/uptodate.c 	struct rb_node *node;
node               95 fs/ocfs2/uptodate.c 	while ((node = rb_last(root)) != NULL) {
node               96 fs/ocfs2/uptodate.c 		item = rb_entry(node, struct ocfs2_meta_cache_item, c_node);
node               31 fs/openpromfs/inode.c 	struct device_node	*node;
node              192 fs/openpromfs/inode.c 	dp = oi->u.node;
node              206 fs/openpromfs/inode.c 			ent_data.node = child;
node              269 fs/openpromfs/inode.c 	struct device_node *dp = oi->u.node;
node              405 fs/openpromfs/inode.c 	oi->u.node = of_find_node_by_path("/");
node              255 fs/ubifs/debug.c 	const struct ubifs_ch *ch = node;
node              264 fs/ubifs/debug.c 			       (void *)node, UBIFS_CH_SZ, 1);
node              269 fs/ubifs/debug.c 	dump_ch(node);
node              274 fs/ubifs/debug.c 		const struct ubifs_pad_node *pad = node;
node              282 fs/ubifs/debug.c 		const struct ubifs_sb_node *sup = node;
node              337 fs/ubifs/debug.c 		const struct ubifs_mst_node *mst = node;
node              399 fs/ubifs/debug.c 		const struct ubifs_ref_node *ref = node;
node              411 fs/ubifs/debug.c 		const struct ubifs_ino_node *ino = node;
node              453 fs/ubifs/debug.c 		const struct ubifs_dent_node *dent = node;
node              477 fs/ubifs/debug.c 		const struct ubifs_data_node *dn = node;
node              495 fs/ubifs/debug.c 		const struct ubifs_trun_node *trun = node;
node              507 fs/ubifs/debug.c 		const struct ubifs_idx_node *idx = node;
node              530 fs/ubifs/debug.c 		const struct ubifs_orph_node *orph = node;
node              673 fs/ubifs/debug.c 		dbg_dump_node(c, snod->node);
node             1682 fs/ubifs/debug.c 	void *node;
node             1693 fs/ubifs/debug.c 	node = kmalloc(zbr->len, GFP_NOFS);
node             1694 fs/ubifs/debug.c 	if (!node)
node             1697 fs/ubifs/debug.c 	err = ubifs_tnc_read_node(c, zbr, node);
node             1706 fs/ubifs/debug.c 		fscki = add_inode(c, priv, node);
node             1723 fs/ubifs/debug.c 	ch = node;
node             1733 fs/ubifs/debug.c 		struct ubifs_data_node *dn = node;
node             1761 fs/ubifs/debug.c 		struct ubifs_dent_node *dent = node;
node             1808 fs/ubifs/debug.c 	kfree(node);
node             1813 fs/ubifs/debug.c 	dbg_dump_node(c, node);
node             1815 fs/ubifs/debug.c 	kfree(node);
node              197 fs/ubifs/gc.c  			err = ubifs_wbuf_write_nolock(wbuf, snod->node,
node              299 fs/ubifs/gc.c  			struct ubifs_idx_node *idx = snod->node;
node              228 fs/ubifs/io.c  	struct ubifs_ch *ch = node;
node              238 fs/ubifs/io.c  	crc = crc32(UBIFS_CRC32_INIT, node + 8, len - 8);
node              244 fs/ubifs/io.c  		ubifs_pad(c, node + len, pad);
node              261 fs/ubifs/io.c  	struct ubifs_ch *ch = node;
node              274 fs/ubifs/io.c  	crc = crc32(UBIFS_CRC32_INIT, node + 8, len - 8);
node              260 fs/ubifs/journal.c 	ubifs_prepare_node(c, node, len, 0);
node              262 fs/ubifs/journal.c 	return ubifs_wbuf_write_nolock(wbuf, node, len);
node              655 fs/ubifs/log.c 	struct ubifs_ch *ch = node;
node              668 fs/ubifs/log.c 	memcpy(buf + *offs, node, len);
node              707 fs/ubifs/log.c 				struct ubifs_ref_node *ref = snod->node;
node              715 fs/ubifs/log.c 						       &offs, snod->node);
node              725 fs/ubifs/log.c 					       snod->node);
node             1166 fs/ubifs/lprops.c 			struct ubifs_idx_node *idx = snod->node;
node               52 fs/ubifs/master.c 		memcpy(c->mst_node, snod->node, snod->len);
node               72 fs/ubifs/master.c 		   (void *)snod->node + UBIFS_CH_SZ,
node              310 fs/ubifs/misc.h 	return ubifs_tnc_locate(c, key, node, NULL, NULL);
node              570 fs/ubifs/orphan.c 			dbg_dump_node(c, snod->node);
node              574 fs/ubifs/orphan.c 		orph = snod->node;
node              598 fs/ubifs/orphan.c 				dbg_dump_node(c, snod->node);
node              735 fs/ubifs/orphan.c 	struct ubifs_ino_node *node;
node              847 fs/ubifs/orphan.c 		err = ubifs_tnc_read_node(c, zbr, ci->node);
node              852 fs/ubifs/orphan.c 		if (ci->node->nlink == 0)
node              875 fs/ubifs/orphan.c 		orph = snod->node;
node              926 fs/ubifs/orphan.c 	ci.node = kmalloc(UBIFS_MAX_INO_NODE_SZ, GFP_NOFS);
node              927 fs/ubifs/orphan.c 	if (!ci.node) {
node              954 fs/ubifs/orphan.c 	kfree(ci.node);
node              570 fs/ubifs/recovery.c 		ch = snod->node;
node              551 fs/ubifs/replay.c 			struct ubifs_ino_node *ino = snod->node;
node              563 fs/ubifs/replay.c 			struct ubifs_data_node *dn = snod->node;
node              576 fs/ubifs/replay.c 			struct ubifs_dent_node *dent = snod->node;
node              590 fs/ubifs/replay.c 			struct ubifs_trun_node *trun = snod->node;
node              644 fs/ubifs/replay.c 	dbg_dump_node(c, snod->node);
node              836 fs/ubifs/replay.c 	const struct ubifs_cs_node *node;
node              852 fs/ubifs/replay.c 	node = sleb->buf;
node              868 fs/ubifs/replay.c 		if (le64_to_cpu(node->cmt_no) != c->cmt_no) {
node              872 fs/ubifs/replay.c 				(unsigned long long)le64_to_cpu(node->cmt_no),
node              877 fs/ubifs/replay.c 		c->cs_sqnum = le64_to_cpu(node->ch.sqnum);
node              919 fs/ubifs/replay.c 			const struct ubifs_ref_node *ref = snod->node;
node              962 fs/ubifs/replay.c 	dbg_dump_node(c, snod->node);
node              209 fs/ubifs/scan.c 	snod->node = buf;
node              352 fs/ubifs/scan.c 	struct ubifs_scan_node *node;
node              357 fs/ubifs/scan.c 		node = list_entry(head->next, struct ubifs_scan_node, list);
node              358 fs/ubifs/scan.c 		list_del(&node->list);
node              359 fs/ubifs/scan.c 		kfree(node);
node              333 fs/ubifs/tnc.c 	const struct ubifs_dent_node *dent = node;
node              351 fs/ubifs/tnc.c 	memcpy(lnc_node, node, zbr->len);
node              373 fs/ubifs/tnc.c 	err = ubifs_validate_entry(c, node);
node              376 fs/ubifs/tnc.c 		dbg_dump_node(c, node);
node              380 fs/ubifs/tnc.c 	zbr->leaf = node;
node              418 fs/ubifs/tnc.c 		memcpy(node, zbr->leaf, zbr->len);
node              422 fs/ubifs/tnc.c 	err = ubifs_tnc_read_node(c, zbr, node);
node              427 fs/ubifs/tnc.c 	err = lnc_add(c, zbr, node);
node              498 fs/ubifs/tnc.c 	ret = try_read_node(c, node, key_type(c, key), zbr->len, zbr->lnum,
node              502 fs/ubifs/tnc.c 		struct ubifs_dent_node *dent = node;
node             1460 fs/ubifs/tnc.c 		err = tnc_read_node_nm(c, zt, node);
node             1464 fs/ubifs/tnc.c 		err = ubifs_tnc_read_node(c, zt, node);
node             1474 fs/ubifs/tnc.c 		err = ubifs_tnc_read_node(c, &zbr, node);
node             1478 fs/ubifs/tnc.c 	err = fallible_read_node(c, key, &zbr, node);
node             1535 fs/ubifs/tnc.c 	err = tnc_read_node_nm(c, &znode->zbranch[n], node);
node             1559 fs/ubifs/tnc.c 	const struct ubifs_dent_node *dent = node;
node             1565 fs/ubifs/tnc.c 	err = ubifs_tnc_lookup(c, key, node);
node             1577 fs/ubifs/tnc.c 	return do_lookup_nm(c, key, node, nm);
node              258 fs/ubifs/tnc_commit.c 		idx = snod->node;
node              471 fs/ubifs/tnc_misc.c 		err = ubifs_read_node_wbuf(wbuf, node, type, zbr->len,
node              474 fs/ubifs/tnc_misc.c 		err = ubifs_read_node(c, node, type, zbr->len, zbr->lnum,
node              484 fs/ubifs/tnc_misc.c 	if (memcmp(node + UBIFS_KEY_OFFSET, &key1, c->key_len)) {
node              489 fs/ubifs/tnc_misc.c 		dbg_dump_node(c, node);
node              282 fs/ubifs/ubifs.h 	void *node;
node             1828 fs/xfs/xfs_attr.c 	xfs_da_intnode_t *node;
node             1848 fs/xfs/xfs_attr.c 			node = bp->data;
node             1849 fs/xfs/xfs_attr.c 			switch (be16_to_cpu(node->hdr.info.magic)) {
node             1851 fs/xfs/xfs_attr.c 				xfs_attr_trace_l_cn("wrong blk", context, node);
node             1898 fs/xfs/xfs_attr.c 			node = bp->data;
node             1899 fs/xfs/xfs_attr.c 			if (be16_to_cpu(node->hdr.info.magic)
node             1902 fs/xfs/xfs_attr.c 			if (unlikely(be16_to_cpu(node->hdr.info.magic)
node             1907 fs/xfs/xfs_attr.c 						     node);
node             1911 fs/xfs/xfs_attr.c 			btree = node->btree;
node             1912 fs/xfs/xfs_attr.c 			for (i = 0; i < be16_to_cpu(node->hdr.count);
node             1922 fs/xfs/xfs_attr.c 			if (i == be16_to_cpu(node->hdr.count)) {
node             2288 fs/xfs/xfs_attr.c 		(__psunsigned_t)be16_to_cpu(node->hdr.count),
node             2289 fs/xfs/xfs_attr.c 		(__psunsigned_t)be32_to_cpu(node->btree[0].hashval),
node             2290 fs/xfs/xfs_attr.c 		(__psunsigned_t)be32_to_cpu(node->btree[
node             2291 fs/xfs/xfs_attr.c 				    be16_to_cpu(node->hdr.count)-1].hashval));
node              849 fs/xfs/xfs_attr_leaf.c 	xfs_da_intnode_t *node;
node              882 fs/xfs/xfs_attr_leaf.c 	node = bp1->data;
node              886 fs/xfs/xfs_attr_leaf.c 	node->btree[0].hashval =
node              888 fs/xfs/xfs_attr_leaf.c 	node->btree[0].before = cpu_to_be32(blkno);
node              889 fs/xfs/xfs_attr_leaf.c 	node->hdr.count = cpu_to_be16(1);
node             2736 fs/xfs/xfs_attr_leaf.c 	xfs_da_intnode_t *node;
node             2750 fs/xfs/xfs_attr_leaf.c 	node = bp->data;
node             2751 fs/xfs/xfs_attr_leaf.c 	ASSERT(be16_to_cpu(node->hdr.info.magic) == XFS_DA_NODE_MAGIC);
node             2753 fs/xfs/xfs_attr_leaf.c 	count = be16_to_cpu(node->hdr.count);
node             2758 fs/xfs/xfs_attr_leaf.c 	child_fsb = be32_to_cpu(node->btree[0].before);
node             2818 fs/xfs/xfs_attr_leaf.c 			child_fsb = be32_to_cpu(node->btree[i+1].before);
node              113 fs/xfs/xfs_da_btree.c 	xfs_da_intnode_t *node;
node              123 fs/xfs/xfs_da_btree.c 	node = bp->data;
node              124 fs/xfs/xfs_da_btree.c 	node->hdr.info.forw = 0;
node              125 fs/xfs/xfs_da_btree.c 	node->hdr.info.back = 0;
node              126 fs/xfs/xfs_da_btree.c 	node->hdr.info.magic = cpu_to_be16(XFS_DA_NODE_MAGIC);
node              127 fs/xfs/xfs_da_btree.c 	node->hdr.info.pad = 0;
node              128 fs/xfs/xfs_da_btree.c 	node->hdr.count = 0;
node              129 fs/xfs/xfs_da_btree.c 	node->hdr.level = cpu_to_be16(level);
node              132 fs/xfs/xfs_da_btree.c 		XFS_DA_LOGRANGE(node, &node->hdr, sizeof(node->hdr)));
node              146 fs/xfs/xfs_da_btree.c 	xfs_da_intnode_t *node;
node              257 fs/xfs/xfs_da_btree.c 	node = oldblk->bp->data;
node              258 fs/xfs/xfs_da_btree.c 	if (node->hdr.info.forw) {
node              259 fs/xfs/xfs_da_btree.c 		if (be32_to_cpu(node->hdr.info.forw) == addblk->blkno) {
node              265 fs/xfs/xfs_da_btree.c 		node = bp->data;
node              266 fs/xfs/xfs_da_btree.c 		node->hdr.info.back = cpu_to_be32(oldblk->blkno);
node              268 fs/xfs/xfs_da_btree.c 		    XFS_DA_LOGRANGE(node, &node->hdr.info,
node              269 fs/xfs/xfs_da_btree.c 		    sizeof(node->hdr.info)));
node              271 fs/xfs/xfs_da_btree.c 	node = oldblk->bp->data;
node              272 fs/xfs/xfs_da_btree.c 	if (node->hdr.info.back) {
node              273 fs/xfs/xfs_da_btree.c 		if (be32_to_cpu(node->hdr.info.back) == addblk->blkno) {
node              279 fs/xfs/xfs_da_btree.c 		node = bp->data;
node              280 fs/xfs/xfs_da_btree.c 		node->hdr.info.forw = cpu_to_be32(oldblk->blkno);
node              282 fs/xfs/xfs_da_btree.c 		    XFS_DA_LOGRANGE(node, &node->hdr.info,
node              283 fs/xfs/xfs_da_btree.c 		    sizeof(node->hdr.info)));
node              300 fs/xfs/xfs_da_btree.c 	xfs_da_intnode_t *node, *oldroot;
node              326 fs/xfs/xfs_da_btree.c 	node = bp->data;
node              337 fs/xfs/xfs_da_btree.c 	memcpy(node, oldroot, size);
node              348 fs/xfs/xfs_da_btree.c 		be16_to_cpu(node->hdr.level) + 1, &bp, args->whichfork);
node              351 fs/xfs/xfs_da_btree.c 	node = bp->data;
node              352 fs/xfs/xfs_da_btree.c 	node->btree[0].hashval = cpu_to_be32(blk1->hashval);
node              353 fs/xfs/xfs_da_btree.c 	node->btree[0].before = cpu_to_be32(blk1->blkno);
node              354 fs/xfs/xfs_da_btree.c 	node->btree[1].hashval = cpu_to_be32(blk2->hashval);
node              355 fs/xfs/xfs_da_btree.c 	node->btree[1].before = cpu_to_be32(blk2->blkno);
node              356 fs/xfs/xfs_da_btree.c 	node->hdr.count = cpu_to_be16(2);
node              369 fs/xfs/xfs_da_btree.c 		XFS_DA_LOGRANGE(node, node->btree,
node              385 fs/xfs/xfs_da_btree.c 	xfs_da_intnode_t *node;
node              390 fs/xfs/xfs_da_btree.c 	node = oldblk->bp->data;
node              391 fs/xfs/xfs_da_btree.c 	ASSERT(be16_to_cpu(node->hdr.info.magic) == XFS_DA_NODE_MAGIC);
node              401 fs/xfs/xfs_da_btree.c 	if ((be16_to_cpu(node->hdr.count) + newcount) > state->node_ents) {
node              437 fs/xfs/xfs_da_btree.c 	node = oldblk->bp->data;
node              438 fs/xfs/xfs_da_btree.c 	if (oldblk->index <= be16_to_cpu(node->hdr.count)) {
node              580 fs/xfs/xfs_da_btree.c 	xfs_da_intnode_t *node;
node              585 fs/xfs/xfs_da_btree.c 	node = oldblk->bp->data;
node              587 fs/xfs/xfs_da_btree.c 	ASSERT(be16_to_cpu(node->hdr.info.magic) == XFS_DA_NODE_MAGIC);
node              588 fs/xfs/xfs_da_btree.c 	ASSERT((oldblk->index >= 0) && (oldblk->index <= be16_to_cpu(node->hdr.count)));
node              598 fs/xfs/xfs_da_btree.c 	btree = &node->btree[ oldblk->index ];
node              599 fs/xfs/xfs_da_btree.c 	if (oldblk->index < be16_to_cpu(node->hdr.count)) {
node              600 fs/xfs/xfs_da_btree.c 		tmp = (be16_to_cpu(node->hdr.count) - oldblk->index) * (uint)sizeof(*btree);
node              606 fs/xfs/xfs_da_btree.c 		XFS_DA_LOGRANGE(node, btree, tmp + sizeof(*btree)));
node              607 fs/xfs/xfs_da_btree.c 	be16_add_cpu(&node->hdr.count, 1);
node              609 fs/xfs/xfs_da_btree.c 		XFS_DA_LOGRANGE(node, &node->hdr, sizeof(node->hdr)));
node              614 fs/xfs/xfs_da_btree.c 	oldblk->hashval = be32_to_cpu(node->btree[be16_to_cpu(node->hdr.count)-1 ].hashval);
node              771 fs/xfs/xfs_da_btree.c 	xfs_da_intnode_t *node;
node              786 fs/xfs/xfs_da_btree.c 	node = (xfs_da_intnode_t *)info;
node              787 fs/xfs/xfs_da_btree.c 	count = be16_to_cpu(node->hdr.count);
node              840 fs/xfs/xfs_da_btree.c 		node = (xfs_da_intnode_t *)info;
node              843 fs/xfs/xfs_da_btree.c 		count -= be16_to_cpu(node->hdr.count);
node              844 fs/xfs/xfs_da_btree.c 		node = bp->data;
node              845 fs/xfs/xfs_da_btree.c 		ASSERT(be16_to_cpu(node->hdr.info.magic) == XFS_DA_NODE_MAGIC);
node              846 fs/xfs/xfs_da_btree.c 		count -= be16_to_cpu(node->hdr.count);
node              894 fs/xfs/xfs_da_btree.c 	xfs_da_intnode_t *node;
node              919 fs/xfs/xfs_da_btree.c 		node = blk->bp->data;
node              920 fs/xfs/xfs_da_btree.c 		ASSERT(be16_to_cpu(node->hdr.info.magic) == XFS_DA_NODE_MAGIC);
node              921 fs/xfs/xfs_da_btree.c 		btree = &node->btree[ blk->index ];
node              927 fs/xfs/xfs_da_btree.c 				  XFS_DA_LOGRANGE(node, btree, sizeof(*btree)));
node              929 fs/xfs/xfs_da_btree.c 		lasthash = be32_to_cpu(node->btree[be16_to_cpu(node->hdr.count)-1].hashval);
node              939 fs/xfs/xfs_da_btree.c 	xfs_da_intnode_t *node;
node              943 fs/xfs/xfs_da_btree.c 	node = drop_blk->bp->data;
node              944 fs/xfs/xfs_da_btree.c 	ASSERT(drop_blk->index < be16_to_cpu(node->hdr.count));
node              950 fs/xfs/xfs_da_btree.c 	btree = &node->btree[drop_blk->index];
node              951 fs/xfs/xfs_da_btree.c 	if (drop_blk->index < (be16_to_cpu(node->hdr.count)-1)) {
node              952 fs/xfs/xfs_da_btree.c 		tmp  = be16_to_cpu(node->hdr.count) - drop_blk->index - 1;
node              956 fs/xfs/xfs_da_btree.c 		    XFS_DA_LOGRANGE(node, btree, tmp));
node              957 fs/xfs/xfs_da_btree.c 		btree = &node->btree[be16_to_cpu(node->hdr.count)-1];
node              961 fs/xfs/xfs_da_btree.c 	    XFS_DA_LOGRANGE(node, btree, sizeof(*btree)));
node              962 fs/xfs/xfs_da_btree.c 	be16_add_cpu(&node->hdr.count, -1);
node              964 fs/xfs/xfs_da_btree.c 	    XFS_DA_LOGRANGE(node, &node->hdr, sizeof(node->hdr)));
node             1053 fs/xfs/xfs_da_btree.c 	xfs_da_intnode_t *node;
node             1091 fs/xfs/xfs_da_btree.c 			node = blk->bp->data;
node             1092 fs/xfs/xfs_da_btree.c 			max = be16_to_cpu(node->hdr.count);
node             1093 fs/xfs/xfs_da_btree.c 			blk->hashval = be32_to_cpu(node->btree[max-1].hashval);
node             1100 fs/xfs/xfs_da_btree.c 			for (btree = &node->btree[probe]; span > 4;
node             1101 fs/xfs/xfs_da_btree.c 				   btree = &node->btree[probe]) {
node             1132 fs/xfs/xfs_da_btree.c 				blkno = be32_to_cpu(node->btree[max-1].before);
node             1305 fs/xfs/xfs_da_btree.c 	xfs_da_intnode_t *node;
node             1307 fs/xfs/xfs_da_btree.c 	node = bp->data;
node             1308 fs/xfs/xfs_da_btree.c 	ASSERT(be16_to_cpu(node->hdr.info.magic) == XFS_DA_NODE_MAGIC);
node             1310 fs/xfs/xfs_da_btree.c 		*count = be16_to_cpu(node->hdr.count);
node             1311 fs/xfs/xfs_da_btree.c 	if (!node->hdr.count)
node             1313 fs/xfs/xfs_da_btree.c 	return be32_to_cpu(node->btree[be16_to_cpu(node->hdr.count)-1].hashval);
node             1403 fs/xfs/xfs_da_btree.c 	xfs_da_intnode_t *node;
node             1420 fs/xfs/xfs_da_btree.c 		node = blk->bp->data;
node             1421 fs/xfs/xfs_da_btree.c 		ASSERT(be16_to_cpu(node->hdr.info.magic) == XFS_DA_NODE_MAGIC);
node             1422 fs/xfs/xfs_da_btree.c 		if (forward && (blk->index < be16_to_cpu(node->hdr.count)-1)) {
node             1424 fs/xfs/xfs_da_btree.c 			blkno = be32_to_cpu(node->btree[blk->index].before);
node             1428 fs/xfs/xfs_da_btree.c 			blkno = be32_to_cpu(node->btree[blk->index].before);
node             1465 fs/xfs/xfs_da_btree.c 			node = (xfs_da_intnode_t *)info;
node             1466 fs/xfs/xfs_da_btree.c 			blk->hashval = be32_to_cpu(node->btree[be16_to_cpu(node->hdr.count)-1].hashval);
node             1470 fs/xfs/xfs_da_btree.c 				blk->index = be16_to_cpu(node->hdr.count)-1;
node             1471 fs/xfs/xfs_da_btree.c 			blkno = be32_to_cpu(node->btree[blk->index].before);
node              265 include/acpi/aclocal.h 	struct acpi_namespace_node *node;
node              388 include/acpi/aclocal.h 	struct acpi_namespace_node *node;
node              498 include/acpi/aclocal.h 	ACPI_STATE_COMMON struct acpi_namespace_node *node;
node              540 include/acpi/aclocal.h 	ACPI_STATE_COMMON struct acpi_namespace_node *node;
node              604 include/acpi/aclocal.h 	struct acpi_namespace_node      *node;          /* For use by interpreter */\
node              134 include/acpi/acobject.h 	struct acpi_namespace_node *node;	/* Link back to parent node */
node              138 include/acpi/acobject.h 	ACPI_OBJECT_COMMON_HEADER struct acpi_namespace_node *node;	/* Link back to parent node */
node              163 include/acpi/acobject.h 	struct acpi_namespace_node *node;	/* Containing namespace node */
node              169 include/acpi/acobject.h 	struct acpi_namespace_node *node;	/* Containing namespace node */
node              241 include/acpi/acobject.h 	struct acpi_namespace_node      *node;              /* Link back to parent node */\
node              284 include/acpi/acobject.h 	ACPI_OBJECT_COMMON_HEADER struct acpi_namespace_node *node;	/* Parent device */
node              293 include/acpi/acobject.h 	struct acpi_namespace_node *node;	/* Parent device */
node              318 include/acpi/acobject.h 	struct acpi_namespace_node *node;
node              417 include/acpi/acobject.h 	struct acpi_namespace_node node;
node              285 include/acpi/acpi_bus.h 	struct list_head node;
node              316 include/acpi/acpi_bus.h 	struct list_head node;
node               38 include/asm-generic/topology.h #define parent_node(node)	((void)(node),0)
node               41 include/asm-generic/topology.h #define node_to_cpumask(node)	((void)node, cpu_online_map)
node               44 include/asm-generic/topology.h #define node_to_first_cpu(node)	((void)(node),0)
node               63 include/asm-generic/topology.h 		cpumask_t _##v = node_to_cpumask(node);			\
node               67 include/asm-generic/topology.h 			  _##v = node_to_cpumask(node)
node               44 include/asm-m32r/mmzone.h 	int node;
node               46 include/asm-m32r/mmzone.h 	for (node = 0 ; node < MAX_NUMNODES ; node++)
node               47 include/asm-m32r/mmzone.h 		if (pfn >= node_start_pfn(node) && pfn <= node_end_pfn(node))
node               50 include/asm-m32r/mmzone.h 	return node;
node              115 include/asm-m68k/dvma.h 	int node;                /* Prom node for this DMA device */
node              277 include/asm-m68k/openprom.h 	int (*no_nextnode)(int node);
node              278 include/asm-m68k/openprom.h 	int (*no_child)(int node);
node              279 include/asm-m68k/openprom.h 	int (*no_proplen)(int node, char *name);
node              280 include/asm-m68k/openprom.h 	int (*no_getprop)(int node, char *name, char *val);
node              281 include/asm-m68k/openprom.h 	int (*no_setprop)(int node, char *name, char *val, int len);
node              282 include/asm-m68k/openprom.h 	char * (*no_nextprop)(int node, char *name);
node               93 include/asm-x86/numaq/apic.h 	int node = apicid_to_node(logical_apicid);
node               96 include/asm-x86/numaq/apic.h 	return physid_mask_of_physid(cpu + 4*node);
node               15 include/asm-x86/pci.h 	int		node;		/* NUMA node */
node              105 include/asm-x86/pci.h 	return sd->node;
node               68 include/asm-x86/topology.h 	return node_to_cpumask_map[node];
node              108 include/asm-x86/topology.h 	return &node_to_cpumask_map[node];
node              114 include/asm-x86/topology.h 	return node_to_cpumask_map[node];
node              121 include/asm-x86/topology.h 		const cpumask_t *v = _node_to_cpumask_ptr(node)
node              124 include/asm-x86/topology.h 			   v = _node_to_cpumask_ptr(node)
node              132 include/asm-x86/topology.h #define parent_node(node) (node)
node              204 include/asm-x86/topology.h 		const cpumask_t *v = _node_to_cpumask_ptr(node)
node              207 include/asm-x86/topology.h 			   v = _node_to_cpumask_ptr(node)
node              216 include/asm-x86/topology.h 	node_to_cpumask_ptr(mask, node);
node              309 include/asm-x86/uv/uv_bau.h 	return constant_test_bit(node, &dstp->bits[0]);
node              313 include/asm-x86/uv/uv_bau.h 	__set_bit(node, &dstp->bits[0]);
node              143 include/linux/acpi.h 	struct list_head	node;
node               35 include/linux/async_tx.h 	struct list_head node;
node               59 include/linux/atmel_tc.h 	struct list_head	node;
node               17 include/linux/attribute_container.h 	struct list_head	node;
node              413 include/linux/blkdev.h 	int			node;
node               13 include/linux/dca.h 	struct list_head	node;
node               27 include/linux/debugobjects.h 	struct hlist_node	node;
node              250 include/linux/device.h 	struct list_head	node;
node              438 include/linux/device.h 	dev->numa_node = node;
node               86 include/linux/dio.h 	struct list_head node;
node              174 include/linux/elevator.h #define rb_entry_rq(node)	rb_entry((node), struct request, rb_node)
node               97 include/linux/enclosure.h 	struct list_head node;
node              809 include/linux/fb.h 	int node;
node               56 include/linux/ftrace.h 	struct hlist_node node;
node               50 include/linux/gameport.h 	struct list_head node;
node               67 include/linux/hidraw.h 	struct list_head node;
node              123 include/linux/hrtimer.h 	struct rb_node			node;
node               75 include/linux/init_task.h 		{ .first = &init_task.pids[PIDTYPE_PID].node },		\
node               76 include/linux/init_task.h 		{ .first = &init_task.pids[PIDTYPE_PGID].node },	\
node               77 include/linux/init_task.h 		{ .first = &init_task.pids[PIDTYPE_SID].node },		\
node               90 include/linux/init_task.h 	.node = {						\
node             1091 include/linux/input.h 	struct list_head	node;
node             1193 include/linux/input.h 	struct list_head	node;
node               50 include/linux/leds.h 	struct list_head	 node;			/* LED Device list */
node              511 include/linux/mm.h 	return zone->node;
node              547 include/linux/mm.h 	page->flags |= (node & NODES_MASK) << NODES_PGSHIFT;
node              560 include/linux/mm.h 	set_page_node(page, node);
node              135 include/linux/mmc/card.h #define mmc_list_to_card(l)	container_of(l, struct mmc_card, node)
node              221 include/linux/mmzone.h 	int node;
node              739 include/linux/mmzone.h 	return zoneref->zone->node;
node               29 include/linux/node.h extern struct node node_devices[];
node               57 include/linux/node.h #define to_node(sys_device) container_of(sys_device, struct node, sysdev)
node               95 include/linux/nodemask.h #define node_set(node, dst) __node_set((node), &(dst))
node               98 include/linux/nodemask.h 	set_bit(node, dstp->bits);
node              101 include/linux/nodemask.h #define node_clear(node, dst) __node_clear((node), &(dst))
node              104 include/linux/nodemask.h 	clear_bit(node, dstp->bits);
node              120 include/linux/nodemask.h #define node_isset(node, nodemask) test_bit((node), (nodemask).bits)
node              123 include/linux/nodemask.h 			__node_test_and_set((node), &(nodemask))
node              126 include/linux/nodemask.h 	return test_and_set_bit(node, addr->bits);
node              246 include/linux/nodemask.h 		m.bits[0] = 1UL<<(node);				\
node              249 include/linux/nodemask.h 		node_set((node), m);					\
node              351 include/linux/nodemask.h 	for ((node) = first_node(mask);			\
node              352 include/linux/nodemask.h 		(node) < MAX_NUMNODES;			\
node              353 include/linux/nodemask.h 		(node) = next_node((node), (mask)))
node              357 include/linux/nodemask.h 		for ((node) = 0; (node) < 1; (node)++)
node              386 include/linux/nodemask.h 	return node_isset(node, node_states[state]);
node              391 include/linux/nodemask.h 	__node_set(node, &node_states[state]);
node              396 include/linux/nodemask.h 	__node_clear(node, &node_states[state]);
node              415 include/linux/nodemask.h 	return node == 0;
node              432 include/linux/nodemask.h 	for ( (node) = 0; (node) == 0; (node) = 1)
node              445 include/linux/nodemask.h 	int node;				\
node              446 include/linux/nodemask.h 	for_each_node_mask(node, (mask))	\
node              447 include/linux/nodemask.h 		if (node_online(node))		\
node              449 include/linux/nodemask.h 	node;					\
node              454 include/linux/nodemask.h #define node_online(node)	node_state((node), N_ONLINE)
node              455 include/linux/nodemask.h #define node_possible(node)	node_state((node), N_POSSIBLE)
node              457 include/linux/nodemask.h #define node_set_online(node)	   node_set_state((node), N_ONLINE)
node              458 include/linux/nodemask.h #define node_set_offline(node)	   node_clear_state((node), N_ONLINE)
node              460 include/linux/nodemask.h #define for_each_node(node)	   for_each_node_state(node, N_POSSIBLE)
node              461 include/linux/nodemask.h #define for_each_online_node(node) for_each_node_state(node, N_ONLINE)
node              283 include/linux/pci.h 	struct list_head node;		/* node in list of buses */
node              312 include/linux/pci.h #define pci_bus_b(n)	list_entry(n, struct pci_bus, node)
node              402 include/linux/pci.h 	struct list_head node;
node               71 include/linux/pid.h 	struct hlist_node node;
node              152 include/linux/pid.h 				&pid->tasks[type], pids[type].node) {
node              119 include/linux/plist.h 	.plist = PLIST_HEAD_INIT((node).plist, NULL),	\
node              144 include/linux/plist.h 	node->prio = prio;
node              145 include/linux/plist.h 	plist_head_init(&node->plist, NULL);
node              206 include/linux/plist.h 	return plist_head_empty(&node->plist);
node               60 include/linux/posix-timers.h 			unsigned int node;
node               91 include/linux/prio_tree.h 	return node->parent == node;
node               96 include/linux/prio_tree.h 	return node->left == node;
node              101 include/linux/prio_tree.h 	return node->right == node;
node              114 include/linux/prio_tree.h 	    (struct prio_tree_node *) (node))
node              116 include/linux/prio_tree.h 	prio_tree_insert(root, (struct prio_tree_node *) (node))
node              118 include/linux/prio_tree.h 	prio_tree_remove(root, (struct prio_tree_node *) (node))
node              136 include/linux/rbtree.h #define RB_EMPTY_NODE(node)	(rb_parent(node) == node)
node              137 include/linux/rbtree.h #define RB_CLEAR_NODE(node)	(rb_set_parent(node, node))
node              155 include/linux/rbtree.h 	node->rb_parent_color = (unsigned long )parent;
node              156 include/linux/rbtree.h 	node->rb_left = node->rb_right = NULL;
node              158 include/linux/rbtree.h 	*rb_link = node;
node              110 include/linux/rfkill.h 	struct list_head node;
node              140 include/linux/rio.h 	struct list_head node;
node              171 include/linux/rio.h 	struct list_head node;	/* node in global list of ports */
node              200 include/linux/rio.h 	struct list_head node;	/* node in list of networks */
node              218 include/linux/rio.h 	struct list_head node;
node              276 include/linux/rio.h 	struct list_head node;
node               52 include/linux/serio.h 	struct list_head node;
node              248 include/linux/slab.h 	__kmalloc_node_track_caller(size, flags, node, \
node              252 include/linux/slab.h 	__kmalloc_node(size, flags, node)
node              288 include/linux/slab.h 	return kmalloc_node(size, flags | __GFP_ZERO, node);
node               88 include/linux/slab_def.h 						flags, node);
node               91 include/linux/slab_def.h 						flags, node);
node               93 include/linux/slab_def.h 	return __kmalloc_node(size, flags, node);
node               15 include/linux/slob_def.h 	return __kmalloc_node(size, flags, node);
node               38 include/linux/slub_def.h 	int node;		/* The node of the page (or -1 for debug) */
node              103 include/linux/slub_def.h 	struct kmem_cache_node *node[MAX_NUMNODES];
node              243 include/linux/slub_def.h 		return kmem_cache_alloc_node(s, flags, node);
node              245 include/linux/slub_def.h 	return __kmalloc_node(size, flags, node);
node               81 include/linux/tc.h 	struct list_head node;		/* Node in list of all TC devices. */
node              105 include/linux/tc.h 	struct list_head node;
node               62 include/linux/thermal.h 	struct list_head node;
node               76 include/linux/thermal.h 	struct list_head node;
node               95 include/linux/thermal.h 	struct list_head node;
node               48 include/linux/tipc.h 	__u32 node;
node               66 include/linux/tipc.h 	return (zone << 24) | (cluster << 12) | node;
node               43 include/linux/topology.h 		node_to_cpumask_ptr(__tmp__, node);	\
node               49 include/linux/topology.h 	for_each_online_node(node)			\
node               50 include/linux/topology.h 		if (nr_cpus_node(node))
node              925 include/linux/usb.h 	struct list_head node;
node              171 include/linux/vmstat.h 	struct zone *zones = NODE_DATA(node)->node_zones;
node              172 include/linux/zorro.h     struct list_head node;
node              164 include/net/ax25.h 	hlist_for_each_entry(__ax25, node, list, uid_node)
node              250 include/net/ax25.h 	hlist_for_each_entry(__ax25, node, list, ax25_node)
node               15 include/net/datalink.h 	struct list_head node;
node               79 include/net/garp.h 	struct rb_node			node;
node               83 include/net/inet_hashtables.h 	struct hlist_node	node;
node               88 include/net/inet_hashtables.h 	hlist_for_each_entry(tb, node, head, node)
node              178 include/net/inet_timewait_sock.h 	hlist_for_each_entry(tw, node, head, tw_node)
node              181 include/net/inet_timewait_sock.h 	hlist_for_each_entry(tw, node, jail, tw_death_node)
node              184 include/net/inet_timewait_sock.h 	hlist_for_each_entry_safe(tw, node, safe, jail, tw_death_node)
node              127 include/net/ip6_fib.h 	struct fib6_node *root, *node;
node               19 include/net/ipx.h 	__u8    node[IPX_NODE_LEN]; 
node               70 include/net/ipx.h 	struct list_head	node; /* node in ipx_interfaces list */
node               78 include/net/ipx.h 	struct list_head	node; /* node in ipx_routes list */
node              102 include/net/ipx.h 	unsigned char		node[IPX_NODE_LEN];
node               82 include/net/lapb.h 	struct list_head	node;
node               55 include/net/llc.h 	struct list_head node;
node              159 include/net/netrom.h 	hlist_for_each_entry(__nr_neigh, node, list, neigh_node)
node              162 include/net/netrom.h 	hlist_for_each_entry_safe(__nr_neigh, node, node2, list, neigh_node)
node              165 include/net/netrom.h 	hlist_for_each_entry(__nr_node, node, list, node_node)
node              168 include/net/netrom.h 	hlist_for_each_entry_safe(__nr_node, node, node2, list, node_node)
node               15 include/net/pkt_cls.h 	int	(*fn)(struct tcf_proto *, unsigned long node, struct tcf_walker *);
node              653 include/net/sctp/sctp.h 	hlist_for_each_entry(epb, node, head, node)
node              103 include/net/sctp/structs.h 	struct hlist_node	node;
node             1259 include/net/sctp/structs.h 	struct hlist_node node;
node              315 include/net/sock.h 	node->pprev = NULL;
node              387 include/net/sock.h 	hlist_for_each_entry(__sk, node, list, sk_node)
node              389 include/net/sock.h 	if (__sk && ({ node = &(__sk)->sk_node; 1; })) \
node              390 include/net/sock.h 		hlist_for_each_entry_from(__sk, node, sk_node)
node              392 include/net/sock.h 	if (__sk && ({ node = &(__sk)->sk_node; 1; })) \
node              393 include/net/sock.h 		hlist_for_each_entry_continue(__sk, node, sk_node)
node              395 include/net/sock.h 	hlist_for_each_entry_safe(__sk, node, tmp, list, sk_node)
node              397 include/net/sock.h 	hlist_for_each_entry(__sk, node, list, sk_bind_node)
node              608 include/net/sock.h 	struct list_head	node;
node              124 include/net/x25.h 	struct list_head	node;		
node              132 include/net/x25.h 	struct list_head	node;
node              165 include/net/x25.h 	struct list_head	node;
node               61 include/scsi/scsi_device.h 	struct list_head	node;
node               53 include/xen/xenbus.h 	const char *node;
node              116 kernel/audit_tree.c 	size = offsetof(struct audit_chunk, owners) + count * sizeof(struct node);
node              323 kernel/audit_tree.c 	struct node *p;
node              433 kernel/audit_tree.c 		struct node *p;
node              436 kernel/audit_tree.c 		p = list_entry(victim->chunks.next, struct node, list);
node              462 kernel/audit_tree.c 		struct node *node = list_entry(p, struct node, list);
node              464 kernel/audit_tree.c 		if (node->index & (1U<<31)) {
node              471 kernel/audit_tree.c 		struct node *node;
node              474 kernel/audit_tree.c 		node = list_entry(tree->chunks.next, struct node, list);
node              477 kernel/audit_tree.c 		if (!(node->index & (1U<<31)))
node              480 kernel/audit_tree.c 		chunk = find_chunk(node);
node              484 kernel/audit_tree.c 		untag_chunk(chunk, node);
node              537 kernel/audit_tree.c 		struct node *node;
node              558 kernel/audit_tree.c 		list_for_each_entry(node, &tree->chunks, list) {
node              559 kernel/audit_tree.c 			struct audit_chunk *chunk = find_chunk(node);
node              562 kernel/audit_tree.c 			node->index |= 1U<<31;
node              565 kernel/audit_tree.c 					node->index &= ~(1U<<31);
node              662 kernel/audit_tree.c 		struct node *node;
node              664 kernel/audit_tree.c 		list_for_each_entry(node, &tree->chunks, list)
node              665 kernel/audit_tree.c 			node->index &= ~(1U<<31);
node              786 kernel/audit_tree.c 			struct node *node;
node              788 kernel/audit_tree.c 			list_for_each_entry(node, &tree->chunks, list)
node              789 kernel/audit_tree.c 				node->index &= ~(1U<<31);
node              328 kernel/cgroup.c 	struct hlist_node *node;
node              347 kernel/cgroup.c 	hlist_for_each_entry(cg, node, hhead, hlist) {
node              645 kernel/cgroup.c 	struct list_head *node;
node              649 kernel/cgroup.c 	node = dentry->d_subdirs.next;
node              650 kernel/cgroup.c 	while (node != &dentry->d_subdirs) {
node              651 kernel/cgroup.c 		struct dentry *d = list_entry(node, struct dentry, d_u.d_child);
node              652 kernel/cgroup.c 		list_del_init(node);
node              664 kernel/cgroup.c 		node = dentry->d_subdirs.next;
node             1041 kernel/cgroup.c 			struct hlist_node *node;
node             1044 kernel/cgroup.c 			hlist_for_each_entry(cg, node, hhead, hlist) {
node             2185 kernel/cpuset.c 	int node;			/* node that zone z is on */
node             2191 kernel/cpuset.c 	node = zone_to_nid(z);
node             2193 kernel/cpuset.c 	if (node_isset(node, current->mems_allowed))
node             2214 kernel/cpuset.c 	allowed = node_isset(node, cs->mems_allowed);
node             2245 kernel/cpuset.c 	int node;			/* node that zone z is on */
node             2249 kernel/cpuset.c 	node = zone_to_nid(z);
node             2250 kernel/cpuset.c 	if (node_isset(node, current->mems_allowed))
node             2316 kernel/cpuset.c 	int node;
node             2318 kernel/cpuset.c 	node = next_node(current->cpuset_mem_spread_rotor, current->mems_allowed);
node             2319 kernel/cpuset.c 	if (node == MAX_NUMNODES)
node             2320 kernel/cpuset.c 		node = first_node(current->mems_allowed);
node             2321 kernel/cpuset.c 	current->cpuset_mem_spread_rotor = node;
node             2322 kernel/cpuset.c 	return node;
node              519 kernel/hrtimer.c 		timer = rb_entry(base->first, struct hrtimer, node);
node              846 kernel/hrtimer.c 		entry = rb_entry(parent, struct hrtimer, node);
node              875 kernel/hrtimer.c 		base->first = &timer->node;
node              878 kernel/hrtimer.c 	rb_link_node(&timer->node, parent, link);
node              879 kernel/hrtimer.c 	rb_insert_color(&timer->node, &base->active);
node              909 kernel/hrtimer.c 		if (base->first == &timer->node) {
node              910 kernel/hrtimer.c 			base->first = rb_next(&timer->node);
node              915 kernel/hrtimer.c 		rb_erase(&timer->node, &base->active);
node             1111 kernel/hrtimer.c 			timer = rb_entry(base->first, struct hrtimer, node);
node             1221 kernel/hrtimer.c 			if (base->first == &timer->node &&
node             1302 kernel/hrtimer.c 		struct rb_node *node;
node             1308 kernel/hrtimer.c 		while ((node = base->first)) {
node             1311 kernel/hrtimer.c 			timer = rb_entry(node, struct hrtimer, node);
node             1392 kernel/hrtimer.c 	struct rb_node *node;
node             1415 kernel/hrtimer.c 		while ((node = base->first)) {
node             1418 kernel/hrtimer.c 			timer = rb_entry(node, struct hrtimer, node);
node             1600 kernel/hrtimer.c 	struct rb_node *node;
node             1603 kernel/hrtimer.c 	while ((node = rb_first(&old_base->active))) {
node             1604 kernel/hrtimer.c 		timer = rb_entry(node, struct hrtimer, node);
node              293 kernel/kprobes.c 	struct hlist_node *node;
node              297 kernel/kprobes.c 	hlist_for_each_entry_rcu(p, node, head, hlist) {
node              438 kernel/kprobes.c 	struct hlist_node *node, *tmp;
node              448 kernel/kprobes.c 	hlist_for_each_entry_safe(ri, node, tmp, head, hlist) {
node              454 kernel/kprobes.c 	hlist_for_each_entry_safe(ri, node, tmp, &empty_rp, hlist) {
node             1150 kernel/kprobes.c 	struct hlist_node *node;
node             1159 kernel/kprobes.c 	hlist_for_each_entry_rcu(p, node, head, hlist) {
node             1194 kernel/kprobes.c 	struct hlist_node *node;
node             1206 kernel/kprobes.c 		hlist_for_each_entry_rcu(p, node, head, hlist)
node             1221 kernel/kprobes.c 	struct hlist_node *node;
node             1235 kernel/kprobes.c 		hlist_for_each_entry_rcu(p, node, head, hlist) {
node              358 kernel/marker.c 	struct hlist_node *node;
node              363 kernel/marker.c 	hlist_for_each_entry(e, node, head, hlist) {
node              377 kernel/marker.c 	struct hlist_node *node;
node              386 kernel/marker.c 	hlist_for_each_entry(e, node, head, hlist) {
node              432 kernel/marker.c 	struct hlist_node *node;
node              439 kernel/marker.c 	hlist_for_each_entry(e, node, head, hlist) {
node              739 kernel/marker.c 	struct hlist_node *node;
node              743 kernel/marker.c 		hlist_for_each_entry(entry, node, head, hlist) {
node              827 kernel/marker.c 	struct hlist_node *node;
node              834 kernel/marker.c 	hlist_for_each_entry(e, node, head, hlist) {
node              322 kernel/pid.c   	hlist_add_head_rcu(&link->node, &pid->tasks[type]);
node              335 kernel/pid.c   	hlist_del_rcu(&link->node);
node              362 kernel/pid.c   	hlist_replace_rcu(&old->pids[type].node, &new->pids[type].node);
node              372 kernel/pid.c   			result = hlist_entry(first, struct task_struct, pids[(type)].node);
node              142 kernel/pm_qos_params.c 	struct requirement_list *node;
node              148 kernel/pm_qos_params.c 	list_for_each_entry(node,
node              151 kernel/pm_qos_params.c 				extreme_value, node->value);
node              254 kernel/pm_qos_params.c 	struct requirement_list *node;
node              258 kernel/pm_qos_params.c 	list_for_each_entry(node,
node              260 kernel/pm_qos_params.c 		if (strcmp(node->name, name) == 0) {
node              262 kernel/pm_qos_params.c 				node->value =
node              265 kernel/pm_qos_params.c 				node->value = new_value;
node              289 kernel/pm_qos_params.c 	struct requirement_list *node;
node              293 kernel/pm_qos_params.c 	list_for_each_entry(node,
node              295 kernel/pm_qos_params.c 		if (strcmp(node->name, name) == 0) {
node              296 kernel/pm_qos_params.c 			kfree(node->name);
node              297 kernel/pm_qos_params.c 			list_del(&node->list);
node              298 kernel/pm_qos_params.c 			kfree(node);
node               73 kernel/power/swsusp.c 	struct rb_node node;
node               88 kernel/power/swsusp.c 		ext = container_of(*new, struct swsusp_extent, node);
node              116 kernel/power/swsusp.c 	rb_link_node(&ext->node, parent, new);
node              117 kernel/power/swsusp.c 	rb_insert_color(&ext->node, &swsusp_extents);
node              148 kernel/power/swsusp.c 	struct rb_node *node;
node              150 kernel/power/swsusp.c 	while ((node = swsusp_extents.rb_node)) {
node              154 kernel/power/swsusp.c 		ext = container_of(node, struct swsusp_extent, node);
node              155 kernel/power/swsusp.c 		rb_erase(node, &swsusp_extents);
node              336 kernel/profile.c 	int node, cpu = (unsigned long)__cpu;
node              342 kernel/profile.c 		node = cpu_to_node(cpu);
node              345 kernel/profile.c 			page = alloc_pages_node(node,
node              353 kernel/profile.c 			page = alloc_pages_node(node,
node              535 kernel/profile.c 		int node = cpu_to_node(cpu);
node              538 kernel/profile.c 		page = alloc_pages_node(node,
node              545 kernel/profile.c 		page = alloc_pages_node(node,
node             2473 kernel/sched.c 	struct hlist_node *node;
node             2475 kernel/sched.c 	hlist_for_each_entry(notifier, node, &curr->preempt_notifiers, link)
node             2484 kernel/sched.c 	struct hlist_node *node;
node             2486 kernel/sched.c 	hlist_for_each_entry(notifier, node, &curr->preempt_notifiers, link)
node             6989 kernel/sched.c 		n = (node + i) % nr_node_ids;
node             6999 kernel/sched.c 		val = node_distance(node, n);
node             7023 kernel/sched.c 	node_to_cpumask_ptr(nodemask, node);
node             7030 kernel/sched.c 	node_set(node, used_nodes);
node             7033 kernel/sched.c 		int next_node = find_next_best_node(node, &used_nodes);
node              230 kernel/time/clockevents.c 	struct list_head *node, *tmp;
node              241 kernel/time/clockevents.c 		list_for_each_safe(node, tmp, &clockevents_released)
node              242 kernel/time/clockevents.c 			list_del(node);
node               98 kernel/time/timer_list.c 		timer = rb_entry(curr, struct hrtimer, node);
node              240 kernel/trace/ftrace.c 	hlist_for_each_entry_rcu(rec, t, head, node) {
node              273 kernel/trace/ftrace.c 	hlist_for_each_entry_rcu(p, t, &ftrace_hash[key], node) {
node              286 kernel/trace/ftrace.c 	hlist_add_head_rcu(&node->node, &ftrace_hash[key]);
node              292 kernel/trace/ftrace.c 	hlist_del(&node->node);
node              337 kernel/trace/ftrace.c 	struct dyn_ftrace *node;
node              379 kernel/trace/ftrace.c 	node = ftrace_alloc_dyn_node(ip);
node              380 kernel/trace/ftrace.c 	if (!node)
node              383 kernel/trace/ftrace.c 	node->ip = ip;
node              385 kernel/trace/ftrace.c 	ftrace_add_hash(node, key);
node              724 kernel/trace/ftrace.c 		hlist_for_each_entry_safe(p, t, n, head, node) {
node              745 kernel/trace/ftrace.c 				INIT_HLIST_NODE(&p->node);
node              746 kernel/trace/ftrace.c 				hlist_add_head(&p->node, &temp_list);
node              766 kernel/trace/ftrace.c 		hlist_for_each_entry_safe(p, t, n, &temp_list, node) {
node              767 kernel/trace/ftrace.c 			hlist_del(&p->node);
node              768 kernel/trace/ftrace.c 			INIT_HLIST_NODE(&p->node);
node              769 kernel/trace/ftrace.c 			hlist_add_head(&p->node, head);
node               86 lib/debugobjects.c 		hlist_add_head(&new->node, &obj_pool);
node               98 lib/debugobjects.c 	struct hlist_node *node;
node              102 lib/debugobjects.c 	hlist_for_each_entry(obj, node, &b->list, node) {
node              124 lib/debugobjects.c 		obj	    = hlist_entry(obj_pool.first, typeof(*obj), node);
node              129 lib/debugobjects.c 		hlist_del(&obj->node);
node              131 lib/debugobjects.c 		hlist_add_head(&obj->node, &b->list);
node              156 lib/debugobjects.c 		hlist_add_head(&obj->node, &obj_pool);
node              175 lib/debugobjects.c 	struct hlist_node *node, *tmp;
node              189 lib/debugobjects.c 		hlist_for_each_entry_safe(obj, node, tmp, &freelist, node) {
node              190 lib/debugobjects.c 			hlist_del(&obj->node);
node              506 lib/debugobjects.c 		hlist_del(&obj->node);
node              519 lib/debugobjects.c 	struct hlist_node *node, *tmp;
node              539 lib/debugobjects.c 		hlist_for_each_entry_safe(obj, node, tmp, &db->list, node) {
node              555 lib/debugobjects.c 				hlist_del(&obj->node);
node              556 lib/debugobjects.c 				hlist_add_head(&obj->node, &freelist);
node              563 lib/debugobjects.c 		hlist_for_each_entry_safe(obj, node, tmp, &freelist, node) {
node              564 lib/debugobjects.c 			hlist_del(&obj->node);
node              873 lib/debugobjects.c 		hlist_add_head(&obj_static_pool[i].node, &obj_pool);
node               79 lib/plist.c    	WARN_ON(!plist_node_empty(node));
node               82 lib/plist.c    		if (node->prio < iter->prio)
node               84 lib/plist.c    		else if (node->prio == iter->prio) {
node               92 lib/plist.c    	list_add_tail(&node->plist.prio_list, &iter->plist.prio_list);
node               94 lib/plist.c    	list_add_tail(&node->plist.node_list, &iter->plist.node_list);
node              109 lib/plist.c    	if (!list_empty(&node->plist.prio_list)) {
node              110 lib/plist.c    		struct plist_node *next = plist_first(&node->plist);
node              112 lib/plist.c    		list_move_tail(&next->plist.prio_list, &node->plist.prio_list);
node              113 lib/plist.c    		list_del_init(&node->plist.prio_list);
node              116 lib/plist.c    	list_del_init(&node->plist.node_list);
node               58 lib/prio_tree.c 		    node, struct vm_area_struct, shared.prio_tree_node);
node               64 lib/prio_tree.c 		*radix = node->start;
node               65 lib/prio_tree.c 		*heap = node->last;
node              123 lib/prio_tree.c 	INIT_PRIO_TREE_NODE(node);
node              126 lib/prio_tree.c 		node->left = first;
node              127 lib/prio_tree.c 		first->parent = node;
node              129 lib/prio_tree.c 		last = node;
node              136 lib/prio_tree.c 	root->prio_tree_node = node;
node              137 lib/prio_tree.c 	return node;
node              146 lib/prio_tree.c 	INIT_PRIO_TREE_NODE(node);
node              154 lib/prio_tree.c 		node->parent = node;
node              155 lib/prio_tree.c 		root->prio_tree_node = node;
node              157 lib/prio_tree.c 		node->parent = old->parent;
node              159 lib/prio_tree.c 			old->parent->left = node;
node              161 lib/prio_tree.c 			old->parent->right = node;
node              165 lib/prio_tree.c 		node->left = old->left;
node              166 lib/prio_tree.c 		old->left->parent = node;
node              170 lib/prio_tree.c 		node->right = old->right;
node              171 lib/prio_tree.c 		old->right->parent = node;
node              190 lib/prio_tree.c 	struct prio_tree_node *cur, *res = node;
node              195 lib/prio_tree.c 	get_index(root, node, &radix_index, &heap_index);
node              199 lib/prio_tree.c 		return prio_tree_expand(root, node, heap_index);
node              212 lib/prio_tree.c 			struct prio_tree_node *tmp = node;
node              213 lib/prio_tree.c 			node = prio_tree_replace(root, cur, node);
node              231 lib/prio_tree.c 				INIT_PRIO_TREE_NODE(node);
node              232 lib/prio_tree.c 				cur->right = node;
node              233 lib/prio_tree.c 				node->parent = cur;
node              239 lib/prio_tree.c 				INIT_PRIO_TREE_NODE(node);
node              240 lib/prio_tree.c 				cur->left = node;
node              241 lib/prio_tree.c 				node->parent = cur;
node              269 lib/prio_tree.c 	cur = node;
node              304 lib/prio_tree.c 	while (cur != node)
node               58 lib/radix-tree.c 	struct radix_tree_node *node;
node               94 lib/radix-tree.c 	__set_bit(offset, node->tags[tag]);
node              100 lib/radix-tree.c 	__clear_bit(offset, node->tags[tag]);
node              106 lib/radix-tree.c 	return test_bit(offset, node->tags[tag]);
node              137 lib/radix-tree.c 		if (node->tags[tag][idx])
node              176 lib/radix-tree.c 	struct radix_tree_node *node =
node              184 lib/radix-tree.c 	tag_clear(node, 0, 0);
node              185 lib/radix-tree.c 	tag_clear(node, 1, 0);
node              186 lib/radix-tree.c 	node->slots[0] = NULL;
node              187 lib/radix-tree.c 	node->count = 0;
node              189 lib/radix-tree.c 	kmem_cache_free(radix_tree_node_cachep, node);
node              195 lib/radix-tree.c 	call_rcu(&node->rcu_head, radix_tree_node_rcu_free);
node              207 lib/radix-tree.c 	struct radix_tree_node *node;
node              214 lib/radix-tree.c 		node = kmem_cache_alloc(radix_tree_node_cachep, gfp_mask);
node              215 lib/radix-tree.c 		if (node == NULL)
node              220 lib/radix-tree.c 			rtp->nodes[rtp->nr++] = node;
node              222 lib/radix-tree.c 			kmem_cache_free(radix_tree_node_cachep, node);
node              244 lib/radix-tree.c 	struct radix_tree_node *node;
node              260 lib/radix-tree.c 		if (!(node = radix_tree_node_alloc(root)))
node              264 lib/radix-tree.c 		node->slots[0] = radix_tree_indirect_to_ptr(root->rnode);
node              269 lib/radix-tree.c 				tag_set(node, tag, 0);
node              273 lib/radix-tree.c 		node->height = newheight;
node              274 lib/radix-tree.c 		node->count = 1;
node              275 lib/radix-tree.c 		node = radix_tree_ptr_to_indirect(node);
node              276 lib/radix-tree.c 		rcu_assign_pointer(root->rnode, node);
node              294 lib/radix-tree.c 	struct radix_tree_node *node = NULL, *slot;
node              320 lib/radix-tree.c 			if (node) {
node              321 lib/radix-tree.c 				rcu_assign_pointer(node->slots[offset], slot);
node              322 lib/radix-tree.c 				node->count++;
node              330 lib/radix-tree.c 		node = slot;
node              331 lib/radix-tree.c 		slot = node->slots[offset];
node              339 lib/radix-tree.c 	if (node) {
node              340 lib/radix-tree.c 		node->count++;
node              341 lib/radix-tree.c 		rcu_assign_pointer(node->slots[offset], item);
node              342 lib/radix-tree.c 		BUG_ON(tag_get(node, 0, offset));
node              343 lib/radix-tree.c 		BUG_ON(tag_get(node, 1, offset));
node              370 lib/radix-tree.c 	struct radix_tree_node *node, **slot;
node              372 lib/radix-tree.c 	node = rcu_dereference(root->rnode);
node              373 lib/radix-tree.c 	if (node == NULL)
node              376 lib/radix-tree.c 	if (!radix_tree_is_indirect_ptr(node)) {
node              381 lib/radix-tree.c 	node = radix_tree_indirect_to_ptr(node);
node              383 lib/radix-tree.c 	height = node->height;
node              391 lib/radix-tree.c 			(node->slots + ((index>>shift) & RADIX_TREE_MAP_MASK));
node              392 lib/radix-tree.c 		node = rcu_dereference(*slot);
node              393 lib/radix-tree.c 		if (node == NULL)
node              419 lib/radix-tree.c 	struct radix_tree_node *node, **slot;
node              421 lib/radix-tree.c 	node = rcu_dereference(root->rnode);
node              422 lib/radix-tree.c 	if (node == NULL)
node              425 lib/radix-tree.c 	if (!radix_tree_is_indirect_ptr(node)) {
node              428 lib/radix-tree.c 		return node;
node              430 lib/radix-tree.c 	node = radix_tree_indirect_to_ptr(node);
node              432 lib/radix-tree.c 	height = node->height;
node              440 lib/radix-tree.c 			(node->slots + ((index>>shift) & RADIX_TREE_MAP_MASK));
node              441 lib/radix-tree.c 		node = rcu_dereference(*slot);
node              442 lib/radix-tree.c 		if (node == NULL)
node              449 lib/radix-tree.c 	return node;
node              528 lib/radix-tree.c 	pathp->node = NULL;
node              539 lib/radix-tree.c 		pathp[1].node = slot;
node              549 lib/radix-tree.c 	while (pathp->node) {
node              550 lib/radix-tree.c 		if (!tag_get(pathp->node, tag, pathp->offset))
node              552 lib/radix-tree.c 		tag_clear(pathp->node, tag, pathp->offset);
node              553 lib/radix-tree.c 		if (any_tag_set(pathp->node, tag))
node              583 lib/radix-tree.c 	struct radix_tree_node *node;
node              590 lib/radix-tree.c 	node = rcu_dereference(root->rnode);
node              591 lib/radix-tree.c 	if (node == NULL)
node              594 lib/radix-tree.c 	if (!radix_tree_is_indirect_ptr(node))
node              596 lib/radix-tree.c 	node = radix_tree_indirect_to_ptr(node);
node              598 lib/radix-tree.c 	height = node->height;
node              607 lib/radix-tree.c 		if (node == NULL)
node              616 lib/radix-tree.c 		if (!tag_get(node, tag, offset))
node              619 lib/radix-tree.c 			int ret = tag_get(node, tag, offset);
node              624 lib/radix-tree.c 		node = rcu_dereference(node->slots[offset]);
node              739 lib/radix-tree.c 	struct radix_tree_node *node;
node              743 lib/radix-tree.c 	node = rcu_dereference(root->rnode);
node              744 lib/radix-tree.c 	if (!node)
node              747 lib/radix-tree.c 	if (!radix_tree_is_indirect_ptr(node)) {
node              750 lib/radix-tree.c 		results[0] = node;
node              753 lib/radix-tree.c 	node = radix_tree_indirect_to_ptr(node);
node              755 lib/radix-tree.c 	max_index = radix_tree_maxindex(node->height);
node              764 lib/radix-tree.c 		slots_found = __lookup(node, (void ***)results + ret, cur_index,
node              807 lib/radix-tree.c 	struct radix_tree_node *node;
node              811 lib/radix-tree.c 	node = rcu_dereference(root->rnode);
node              812 lib/radix-tree.c 	if (!node)
node              815 lib/radix-tree.c 	if (!radix_tree_is_indirect_ptr(node)) {
node              821 lib/radix-tree.c 	node = radix_tree_indirect_to_ptr(node);
node              823 lib/radix-tree.c 	max_index = radix_tree_maxindex(node->height);
node              832 lib/radix-tree.c 		slots_found = __lookup(node, results + ret, cur_index,
node              927 lib/radix-tree.c 	struct radix_tree_node *node;
node              936 lib/radix-tree.c 	node = rcu_dereference(root->rnode);
node              937 lib/radix-tree.c 	if (!node)
node              940 lib/radix-tree.c 	if (!radix_tree_is_indirect_ptr(node)) {
node              943 lib/radix-tree.c 		results[0] = node;
node              946 lib/radix-tree.c 	node = radix_tree_indirect_to_ptr(node);
node              948 lib/radix-tree.c 	max_index = radix_tree_maxindex(node->height);
node              957 lib/radix-tree.c 		slots_found = __lookup_tag(node, (void ***)results + ret,
node              996 lib/radix-tree.c 	struct radix_tree_node *node;
node             1005 lib/radix-tree.c 	node = rcu_dereference(root->rnode);
node             1006 lib/radix-tree.c 	if (!node)
node             1009 lib/radix-tree.c 	if (!radix_tree_is_indirect_ptr(node)) {
node             1015 lib/radix-tree.c 	node = radix_tree_indirect_to_ptr(node);
node             1017 lib/radix-tree.c 	max_index = radix_tree_maxindex(node->height);
node             1026 lib/radix-tree.c 		slots_found = __lookup_tag(node, results + ret,
node             1113 lib/radix-tree.c 	pathp->node = NULL;
node             1122 lib/radix-tree.c 		pathp->node = slot;
node             1135 lib/radix-tree.c 		if (tag_get(pathp->node, tag, pathp->offset))
node             1141 lib/radix-tree.c 	while (pathp->node) {
node             1142 lib/radix-tree.c 		pathp->node->slots[pathp->offset] = NULL;
node             1143 lib/radix-tree.c 		pathp->node->count--;
node             1151 lib/radix-tree.c 		if (pathp->node->count) {
node             1152 lib/radix-tree.c 			if (pathp->node ==
node             1159 lib/radix-tree.c 		to_free = pathp->node;
node             1188 lib/radix-tree.c 	memset(node, 0, sizeof(struct radix_tree_node));
node               28 lib/rbtree.c   	struct rb_node *right = node->rb_right;
node               29 lib/rbtree.c   	struct rb_node *parent = rb_parent(node);
node               31 lib/rbtree.c   	if ((node->rb_right = right->rb_left))
node               32 lib/rbtree.c   		rb_set_parent(right->rb_left, node);
node               33 lib/rbtree.c   	right->rb_left = node;
node               39 lib/rbtree.c   		if (node == parent->rb_left)
node               46 lib/rbtree.c   	rb_set_parent(node, right);
node               51 lib/rbtree.c   	struct rb_node *left = node->rb_left;
node               52 lib/rbtree.c   	struct rb_node *parent = rb_parent(node);
node               54 lib/rbtree.c   	if ((node->rb_left = left->rb_right))
node               55 lib/rbtree.c   		rb_set_parent(left->rb_right, node);
node               56 lib/rbtree.c   	left->rb_right = node;
node               62 lib/rbtree.c   		if (node == parent->rb_right)
node               69 lib/rbtree.c   	rb_set_parent(node, left);
node               76 lib/rbtree.c   	while ((parent = rb_parent(node)) && rb_is_red(parent))
node               89 lib/rbtree.c   					node = gparent;
node               94 lib/rbtree.c   			if (parent->rb_right == node)
node               99 lib/rbtree.c   				parent = node;
node              100 lib/rbtree.c   				node = tmp;
node              114 lib/rbtree.c   					node = gparent;
node              119 lib/rbtree.c   			if (parent->rb_left == node)
node              124 lib/rbtree.c   				parent = node;
node              125 lib/rbtree.c   				node = tmp;
node              143 lib/rbtree.c   	while ((!node || rb_is_black(node)) && node != root->rb_node)
node              145 lib/rbtree.c   		if (parent->rb_left == node)
node              159 lib/rbtree.c   				node = parent;
node              160 lib/rbtree.c   				parent = rb_parent(node);
node              178 lib/rbtree.c   				node = root->rb_node;
node              196 lib/rbtree.c   				node = parent;
node              197 lib/rbtree.c   				parent = rb_parent(node);
node              215 lib/rbtree.c   				node = root->rb_node;
node              220 lib/rbtree.c   	if (node)
node              221 lib/rbtree.c   		rb_set_black(node);
node              229 lib/rbtree.c   	if (!node->rb_left)
node              230 lib/rbtree.c   		child = node->rb_right;
node              231 lib/rbtree.c   	else if (!node->rb_right)
node              232 lib/rbtree.c   		child = node->rb_left;
node              235 lib/rbtree.c   		struct rb_node *old = node, *left;
node              237 lib/rbtree.c   		node = node->rb_right;
node              238 lib/rbtree.c   		while ((left = node->rb_left) != NULL)
node              239 lib/rbtree.c   			node = left;
node              240 lib/rbtree.c   		child = node->rb_right;
node              241 lib/rbtree.c   		parent = rb_parent(node);
node              242 lib/rbtree.c   		color = rb_color(node);
node              248 lib/rbtree.c   			parent = node;
node              252 lib/rbtree.c   		node->rb_parent_color = old->rb_parent_color;
node              253 lib/rbtree.c   		node->rb_right = old->rb_right;
node              254 lib/rbtree.c   		node->rb_left = old->rb_left;
node              259 lib/rbtree.c   				rb_parent(old)->rb_left = node;
node              261 lib/rbtree.c   				rb_parent(old)->rb_right = node;
node              263 lib/rbtree.c   			root->rb_node = node;
node              265 lib/rbtree.c   		rb_set_parent(old->rb_left, node);
node              267 lib/rbtree.c   			rb_set_parent(old->rb_right, node);
node              271 lib/rbtree.c   	parent = rb_parent(node);
node              272 lib/rbtree.c   	color = rb_color(node);
node              278 lib/rbtree.c   		if (parent->rb_left == node)
node              325 lib/rbtree.c   	if (rb_parent(node) == node)
node              330 lib/rbtree.c   	if (node->rb_right) {
node              331 lib/rbtree.c   		node = node->rb_right; 
node              332 lib/rbtree.c   		while (node->rb_left)
node              333 lib/rbtree.c   			node=node->rb_left;
node              334 lib/rbtree.c   		return node;
node              343 lib/rbtree.c   	while ((parent = rb_parent(node)) && node == parent->rb_right)
node              344 lib/rbtree.c   		node = parent;
node              354 lib/rbtree.c   	if (rb_parent(node) == node)
node              359 lib/rbtree.c   	if (node->rb_left) {
node              360 lib/rbtree.c   		node = node->rb_left; 
node              361 lib/rbtree.c   		while (node->rb_right)
node              362 lib/rbtree.c   			node=node->rb_right;
node              363 lib/rbtree.c   		return node;
node              368 lib/rbtree.c   	while ((parent = rb_parent(node)) && node == parent->rb_left)
node              369 lib/rbtree.c   		node = parent;
node              548 lib/zlib_deflate/deftree.c     int node;          /* new node being created */
node              571 lib/zlib_deflate/deftree.c         node = s->heap[++(s->heap_len)] = (max_code < 2 ? ++max_code : 0);
node              572 lib/zlib_deflate/deftree.c         tree[node].Freq = 1;
node              573 lib/zlib_deflate/deftree.c         s->depth[node] = 0;
node              574 lib/zlib_deflate/deftree.c         s->opt_len--; if (stree) s->static_len -= stree[node].Len;
node              587 lib/zlib_deflate/deftree.c     node = elems;              /* next internal node of the tree */
node              596 lib/zlib_deflate/deftree.c         tree[node].Freq = tree[n].Freq + tree[m].Freq;
node              597 lib/zlib_deflate/deftree.c         s->depth[node] = (uch) (max(s->depth[n], s->depth[m]) + 1);
node              598 lib/zlib_deflate/deftree.c         tree[n].Dad = tree[m].Dad = (ush)node;
node              602 lib/zlib_deflate/deftree.c                     node, tree[node].Freq, n, tree[n].Freq, m, tree[m].Freq);
node              606 lib/zlib_deflate/deftree.c         s->heap[SMALLEST] = node++;
node               58 mm/allocpercpu.c 	int node = cpu_to_node(cpu);
node               66 mm/allocpercpu.c 	if (node_online(node))
node               67 mm/allocpercpu.c 		pdata->ptrs[cpu] = kmalloc_node(size, gfp|__GFP_ZERO, node);
node             1391 mm/hugetlb.c   	int node;
node             1394 mm/hugetlb.c   	for_each_node_mask(node, cpuset_current_mems_allowed)
node             1395 mm/hugetlb.c   		nr += array[node];
node              911 mm/memcontrol.c 	int node, zid;
node              922 mm/memcontrol.c 		for_each_node_state(node, N_POSSIBLE)
node              925 mm/memcontrol.c 				mz = mem_cgroup_zoneinfo(mem, node, zid);
node             1065 mm/memcontrol.c 	int zone, tmp = node;
node             1074 mm/memcontrol.c 	if (!node_state(node, N_NORMAL_MEMORY))
node             1080 mm/memcontrol.c 	mem->info.nodeinfo[node] = pn;
node             1094 mm/memcontrol.c 	kfree(mem->info.nodeinfo[node]);
node             1124 mm/memcontrol.c 	int node;
node             1137 mm/memcontrol.c 	for_each_node_state(node, N_POSSIBLE)
node             1138 mm/memcontrol.c 		if (alloc_mem_cgroup_per_zone_info(mem, node))
node             1143 mm/memcontrol.c 	for_each_node_state(node, N_POSSIBLE)
node             1144 mm/memcontrol.c 		free_mem_cgroup_per_zone_info(mem, node);
node             1160 mm/memcontrol.c 	int node;
node             1163 mm/memcontrol.c 	for_each_node_state(node, N_POSSIBLE)
node             1164 mm/memcontrol.c 		free_mem_cgroup_per_zone_info(mem, node);
node              129 mm/memory_hotplug.c 	int node = pgdat->node_id;
node              137 mm/memory_hotplug.c 		get_page_bootmem(node, page, NODE_INFO);
node              148 mm/memory_hotplug.c 				get_page_bootmem(node, page, NODE_INFO);
node              743 mm/memory_hotplug.c 	int ret, drain, retry_max, node;
node              759 mm/memory_hotplug.c 	node = zone_to_nid(zone);
node              770 mm/memory_hotplug.c 	if (nr_pages >= node_present_pages(node))
node              771 mm/memory_hotplug.c 		arg.status_change_nid = node;
node              292 mm/mempolicy.c 		int node = first_node(pol->w.user_nodemask);
node              294 mm/mempolicy.c 		if (node_isset(node, *nodes)) {
node              295 mm/mempolicy.c 			pol->v.preferred_node = node;
node              771 mm/mempolicy.c 	return alloc_pages_node(node, GFP_HIGHUSER_MOVABLE, 0);
node             1408 mm/mempolicy.c 		return zone->node;
node             2185 mm/mempolicy.c 	unsigned long node[MAX_NUMNODES];
node             2212 mm/mempolicy.c 	md->node[page_to_nid(page)]++;
node             2322 mm/mempolicy.c 		if (md->node[n])
node             2323 mm/mempolicy.c 			seq_printf(m, " N%d=%lu", n, md->node[n]);
node              836 mm/migrate.c   	int node;
node              845 mm/migrate.c   	while (pm->node != MAX_NUMNODES && pm->page != p)
node              848 mm/migrate.c   	if (pm->node == MAX_NUMNODES)
node              853 mm/migrate.c   	return alloc_pages_node(pm->node,
node              875 mm/migrate.c   	for (pp = pm; pp->node != MAX_NUMNODES; pp++) {
node              906 mm/migrate.c   		if (err == pp->node)
node              948 mm/migrate.c   	for ( ; pm->node != MAX_NUMNODES; pm++) {
node             1059 mm/migrate.c   			int node;
node             1061 mm/migrate.c   			if (get_user(node, nodes + i))
node             1065 mm/migrate.c   			if (!node_state(node, N_HIGH_MEMORY))
node             1069 mm/migrate.c   			if (!node_isset(node, task_nodes))
node             1072 mm/migrate.c   			pm[i].node = node;
node             1074 mm/migrate.c   			pm[i].node = 0;	/* anything to not match MAX_NUMNODES */
node             1077 mm/migrate.c   	pm[nr_pages].node = MAX_NUMNODES;
node               56 mm/mm_init.c   					zone->node, zone->name);
node              325 mm/page-writeback.c 	int node;
node              328 mm/page-writeback.c 	for_each_node_state(node, N_HIGH_MEMORY) {
node              330 mm/page-writeback.c 			&NODE_DATA(node)->node_zones[ZONE_HIGHMEM];
node             2089 mm/page_alloc.c 	if (!node_isset(node, *used_node_mask)) {
node             2090 mm/page_alloc.c 		node_set(node, *used_node_mask);
node             2091 mm/page_alloc.c 		return node;
node             2101 mm/page_alloc.c 		val = node_distance(node, n);
node             2104 mm/page_alloc.c 		val += (n < node);
node             2141 mm/page_alloc.c 	j = build_zonelists_node(NODE_DATA(node), zonelist, j,
node             2171 mm/page_alloc.c 	int pos, j, node;
node             2180 mm/page_alloc.c 			node = node_order[j];
node             2181 mm/page_alloc.c 			z = &NODE_DATA(node)->node_zones[zone_type];
node             2257 mm/page_alloc.c 	int j, node, load;
node             2281 mm/page_alloc.c 	while ((node = find_next_best_node(local_node, &used_mask)) >= 0) {
node             2282 mm/page_alloc.c 		int distance = node_distance(local_node, node);
node             2297 mm/page_alloc.c 			node_load[node] = load;
node             2299 mm/page_alloc.c 		prev_node = node;
node             2302 mm/page_alloc.c 			build_zonelists_in_node_order(pgdat, node);
node             2304 mm/page_alloc.c 			node_order[j++] = node;	/* remember order */
node             2339 mm/page_alloc.c 	int node, local_node;
node             2356 mm/page_alloc.c 	for (node = local_node + 1; node < MAX_NUMNODES; node++) {
node             2357 mm/page_alloc.c 		if (!node_online(node))
node             2359 mm/page_alloc.c 		j = build_zonelists_node(NODE_DATA(node), zonelist, j,
node             2362 mm/page_alloc.c 	for (node = 0; node < local_node; node++) {
node             2363 mm/page_alloc.c 		if (!node_online(node))
node             2365 mm/page_alloc.c 		j = build_zonelists_node(NODE_DATA(node), zonelist, j,
node             2724 mm/page_alloc.c 	int node = cpu_to_node(cpu);
node             2726 mm/page_alloc.c 	node_set_state(node, N_CPU);	/* this node has a cpu */
node             2734 mm/page_alloc.c 					 GFP_KERNEL, node);
node             3454 mm/page_alloc.c 		zone->node = nid;
node             3555 mm/page_alloc.c 	unsigned int node;
node             3558 mm/page_alloc.c 	for_each_node_mask(node, node_possible_map)
node             3559 mm/page_alloc.c 		highest = node;
node              116 mm/prio_tree.c 	struct vm_area_struct *node, *head, *new_head;
node              144 mm/prio_tree.c 			node = vma->shared.vm_set.head;
node              151 mm/prio_tree.c 				node->shared.vm_set.head = new_head;
node              152 mm/prio_tree.c 				new_head->shared.vm_set.head = node;
node              154 mm/prio_tree.c 				node->shared.vm_set.head = NULL;
node               29 mm/quicklist.c 	int node = numa_node_id();
node               30 mm/quicklist.c 	struct zone *zones = NODE_DATA(node)->node_zones;
node               32 mm/quicklist.c 	node_to_cpumask_ptr(cpumask_on_node, node);
node              902 mm/slab.c      	int node;
node              904 mm/slab.c      	node = next_node(cpu_to_node(cpu), node_online_map);
node              905 mm/slab.c      	if (node == MAX_NUMNODES)
node              906 mm/slab.c      		node = first_node(node_online_map);
node              908 mm/slab.c      	per_cpu(reap_node, cpu) = node;
node              913 mm/slab.c      	int node = __get_cpu_var(reap_node);
node              915 mm/slab.c      	node = next_node(node, node_online_map);
node              916 mm/slab.c      	if (unlikely(node >= MAX_NUMNODES))
node              917 mm/slab.c      		node = first_node(node_online_map);
node              918 mm/slab.c      	__get_cpu_var(reap_node) = node;
node              956 mm/slab.c      	nc = kmalloc_node(memsize, GFP_KERNEL, node);
node             1035 mm/slab.c      	ac_ptr = kmalloc_node(memsize, GFP_KERNEL, node);
node             1038 mm/slab.c      			if (i == node || !node_online(i)) {
node             1042 mm/slab.c      			ac_ptr[i] = alloc_arraycache(node, limit, 0xbaadf00d);
node             1068 mm/slab.c      	struct kmem_list3 *rl3 = cachep->nodelists[node];
node             1080 mm/slab.c      		free_block(cachep, ac->entry, ac->avail, node);
node             1091 mm/slab.c      	int node = __get_cpu_var(reap_node);
node             1094 mm/slab.c      		struct array_cache *ac = l3->alien[node];
node             1097 mm/slab.c      			__drain_alien_cache(cachep, ac, node);
node             1126 mm/slab.c      	int node;
node             1128 mm/slab.c      	node = numa_node_id();
node             1134 mm/slab.c      	if (likely(slabp->nodeid == node))
node             1137 mm/slab.c      	l3 = cachep->nodelists[node];
node             1161 mm/slab.c      	int node = cpu_to_node(cpu);
node             1162 mm/slab.c      	node_to_cpumask_ptr(mask, node);
node             1172 mm/slab.c      		l3 = cachep->nodelists[node];
node             1182 mm/slab.c      			free_block(cachep, nc->entry, nc->avail, node);
node             1192 mm/slab.c      				   shared->avail, node);
node             1215 mm/slab.c      		l3 = cachep->nodelists[node];
node             1226 mm/slab.c      	int node = cpu_to_node(cpu);
node             1242 mm/slab.c      		if (!cachep->nodelists[node]) {
node             1243 mm/slab.c      			l3 = kmalloc_node(memsize, GFP_KERNEL, node);
node             1255 mm/slab.c      			cachep->nodelists[node] = l3;
node             1258 mm/slab.c      		spin_lock_irq(&cachep->nodelists[node]->list_lock);
node             1259 mm/slab.c      		cachep->nodelists[node]->free_limit =
node             1260 mm/slab.c      			(1 + nr_cpus_node(node)) *
node             1262 mm/slab.c      		spin_unlock_irq(&cachep->nodelists[node]->list_lock);
node             1274 mm/slab.c      		nc = alloc_arraycache(node, cachep->limit,
node             1279 mm/slab.c      			shared = alloc_arraycache(node,
node             1288 mm/slab.c      			alien = alloc_alien_cache(node, cachep->limit);
node             1296 mm/slab.c      		l3 = cachep->nodelists[node];
node             1413 mm/slab.c      	int node;
node             1415 mm/slab.c      	for_each_online_node(node) {
node             1416 mm/slab.c      		cachep->nodelists[node] = &initkmem_list3[index + node];
node             1417 mm/slab.c      		cachep->nodelists[node]->next_reap = jiffies +
node             1434 mm/slab.c      	int node;
node             1475 mm/slab.c      	node = numa_node_id();
node             1482 mm/slab.c      	cache_cache.nodelists[node] = &initkmem_list3[CACHE_CACHE + node];
node             2088 mm/slab.c      			int node;
node             2089 mm/slab.c      			for_each_online_node(node) {
node             2090 mm/slab.c      				cachep->nodelists[node] =
node             2092 mm/slab.c      						GFP_KERNEL, node);
node             2093 mm/slab.c      				BUG_ON(!cachep->nodelists[node]);
node             2094 mm/slab.c      				kmem_list3_init(cachep->nodelists[node]);
node             2414 mm/slab.c      	assert_spin_locked(&cachep->nodelists[node]->list_lock);
node             2433 mm/slab.c      	int node = numa_node_id();
node             2437 mm/slab.c      	spin_lock(&cachep->nodelists[node]->list_lock);
node             2438 mm/slab.c      	free_block(cachep, ac->entry, ac->avail, node);
node             2439 mm/slab.c      	spin_unlock(&cachep->nodelists[node]->list_lock);
node             2446 mm/slab.c      	int node;
node             2450 mm/slab.c      	for_each_online_node(node) {
node             2451 mm/slab.c      		l3 = cachep->nodelists[node];
node             2456 mm/slab.c      	for_each_online_node(node) {
node             2457 mm/slab.c      		l3 = cachep->nodelists[node];
node             2459 mm/slab.c      			drain_array(cachep, l3, l3->shared, 1, node);
node             2954 mm/slab.c      	int node;
node             2958 mm/slab.c      	node = numa_node_id();
node             2969 mm/slab.c      	l3 = cachep->nodelists[node];
node             3007 mm/slab.c      							    node);
node             3026 mm/slab.c      		x = cache_grow(cachep, flags | GFP_THISNODE, node, NULL);
node             3489 mm/slab.c      		l3 = cachep->nodelists[node];
node             3491 mm/slab.c      		check_spinlock_acquired_node(cachep, node);
node             3493 mm/slab.c      		slab_put_obj(cachep, slabp, objp, node);
node             3526 mm/slab.c      	int node = numa_node_id();
node             3533 mm/slab.c      	l3 = cachep->nodelists[node];
node             3548 mm/slab.c      	free_block(cachep, ac->entry, batchcount, node);
node             3676 mm/slab.c      	return kmem_cache_alloc_node(cachep, flags, node);
node             3682 mm/slab.c      	return __do_kmalloc_node(size, flags, node,
node             3690 mm/slab.c      	return __do_kmalloc_node(size, flags, node, caller);
node             3696 mm/slab.c      	return __do_kmalloc_node(size, flags, node, NULL);
node             3810 mm/slab.c      	int node;
node             3815 mm/slab.c      	for_each_online_node(node) {
node             3818 mm/slab.c                              new_alien = alloc_alien_cache(node, cachep->limit);
node             3825 mm/slab.c      			new_shared = alloc_arraycache(node,
node             3834 mm/slab.c      		l3 = cachep->nodelists[node];
node             3842 mm/slab.c      						shared->avail, node);
node             3849 mm/slab.c      			l3->free_limit = (1 + nr_cpus_node(node)) *
node             3856 mm/slab.c      		l3 = kmalloc_node(sizeof(struct kmem_list3), GFP_KERNEL, node);
node             3868 mm/slab.c      		l3->free_limit = (1 + nr_cpus_node(node)) *
node             3870 mm/slab.c      		cachep->nodelists[node] = l3;
node             3877 mm/slab.c      		node--;
node             3878 mm/slab.c      		while (node >= 0) {
node             3879 mm/slab.c      			if (cachep->nodelists[node]) {
node             3880 mm/slab.c      				l3 = cachep->nodelists[node];
node             3885 mm/slab.c      				cachep->nodelists[node] = NULL;
node             3887 mm/slab.c      			node--;
node             4027 mm/slab.c      			free_block(cachep, ac->entry, tofree, node);
node             4052 mm/slab.c      	int node = numa_node_id();
node             4068 mm/slab.c      		l3 = searchp->nodelists[node];
node             4072 mm/slab.c      		drain_array(searchp, l3, cpu_cache_get(searchp), 0, node);
node             4083 mm/slab.c      		drain_array(searchp, l3, l3->shared, 0, node);
node             4161 mm/slab.c      	int node;
node             4166 mm/slab.c      	for_each_online_node(node) {
node             4167 mm/slab.c      		l3 = cachep->nodelists[node];
node             4393 mm/slab.c      	int node;
node             4405 mm/slab.c      	for_each_online_node(node) {
node             4406 mm/slab.c      		l3 = cachep->nodelists[node];
node              238 mm/slob.c      	if (node != -1)
node              239 mm/slob.c      		page = alloc_pages_node(node, gfp, order);
node              327 mm/slob.c      		if (node != -1 && page_to_nid(&sp->page) != node)
node              352 mm/slob.c      		b = slob_new_page(gfp & ~__GFP_ZERO, 0, node);
node              471 mm/slob.c      		m = slob_alloc(size + align, gfp, align, node);
node              479 mm/slob.c      		ret = slob_new_page(gfp | __GFP_COMP, get_order(size), node);
node              573 mm/slob.c      		b = slob_alloc(c->size, flags, c->align, node);
node              575 mm/slob.c      		b = slob_new_page(flags, get_order(c->size), node);
node              223 mm/slub.c      	return s->node[node];
node              828 mm/slub.c      	struct kmem_cache_node *n = get_node(s, node);
node              835 mm/slub.c      	struct kmem_cache_node *n = get_node(s, node);
node              850 mm/slub.c      	struct kmem_cache_node *n = get_node(s, node);
node             1065 mm/slub.c      	if (node == -1)
node             1068 mm/slub.c      		return alloc_pages_node(node, flags, order);
node             1078 mm/slub.c      	page = alloc_slab_page(flags | __GFP_NOWARN | __GFP_NORETRY, node,
node             1086 mm/slub.c      		page = alloc_slab_page(flags, node, oo);
node             1119 mm/slub.c      		flags & (GFP_RECLAIM_MASK | GFP_CONSTRAINT_MASK), node);
node             1348 mm/slub.c      	int searchnode = (node == -1) ? numa_node_id() : node;
node             1477 mm/slub.c      	if (node != -1 && c->node != node)
node             1514 mm/slub.c      	if (unlikely(!node_match(c, node)))
node             1529 mm/slub.c      	c->node = page_to_nid(c->page);
node             1539 mm/slub.c      	new = get_partial(s, gfpflags, node);
node             1549 mm/slub.c      	new = new_slab(s, gfpflags, node);
node             1571 mm/slub.c      	c->node = -1;
node             1596 mm/slub.c      	if (unlikely(!c->freelist || !node_match(c, node)))
node             1598 mm/slub.c      		object = __slab_alloc(s, gfpflags, node, addr, c);
node             1622 mm/slub.c      	return slab_alloc(s, gfpflags, node, __builtin_return_address(0));
node             1717 mm/slub.c      	if (likely(page == c->page && c->node >= 0)) {
node             1908 mm/slub.c      	c->node = 0;
node             2084 mm/slub.c      	page = new_slab(kmalloc_caches, gfpflags, node);
node             2087 mm/slub.c      	if (page_to_nid(page) != node) {
node             2089 mm/slub.c      				"node %d\n", node);
node             2098 mm/slub.c      	kmalloc_caches->node[node] = n;
node             2104 mm/slub.c      	inc_slabs_node(kmalloc_caches, node, page->objects);
node             2119 mm/slub.c      	int node;
node             2121 mm/slub.c      	for_each_node_state(node, N_NORMAL_MEMORY) {
node             2122 mm/slub.c      		struct kmem_cache_node *n = s->node[node];
node             2125 mm/slub.c      		s->node[node] = NULL;
node             2131 mm/slub.c      	int node;
node             2139 mm/slub.c      	for_each_node_state(node, N_NORMAL_MEMORY) {
node             2142 mm/slub.c      		if (local_node == node)
node             2147 mm/slub.c      								node);
node             2151 mm/slub.c      							gfpflags, node);
node             2159 mm/slub.c      		s->node[node] = n;
node             2427 mm/slub.c      	int node;
node             2433 mm/slub.c      	for_each_node_state(node, N_NORMAL_MEMORY) {
node             2434 mm/slub.c      		struct kmem_cache_node *n = get_node(s, node);
node             2437 mm/slub.c      		if (n->nr_partial || slabs_node(s, node))
node             2667 mm/slub.c      	struct page *page = alloc_pages_node(node, flags | __GFP_COMP,
node             2682 mm/slub.c      		return kmalloc_large_node(size, flags, node);
node             2689 mm/slub.c      	return slab_alloc(s, flags, node, __builtin_return_address(0));
node             2762 mm/slub.c      	int node;
node             2776 mm/slub.c      	for_each_node_state(node, N_NORMAL_MEMORY) {
node             2777 mm/slub.c      		n = get_node(s, node);
node             2866 mm/slub.c      			s->node[offline_node] = NULL;
node             2906 mm/slub.c      		s->node[nid] = n;
node             3223 mm/slub.c      		return kmalloc_large_node(size, gfpflags, node);
node             3230 mm/slub.c      	return slab_alloc(s, gfpflags, node, caller);
node             3346 mm/slub.c      	int node;
node             3355 mm/slub.c      	for_each_node_state(node, N_NORMAL_MEMORY) {
node             3356 mm/slub.c      		struct kmem_cache_node *n = get_node(s, node);
node             3569 mm/slub.c      	int node;
node             3578 mm/slub.c      	for_each_node_state(node, N_NORMAL_MEMORY) {
node             3579 mm/slub.c      		struct kmem_cache_node *n = get_node(s, node);
node             3663 mm/slub.c      	int node;
node             3679 mm/slub.c      			if (!c || c->node < 0)
node             3691 mm/slub.c      				nodes[c->node] += x;
node             3693 mm/slub.c      			per_cpu[c->node]++;
node             3698 mm/slub.c      		for_each_node_state(node, N_NORMAL_MEMORY) {
node             3699 mm/slub.c      			struct kmem_cache_node *n = get_node(s, node);
node             3710 mm/slub.c      			nodes[node] += x;
node             3714 mm/slub.c      		for_each_node_state(node, N_NORMAL_MEMORY) {
node             3715 mm/slub.c      			struct kmem_cache_node *n = get_node(s, node);
node             3724 mm/slub.c      			nodes[node] += x;
node             3729 mm/slub.c      	for_each_node_state(node, N_NORMAL_MEMORY)
node             3730 mm/slub.c      		if (nodes[node])
node             3732 mm/slub.c      					node, nodes[node]);
node             3740 mm/slub.c      	int node;
node             3742 mm/slub.c      	for_each_online_node(node) {
node             3743 mm/slub.c      		struct kmem_cache_node *n = get_node(s, node);
node             4467 mm/slub.c      	int node;
node             4471 mm/slub.c      	for_each_online_node(node) {
node             4472 mm/slub.c      		struct kmem_cache_node *n = get_node(s, node);
node               43 mm/sparse-vmemmap.c 	return __alloc_bootmem_node(NODE_DATA(node), size, align, goal);
node               51 mm/sparse-vmemmap.c 		struct page *page = alloc_pages_node(node,
node               57 mm/sparse-vmemmap.c 		return __earlyonly_bootmem_alloc(node, size, size,
node               67 mm/sparse-vmemmap.c 	if (actual_node != node)
node               77 mm/sparse-vmemmap.c 		void *p = vmemmap_alloc_block(PAGE_SIZE, node);
node               90 mm/sparse-vmemmap.c 		void *p = vmemmap_alloc_block(PAGE_SIZE, node);
node              102 mm/sparse-vmemmap.c 		void *p = vmemmap_alloc_block(PAGE_SIZE, node);
node              114 mm/sparse-vmemmap.c 		void *p = vmemmap_alloc_block(PAGE_SIZE, node);
node              133 mm/sparse-vmemmap.c 		pgd = vmemmap_pgd_populate(addr, node);
node              136 mm/sparse-vmemmap.c 		pud = vmemmap_pud_populate(pgd, addr, node);
node              139 mm/sparse-vmemmap.c 		pmd = vmemmap_pmd_populate(pud, addr, node);
node              142 mm/sparse-vmemmap.c 		pte = vmemmap_pte_populate(pmd, addr, node);
node              145 mm/sparse-vmemmap.c 		vmemmap_verify(pte, node, addr, addr + PAGE_SIZE);
node              240 mm/vmalloc.c   	area = kmalloc_node(sizeof(*area), gfp_mask & GFP_RECLAIM_MASK, node);
node              327 mm/vmalloc.c   	return __get_vm_area_node(size, flags, VMALLOC_START, VMALLOC_END, node,
node              503 mm/vmalloc.c   				PAGE_KERNEL, node, caller);
node              508 mm/vmalloc.c   				node);
node              521 mm/vmalloc.c   		if (node < 0)
node              524 mm/vmalloc.c   			page = alloc_pages_node(node, gfp_mask, 0);
node              571 mm/vmalloc.c   						node, gfp_mask, caller);
node              576 mm/vmalloc.c   	return __vmalloc_area_node(area, gfp_mask, prot, node, caller);
node              639 mm/vmalloc.c   					node, __builtin_return_address(0));
node              379 mm/vmstat.c    	if (z->node == numa_node_id())
node              401 mm/vmstat.c    	loff_t node = *pos;
node              403 mm/vmstat.c    	     pgdat && node;
node              405 mm/vmstat.c    		--node;
node              156 net/802/garp.c 		attr = rb_entry(parent, struct garp_attr, node);
node              176 net/802/garp.c 		attr = rb_entry(parent, struct garp_attr, node);
node              183 net/802/garp.c 	rb_link_node(&new->node, parent, p);
node              184 net/802/garp.c 	rb_insert_color(&new->node, &app->gid);
node              205 net/802/garp.c 	rb_erase(&attr->node, &app->gid);
node              385 net/802/garp.c 	struct rb_node *node, *next;
node              388 net/802/garp.c 	for (node = rb_first(&app->gid);
node              389 net/802/garp.c 	     next = node ? rb_next(node) : NULL, node != NULL;
node              390 net/802/garp.c 	     node = next) {
node              391 net/802/garp.c 		attr = rb_entry(node, struct garp_attr, node);
node               36 net/802/psnap.c 	list_for_each_entry_rcu(p, &snap_list, node) {
node              143 net/802/psnap.c 		list_add_rcu(&proto->node, &snap_list);
node              158 net/802/psnap.c 	list_del_rcu(&proto->node);
node              150 net/appletalk/atalk_proc.c 	struct hlist_node *node;
node              152 net/appletalk/atalk_proc.c 	sk_for_each(s, node, &atalk_sockets)
node               94 net/appletalk/ddp.c 	struct hlist_node *node;
node               97 net/appletalk/ddp.c 	sk_for_each(s, node, &atalk_sockets) {
node              142 net/appletalk/ddp.c 	struct hlist_node *node;
node              146 net/appletalk/ddp.c 	sk_for_each(s, node, &atalk_sockets) {
node              390 net/appletalk/ddp.c 	if (node != ATADDR_BCAST &&
node              391 net/appletalk/ddp.c 	    iface->address.s_node != node &&
node              392 net/appletalk/ddp.c 	    node != ATADDR_ANYNODE)
node              408 net/appletalk/ddp.c 		if ((node == ATADDR_BCAST ||
node              409 net/appletalk/ddp.c 		     node == ATADDR_ANYNODE ||
node              410 net/appletalk/ddp.c 		     iface->address.s_node == node) &&
node              416 net/appletalk/ddp.c 		if (node == ATADDR_ANYNODE && net != ATADDR_ANYNET &&
node             1092 net/appletalk/ddp.c 		struct hlist_node *node;
node             1094 net/appletalk/ddp.c 		sk_for_each(s, node, &atalk_sockets) {
node              223 net/atm/common.c 		struct hlist_node *node, *tmp;
node              227 net/atm/common.c 		sk_for_each_safe(s, node, tmp, head) {
node              269 net/atm/common.c 	struct hlist_node *node;
node              273 net/atm/common.c 	sk_for_each(s, node, head) {
node             1010 net/atm/lec.c  	struct hlist_node *node;
node             1020 net/atm/lec.c  	struct hlist_node *e = state->node;
node             1034 net/atm/lec.c  	state->node = e;
node             1127 net/atm/lec.c  	state->node = SEQ_START_TOKEN;
node             1163 net/atm/lec.c  		struct lec_arp_table *entry = hlist_entry(state->node, struct lec_arp_table, next);
node             1500 net/atm/lec.c  	struct hlist_node *node;
node             1517 net/atm/lec.c  			hlist_for_each_entry(entry, node, &priv->lec_arp_tables[i], next) {
node             1558 net/atm/lec.c  	struct hlist_node *node;
node             1565 net/atm/lec.c  		hlist_for_each_entry(rulla, node, &priv->lec_arp_tables[i], next) {
node             1600 net/atm/lec.c  	hlist_for_each_entry(rulla, node, &priv->lec_no_forward, next) {
node             1629 net/atm/lec.c  	hlist_for_each_entry(rulla, node, &priv->lec_arp_empty_ones, next) {
node             1658 net/atm/lec.c  	hlist_for_each_entry(rulla, node, &priv->mcast_fwds, next) {
node             1696 net/atm/lec.c  	struct hlist_node *node, *next;
node             1708 net/atm/lec.c  		hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_tables[i], next) {
node             1715 net/atm/lec.c  	hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_empty_ones, next) {
node             1723 net/atm/lec.c  	hlist_for_each_entry_safe(entry, node, next, &priv->lec_no_forward, next) {
node             1731 net/atm/lec.c  	hlist_for_each_entry_safe(entry, node, next, &priv->mcast_fwds, next) {
node             1748 net/atm/lec.c  	struct hlist_node *node;
node             1757 net/atm/lec.c  	hlist_for_each_entry(entry, node, head, next) {
node             1851 net/atm/lec.c  	struct hlist_node *node, *next;
node             1862 net/atm/lec.c  		hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_tables[i], next) {
node             2022 net/atm/lec.c  	struct hlist_node *node, *next;
node             2029 net/atm/lec.c  		hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_tables[i], next) {
node             2053 net/atm/lec.c  	struct hlist_node *node, *next;
node             2070 net/atm/lec.c  		hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_empty_ones, next) {
node             2114 net/atm/lec.c  		hlist_for_each_entry(tmp, node, &priv->lec_arp_tables[i], next) {
node             2154 net/atm/lec.c  	struct hlist_node *node;
node             2229 net/atm/lec.c  		hlist_for_each_entry(entry, node, &priv->lec_arp_tables[i], next) {
node             2305 net/atm/lec.c  	struct hlist_node *node;
node             2313 net/atm/lec.c  		hlist_for_each_entry(entry, node, &priv->lec_arp_tables[i], next) {
node             2340 net/atm/lec.c  	struct hlist_node *node;
node             2346 net/atm/lec.c  		hlist_for_each_entry(entry, node, &priv->lec_arp_tables[i], next) {
node             2396 net/atm/lec.c  	struct hlist_node *node, *next;
node             2406 net/atm/lec.c  		hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_tables[i], next) {
node             2417 net/atm/lec.c  	hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_empty_ones, next) {
node             2426 net/atm/lec.c  	hlist_for_each_entry_safe(entry, node, next, &priv->lec_no_forward, next) {
node             2435 net/atm/lec.c  	hlist_for_each_entry_safe(entry, node, next, &priv->mcast_fwds, next) {
node             2453 net/atm/lec.c  	struct hlist_node *node, *next;
node             2467 net/atm/lec.c  	hlist_for_each_entry_safe(entry, node, next, &priv->lec_arp_empty_ones, next) {
node              218 net/atm/signaling.c 	struct hlist_node *node;
node              232 net/atm/signaling.c 		sk_for_each(s, node, head) {
node               84 net/ax25/af_ax25.c 	struct hlist_node *node;
node               91 net/ax25/af_ax25.c 	ax25_for_each(s, node, &ax25_list) {
node              161 net/ax25/af_ax25.c 	struct hlist_node *node;
node              164 net/ax25/af_ax25.c 	ax25_for_each(s, node, &ax25_list) {
node              190 net/ax25/af_ax25.c 	struct hlist_node *node;
node              193 net/ax25/af_ax25.c 	ax25_for_each(s, node, &ax25_list) {
node              216 net/ax25/af_ax25.c 	struct hlist_node *node;
node              219 net/ax25/af_ax25.c 	ax25_for_each(s, node, &ax25_list) {
node              251 net/ax25/af_ax25.c 	struct hlist_node *node;
node              254 net/ax25/af_ax25.c 	ax25_for_each(s, node, &ax25_list) {
node             1861 net/ax25/af_ax25.c 	struct hlist_node *node;
node             1865 net/ax25/af_ax25.c 	ax25_for_each(ax25, node, &ax25_list) {
node               42 net/ax25/ax25_ds_subr.c 	struct hlist_node *node;
node               83 net/ax25/ax25_ds_subr.c 	ax25_for_each(ax25o, node, &ax25_list) {
node              162 net/ax25/ax25_ds_subr.c 	struct hlist_node *node;
node              165 net/ax25/ax25_ds_subr.c 	ax25_for_each(ax25, node, &ax25_list)
node               74 net/ax25/ax25_ds_timer.c 	struct hlist_node *node;
node               85 net/ax25/ax25_ds_timer.c 	ax25_for_each(ax25, node, &ax25_list) {
node              200 net/ax25/ax25_iface.c 	struct hlist_node *node;
node              203 net/ax25/ax25_iface.c 	hlist_for_each_entry(lf, node, &ax25_linkfail_list, lf_node)
node               56 net/ax25/ax25_uid.c 	struct hlist_node *node;
node               59 net/ax25/ax25_uid.c 	ax25_uid_for_each(ax25_uid, node, &ax25_uid_list) {
node               76 net/ax25/ax25_uid.c 	struct hlist_node *node;
node               84 net/ax25/ax25_uid.c 		ax25_uid_for_each(ax25_uid, node, &ax25_uid_list) {
node              123 net/ax25/ax25_uid.c 		ax25_uid_for_each(ax25_uid, node, &ax25_uid_list) {
node              150 net/ax25/ax25_uid.c 	struct hlist_node *node;
node              154 net/ax25/ax25_uid.c 	ax25_uid_for_each(pt, node, &ax25_uid_list) {
node              218 net/ax25/ax25_uid.c 	struct hlist_node *node;
node              222 net/ax25/ax25_uid.c 	ax25_uid_for_each(ax25_uid, node, &ax25_uid_list) {
node               94 net/bluetooth/hci_sock.c 	struct hlist_node *node;
node               99 net/bluetooth/hci_sock.c 	sk_for_each(sk, node, &hci_sk_list.head) {
node              671 net/bluetooth/hci_sock.c 		struct hlist_node *node;
node              675 net/bluetooth/hci_sock.c 		sk_for_each(sk, node, &hci_sk_list.head) {
node              521 net/bluetooth/l2cap.c 	struct hlist_node *node;
node              522 net/bluetooth/l2cap.c 	sk_for_each(sk, node, &l2cap_sk_list.head)
node              536 net/bluetooth/l2cap.c 	struct hlist_node *node;
node              538 net/bluetooth/l2cap.c 	sk_for_each(sk, node, &l2cap_sk_list.head) {
node              552 net/bluetooth/l2cap.c 	return node ? sk : sk1;
node             2137 net/bluetooth/l2cap.c 	struct hlist_node *node;
node             2146 net/bluetooth/l2cap.c 	sk_for_each(sk, node, &l2cap_sk_list.head) {
node             2424 net/bluetooth/l2cap.c 	struct hlist_node *node;
node             2429 net/bluetooth/l2cap.c 	sk_for_each(sk, node, &l2cap_sk_list.head) {
node              128 net/bluetooth/rfcomm/sock.c 	struct hlist_node *node;
node              130 net/bluetooth/rfcomm/sock.c 	sk_for_each(sk, node, &rfcomm_sk_list.head) {
node              136 net/bluetooth/rfcomm/sock.c 	return node ? sk : NULL;
node              145 net/bluetooth/rfcomm/sock.c 	struct hlist_node *node;
node              147 net/bluetooth/rfcomm/sock.c 	sk_for_each(sk, node, &rfcomm_sk_list.head) {
node              161 net/bluetooth/rfcomm/sock.c 	return node ? sk : sk1;
node              902 net/bluetooth/rfcomm/sock.c 	struct hlist_node *node;
node              907 net/bluetooth/rfcomm/sock.c 	sk_for_each(sk, node, &rfcomm_sk_list.head) {
node              289 net/bluetooth/sco.c 	struct hlist_node *node;
node              291 net/bluetooth/sco.c 	sk_for_each(sk, node, &sco_sk_list.head)
node              305 net/bluetooth/sco.c 	struct hlist_node *node;
node              309 net/bluetooth/sco.c 	sk_for_each(sk, node, &sco_sk_list.head) {
node              324 net/bluetooth/sco.c 	return node ? sk : sk1;
node              899 net/bluetooth/sco.c 	struct hlist_node *node;
node              904 net/bluetooth/sco.c 	sk_for_each(sk, node, &sco_sk_list.head) {
node              190 net/core/skbuff.c 	skb = kmem_cache_alloc_node(cache, gfp_mask & ~__GFP_DMA, node);
node              196 net/core/skbuff.c 			gfp_mask, node);
node              255 net/core/skbuff.c 	int node = dev->dev.parent ? dev_to_node(dev->dev.parent) : -1;
node              258 net/core/skbuff.c 	skb = __alloc_skb(length + NET_SKB_PAD, gfp_mask, 0, node);
node              268 net/core/skbuff.c 	int node = dev->dev.parent ? dev_to_node(dev->dev.parent) : -1;
node              271 net/core/skbuff.c 	page = alloc_pages_node(node, gfp_mask, 0);
node             2092 net/core/sock.c 	list_add(&prot->node, &proto_list);
node             2119 net/core/sock.c 	list_del(&prot->node);
node             2217 net/core/sock.c 		proto_seq_printf(seq, list_entry(v, struct proto, node));
node              179 net/decnet/af_decnet.c 	struct hlist_node *node;
node              184 net/decnet/af_decnet.c 	sk_for_each(sk, node, &dn_sk_hash[dn_ntohs(port) & DN_SK_HASH_MASK]) {
node              378 net/decnet/af_decnet.c 	struct hlist_node *node;
node              382 net/decnet/af_decnet.c 	sk_for_each(sk, node, list) {
node              418 net/decnet/af_decnet.c 	struct hlist_node *node;
node              422 net/decnet/af_decnet.c 	sk_for_each(sk, node, &dn_sk_hash[dn_ntohs(cb->dst_port) & DN_SK_HASH_MASK]) {
node              667 net/decnet/af_decnet.c 	unsigned short node, area;
node              669 net/decnet/af_decnet.c 	node = addr & 0x03ff;
node              671 net/decnet/af_decnet.c 	sprintf(buf, "%hd.%hd", area, node);
node              470 net/decnet/dn_table.c 	struct hlist_node *node;
node              485 net/decnet/dn_table.c 		hlist_for_each_entry(tb, node, &dn_fib_table_hash[h], hlist) {
node              814 net/decnet/dn_table.c 	struct hlist_node *node;
node              825 net/decnet/dn_table.c 	hlist_for_each_entry_rcu(t, node, &dn_fib_table_hash[h], hlist) {
node              871 net/decnet/dn_table.c 	struct hlist_node *node;
node              875 net/decnet/dn_table.c 		hlist_for_each_entry(tb, node, &dn_fib_table_hash[h], hlist)
node              894 net/decnet/dn_table.c 	struct hlist_node *node, *next;
node              899 net/decnet/dn_table.c 		hlist_for_each_entry_safe(t, node, next, &dn_fib_table_hash[h],
node               90 net/decnet/sysctl_net_decnet.c 	__u16 area, node;
node              109 net/decnet/sysctl_net_decnet.c 	node = *str++ - '0';
node              111 net/decnet/sysctl_net_decnet.c 		node *= 10;
node              112 net/decnet/sysctl_net_decnet.c 		node += (*str++ - '0');
node              115 net/decnet/sysctl_net_decnet.c 		node *= 10;
node              116 net/decnet/sysctl_net_decnet.c 		node += (*str++ - '0');
node              119 net/decnet/sysctl_net_decnet.c 		node *= 10;
node              120 net/decnet/sysctl_net_decnet.c 		node += (*str++ - '0');
node              123 net/decnet/sysctl_net_decnet.c 	if ((node > 1023) || (area > 63))
node              129 net/decnet/sysctl_net_decnet.c 	*addr = dn_htons((area << 10) | node);
node              782 net/econet/af_econet.c 	struct hlist_node *node;
node              784 net/econet/af_econet.c 	sk_for_each(sk, node, &econet_sklist) {
node               96 net/ipv4/fib_frontend.c 	struct hlist_node *node;
node              106 net/ipv4/fib_frontend.c 	hlist_for_each_entry_rcu(tb, node, head, tb_hlist) {
node              136 net/ipv4/fib_frontend.c 	struct hlist_node *node;
node              142 net/ipv4/fib_frontend.c 		hlist_for_each_entry(tb, node, head, tb_hlist)
node              631 net/ipv4/fib_frontend.c 	struct hlist_node *node;
node              645 net/ipv4/fib_frontend.c 		hlist_for_each_entry(tb, node, head, tb_hlist) {
node             1008 net/ipv4/fib_frontend.c 		struct hlist_node *node, *tmp;
node             1011 net/ipv4/fib_frontend.c 		hlist_for_each_entry_safe(tb, node, tmp, head, tb_hlist) {
node             1012 net/ipv4/fib_frontend.c 			hlist_del(node);
node              119 net/ipv4/fib_hash.c 		struct hlist_node *node, *n;
node              122 net/ipv4/fib_hash.c 		hlist_for_each_entry_safe(f, node, n, &old_ht[i], fn_hash) {
node              255 net/ipv4/fib_hash.c 		struct hlist_node *node;
node              260 net/ipv4/fib_hash.c 		hlist_for_each_entry(f, node, head, fn_hash) {
node              282 net/ipv4/fib_hash.c 	struct hlist_node *node;
node              297 net/ipv4/fib_hash.c 	hlist_for_each_entry(f, node, &fz->fz_hash[0], fn_hash) {
node              359 net/ipv4/fib_hash.c 	struct hlist_node *node;
node              362 net/ipv4/fib_hash.c 	hlist_for_each_entry(f, node, head, fn_hash) {
node              632 net/ipv4/fib_hash.c 	struct hlist_node *node, *n;
node              636 net/ipv4/fib_hash.c 	hlist_for_each_entry_safe(f, node, n, head, fn_hash) {
node              688 net/ipv4/fib_hash.c 	struct hlist_node *node;
node              694 net/ipv4/fib_hash.c 	hlist_for_each_entry(f, node, head, fn_hash) {
node              845 net/ipv4/fib_hash.c 			struct hlist_node *node;
node              848 net/ipv4/fib_hash.c 			hlist_for_each_entry(fn,node,iter->hash_head,fn_hash) {
node              884 net/ipv4/fib_hash.c 		struct hlist_node *node = &fn->fn_hash;
node              885 net/ipv4/fib_hash.c 		hlist_for_each_entry_continue(fn, node, fn_hash) {
node              902 net/ipv4/fib_hash.c 		struct hlist_node *node;
node              910 net/ipv4/fib_hash.c 			hlist_for_each_entry(fn, node, iter->hash_head, fn_hash) {
node              927 net/ipv4/fib_hash.c 		hlist_for_each_entry(fn, node, iter->hash_head, fn_hash) {
node              223 net/ipv4/fib_semantics.c 	struct hlist_node *node;
node              230 net/ipv4/fib_semantics.c 	hlist_for_each_entry(fi, node, head, fib_hash) {
node              255 net/ipv4/fib_semantics.c 	struct hlist_node *node;
node              263 net/ipv4/fib_semantics.c 	hlist_for_each_entry(nh, node, head, nh_hash) {
node              644 net/ipv4/fib_semantics.c 		struct hlist_node *node, *n;
node              647 net/ipv4/fib_semantics.c 		hlist_for_each_entry_safe(fi, node, n, head, fib_hash) {
node              662 net/ipv4/fib_semantics.c 		struct hlist_node *node, *n;
node              665 net/ipv4/fib_semantics.c 		hlist_for_each_entry_safe(fi, node, n, lhead, fib_lhash) {
node             1043 net/ipv4/fib_semantics.c 	struct hlist_node *node;
node             1049 net/ipv4/fib_semantics.c 	hlist_for_each_entry(fi, node, head, fib_lhash) {
node             1067 net/ipv4/fib_semantics.c 	struct hlist_node *node;
node             1073 net/ipv4/fib_semantics.c 	hlist_for_each_entry(nh, node, head, nh_hash) {
node             1124 net/ipv4/fib_semantics.c 	struct hlist_node *node;
node             1136 net/ipv4/fib_semantics.c 	hlist_for_each_entry(nh, node, head, nh_hash) {
node               92 net/ipv4/fib_trie.c #define NODE_TYPE(node) ((node)->parent & NODE_TYPE_MASK)
node              127 net/ipv4/fib_trie.c 	struct node *child[0];
node              152 net/ipv4/fib_trie.c 	struct node *trie;
node              161 net/ipv4/fib_trie.c static struct node *resize(struct trie *t, struct tnode *tn);
node              170 net/ipv4/fib_trie.c 	return (struct tnode *)(node->parent & ~NODE_TYPE_MASK);
node              175 net/ipv4/fib_trie.c 	struct tnode *ret = node_parent(node);
node              186 net/ipv4/fib_trie.c 	node->parent = (unsigned long)ptr | NODE_TYPE(node);
node              189 net/ipv4/fib_trie.c static inline struct node *tnode_get_child(struct tnode *tn, unsigned int i)
node              196 net/ipv4/fib_trie.c static inline struct node *tnode_get_child_rcu(struct tnode *tn, unsigned int i)
node              198 net/ipv4/fib_trie.c 	struct node *ret = tnode_get_child(tn, i);
node              370 net/ipv4/fib_trie.c 		      (sizeof(struct node *) << tn->bits);
node              410 net/ipv4/fib_trie.c 	size_t sz = sizeof(struct tnode) + (sizeof(struct node *) << bits);
node              423 net/ipv4/fib_trie.c 		 (unsigned long) (sizeof(struct node) << bits));
node              454 net/ipv4/fib_trie.c 	struct node *chi = tn->child[i];
node              481 net/ipv4/fib_trie.c static struct node *resize(struct trie *t, struct tnode *tn)
node              504 net/ipv4/fib_trie.c 			struct node *n;
node              664 net/ipv4/fib_trie.c 			struct node *n;
node              677 net/ipv4/fib_trie.c 	return (struct node *) tn;
node              724 net/ipv4/fib_trie.c 			put_child(t, tn, 2*i, (struct node *) left);
node              725 net/ipv4/fib_trie.c 			put_child(t, tn, 2*i+1, (struct node *) right);
node              731 net/ipv4/fib_trie.c 		struct node *node = tnode_get_child(oldtnode, i);
node              736 net/ipv4/fib_trie.c 		if (node == NULL)
node              741 net/ipv4/fib_trie.c 		if (IS_LEAF(node) || ((struct tnode *) node)->pos >
node              743 net/ipv4/fib_trie.c 			if (tkey_extract_bits(node->key,
node              746 net/ipv4/fib_trie.c 				put_child(t, tn, 2*i, node);
node              748 net/ipv4/fib_trie.c 				put_child(t, tn, 2*i+1, node);
node              753 net/ipv4/fib_trie.c 		inode = (struct tnode *) node;
node              826 net/ipv4/fib_trie.c 	struct node *left, *right;
node              857 net/ipv4/fib_trie.c 			put_child(t, tn, i/2, (struct node *)newn);
node              911 net/ipv4/fib_trie.c 	struct hlist_node *node;
node              914 net/ipv4/fib_trie.c 	hlist_for_each_entry_rcu(li, node, head, hlist)
node              934 net/ipv4/fib_trie.c 	struct hlist_node *node;
node              939 net/ipv4/fib_trie.c 		hlist_for_each_entry(li, node, head, hlist) {
node              959 net/ipv4/fib_trie.c 	struct node *n;
node              986 net/ipv4/fib_trie.c static struct node *trie_rebalance(struct trie *t, struct tnode *tn)
node              992 net/ipv4/fib_trie.c 	while (tn != NULL && (tp = node_parent((struct node *)tn)) != NULL) {
node              998 net/ipv4/fib_trie.c 				      (struct node *)tn, wasfull);
node             1000 net/ipv4/fib_trie.c 		tp = node_parent((struct node *) tn);
node             1010 net/ipv4/fib_trie.c 	return (struct node *)tn;
node             1019 net/ipv4/fib_trie.c 	struct node *n;
node             1105 net/ipv4/fib_trie.c 		node_set_parent((struct node *)l, tp);
node             1108 net/ipv4/fib_trie.c 		put_child(t, (struct tnode *)tp, cindex, (struct node *)l);
node             1135 net/ipv4/fib_trie.c 		node_set_parent((struct node *)tn, tp);
node             1138 net/ipv4/fib_trie.c 		put_child(t, tn, missbit, (struct node *)l);
node             1144 net/ipv4/fib_trie.c 				  (struct node *)tn);
node             1146 net/ipv4/fib_trie.c 			rcu_assign_pointer(t->trie, (struct node *)tn);
node             1340 net/ipv4/fib_trie.c 	struct hlist_node *node;
node             1342 net/ipv4/fib_trie.c 	hlist_for_each_entry_rcu(li, node, hhead, hlist) {
node             1371 net/ipv4/fib_trie.c 	struct node *n;
node             1544 net/ipv4/fib_trie.c 			struct tnode *parent = node_parent((struct node *) pn);
node             1571 net/ipv4/fib_trie.c 	struct tnode *tp = node_parent((struct node *) l);
node             1691 net/ipv4/fib_trie.c 	struct hlist_node *node, *tmp;
node             1694 net/ipv4/fib_trie.c 	hlist_for_each_entry_safe(li, node, tmp, lih, hlist) {
node             1735 net/ipv4/fib_trie.c 		c = (struct node *) p;
node             1756 net/ipv4/fib_trie.c 	struct node *c = (struct node *) l;
node             1915 net/ipv4/fib_trie.c 	struct hlist_node *node;
node             1922 net/ipv4/fib_trie.c 	hlist_for_each_entry_rcu(li, node, &l->list, hlist) {
node             2039 net/ipv4/fib_trie.c static struct node *fib_trie_get_next(struct fib_trie_iter *iter)
node             2053 net/ipv4/fib_trie.c 		struct node *n = tnode_get_child_rcu(tn, cindex);
node             2072 net/ipv4/fib_trie.c 	p = node_parent_rcu((struct node *)tn);
node             2084 net/ipv4/fib_trie.c static struct node *fib_trie_get_first(struct fib_trie_iter *iter,
node             2087 net/ipv4/fib_trie.c 	struct node *n;
node             2111 net/ipv4/fib_trie.c 	struct node *n;
node             2184 net/ipv4/fib_trie.c 	bytes += sizeof(struct node *) * pointers;
node             2229 net/ipv4/fib_trie.c 		struct hlist_node *node;
node             2232 net/ipv4/fib_trie.c 		hlist_for_each_entry_rcu(tb, node, head, tb_hlist) {
node             2265 net/ipv4/fib_trie.c static struct node *fib_trie_get_idx(struct seq_file *seq, loff_t pos)
node             2274 net/ipv4/fib_trie.c 		struct hlist_node *node;
node             2277 net/ipv4/fib_trie.c 		hlist_for_each_entry_rcu(tb, node, head, tb_hlist) {
node             2278 net/ipv4/fib_trie.c 			struct node *n;
node             2307 net/ipv4/fib_trie.c 	struct node *n;
node             2392 net/ipv4/fib_trie.c 	struct node *n = v;
node             2409 net/ipv4/fib_trie.c 		struct hlist_node *node;
node             2415 net/ipv4/fib_trie.c 		hlist_for_each_entry_rcu(li, node, &l->list, hlist) {
node             2561 net/ipv4/fib_trie.c 	struct hlist_node *node;
node             2570 net/ipv4/fib_trie.c 	hlist_for_each_entry_rcu(li, node, &l->list, hlist) {
node               57 net/ipv4/inet_connection_sock.c 	struct hlist_node *node;
node               67 net/ipv4/inet_connection_sock.c 	sk_for_each_bound(sk2, node, &tb->owners) {
node               82 net/ipv4/inet_connection_sock.c 	return node != NULL;
node               94 net/ipv4/inet_connection_sock.c 	struct hlist_node *node;
node              111 net/ipv4/inet_connection_sock.c 			inet_bind_bucket_for_each(tb, node, &head->chain)
node              139 net/ipv4/inet_connection_sock.c 		inet_bind_bucket_for_each(tb, node, &head->chain)
node              726 net/ipv4/inet_diag.c 			struct hlist_node *node;
node              729 net/ipv4/inet_diag.c 			sk_for_each(sk, node, &hashinfo->listening_hash[i]) {
node              783 net/ipv4/inet_diag.c 		struct hlist_node *node;
node              794 net/ipv4/inet_diag.c 		sk_for_each(sk, node, &head->chain) {
node              818 net/ipv4/inet_diag.c 			inet_twsk_for_each(tw, node,
node               42 net/ipv4/inet_hashtables.c 		hlist_add_head(&tb->node, &head->chain);
node               53 net/ipv4/inet_hashtables.c 		__hlist_del(&tb->node);
node              155 net/ipv4/inet_hashtables.c 	const struct hlist_node *node;
node              158 net/ipv4/inet_hashtables.c 	sk_for_each(sk, node, head) {
node              226 net/ipv4/inet_hashtables.c 	const struct hlist_node *node;
node              236 net/ipv4/inet_hashtables.c 	sk_for_each(sk, node, &head->chain) {
node              243 net/ipv4/inet_hashtables.c 	sk_for_each(sk, node, &head->twchain) {
node              275 net/ipv4/inet_hashtables.c 	const struct hlist_node *node;
node              282 net/ipv4/inet_hashtables.c 	sk_for_each(sk2, node, &head->twchain) {
node              296 net/ipv4/inet_hashtables.c 	sk_for_each(sk2, node, &head->chain) {
node              434 net/ipv4/inet_hashtables.c 		struct hlist_node *node;
node              451 net/ipv4/inet_hashtables.c 			inet_bind_bucket_for_each(tb, node, &head->chain) {
node              147 net/ipv4/inet_timewait_sock.c 	struct hlist_node *node;
node              160 net/ipv4/inet_timewait_sock.c 	inet_twsk_for_each_inmate(tw, node, &twdr->cells[slot]) {
node              373 net/ipv4/inet_timewait_sock.c 			struct hlist_node *node, *safe;
node              376 net/ipv4/inet_timewait_sock.c 			inet_twsk_for_each_inmate_safe(tw, node, safe,
node              419 net/ipv4/inet_timewait_sock.c 	struct hlist_node *node;
node              429 net/ipv4/inet_timewait_sock.c 		sk_for_each(sk, node, &head->twchain) {
node              188 net/ipv4/inetpeer.c 	struct inet_peer **nodep, *node, *l, *r;
node              193 net/ipv4/inetpeer.c 		node = *nodep;
node              194 net/ipv4/inetpeer.c 		l = node->avl_left;
node              195 net/ipv4/inetpeer.c 		r = node->avl_right;
node              205 net/ipv4/inetpeer.c 				node->avl_left = lr;	/* lr: RH or RH+1 */
node              206 net/ipv4/inetpeer.c 				node->avl_right = r;	/* r: RH */
node              207 net/ipv4/inetpeer.c 				node->avl_height = lrh + 1; /* RH+1 or RH+2 */
node              209 net/ipv4/inetpeer.c 				l->avl_right = node;	/* node: RH+1 or RH+2 */
node              210 net/ipv4/inetpeer.c 				l->avl_height = node->avl_height + 1;
node              215 net/ipv4/inetpeer.c 				node->avl_left = lrr;	/* lrr: RH or RH-1 */
node              216 net/ipv4/inetpeer.c 				node->avl_right = r;	/* r: RH */
node              217 net/ipv4/inetpeer.c 				node->avl_height = rh + 1; /* node: RH+1 */
node              222 net/ipv4/inetpeer.c 				lr->avl_right = node;	/* node: RH+1 */
node              233 net/ipv4/inetpeer.c 				node->avl_right = rl;	/* rl: LH or LH+1 */
node              234 net/ipv4/inetpeer.c 				node->avl_left = l;	/* l: LH */
node              235 net/ipv4/inetpeer.c 				node->avl_height = rlh + 1; /* LH+1 or LH+2 */
node              237 net/ipv4/inetpeer.c 				r->avl_left = node;	/* node: LH+1 or LH+2 */
node              238 net/ipv4/inetpeer.c 				r->avl_height = node->avl_height + 1;
node              243 net/ipv4/inetpeer.c 				node->avl_right = rll;	/* rll: LH or LH-1 */
node              244 net/ipv4/inetpeer.c 				node->avl_left = l;	/* l: LH */
node              245 net/ipv4/inetpeer.c 				node->avl_height = lh + 1; /* node: LH+1 */
node              250 net/ipv4/inetpeer.c 				rl->avl_left = node;	/* node: LH+1 */
node              255 net/ipv4/inetpeer.c 			node->avl_height = (lh > rh ? lh : rh) + 1;
node              113 net/ipv4/raw.c 	struct hlist_node *node;
node              115 net/ipv4/raw.c 	sk_for_each_from(sk, node) {
node              858 net/ipv4/raw.c 		struct hlist_node *node;
node              860 net/ipv4/raw.c 		sk_for_each(sk, node, &state->h->ht[state->bucket])
node             1878 net/ipv4/tcp_ipv4.c 	struct hlist_node *node;
node             1921 net/ipv4/tcp_ipv4.c 	sk_for_each_from(sk, node) {
node             1972 net/ipv4/tcp_ipv4.c 		struct hlist_node *node;
node             1981 net/ipv4/tcp_ipv4.c 		sk_for_each(sk, node, &tcp_hashinfo.ehash[st->bucket].chain) {
node             1990 net/ipv4/tcp_ipv4.c 		inet_twsk_for_each(tw, node,
node             2010 net/ipv4/tcp_ipv4.c 	struct hlist_node *node;
node             2042 net/ipv4/tcp_ipv4.c 	sk_for_each_from(sk, node) {
node              132 net/ipv4/udp.c 	struct hlist_node *node;
node              134 net/ipv4/udp.c 	sk_for_each(sk2, node, &udptable[udp_hashfn(net, num)])
node              219 net/ipv4/udp.c 	struct hlist_node *node;
node              224 net/ipv4/udp.c 	sk_for_each(sk, node, &udptable[udp_hashfn(net, hnum)]) {
node              292 net/ipv4/udp.c 	struct hlist_node *node;
node              296 net/ipv4/udp.c 	sk_for_each_from(s, node) {
node             1508 net/ipv4/udp.c 		struct hlist_node *node;
node             1509 net/ipv4/udp.c 		sk_for_each(sk, node, state->hashtable + state->bucket) {
node               33 net/ipv6/inet6_connection_sock.c 	const struct hlist_node *node;
node               40 net/ipv6/inet6_connection_sock.c 	sk_for_each_bound(sk2, node, &tb->owners) {
node               51 net/ipv6/inet6_connection_sock.c 	return node != NULL;
node               66 net/ipv6/inet6_hashtables.c 	const struct hlist_node *node;
node               77 net/ipv6/inet6_hashtables.c 	sk_for_each(sk, node, &head->chain) {
node               83 net/ipv6/inet6_hashtables.c 	sk_for_each(sk, node, &head->twchain) {
node              102 net/ipv6/inet6_hashtables.c 	const struct hlist_node *node;
node              107 net/ipv6/inet6_hashtables.c 	sk_for_each(sk, node,
node              175 net/ipv6/inet6_hashtables.c 	const struct hlist_node *node;
node              182 net/ipv6/inet6_hashtables.c 	sk_for_each(sk2, node, &head->twchain) {
node              195 net/ipv6/inet6_hashtables.c 	sk_for_each(sk2, node, &head->chain) {
node              229 net/ipv6/ip6_fib.c 	struct hlist_node *node;
node              237 net/ipv6/ip6_fib.c 	hlist_for_each_entry_rcu(tb, node, head, tb6_hlist) {
node              353 net/ipv6/ip6_fib.c 	struct hlist_node *node;
node              387 net/ipv6/ip6_fib.c 		hlist_for_each_entry(tb, node, head, tb6_hlist) {
node             1052 net/ipv6/ip6_fib.c 					w->root = w->node = NULL;
node             1054 net/ipv6/ip6_fib.c 				} else if (w->node == fn) {
node             1056 net/ipv6/ip6_fib.c 					w->node = pn;
node             1064 net/ipv6/ip6_fib.c 				if (w->node == fn) {
node             1065 net/ipv6/ip6_fib.c 					w->node = child;
node             1223 net/ipv6/ip6_fib.c 		fn = w->node;
node             1236 net/ipv6/ip6_fib.c 				w->node = FIB6_SUBTREE(fn);
node             1243 net/ipv6/ip6_fib.c 				w->node = fn->left;
node             1250 net/ipv6/ip6_fib.c 				w->node = fn->right;
node             1268 net/ipv6/ip6_fib.c 			w->node = pn;
node             1282 net/ipv6/ip6_fib.c 				w->leaf = w->node->leaf;
node             1297 net/ipv6/ip6_fib.c 	w->node = w->root;
node             1365 net/ipv6/ip6_fib.c 	struct hlist_node *node;
node             1372 net/ipv6/ip6_fib.c 		hlist_for_each_entry_rcu(table, node, head, tb6_hlist) {
node               71 net/ipv6/raw.c 	struct hlist_node *node;
node               74 net/ipv6/raw.c 	sk_for_each_from(sk, node)
node               63 net/ipv6/udp.c 	struct hlist_node *node;
node               68 net/ipv6/udp.c 	sk_for_each(sk, node, &udptable[udp_hashfn(net, hnum)]) {
node              336 net/ipv6/udp.c 	struct hlist_node *node;
node              340 net/ipv6/udp.c 	sk_for_each_from(s, node) {
node              101 net/ipx/af_ipx.c 				struct ipx_interface, node);
node              172 net/ipx/af_ipx.c 	list_for_each_entry(i, &ipx_interfaces, node)
node              199 net/ipx/af_ipx.c 		list_for_each_entry(i, &ipx_interfaces, node)
node              231 net/ipx/af_ipx.c 	struct hlist_node *node;
node              233 net/ipx/af_ipx.c 	sk_for_each(s, node, &intrfc->if_sklist)
node              262 net/ipx/af_ipx.c 	struct hlist_node *node;
node              267 net/ipx/af_ipx.c 	sk_for_each(s, node, &intrfc->if_sklist) {
node              271 net/ipx/af_ipx.c 		    !memcmp(ipx_node, ipxs->node, IPX_NODE_LEN))
node              285 net/ipx/af_ipx.c 	struct hlist_node *node, *t;
node              292 net/ipx/af_ipx.c 	sk_for_each_safe(s, node, t, &intrfc->if_sklist) {
node              306 net/ipx/af_ipx.c 	list_del(&intrfc->node);
node              345 net/ipx/af_ipx.c 	list_for_each_entry_safe(i, tmp, &ipx_interfaces, node)
node              363 net/ipx/af_ipx.c 	list_for_each_entry_safe(i, tmp, &ipx_interfaces, node)
node              385 net/ipx/af_ipx.c 	int is_broadcast = !memcmp(ipx->ipx_dest.node, ipx_broadcast_node,
node              388 net/ipx/af_ipx.c 	struct hlist_node *node;
node              393 net/ipx/af_ipx.c 	sk_for_each(s, node, &intrfc->if_sklist) {
node              397 net/ipx/af_ipx.c 		    (is_broadcast || !memcmp(ipx->ipx_dest.node,
node              398 net/ipx/af_ipx.c 					     ipxs->node, IPX_NODE_LEN))) {
node              449 net/ipx/af_ipx.c 		struct hlist_node *node;
node              454 net/ipx/af_ipx.c 		sk_for_each(sk, node, &intrfc->if_sklist)
node              625 net/ipx/af_ipx.c 		    !memcmp(intrfc->if_node, node, IPX_NODE_LEN)) {
node              634 net/ipx/af_ipx.c 		if (!memcmp(ipx_broadcast_node, node, IPX_NODE_LEN)) {
node              667 net/ipx/af_ipx.c 	if (!memcmp(ipx_broadcast_node, node, IPX_NODE_LEN))
node              670 net/ipx/af_ipx.c 		memcpy(dest_node, &(node[IPX_NODE_LEN-addr_len]), addr_len);
node              736 net/ipx/af_ipx.c 	if (!memcmp(ipx_broadcast_node, ipx->ipx_dest.node, IPX_NODE_LEN) ||
node              737 net/ipx/af_ipx.c 	    !memcmp(intrfc->if_node, ipx->ipx_dest.node, IPX_NODE_LEN)) {
node              843 net/ipx/af_ipx.c 	list_for_each_entry(ifcs, &ipx_interfaces, node) {
node              874 net/ipx/af_ipx.c 	list_add_tail(&intrfc->node, &ipx_interfaces);
node             1468 net/ipx/af_ipx.c 			memcpy(ipxs->node, intrfc->if_node, IPX_NODE_LEN);
node             1470 net/ipx/af_ipx.c 			memcpy(ipxs->node, addr->sipx_node, IPX_NODE_LEN);
node             1473 net/ipx/af_ipx.c 		if (ipxitf_find_internal_socket(intrfc, ipxs->node,
node             1486 net/ipx/af_ipx.c 		memcpy(ipxs->node, intrfc->if_node, IPX_NODE_LEN);
node             1567 net/ipx/af_ipx.c 	memcpy(ipxs->dest_addr.node, addr->sipx_node, IPX_NODE_LEN);
node             1602 net/ipx/af_ipx.c 		memcpy(sipx.sipx_node, addr->node, IPX_NODE_LEN);
node             1607 net/ipx/af_ipx.c 			memcpy(sipx.sipx_node, ipxs->node, IPX_NODE_LEN);
node             1745 net/ipx/af_ipx.c 		memcpy(usipx->sipx_node, ipxs->dest_addr.node, IPX_NODE_LEN);
node             1815 net/ipx/af_ipx.c 		memcpy(sipx->sipx_node, ipx->ipx_source.node, IPX_NODE_LEN);
node               20 net/ipx/ipx_proc.c 	list_for_each_entry(i, &ipx_interfaces, node)
node               32 net/ipx/ipx_proc.c 	if (i->node.next != &ipx_interfaces)
node               33 net/ipx/ipx_proc.c 		rc = list_entry(i->node.next, struct ipx_interface, node);
node               97 net/ipx/ipx_proc.c 		rc = list_entry(ipx_routes.next, struct ipx_route, node);
node              105 net/ipx/ipx_proc.c 	if (r->node.next != &ipx_routes)
node              106 net/ipx/ipx_proc.c 		rc = list_entry(r->node.next, struct ipx_route, node);
node              114 net/ipx/ipx_proc.c 	list_for_each_entry(r, &ipx_routes, node)
node              171 net/ipx/ipx_proc.c 	struct hlist_node *node;
node              174 net/ipx/ipx_proc.c 	list_for_each_entry(i, &ipx_interfaces, node) {
node              176 net/ipx/ipx_proc.c 		sk_for_each(s, node, &i->if_sklist) {
node              183 net/ipx/ipx_proc.c 			if (node)
node              265 net/ipx/ipx_proc.c 		   ipxs->node[0], ipxs->node[1], ipxs->node[2], ipxs->node[3],
node              266 net/ipx/ipx_proc.c 		   ipxs->node[4], ipxs->node[5], ntohs(ipxs->port));
node              276 net/ipx/ipx_proc.c 			   ipxs->dest_addr.node[0], ipxs->dest_addr.node[1],
node              277 net/ipx/ipx_proc.c 			   ipxs->dest_addr.node[2], ipxs->dest_addr.node[3],
node              278 net/ipx/ipx_proc.c 			   ipxs->dest_addr.node[4], ipxs->dest_addr.node[5],
node               37 net/ipx/ipx_route.c 	list_for_each_entry(r, &ipx_routes, node)
node               68 net/ipx/ipx_route.c 		list_add(&rt->node, &ipx_routes);
node               78 net/ipx/ipx_route.c 	if (!node) {
node               82 net/ipx/ipx_route.c 		memcpy(rt->ir_router_node, node, IPX_NODE_LEN);
node               98 net/ipx/ipx_route.c 	list_for_each_entry_safe(r, tmp, &ipx_routes, node)
node              100 net/ipx/ipx_route.c 			list_del(&r->node);
node              127 net/ipx/ipx_route.c 	list_for_each_entry_safe(r, tmp, &ipx_routes, node)
node              133 net/ipx/ipx_route.c 			list_del(&r->node);
node              160 net/ipx/ipx_route.c 			r->ir_router_node : ipx->ipx_dest.node);
node              217 net/ipx/ipx_route.c 	memcpy(ipx->ipx_source.node, ipxs->node, IPX_NODE_LEN);
node              223 net/ipx/ipx_route.c 		memcpy(ipx->ipx_source.node, intrfc->if_node, IPX_NODE_LEN);
node              226 net/ipx/ipx_route.c 		memcpy(ipx->ipx_source.node, ipxs->intrfc->if_node,
node              232 net/ipx/ipx_route.c 	memcpy(ipx->ipx_dest.node, usipx->sipx_node, IPX_NODE_LEN);
node              248 net/ipx/ipx_route.c 			 rt->ir_router_node : ipx->ipx_dest.node);
node               62 net/irda/discovery.c 	discovery_t *discovery, *node;
node               78 net/irda/discovery.c 		node = discovery;
node               83 net/irda/discovery.c 		if ((node->data.saddr == new->data.saddr) &&
node               84 net/irda/discovery.c 		    ((node->data.daddr == new->data.daddr) ||
node               85 net/irda/discovery.c 		     (strcmp(node->data.info, new->data.info) == 0)))
node               90 net/irda/discovery.c 			hashbin_remove_this(cachelog, (irda_queue_t *) node);
node               92 net/irda/discovery.c 			if (get_unaligned((__u16 *)node->data.hints) == get_unaligned((__u16 *)new->data.hints))
node               94 net/irda/discovery.c 				new->firststamp = node->firststamp;
node               95 net/irda/discovery.c 			kfree(node);
node              128 net/irda/irias_object.c 	struct ias_object *node;
node              134 net/irda/irias_object.c 	node = hashbin_remove_this(irias_objects, (irda_queue_t *) obj);
node              135 net/irda/irias_object.c 	if (!node)
node              156 net/irda/irias_object.c 	struct ias_attrib *node;
node              163 net/irda/irias_object.c 	node = hashbin_remove_this(obj->attribs, (irda_queue_t *) attrib);
node              164 net/irda/irias_object.c 	if (!node)
node              168 net/irda/irias_object.c 	__irias_delete_attrib(node);
node              175 net/irda/irias_object.c 	node = (struct ias_attrib *) hashbin_get_first(obj->attribs);
node              176 net/irda/irias_object.c 	if (cleanobject && !node)
node              100 net/iucv/af_iucv.c 	struct hlist_node *node;
node              102 net/iucv/af_iucv.c 	sk_for_each(sk, node, &iucv_sk_list.head)
node              978 net/iucv/af_iucv.c 	struct hlist_node *node;
node              989 net/iucv/af_iucv.c 	sk_for_each(sk, node, &iucv_sk_list.head)
node              261 net/key/af_key.c 	struct hlist_node *node;
node              272 net/key/af_key.c 	sk_for_each(sk, node, &pfkey_table) {
node             3649 net/key/af_key.c 	struct hlist_node *node;
node             3656 net/key/af_key.c 	sk_for_each(s, node, &pfkey_table)
node               69 net/lapb/lapb_iface.c 	if (lapb->node.next) {
node               70 net/lapb/lapb_iface.c 		list_del(&lapb->node);
node               80 net/lapb/lapb_iface.c 	list_add(&lapb->node, &lapb_list);
node               90 net/lapb/lapb_iface.c 		lapb = list_entry(entry, struct lapb_cb, node);
node              492 net/llc/llc_conn.c 	struct hlist_node *node;
node              495 net/llc/llc_conn.c 	sk_for_each(rc, node, &sap->sk_list.list) {
node              538 net/llc/llc_conn.c 	struct hlist_node *node;
node              541 net/llc/llc_conn.c 	sk_for_each(rc, node, &sap->sk_list.list) {
node               54 net/llc/llc_core.c 	list_add_tail(&sap->node, &llc_sap_list);
node               66 net/llc/llc_core.c 	list_del(&sap->node);
node               74 net/llc/llc_core.c 	list_for_each_entry(sap, &llc_sap_list, node)
node               38 net/llc/llc_proc.c 	struct hlist_node *node;
node               42 net/llc/llc_proc.c 		sap = list_entry(sap_entry, struct llc_sap, node);
node               45 net/llc/llc_proc.c 		sk_for_each(sk, node, &sap->sk_list.list) {
node               87 net/llc/llc_proc.c 		if (sap->node.next == &llc_sap_list)
node               89 net/llc/llc_proc.c 		sap = list_entry(sap->node.next, struct llc_sap, node);
node              312 net/llc/llc_sap.c 	struct hlist_node *node;
node              315 net/llc/llc_sap.c 	sk_for_each(rc, node, &sap->sk_list.list) {
node              344 net/llc/llc_sap.c 	struct hlist_node *node;
node              347 net/llc/llc_sap.c 	sk_for_each(sk, node, &sap->sk_list.list) {
node               80 net/llc/llc_station.c 	struct list_head node; /* node in station->ev_q.list */
node              298 net/mac80211/mesh.h 		hlist_for_each_entry_rcu(node, p, &x->hash_buckets[i], list)
node               77 net/mac80211/mesh_pathtbl.c 	struct mpath_node *node;
node               82 net/mac80211/mesh_pathtbl.c 	hlist_for_each_entry_rcu(node, n, bucket, list) {
node               83 net/mac80211/mesh_pathtbl.c 		mpath = node->mpath;
node              104 net/mac80211/mesh_pathtbl.c 	struct mpath_node *node;
node              109 net/mac80211/mesh_pathtbl.c 	hlist_for_each_entry_rcu(node, n, bucket, list) {
node              110 net/mac80211/mesh_pathtbl.c 		mpath = node->mpath;
node              137 net/mac80211/mesh_pathtbl.c 	struct mpath_node *node;
node              142 net/mac80211/mesh_pathtbl.c 	for_each_mesh_entry(mesh_paths, p, node, i) {
node              143 net/mac80211/mesh_pathtbl.c 		if (sdata && node->mpath->sdata != sdata)
node              146 net/mac80211/mesh_pathtbl.c 			if (MPATH_EXPIRED(node->mpath)) {
node              147 net/mac80211/mesh_pathtbl.c 				spin_lock_bh(&node->mpath->state_lock);
node              148 net/mac80211/mesh_pathtbl.c 				if (MPATH_EXPIRED(node->mpath))
node              149 net/mac80211/mesh_pathtbl.c 					node->mpath->flags &= ~MESH_PATH_ACTIVE;
node              150 net/mac80211/mesh_pathtbl.c 				spin_unlock_bh(&node->mpath->state_lock);
node              152 net/mac80211/mesh_pathtbl.c 			return node->mpath;
node              171 net/mac80211/mesh_pathtbl.c 	struct mpath_node *node, *new_node;
node              215 net/mac80211/mesh_pathtbl.c 	hlist_for_each_entry(node, n, bucket, list) {
node              216 net/mac80211/mesh_pathtbl.c 		mpath = node->mpath;
node              261 net/mac80211/mesh_pathtbl.c 	struct mpath_node *node, *new_node;
node              301 net/mac80211/mesh_pathtbl.c 	hlist_for_each_entry(node, n, bucket, list) {
node              302 net/mac80211/mesh_pathtbl.c 		mpath = node->mpath;
node              354 net/mac80211/mesh_pathtbl.c 	struct mpath_node *node;
node              360 net/mac80211/mesh_pathtbl.c 	for_each_mesh_entry(mesh_paths, p, node, i) {
node              361 net/mac80211/mesh_pathtbl.c 		mpath = node->mpath;
node              392 net/mac80211/mesh_pathtbl.c 	struct mpath_node *node;
node              396 net/mac80211/mesh_pathtbl.c 	for_each_mesh_entry(mesh_paths, p, node, i) {
node              397 net/mac80211/mesh_pathtbl.c 		mpath = node->mpath;
node              406 net/mac80211/mesh_pathtbl.c 	struct mpath_node *node;
node              410 net/mac80211/mesh_pathtbl.c 	for_each_mesh_entry(mesh_paths, p, node, i) {
node              411 net/mac80211/mesh_pathtbl.c 		mpath = node->mpath;
node              419 net/mac80211/mesh_pathtbl.c 	struct mpath_node *node = container_of(rp, struct mpath_node, rcu);
node              420 net/mac80211/mesh_pathtbl.c 	struct ieee80211_sub_if_data *sdata = node->mpath->sdata;
node              422 net/mac80211/mesh_pathtbl.c 	del_timer_sync(&node->mpath->timer);
node              424 net/mac80211/mesh_pathtbl.c 	kfree(node->mpath);
node              425 net/mac80211/mesh_pathtbl.c 	kfree(node);
node              439 net/mac80211/mesh_pathtbl.c 	struct mpath_node *node;
node              450 net/mac80211/mesh_pathtbl.c 	hlist_for_each_entry(node, n, bucket, list) {
node              451 net/mac80211/mesh_pathtbl.c 		mpath = node->mpath;
node              456 net/mac80211/mesh_pathtbl.c 			hlist_del_rcu(&node->list);
node              457 net/mac80211/mesh_pathtbl.c 			call_rcu(&node->rcu, mesh_path_node_reclaim);
node              562 net/mac80211/mesh_pathtbl.c 	struct mpath_node *node = hlist_entry(p, struct mpath_node, list);
node              563 net/mac80211/mesh_pathtbl.c 	mpath = node->mpath;
node              567 net/mac80211/mesh_pathtbl.c 	kfree(node);
node              573 net/mac80211/mesh_pathtbl.c 	struct mpath_node *node, *new_node;
node              580 net/mac80211/mesh_pathtbl.c 	node = hlist_entry(p, struct mpath_node, list);
node              581 net/mac80211/mesh_pathtbl.c 	mpath = node->mpath;
node              613 net/mac80211/mesh_pathtbl.c 	struct mpath_node *node;
node              618 net/mac80211/mesh_pathtbl.c 	for_each_mesh_entry(mesh_paths, p, node, i) {
node              619 net/mac80211/mesh_pathtbl.c 		if (node->mpath->sdata != sdata)
node              621 net/mac80211/mesh_pathtbl.c 		mpath = node->mpath;
node               68 net/netfilter/xt_hashlimit.c 	struct hlist_node node;
node               81 net/netfilter/xt_hashlimit.c 	struct hlist_node node;		/* global list of all htables */
node              135 net/netfilter/xt_hashlimit.c 		hlist_for_each_entry(ent, pos, &ht->hash[hash], node)
node              174 net/netfilter/xt_hashlimit.c 	hlist_add_head(&ent->node, &ht->hash[hash_dst(ht, dst)]);
node              182 net/netfilter/xt_hashlimit.c 	hlist_del(&ent->node);
node              254 net/netfilter/xt_hashlimit.c 	hlist_add_head(&hinfo->node, &hashlimit_htables);
node              316 net/netfilter/xt_hashlimit.c 	hlist_add_head(&hinfo->node, &hashlimit_htables);
node              345 net/netfilter/xt_hashlimit.c 		hlist_for_each_entry_safe(dh, pos, n, &ht->hash[i], node) {
node              384 net/netfilter/xt_hashlimit.c 	hlist_for_each_entry(hinfo, pos, &hashlimit_htables, node) {
node              400 net/netfilter/xt_hashlimit.c 		hlist_del(&hinfo->node);
node              932 net/netfilter/xt_hashlimit.c 		hlist_for_each_entry(ent, pos, &htable->hash[*bucket], node)
node              226 net/netlink/af_netlink.c 	struct hlist_node *node;
node              230 net/netlink/af_netlink.c 	sk_for_each(sk, node, head) {
node              290 net/netlink/af_netlink.c 		struct hlist_node *node, *tmp;
node              292 net/netlink/af_netlink.c 		sk_for_each_safe(sk, node, tmp, &otable[i])
node              322 net/netlink/af_netlink.c 	struct hlist_node *node;
node              328 net/netlink/af_netlink.c 		sk_for_each_bound(sk, node, &tbl->mc_list) {
node              344 net/netlink/af_netlink.c 	struct hlist_node *node;
node              350 net/netlink/af_netlink.c 	sk_for_each(osk, node, head) {
node              355 net/netlink/af_netlink.c 	if (node)
node              525 net/netlink/af_netlink.c 	struct hlist_node *node;
node              534 net/netlink/af_netlink.c 	sk_for_each(osk, node, head) {
node             1011 net/netlink/af_netlink.c 	struct hlist_node *node;
node             1031 net/netlink/af_netlink.c 	sk_for_each_bound(sk, node, &nl_table[ssk->sk_protocol].mc_list)
node             1083 net/netlink/af_netlink.c 	struct hlist_node *node;
node             1093 net/netlink/af_netlink.c 	sk_for_each_bound(sk, node, &nl_table[ssk->sk_protocol].mc_list)
node             1492 net/netlink/af_netlink.c 	struct hlist_node *node;
node             1497 net/netlink/af_netlink.c 	sk_for_each_bound(sk, node, &tbl->mc_list)
node             1754 net/netlink/af_netlink.c 	struct hlist_node *node;
node             1761 net/netlink/af_netlink.c 			sk_for_each(s, node, &hash->table[j]) {
node              107 net/netrom/af_netrom.c 	struct hlist_node *node;
node              110 net/netrom/af_netrom.c 	sk_for_each(s, node, &nr_list)
node              152 net/netrom/af_netrom.c 	struct hlist_node *node;
node              155 net/netrom/af_netrom.c 	sk_for_each(s, node, &nr_list)
node              173 net/netrom/af_netrom.c 	struct hlist_node *node;
node              176 net/netrom/af_netrom.c 	sk_for_each(s, node, &nr_list) {
node              197 net/netrom/af_netrom.c 	struct hlist_node *node;
node              200 net/netrom/af_netrom.c 	sk_for_each(s, node, &nr_list) {
node             1264 net/netrom/af_netrom.c 	struct hlist_node *node;
node             1271 net/netrom/af_netrom.c 	sk_for_each(s, node, &nr_list) {
node               51 net/netrom/nr_route.c 	struct hlist_node *node;
node               54 net/netrom/nr_route.c 	nr_node_for_each(nr_node, node, &nr_node_list)
node               69 net/netrom/nr_route.c 	struct hlist_node *node;
node               72 net/netrom/nr_route.c 	nr_neigh_for_each(nr_neigh, node, &nr_neigh_list)
node              116 net/netrom/nr_route.c 		struct hlist_node *node;
node              119 net/netrom/nr_route.c 		nr_node_for_each(nr_nodet, node, &nr_node_list) {
node              484 net/netrom/nr_route.c 	struct hlist_node *node, *nodet;
node              488 net/netrom/nr_route.c 	nr_node_for_each_safe(s, node, nodet, &nr_node_list) {
node              538 net/netrom/nr_route.c 	struct hlist_node *node, *nodet, *node2, *node2t;
node              543 net/netrom/nr_route.c 	nr_neigh_for_each_safe(s, node, nodet, &nr_neigh_list) {
node              728 net/netrom/nr_route.c 	struct hlist_node *node;
node              732 net/netrom/nr_route.c 	nr_neigh_for_each(s, node, &nr_neigh_list) {
node              752 net/netrom/nr_route.c 	nr_node_for_each(nr_node, node, &nr_node_list) {
node              862 net/netrom/nr_route.c 	struct hlist_node *node;
node              869 net/netrom/nr_route.c 	nr_node_for_each(nr_node, node, &nr_node_list) {
node              880 net/netrom/nr_route.c 	struct hlist_node *node;
node              883 net/netrom/nr_route.c 	node = (v == SEQ_START_TOKEN)
node              887 net/netrom/nr_route.c 	return hlist_entry(node, struct nr_node, node_node);
node              948 net/netrom/nr_route.c 	struct hlist_node *node;
node              955 net/netrom/nr_route.c 	nr_neigh_for_each(nr_neigh, node, &nr_neigh_list) {
node              964 net/netrom/nr_route.c 	struct hlist_node *node;
node              967 net/netrom/nr_route.c 	node = (v == SEQ_START_TOKEN)
node              971 net/netrom/nr_route.c 	return hlist_entry(node, struct nr_neigh, neigh_node);
node             1038 net/netrom/nr_route.c 	struct hlist_node *node, *nodet;
node             1042 net/netrom/nr_route.c 	nr_node_for_each_safe(t, node, nodet, &nr_node_list) {
node             1047 net/netrom/nr_route.c 	nr_neigh_for_each_safe(s, node, nodet, &nr_neigh_list) {
node             1588 net/packet/af_packet.c 	struct hlist_node *node;
node             1593 net/packet/af_packet.c 	sk_for_each(sk, node, &net->packet.sklist) {
node             2004 net/packet/af_packet.c 	struct hlist_node *node;
node             2006 net/packet/af_packet.c 	sk_for_each(s, node, &net->packet.sklist) {
node              527 net/phonet/pep.c 	struct hlist_node *node;
node              531 net/phonet/pep.c 	sk_for_each(sknode, node, hlist) {
node               62 net/phonet/socket.c 	struct hlist_node *node;
node               70 net/phonet/socket.c 	sk_for_each(sknode, node, &pnsocks.hlist) {
node              251 net/rfkill/rfkill.c 	list_for_each_entry(rfkill, &rfkill_list, node) {
node              292 net/rfkill/rfkill.c 	list_for_each_entry(rfkill, &rfkill_list, node) {
node              597 net/rfkill/rfkill.c 	list_for_each_entry(p, &rfkill_list, node) {
node              630 net/rfkill/rfkill.c 	list_add_tail(&rfkill->node, &rfkill_list);
node              642 net/rfkill/rfkill.c 	list_del_init(&rfkill->node);
node              680 net/rfkill/rfkill.c 	INIT_LIST_HEAD(&rfkill->node);
node              170 net/rose/af_rose.c 	struct hlist_node *node;
node              173 net/rose/af_rose.c 	sk_for_each(s, node, &rose_list) {
node              191 net/rose/af_rose.c 	struct hlist_node *node;
node              194 net/rose/af_rose.c 	sk_for_each(s, node, &rose_list) {
node              251 net/rose/af_rose.c 	struct hlist_node *node;
node              254 net/rose/af_rose.c 	sk_for_each(s, node, &rose_list) {
node              263 net/rose/af_rose.c 	sk_for_each(s, node, &rose_list) {
node              283 net/rose/af_rose.c 	struct hlist_node *node;
node              286 net/rose/af_rose.c 	sk_for_each(s, node, &rose_list) {
node             1405 net/rose/af_rose.c 	struct hlist_node *node;
node             1412 net/rose/af_rose.c 	sk_for_each(s, node, &rose_list) {
node              671 net/rose/rose_route.c 	struct rose_node *node;
node              676 net/rose/rose_route.c 	for (node = rose_node_list; node != NULL; node = node->next) {
node              677 net/rose/rose_route.c 		if (rosecmpm(addr, &node->address, node->mask) == 0) {
node              678 net/rose/rose_route.c 			for (i = 0; i < node->count; i++) {
node              680 net/rose/rose_route.c 					if (node->neighbour[i]->restarted) {
node              681 net/rose/rose_route.c 						res = node->neighbour[i];
node              686 net/rose/rose_route.c 					if (!rose_ftimer_running(node->neighbour[i])) {
node              687 net/rose/rose_route.c 						res = node->neighbour[i];
node               87 net/rxrpc/ar-connection.c 		bundle = rb_entry(p, struct rxrpc_conn_bundle, node);
node              116 net/rxrpc/ar-connection.c 		bundle = rb_entry(parent, struct rxrpc_conn_bundle, node);
node              130 net/rxrpc/ar-connection.c 	rb_link_node(&bundle->node, parent, pp);
node              131 net/rxrpc/ar-connection.c 	rb_insert_color(&bundle->node, &trans->bundles);
node              177 net/rxrpc/ar-connection.c 		rb_erase(&bundle->node, &trans->bundles);
node              247 net/rxrpc/ar-connection.c 		xconn = rb_entry(parent, struct rxrpc_connection, node);
node              263 net/rxrpc/ar-connection.c 	rb_link_node(&conn->node, parent, p);
node              264 net/rxrpc/ar-connection.c 	rb_insert_color(&conn->node, &conn->trans->client_conns);
node              287 net/rxrpc/ar-connection.c 		xconn = rb_entry(parent, struct rxrpc_connection, node);
node              627 net/rxrpc/ar-connection.c 		conn = rb_entry(p, struct rxrpc_connection, node);
node              670 net/rxrpc/ar-connection.c 		conn = rb_entry(p, struct rxrpc_connection, node);
node              687 net/rxrpc/ar-connection.c 	rb_link_node(&conn->node, p, pp);
node              688 net/rxrpc/ar-connection.c 	rb_insert_color(&conn->node, &trans->server_conns);
node              757 net/rxrpc/ar-connection.c 		conn = rb_entry(p, struct rxrpc_connection, node);
node              859 net/rxrpc/ar-connection.c 				rb_erase(&conn->node,
node              862 net/rxrpc/ar-connection.c 				rb_erase(&conn->node,
node              222 net/rxrpc/ar-internal.h 	struct rb_node		node;		/* node in transport's lookup tree */
node              244 net/rxrpc/ar-internal.h 	struct rb_node		node;		/* node in transport's lookup tree */
node              106 net/sched/sch_htb.c 	struct rb_node node[TC_HTB_NUMPRIO];	/* node for self or feed tree */
node              262 net/sched/sch_htb.c 		c = rb_entry(parent, struct htb_class, node[prio]);
node              269 net/sched/sch_htb.c 	rb_link_node(&cl->node[prio], parent, p);
node              270 net/sched/sch_htb.c 	rb_insert_color(&cl->node[prio], root);
node              361 net/sched/sch_htb.c 		if (q->ptr[cl->level][prio] == cl->node + prio)
node              364 net/sched/sch_htb.c 		htb_safe_rb_erase(cl->node + prio, q->row[cl->level] + prio);
node              424 net/sched/sch_htb.c 			if (p->un.inner.ptr[prio] == cl->node + prio) {
node              432 net/sched/sch_htb.c 			htb_safe_rb_erase(cl->node + prio, p->un.inner.feed + prio);
node              735 net/sched/sch_htb.c 		    rb_entry(n, struct htb_class, node[prio]);
node              791 net/sched/sch_htb.c 			cl = rb_entry(*sp->pptr, struct htb_class, node[prio]);
node             1383 net/sched/sch_htb.c 			RB_CLEAR_NODE(&cl->node[prio]);
node              331 net/sctp/endpointola.c 	struct hlist_node *node;
node              341 net/sctp/endpointola.c 	sctp_for_each_hentry(epb, node, &head->chain) {
node              693 net/sctp/input.c 	hlist_add_head(&epb->node, &head->chain);
node              713 net/sctp/input.c 	if (hlist_unhashed(&epb->node))
node              721 net/sctp/input.c 	__hlist_del(&epb->node);
node              739 net/sctp/input.c 	struct hlist_node *node;
node              745 net/sctp/input.c 	sctp_for_each_hentry(epb, node, &head->chain) {
node              773 net/sctp/input.c 	hlist_add_head(&epb->node, &head->chain);
node              802 net/sctp/input.c 	__hlist_del(&epb->node);
node              827 net/sctp/input.c 	struct hlist_node *node;
node              836 net/sctp/input.c 	sctp_for_each_hentry(epb, node, &head->chain) {
node              204 net/sctp/proc.c 	struct hlist_node *node;
node              213 net/sctp/proc.c 	sctp_for_each_hentry(epb, node, &head->chain) {
node              308 net/sctp/proc.c 	struct hlist_node *node;
node              317 net/sctp/proc.c 	sctp_for_each_hentry(epb, node, &head->chain) {
node              420 net/sctp/proc.c 	struct hlist_node *node;
node              430 net/sctp/proc.c 	sctp_for_each_hentry(epb, node, &head->chain) {
node             5551 net/sctp/socket.c 	struct hlist_node *node;
node             5576 net/sctp/socket.c 			sctp_for_each_hentry(pp, node, &head->chain)
node             5603 net/sctp/socket.c 		sctp_for_each_hentry(pp, node, &head->chain) {
node             5619 net/sctp/socket.c 		struct hlist_node *node;
node             5636 net/sctp/socket.c 		sk_for_each_bound(sk2, node, &pp->owner) {
node             5954 net/sctp/socket.c 		hlist_add_head(&pp->node, &head->chain);
node             5963 net/sctp/socket.c 		__hlist_del(&pp->node);
node              125 net/sunrpc/svc.c 	unsigned int node;
node              135 net/sunrpc/svc.c 	node = any_online_node(node_online_map);
node              136 net/sunrpc/svc.c 	if (nr_cpus_node(node) > 2) {
node              209 net/sunrpc/svc.c 	unsigned int node;
node              216 net/sunrpc/svc.c 	for_each_node_with_cpus(node) {
node              219 net/sunrpc/svc.c 		m->to_pool[node] = pidx;
node              220 net/sunrpc/svc.c 		m->pool_to[pidx] = node;
node              304 net/sunrpc/svc.c 	unsigned int node = m->pool_to[pidx];
node              315 net/sunrpc/svc.c 		set_cpus_allowed_ptr(task, &cpumask_of_cpu(node));
node              320 net/sunrpc/svc.c 		node_to_cpumask_ptr(nodecpumask, node);
node              113 net/tipc/bcast.c 	struct tipc_node node;
node              421 net/tipc/bcast.c 	struct tipc_node* node = tipc_node_find(msg_prevnode(msg));
node              428 net/tipc/bcast.c 	if (unlikely(!node || !tipc_node_is_up(node) || !node->bclink.supported ||
node              437 net/tipc/bcast.c 			tipc_node_lock(node);
node              438 net/tipc/bcast.c 			tipc_bclink_acknowledge(node, msg_bcast_ack(msg));
node              439 net/tipc/bcast.c 			tipc_node_unlock(node);
node              442 net/tipc/bcast.c 			bcl->owner->next = node;   /* remember requestor */
node              465 net/tipc/bcast.c 	tipc_node_lock(node);
node              467 net/tipc/bcast.c 	deferred = node->bclink.deferred_head;
node              468 net/tipc/bcast.c 	next_in = mod(node->bclink.last_in + 1);
node              473 net/tipc/bcast.c 		node->bclink.last_in++;
node              474 net/tipc/bcast.c 		bclink_set_gap(node);
node              476 net/tipc/bcast.c 			bclink_send_ack(node);
node              480 net/tipc/bcast.c 			tipc_node_unlock(node);
node              485 net/tipc/bcast.c 			tipc_node_unlock(node);
node              489 net/tipc/bcast.c 			if (tipc_link_recv_fragment(&node->bclink.defragm,
node              492 net/tipc/bcast.c 			tipc_node_unlock(node);
node              495 net/tipc/bcast.c 			tipc_node_unlock(node);
node              499 net/tipc/bcast.c 			tipc_node_lock(node);
node              502 net/tipc/bcast.c 			node->bclink.deferred_head = deferred->next;
node              507 net/tipc/bcast.c 		u32 gap_after = node->bclink.gap_after;
node              508 net/tipc/bcast.c 		u32 gap_to = node->bclink.gap_to;
node              510 net/tipc/bcast.c 		if (tipc_link_defer_pkt(&node->bclink.deferred_head,
node              511 net/tipc/bcast.c 					&node->bclink.deferred_tail,
node              513 net/tipc/bcast.c 			node->bclink.nack_sync++;
node              516 net/tipc/bcast.c 				node->bclink.gap_after = seqno;
node              518 net/tipc/bcast.c 				node->bclink.gap_to = seqno;
node              520 net/tipc/bcast.c 		if (bclink_ack_allowed(node->bclink.nack_sync)) {
node              522 net/tipc/bcast.c 				bclink_send_nack(node);
node              523 net/tipc/bcast.c 			bclink_set_gap(node);
node              529 net/tipc/bcast.c 	tipc_node_unlock(node);
node              797 net/tipc/bcast.c 	spin_lock_init(&bclink->node.lock);
node              798 net/tipc/bcast.c 	bcl->owner = &bclink->node;
node               82 net/tipc/bcast.h 	int n = tipc_node(node);
node               98 net/tipc/bcast.h 	int n = tipc_node(node);
node              645 net/tipc/config.c 	rep_buf = tipc_cfg_do_cmd(orig->node,
node             1273 net/tipc/link.c 	struct tipc_node *node;
node             1287 net/tipc/link.c 	node = tipc_node_select(destaddr, selector);
node             1288 net/tipc/link.c 	if (likely(node)) {
node             1289 net/tipc/link.c 		tipc_node_lock(node);
node             1290 net/tipc/link.c 		l_ptr = node->active_links[selector];
node             1298 net/tipc/link.c 				tipc_node_unlock(node);
node             1323 net/tipc/link.c 			tipc_node_unlock(node);
node             1333 net/tipc/link.c 		tipc_node_unlock(node);
node             1367 net/tipc/link.c 	struct tipc_node *node;
node             1477 net/tipc/link.c 	node = tipc_node_select(destaddr, sender->publ.ref & 1);
node             1478 net/tipc/link.c 	if (likely(node)) {
node             1479 net/tipc/link.c 		tipc_node_lock(node);
node             1480 net/tipc/link.c 		l_ptr = node->active_links[sender->publ.ref & 1];
node             1482 net/tipc/link.c 			tipc_node_unlock(node);
node             1487 net/tipc/link.c 			tipc_node_unlock(node);
node             1525 net/tipc/link.c 	tipc_node_unlock(node);
node             2951 net/tipc/link.c 	*node = tipc_node_find(link_name_parts.addr_peer);
node             2952 net/tipc/link.c 	if (!*node)
node             2955 net/tipc/link.c 	l_ptr = (*node)->links[b_ptr->identity];
node             2968 net/tipc/link.c 	struct tipc_node *node;
node             2986 net/tipc/link.c 	l_ptr = link_find_link(args->name, &node);
node             2992 net/tipc/link.c 	tipc_node_lock(node);
node             3021 net/tipc/link.c 	tipc_node_unlock(node);
node             3046 net/tipc/link.c 	struct tipc_node *node;
node             3059 net/tipc/link.c 	l_ptr = link_find_link(link_name, &node);
node             3065 net/tipc/link.c 	tipc_node_lock(node);
node             3067 net/tipc/link.c 	tipc_node_unlock(node);
node             3094 net/tipc/link.c 	struct tipc_node *node;
node             3104 net/tipc/link.c 	l_ptr = link_find_link(name, &node);
node             3109 net/tipc/link.c 	tipc_node_lock(node);
node             3169 net/tipc/link.c 	tipc_node_unlock(node);
node             3210 net/tipc/link.c 	struct tipc_node * node;
node             3215 net/tipc/link.c 	node = tipc_node_find(a);
node             3216 net/tipc/link.c 	if (node) {
node             3217 net/tipc/link.c 		tipc_node_lock(node);
node             3218 net/tipc/link.c 		l_ptr = node->links[bearer_id];
node             3235 net/tipc/link.c 		tipc_node_unlock(node);
node              182 net/tipc/name_distr.c 			buf = named_prepare_buf(PUBLICATION, left, node);
node              193 net/tipc/name_distr.c 			msg_set_link_selector(buf_msg(buf), node);
node              195 net/tipc/name_distr.c 			    "<%u.%u.%u>\n", tipc_zone(node),
node              196 net/tipc/name_distr.c 			    tipc_cluster(node), tipc_node(node));
node              197 net/tipc/name_distr.c 			tipc_link_send(buf, node, node);
node              225 net/tipc/name_distr.c 				     publ->node, publ->ref, publ->key);
node              231 net/tipc/name_distr.c 		    publ->type, publ->lower, publ->node, publ->ref, publ->key);
node              313 net/tipc/name_distr.c 		if (publ->node == tipc_own_addr)
node              315 net/tipc/name_distr.c 		publ->node = tipc_own_addr;
node              140 net/tipc/name_table.c 	publ->node = node;
node              319 net/tipc/name_table.c 	    type, lower, upper, node, port, sseq,
node              324 net/tipc/name_table.c 	publ = publ_create(type, lower, upper, scope, node, port, key);
node              328 net/tipc/name_table.c 	    publ, node, publ->node, publ->subscr.node);
node              338 net/tipc/name_table.c 	if (in_own_cluster(node)) {
node              349 net/tipc/name_table.c 	if (node == tipc_own_addr) {
node              369 net/tipc/name_table.c 					   publ->node,
node              409 net/tipc/name_table.c 	       (publ->node && (publ->node != node))) {
node              431 net/tipc/name_table.c 	if (in_own_cluster(node)) {
node              443 net/tipc/name_table.c 				    publ->type, publ->lower, publ->node,
node              462 net/tipc/name_table.c 	if (node == tipc_own_addr) {
node              474 net/tipc/name_table.c 				    publ->type, publ->lower, publ->node,
node              507 net/tipc/name_table.c 					   publ->node,
node              541 net/tipc/name_table.c 							   crs->node,
node              583 net/tipc/name_table.c 	dbg("Publishing {%u,%u,%u} from 0x%x\n", type, lower, upper, node);
node              592 net/tipc/name_table.c 					scope, node, port, key);
node              604 net/tipc/name_table.c 	dbg("Withdrawing {%u,%u} from 0x%x\n", type, lower, node);
node              605 net/tipc/name_table.c 	publ = tipc_nameseq_remove_publ(seq, lower, node, ref, key);
node              649 net/tipc/name_table.c 			*destnode = publ->node;
node              892 net/tipc/name_table.c 			 tipc_zone(publ->node), tipc_cluster(publ->node),
node              893 net/tipc/name_table.c 			 tipc_node(publ->node), publ->ref);
node              896 net/tipc/name_table.c 			if (publ->node != tipc_own_addr)
node               76 net/tipc/name_table.h 	u32 node;
node              456 net/tipc/node.c 		ns->node = NULL;
node               51 net/tipc/node_subscr.c 		node_sub->node = NULL;
node               55 net/tipc/node_subscr.c 	node_sub->node = tipc_node_find(addr);
node               56 net/tipc/node_subscr.c 	if (!node_sub->node) {
node               63 net/tipc/node_subscr.c 	tipc_node_lock(node_sub->node);
node               64 net/tipc/node_subscr.c 	list_add_tail(&node_sub->nodesub_list, &node_sub->node->nsub);
node               65 net/tipc/node_subscr.c 	tipc_node_unlock(node_sub->node);
node               74 net/tipc/node_subscr.c 	if (!node_sub->node)
node               77 net/tipc/node_subscr.c 	tipc_node_lock(node_sub->node);
node               79 net/tipc/node_subscr.c 	tipc_node_unlock(node_sub->node);
node               53 net/tipc/node_subscr.h 	struct tipc_node *node;
node              821 net/tipc/port.c 		orig.node = msg_orignode(msg);
node             1077 net/tipc/port.c 	id->node = tipc_own_addr;
node             1202 net/tipc/port.c 	msg_set_destnode(msg, peer->node);
node             1208 net/tipc/port.c 	if (!may_route(peer->node))
node             1218 net/tipc/port.c 	tipc_nodesub_subscribe(&p_ptr->subscription,peer->node,
node             1224 net/tipc/port.c 	p_ptr->publ.max_pkt = tipc_link_get_max_pkt(peer->node, ref);
node             1321 net/tipc/port.c 		peer->node = port_peernode(p_ptr);
node             1465 net/tipc/port.c 	msg_set_orignode(msg, orig->node);
node             1508 net/tipc/port.c 	orig.node = tipc_own_addr;
node             1539 net/tipc/port.c 	msg_set_orignode(msg, orig->node);
node             1582 net/tipc/port.c 	orig.node = tipc_own_addr;
node             1608 net/tipc/port.c 	msg_set_orignode(msg, orig->node);
node             1610 net/tipc/port.c 	msg_set_destnode(msg, dest->node);
node             1616 net/tipc/port.c 	if (dest->node == tipc_own_addr)
node             1618 net/tipc/port.c 	res = tipc_link_send_sections_fast(p_ptr, msg_sect, num_sect, dest->node);
node             1640 net/tipc/port.c 	orig.node = tipc_own_addr;
node             1665 net/tipc/port.c 	msg_set_orignode(msg, orig->node);
node             1667 net/tipc/port.c 	msg_set_destnode(msg, dest->node);
node             1680 net/tipc/port.c 	if (dest->node == tipc_own_addr)
node             1682 net/tipc/port.c 	res = tipc_send_buf_fast(buf, dest->node);
node             1702 net/tipc/port.c 	orig.node = tipc_own_addr;
node              401 net/tipc/socket.c 		addr->addr.id.node = tsock->peer_name.node;
node              780 net/tipc/socket.c 	tsock->peer_name.node = msg_orignode(msg);
node              803 net/tipc/socket.c 		addr->addr.id.node = msg_orignode(msg);
node             1551 net/tipc/socket.c 		new_tsock->peer_name.node = msg_orignode(msg);
node              114 net/tipc/subscr.c 	sub->evt.port.node = htohl(node, sub->swap);
node              157 net/tipc/subscr.c 	sub->event_cb(sub, found_lower, found_upper, event, port_ref, node);
node              253 net/unix/af_unix.c 	struct hlist_node *node;
node              255 net/unix/af_unix.c 	sk_for_each(s, node, &unix_socket_table[hash ^ type]) {
node              288 net/unix/af_unix.c 	struct hlist_node *node;
node              291 net/unix/af_unix.c 	sk_for_each(s, node,
node              174 net/x25/af_x25.c 	struct hlist_node *node;
node              178 net/x25/af_x25.c 	sk_for_each(s, node, &x25_list)
node              246 net/x25/af_x25.c 	struct hlist_node *node;
node              251 net/x25/af_x25.c 	sk_for_each(s, node, &x25_list)
node              288 net/x25/af_x25.c 	struct hlist_node *node;
node              290 net/x25/af_x25.c 	sk_for_each(s, node, &x25_list)
node             1623 net/x25/af_x25.c 	struct hlist_node *node;
node             1627 net/x25/af_x25.c 	sk_for_each(s, node, &x25_list)
node               51 net/x25/x25_forward.c 		x25_frwd = list_entry(entry, struct x25_forward, node);
node               70 net/x25/x25_forward.c 		list_add(&new_frwd->node, &x25_forward_list);
node              104 net/x25/x25_forward.c 		frwd = list_entry(entry, struct x25_forward, node);
node              141 net/x25/x25_forward.c 		fwd = list_entry(entry, struct x25_forward, node);
node              143 net/x25/x25_forward.c 			list_del(&fwd->node);
node              159 net/x25/x25_forward.c 		fwd = list_entry(entry, struct x25_forward, node);
node              161 net/x25/x25_forward.c 			list_del(&fwd->node);
node              267 net/x25/x25_link.c 	list_add(&nb->node, &x25_neigh_list);
node              283 net/x25/x25_link.c 	if (nb->node.next) {
node              284 net/x25/x25_link.c 		list_del(&nb->node);
node              300 net/x25/x25_link.c 		nb = list_entry(entry, struct x25_neigh, node);
node              321 net/x25/x25_link.c 		nb = list_entry(entry, struct x25_neigh, node);
node              394 net/x25/x25_link.c 		nb = list_entry(entry, struct x25_neigh, node);
node               34 net/x25/x25_proc.c 		rt = list_entry(route_entry, struct x25_route, node);
node               61 net/x25/x25_proc.c 					struct x25_route, node);
node               65 net/x25/x25_proc.c 	if (rt->node.next != &x25_route_list)
node               66 net/x25/x25_proc.c 		rt = list_entry(rt->node.next, struct x25_route, node);
node               99 net/x25/x25_proc.c 	struct hlist_node *node;
node              101 net/x25/x25_proc.c 	sk_for_each(s, node, &x25_list)
node              179 net/x25/x25_proc.c 		f = list_entry(entry, struct x25_forward, node);
node              207 net/x25/x25_proc.c 					struct x25_forward, node);
node              211 net/x25/x25_proc.c 	if (f->node.next != &x25_forward_list)
node              212 net/x25/x25_proc.c 		f = list_entry(f->node.next, struct x25_forward, node);
node               40 net/x25/x25_route.c 		rt = list_entry(entry, struct x25_route, node);
node               59 net/x25/x25_route.c 	list_add(&rt->node, &x25_route_list);
node               75 net/x25/x25_route.c 	if (rt->node.next) {
node               76 net/x25/x25_route.c 		list_del(&rt->node);
node               91 net/x25/x25_route.c 		rt = list_entry(entry, struct x25_route, node);
node              116 net/x25/x25_route.c 		rt = list_entry(entry, struct x25_route, node);
node              159 net/x25/x25_route.c 		rt = list_entry(entry, struct x25_route, node);
node              219 net/x25/x25_route.c 		rt = list_entry(entry, struct x25_route, node);
node              203 scripts/genksyms/genksyms.c 	free(node->string);
node              204 scripts/genksyms/genksyms.c 	free(node);
node              221 scripts/genksyms/genksyms.c 	newnode->string = xstrdup(node->string);
node              222 scripts/genksyms/genksyms.c 	newnode->tag = node->tag;
node               40 scripts/genksyms/parse.y   struct string_list *node = *p;
node               41 scripts/genksyms/parse.y   *p = node->next;
node               42 scripts/genksyms/parse.y   free_node(node);
node             1295 scripts/kconfig/gconf.c 	gtk_tree_store_set(tree, node,
node             1321 scripts/kconfig/gconf.c 	GtkTreeIter *node = parents[indent];
node             1323 scripts/kconfig/gconf.c 	gtk_tree_store_append(tree, node, parent);
node             1324 scripts/kconfig/gconf.c 	set_node(node, menu, row);
node              174 scripts/kconfig/lxdialog/dialog.h 	struct dialog_item node;
node              561 scripts/kconfig/lxdialog/util.c 	vsnprintf(item_cur->node.str, sizeof(item_cur->node.str), fmt, ap);
node              570 scripts/kconfig/lxdialog/util.c 	avail = sizeof(item_cur->node.str) - strlen(item_cur->node.str);
node              573 scripts/kconfig/lxdialog/util.c 	vsnprintf(item_cur->node.str + strlen(item_cur->node.str),
node              575 scripts/kconfig/lxdialog/util.c 	item_cur->node.str[sizeof(item_cur->node.str) - 1] = '\0';
node              581 scripts/kconfig/lxdialog/util.c 	item_cur->node.tag = tag;
node              585 scripts/kconfig/lxdialog/util.c 	item_cur->node.data = ptr;
node              590 scripts/kconfig/lxdialog/util.c 	item_cur->node.selected = val;
node              603 scripts/kconfig/lxdialog/util.c 	return item_cur->node.data;
node              608 scripts/kconfig/lxdialog/util.c 	return item_cur->node.tag;
node              644 scripts/kconfig/lxdialog/util.c 	return item_cur->node.str;
node              649 scripts/kconfig/lxdialog/util.c 	return (item_cur->node.selected != 0);
node              654 scripts/kconfig/lxdialog/util.c 	return (item_cur->node.tag == tag);
node               49 security/keys/internal.h 	struct rb_node		node;
node               76 security/keys/key.c 		user = rb_entry(parent, struct key_user, node);
node              114 security/keys/key.c 	rb_link_node(&candidate->node, parent, p);
node              115 security/keys/key.c 	rb_insert_color(&candidate->node, &key_user_tree);
node              137 security/keys/key.c 		rb_erase(&user->node, &key_user_tree);
node              999 security/keys/key.c 	rb_link_node(&root_key_user.node,
node             1003 security/keys/key.c 	rb_insert_color(&root_key_user.node,
node              244 security/keys/proc.c 	struct key_user *user = rb_entry(_p, struct key_user, node);
node              250 security/selinux/avc.c 	struct avc_node *node;
node              260 security/selinux/avc.c 			list_for_each_entry_rcu(node, &avc_cache.slots[i], list)
node              277 security/selinux/avc.c 	struct avc_node *node = container_of(rhead, struct avc_node, rhead);
node              278 security/selinux/avc.c 	kmem_cache_free(avc_node_cachep, node);
node              284 security/selinux/avc.c 	list_del_rcu(&node->list);
node              285 security/selinux/avc.c 	call_rcu(&node->rhead, avc_node_free);
node              291 security/selinux/avc.c 	kmem_cache_free(avc_node_cachep, node);
node              305 security/selinux/avc.c 	struct avc_node *node;
node              316 security/selinux/avc.c 		list_for_each_entry(node, &avc_cache.slots[hvalue], list) {
node              317 security/selinux/avc.c 			if (atomic_dec_and_test(&node->ae.used)) {
node              319 security/selinux/avc.c 				avc_node_delete(node);
node              338 security/selinux/avc.c 	struct avc_node *node;
node              340 security/selinux/avc.c 	node = kmem_cache_zalloc(avc_node_cachep, GFP_ATOMIC);
node              341 security/selinux/avc.c 	if (!node)
node              344 security/selinux/avc.c 	INIT_RCU_HEAD(&node->rhead);
node              345 security/selinux/avc.c 	INIT_LIST_HEAD(&node->list);
node              346 security/selinux/avc.c 	atomic_set(&node->ae.used, 1);
node              353 security/selinux/avc.c 	return node;
node              358 security/selinux/avc.c 	node->ae.ssid = ssid;
node              359 security/selinux/avc.c 	node->ae.tsid = tsid;
node              360 security/selinux/avc.c 	node->ae.tclass = tclass;
node              361 security/selinux/avc.c 	memcpy(&node->ae.avd, &ae->avd, sizeof(node->ae.avd));
node              366 security/selinux/avc.c 	struct avc_node *node, *ret = NULL;
node              370 security/selinux/avc.c 	list_for_each_entry_rcu(node, &avc_cache.slots[hvalue], list) {
node              371 security/selinux/avc.c 		if (ssid == node->ae.ssid &&
node              372 security/selinux/avc.c 		    tclass == node->ae.tclass &&
node              373 security/selinux/avc.c 		    tsid == node->ae.tsid) {
node              374 security/selinux/avc.c 			ret = node;
node              407 security/selinux/avc.c 	struct avc_node *node;
node              410 security/selinux/avc.c 	node = avc_search_node(ssid, tsid, tclass);
node              412 security/selinux/avc.c 	if (node && ((node->ae.avd.decided & requested) == requested)) {
node              417 security/selinux/avc.c 	node = NULL;
node              420 security/selinux/avc.c 	return node;
node              464 security/selinux/avc.c 	struct avc_node *pos, *node = NULL;
node              471 security/selinux/avc.c 	node = avc_alloc_node();
node              472 security/selinux/avc.c 	if (node) {
node              474 security/selinux/avc.c 		avc_node_populate(node, ssid, tsid, tclass, ae);
node              481 security/selinux/avc.c 				avc_node_replace(node, pos);
node              485 security/selinux/avc.c 		list_add_rcu(&node->list, &avc_cache.slots[hvalue]);
node              490 security/selinux/avc.c 	return node;
node              753 security/selinux/avc.c 	struct avc_node *pos, *node, *orig = NULL;
node              755 security/selinux/avc.c 	node = avc_alloc_node();
node              756 security/selinux/avc.c 	if (!node) {
node              776 security/selinux/avc.c 		avc_node_kill(node);
node              784 security/selinux/avc.c 	avc_node_populate(node, ssid, tsid, tclass, &orig->ae);
node              788 security/selinux/avc.c 		node->ae.avd.allowed |= perms;
node              792 security/selinux/avc.c 		node->ae.avd.allowed &= ~perms;
node              795 security/selinux/avc.c 		node->ae.avd.auditallow |= perms;
node              798 security/selinux/avc.c 		node->ae.avd.auditallow &= ~perms;
node              801 security/selinux/avc.c 		node->ae.avd.auditdeny |= perms;
node              804 security/selinux/avc.c 		node->ae.avd.auditdeny &= ~perms;
node              807 security/selinux/avc.c 	avc_node_replace(node, orig);
node              823 security/selinux/avc.c 	struct avc_node *node;
node              832 security/selinux/avc.c 		list_for_each_entry(node, &avc_cache.slots[i], list)
node              833 security/selinux/avc.c 			avc_node_delete(node);
node              878 security/selinux/avc.c 	struct avc_node *node;
node              887 security/selinux/avc.c 	node = avc_lookup(ssid, tsid, tclass, requested);
node              888 security/selinux/avc.c 	if (!node) {
node              894 security/selinux/avc.c 		node = avc_insert(ssid, tsid, tclass, &entry);
node              897 security/selinux/avc.c 	p_ae = node ? &node->ae : &entry;
node               82 security/selinux/netnode.c 	struct sel_netnode *node = container_of(p, struct sel_netnode, rcu);
node               83 security/selinux/netnode.c 	kfree(node);
node              132 security/selinux/netnode.c 	struct sel_netnode *node;
node              145 security/selinux/netnode.c 	list_for_each_entry_rcu(node, &sel_netnode_hash[idx].list, list)
node              146 security/selinux/netnode.c 		if (node->nsec.family == family)
node              149 security/selinux/netnode.c 				if (node->nsec.addr.ipv4 == *(__be32 *)addr)
node              150 security/selinux/netnode.c 					return node;
node              153 security/selinux/netnode.c 				if (ipv6_addr_equal(&node->nsec.addr.ipv6,
node              155 security/selinux/netnode.c 					return node;
node              174 security/selinux/netnode.c 	switch (node->nsec.family) {
node              176 security/selinux/netnode.c 		idx = sel_netnode_hashfn_ipv4(node->nsec.addr.ipv4);
node              179 security/selinux/netnode.c 		idx = sel_netnode_hashfn_ipv6(&node->nsec.addr.ipv6);
node              185 security/selinux/netnode.c 	INIT_RCU_HEAD(&node->rcu);
node              189 security/selinux/netnode.c 	list_add_rcu(&node->list, &sel_netnode_hash[idx].list);
node              217 security/selinux/netnode.c 	struct sel_netnode *node;
node              221 security/selinux/netnode.c 	node = sel_netnode_find(addr, family);
node              222 security/selinux/netnode.c 	if (node != NULL) {
node              223 security/selinux/netnode.c 		*sid = node->nsec.sid;
node              278 security/selinux/netnode.c 	struct sel_netnode *node;
node              281 security/selinux/netnode.c 	node = sel_netnode_find(addr, family);
node              282 security/selinux/netnode.c 	if (node != NULL) {
node              283 security/selinux/netnode.c 		*sid = node->nsec.sid;
node              302 security/selinux/netnode.c 	struct sel_netnode *node, *node_tmp;
node              306 security/selinux/netnode.c 		list_for_each_entry_safe(node, node_tmp,
node              308 security/selinux/netnode.c 				list_del_rcu(&node->list);
node              309 security/selinux/netnode.c 				call_rcu(&node->rcu, sel_netnode_free);
node             1012 security/selinux/selinuxfs.c 	struct list_head *node;
node             1015 security/selinux/selinuxfs.c 	node = de->d_subdirs.next;
node             1016 security/selinux/selinuxfs.c 	while (node != &de->d_subdirs) {
node             1017 security/selinux/selinuxfs.c 		struct dentry *d = list_entry(node, struct dentry, d_u.d_child);
node             1018 security/selinux/selinuxfs.c 		list_del_init(node);
node             1028 security/selinux/selinuxfs.c 		node = de->d_subdirs.next;
node              198 security/selinux/ss/avtab.c 	if (!node)
node              202 security/selinux/ss/avtab.c 	for (cur = node->next; cur; cur = cur->next) {
node              203 security/selinux/ss/avtab.c 		if (node->key.source_type == cur->key.source_type &&
node              204 security/selinux/ss/avtab.c 		    node->key.target_type == cur->key.target_type &&
node              205 security/selinux/ss/avtab.c 		    node->key.target_class == cur->key.target_class &&
node              209 security/selinux/ss/avtab.c 		if (node->key.source_type < cur->key.source_type)
node              211 security/selinux/ss/avtab.c 		if (node->key.source_type == cur->key.source_type &&
node              212 security/selinux/ss/avtab.c 		    node->key.target_type < cur->key.target_type)
node              214 security/selinux/ss/avtab.c 		if (node->key.source_type == cur->key.source_type &&
node              215 security/selinux/ss/avtab.c 		    node->key.target_type == cur->key.target_type &&
node              216 security/selinux/ss/avtab.c 		    node->key.target_class < cur->key.target_class)
node               94 security/selinux/ss/conditional.c 	new_state = cond_evaluate_expr(p, node->expr);
node               95 security/selinux/ss/conditional.c 	if (new_state != node->cur_state) {
node               96 security/selinux/ss/conditional.c 		node->cur_state = new_state;
node              100 security/selinux/ss/conditional.c 		for (cur = node->true_list; cur; cur = cur->next) {
node              102 security/selinux/ss/conditional.c 				cur->node->key.specified &= ~AVTAB_ENABLED;
node              104 security/selinux/ss/conditional.c 				cur->node->key.specified |= AVTAB_ENABLED;
node              107 security/selinux/ss/conditional.c 		for (cur = node->false_list; cur; cur = cur->next) {
node              110 security/selinux/ss/conditional.c 				cur->node->key.specified &= ~AVTAB_ENABLED;
node              112 security/selinux/ss/conditional.c 				cur->node->key.specified |= AVTAB_ENABLED;
node              142 security/selinux/ss/conditional.c 	for (cur_expr = node->expr; cur_expr; cur_expr = next_expr) {
node              146 security/selinux/ss/conditional.c 	cond_av_list_destroy(node->true_list);
node              147 security/selinux/ss/conditional.c 	cond_av_list_destroy(node->false_list);
node              148 security/selinux/ss/conditional.c 	kfree(node);
node              295 security/selinux/ss/conditional.c 					if (cur->node == node_ptr) {
node              323 security/selinux/ss/conditional.c 	list->node = node_ptr;
node              396 security/selinux/ss/conditional.c 	node->cur_state = le32_to_cpu(buf[0]);
node              424 security/selinux/ss/conditional.c 			node->expr = expr;
node              430 security/selinux/ss/conditional.c 	if (cond_read_av_list(p, fp, &node->true_list, NULL) != 0)
node              432 security/selinux/ss/conditional.c 	if (cond_read_av_list(p, fp, &node->false_list, node->true_list) != 0)
node              436 security/selinux/ss/conditional.c 	cond_node_destroy(node);
node              442 security/selinux/ss/conditional.c 	struct cond_node *node, *last = NULL;
node              458 security/selinux/ss/conditional.c 		node = kzalloc(sizeof(struct cond_node), GFP_KERNEL);
node              459 security/selinux/ss/conditional.c 		if (!node)
node              462 security/selinux/ss/conditional.c 		if (cond_read_node(p, node, fp) != 0)
node              466 security/selinux/ss/conditional.c 			p->cond_list = node;
node              468 security/selinux/ss/conditional.c 			last->next = node;
node              469 security/selinux/ss/conditional.c 		last = node;
node              483 security/selinux/ss/conditional.c 	struct avtab_node *node;
node              488 security/selinux/ss/conditional.c 	for (node = avtab_search_node(ctab, key); node;
node              489 security/selinux/ss/conditional.c 				node = avtab_search_node_next(node, key->specified)) {
node              491 security/selinux/ss/conditional.c 		    (node->key.specified & (AVTAB_ALLOWED|AVTAB_ENABLED)))
node              492 security/selinux/ss/conditional.c 			avd->allowed |= node->datum.data;
node              494 security/selinux/ss/conditional.c 		    (node->key.specified & (AVTAB_AUDITDENY|AVTAB_ENABLED)))
node              500 security/selinux/ss/conditional.c 			avd->auditdeny &= node->datum.data;
node              502 security/selinux/ss/conditional.c 		    (node->key.specified & (AVTAB_AUDITALLOW|AVTAB_ENABLED)))
node              503 security/selinux/ss/conditional.c 			avd->auditallow |= node->datum.data;
node               43 security/selinux/ss/conditional.h 	struct avtab_node *node;
node               32 security/selinux/ss/ebitmap.c 	n1 = e1->node;
node               33 security/selinux/ss/ebitmap.c 	n2 = e2->node;
node               52 security/selinux/ss/ebitmap.c 	n = src->node;
node               66 security/selinux/ss/ebitmap.c 			dst->node = new;
node               89 security/selinux/ss/ebitmap.c 	struct ebitmap_node *e_iter = ebmap->node;
node              186 security/selinux/ss/ebitmap.c 					ebmap->node = e_iter;
node              222 security/selinux/ss/ebitmap.c 	n1 = e1->node;
node              223 security/selinux/ss/ebitmap.c 	n2 = e2->node;
node              251 security/selinux/ss/ebitmap.c 	n = e->node;
node              266 security/selinux/ss/ebitmap.c 	n = e->node;
node              295 security/selinux/ss/ebitmap.c 					e->node = n->next;
node              322 security/selinux/ss/ebitmap.c 		new->next = e->node;
node              323 security/selinux/ss/ebitmap.c 		e->node = new;
node              336 security/selinux/ss/ebitmap.c 	n = e->node;
node              344 security/selinux/ss/ebitmap.c 	e->node = NULL;
node              378 security/selinux/ss/ebitmap.c 		e->node = NULL;
node              417 security/selinux/ss/ebitmap.c 				e->node = tmp;
node               34 security/selinux/ss/ebitmap.h 	struct ebitmap_node *node;	/* first node in the bitmap */
node               39 security/selinux/ss/ebitmap.h #define ebitmap_startbit(e) ((e)->node ? (e)->node->startbit : 0)
node               46 security/selinux/ss/ebitmap.h 	for (*n = e->node; *n; *n = (*n)->next) {
node               78 security/selinux/ss/ebitmap.h 	(((bit) - (node)->startbit) / EBITMAP_UNIT_SIZE)
node               80 security/selinux/ss/ebitmap.h 	(((bit) - (node)->startbit) % EBITMAP_UNIT_SIZE)
node               40 security/selinux/ss/mls.c 	struct ebitmap_node *node;
node               54 security/selinux/ss/mls.c 		ebitmap_for_each_positive_bit(e, node, i) {
node               94 security/selinux/ss/mls.c 	struct ebitmap_node *node;
node              113 security/selinux/ss/mls.c 		ebitmap_for_each_positive_bit(e, node, i) {
node              163 security/selinux/ss/mls.c 	struct ebitmap_node *node;
node              173 security/selinux/ss/mls.c 	ebitmap_for_each_positive_bit(&l->cat, node, i) {
node              477 security/selinux/ss/mls.c 	struct ebitmap_node *node;
node              492 security/selinux/ss/mls.c 		ebitmap_for_each_positive_bit(&c->range.level[l].cat, node, i) {
node             1522 security/selinux/ss/policydb.c 		struct ebitmap_node *node;
node             1533 security/selinux/ss/policydb.c 		ebitmap_for_each_positive_bit(&user->roles, node, bit) {
node             1559 security/selinux/ss/policydb.c 		struct ebitmap_node *node;
node             1570 security/selinux/ss/policydb.c 		ebitmap_for_each_positive_bit(&role->types, node, bit) {
node             1929 security/selinux/ss/policydb.c 				c->u.node.addr = nodebuf[0]; /* network order */
node             1930 security/selinux/ss/policydb.c 				c->u.node.mask = nodebuf[1]; /* network order */
node              146 security/selinux/ss/policydb.h 		} node;		/* node information */
node              388 security/selinux/ss/services.c 	struct avtab_node *node;
node              457 security/selinux/ss/services.c 			for (node = avtab_search_node(&policydb.te_avtab, &avkey);
node              458 security/selinux/ss/services.c 			     node;
node              459 security/selinux/ss/services.c 			     node = avtab_search_node_next(node, avkey.specified)) {
node              460 security/selinux/ss/services.c 				if (node->key.specified == AVTAB_ALLOWED)
node              461 security/selinux/ss/services.c 					avd->allowed |= node->datum.data;
node              462 security/selinux/ss/services.c 				else if (node->key.specified == AVTAB_AUDITALLOW)
node              463 security/selinux/ss/services.c 					avd->auditallow |= node->datum.data;
node              464 security/selinux/ss/services.c 				else if (node->key.specified == AVTAB_AUDITDENY)
node              465 security/selinux/ss/services.c 					avd->auditdeny &= node->datum.data;
node             1139 security/selinux/ss/services.c 	struct avtab_node *node;
node             1209 security/selinux/ss/services.c 		node = avtab_search_node(&policydb.te_cond_avtab, &avkey);
node             1210 security/selinux/ss/services.c 		for (; node; node = avtab_search_node_next(node, specified)) {
node             1211 security/selinux/ss/services.c 			if (node->key.specified & AVTAB_ENABLED) {
node             1212 security/selinux/ss/services.c 				avdatum = &node->datum;
node             1872 security/selinux/ss/services.c 			if (c->u.node.addr == (addr & c->u.node.mask))
node               71 sound/aoa/aoa-gpio.h 	struct device_node *node;
node               36 sound/aoa/aoa.h 	struct device_node *node;
node             1036 sound/aoa/codecs/snd-aoa-codec-onyx.c 	onyx->codec.node = of_node_get(node);
node             1091 sound/aoa/codecs/snd-aoa-codec-onyx.c 	of_node_put(onyx->codec.node);
node              915 sound/aoa/codecs/snd-aoa-codec-tas.c 	tas->codec.node = of_node_get(node);
node              922 sound/aoa/codecs/snd-aoa-codec-tas.c 	       addr, node->full_name);
node              982 sound/aoa/codecs/snd-aoa-codec-tas.c 	of_node_put(tas->codec.node);
node               20 sound/aoa/core/snd-aoa-gpio-pmf.c 	rc = pmf_call_function(rt->node, #name "-mute", &args);	\
node               43 sound/aoa/core/snd-aoa-gpio-pmf.c 	rc = pmf_call_function(rt->node, "hw-reset", &args);
node              188 sound/aoa/core/snd-aoa-gpio-pmf.c 		err = pmf_register_irq_client(rt->node,
node              229 sound/aoa/core/snd-aoa-gpio-pmf.c 	err = pmf_call_function(rt->node, name, &args);
node              724 sound/aoa/fabrics/snd-aoa-fabric-layout.c 	if (codec->node && (strcmp(codec->node->name, "codec") == 0)) {
node              733 sound/aoa/fabrics/snd-aoa-fabric-layout.c 		if (*ref != codec->node->linux_phandle) {
node              951 sound/aoa/fabrics/snd-aoa-fabric-layout.c 	while ((sound = of_get_next_child(sdev->ofdev.node, sound))) {
node              977 sound/aoa/fabrics/snd-aoa-fabric-layout.c 	ldev->gpio.node = sound->parent;
node               77 sound/aoa/soundbus/core.c 	retval = add_uevent_var(env, "OF_NAME=%s", of->node->name);
node               81 sound/aoa/soundbus/core.c 	retval = add_uevent_var(env, "OF_TYPE=%s", of->node->type);
node               89 sound/aoa/soundbus/core.c 	compat = of_get_property(of->node, "compatible", &cplen);
node              172 sound/aoa/soundbus/core.c 	    !dev->ofdev.node ||
node               44 sound/aoa/soundbus/i2sbus/i2sbus-control.c 	np = i2sdev->sound.ofdev.node;
node              211 sound/aoa/soundbus/i2sbus/i2sbus-core.c 	dev->sound.ofdev.node = np;
node              336 sound/aoa/soundbus/i2sbus/i2sbus-core.c 	while ((np = of_get_next_child(dev->ofdev.node, np))) {
node               12 sound/aoa/soundbus/sysfs.c 	return sprintf (buf, format_string, mdev->ofdev.node->field);	\
node               28 sound/aoa/soundbus/sysfs.c 				 of->node->name, of->node->type);
node               50 sound/pci/hda/hda_generic.c 	struct hda_gnode *node;	/* Node for PCM volume */
node               84 sound/pci/hda/hda_generic.c #define defcfg_type(node) (((node)->def_cfg & AC_DEFCFG_DEVICE) >> \
node               86 sound/pci/hda/hda_generic.c #define defcfg_location(node) (((node)->def_cfg & AC_DEFCFG_LOCATION) >> \
node               88 sound/pci/hda/hda_generic.c #define defcfg_port_conn(node) (((node)->def_cfg & AC_DEFCFG_PORT_CONN) >> \
node               97 sound/pci/hda/hda_generic.c 	struct hda_gnode *node, *n;
node              102 sound/pci/hda/hda_generic.c 	list_for_each_entry_safe(node, n, &spec->nid_list, list) {
node              103 sound/pci/hda/hda_generic.c 		if (node->conn_list != node->slist)
node              104 sound/pci/hda/hda_generic.c 			kfree(node->conn_list);
node              105 sound/pci/hda/hda_generic.c 		kfree(node);
node              116 sound/pci/hda/hda_generic.c 	struct hda_gnode *node;
node              120 sound/pci/hda/hda_generic.c 	node = kzalloc(sizeof(*node), GFP_KERNEL);
node              121 sound/pci/hda/hda_generic.c 	if (node == NULL)
node              123 sound/pci/hda/hda_generic.c 	node->nid = nid;
node              127 sound/pci/hda/hda_generic.c 		kfree(node);
node              130 sound/pci/hda/hda_generic.c 	if (nconns <= ARRAY_SIZE(node->slist))
node              131 sound/pci/hda/hda_generic.c 		node->conn_list = node->slist;
node              133 sound/pci/hda/hda_generic.c 		node->conn_list = kmalloc(sizeof(hda_nid_t) * nconns,
node              135 sound/pci/hda/hda_generic.c 		if (! node->conn_list) {
node              137 sound/pci/hda/hda_generic.c 			kfree(node);
node              141 sound/pci/hda/hda_generic.c 	memcpy(node->conn_list, conn_list, nconns * sizeof(hda_nid_t));
node              142 sound/pci/hda/hda_generic.c 	node->nconns = nconns;
node              143 sound/pci/hda/hda_generic.c 	node->wid_caps = get_wcaps(codec, nid);
node              144 sound/pci/hda/hda_generic.c 	node->type = (node->wid_caps & AC_WCAP_TYPE) >> AC_WCAP_TYPE_SHIFT;
node              146 sound/pci/hda/hda_generic.c 	if (node->type == AC_WID_PIN) {
node              147 sound/pci/hda/hda_generic.c 		node->pin_caps = snd_hda_param_read(codec, node->nid, AC_PAR_PIN_CAP);
node              148 sound/pci/hda/hda_generic.c 		node->pin_ctl = snd_hda_codec_read(codec, node->nid, 0, AC_VERB_GET_PIN_WIDGET_CONTROL, 0);
node              149 sound/pci/hda/hda_generic.c 		node->def_cfg = snd_hda_codec_read(codec, node->nid, 0, AC_VERB_GET_CONFIG_DEFAULT, 0);
node              152 sound/pci/hda/hda_generic.c 	if (node->wid_caps & AC_WCAP_OUT_AMP) {
node              153 sound/pci/hda/hda_generic.c 		if (node->wid_caps & AC_WCAP_AMP_OVRD)
node              154 sound/pci/hda/hda_generic.c 			node->amp_out_caps = snd_hda_param_read(codec, node->nid, AC_PAR_AMP_OUT_CAP);
node              155 sound/pci/hda/hda_generic.c 		if (! node->amp_out_caps)
node              156 sound/pci/hda/hda_generic.c 			node->amp_out_caps = spec->def_amp_out_caps;
node              158 sound/pci/hda/hda_generic.c 	if (node->wid_caps & AC_WCAP_IN_AMP) {
node              159 sound/pci/hda/hda_generic.c 		if (node->wid_caps & AC_WCAP_AMP_OVRD)
node              160 sound/pci/hda/hda_generic.c 			node->amp_in_caps = snd_hda_param_read(codec, node->nid, AC_PAR_AMP_IN_CAP);
node              161 sound/pci/hda/hda_generic.c 		if (! node->amp_in_caps)
node              162 sound/pci/hda/hda_generic.c 			node->amp_in_caps = spec->def_amp_in_caps;
node              164 sound/pci/hda/hda_generic.c 	list_add_tail(&node->list, &spec->nid_list);
node              205 sound/pci/hda/hda_generic.c 	struct hda_gnode *node;
node              207 sound/pci/hda/hda_generic.c 	list_for_each_entry(node, &spec->nid_list, list) {
node              208 sound/pci/hda/hda_generic.c 		if (node->nid == nid)
node              209 sound/pci/hda/hda_generic.c 			return node;
node              220 sound/pci/hda/hda_generic.c 	snd_printdd("UNMUTE OUT: NID=0x%x\n", node->nid);
node              221 sound/pci/hda/hda_generic.c 	val = (node->amp_out_caps & AC_AMPCAP_NUM_STEPS) >> AC_AMPCAP_NUM_STEPS_SHIFT;
node              222 sound/pci/hda/hda_generic.c 	ofs = (node->amp_out_caps & AC_AMPCAP_OFFSET) >> AC_AMPCAP_OFFSET_SHIFT;
node              225 sound/pci/hda/hda_generic.c 	snd_hda_codec_amp_stereo(codec, node->nid, HDA_OUTPUT, 0, 0xff, val);
node              235 sound/pci/hda/hda_generic.c 	snd_printdd("UNMUTE IN: NID=0x%x IDX=0x%x\n", node->nid, index);
node              236 sound/pci/hda/hda_generic.c 	val = (node->amp_in_caps & AC_AMPCAP_NUM_STEPS) >> AC_AMPCAP_NUM_STEPS_SHIFT;
node              237 sound/pci/hda/hda_generic.c 	ofs = (node->amp_in_caps & AC_AMPCAP_OFFSET) >> AC_AMPCAP_OFFSET_SHIFT;
node              240 sound/pci/hda/hda_generic.c 	snd_hda_codec_amp_stereo(codec, node->nid, HDA_INPUT, index, 0xff, val);
node              250 sound/pci/hda/hda_generic.c 	snd_printdd("CONNECT: NID=0x%x IDX=0x%x\n", node->nid, index);
node              251 sound/pci/hda/hda_generic.c 	return snd_hda_codec_write_cache(codec, node->nid, 0,
node              260 sound/pci/hda/hda_generic.c 	struct hda_gnode *node;
node              262 sound/pci/hda/hda_generic.c 	list_for_each_entry(node, &spec->nid_list, list) {
node              263 sound/pci/hda/hda_generic.c 		node->checked = 0;
node              278 sound/pci/hda/hda_generic.c 	if (node->checked)
node              281 sound/pci/hda/hda_generic.c 	node->checked = 1;
node              282 sound/pci/hda/hda_generic.c 	if (node->type == AC_WID_AUD_OUT) {
node              283 sound/pci/hda/hda_generic.c 		if (node->wid_caps & AC_WCAP_DIGITAL) {
node              284 sound/pci/hda/hda_generic.c 			snd_printdd("Skip Digital OUT node %x\n", node->nid);
node              287 sound/pci/hda/hda_generic.c 		snd_printdd("AUD_OUT found %x\n", node->nid);
node              290 sound/pci/hda/hda_generic.c 			return node == spec->dac_node[dac_idx];
node              292 sound/pci/hda/hda_generic.c 		spec->dac_node[dac_idx] = node;
node              293 sound/pci/hda/hda_generic.c 		if ((node->wid_caps & AC_WCAP_OUT_AMP) &&
node              295 sound/pci/hda/hda_generic.c 			spec->pcm_vol[spec->pcm_vol_nodes].node = node;
node              302 sound/pci/hda/hda_generic.c 	for (i = 0; i < node->nconns; i++) {
node              303 sound/pci/hda/hda_generic.c 		child = hda_get_node(spec, node->conn_list[i]);
node              313 sound/pci/hda/hda_generic.c 			if (node->nconns > 1)
node              314 sound/pci/hda/hda_generic.c 				select_input_connection(codec, node, i);
node              315 sound/pci/hda/hda_generic.c 			unmute_input(codec, node, i);
node              316 sound/pci/hda/hda_generic.c 			unmute_output(codec, node);
node              321 sound/pci/hda/hda_generic.c 				if ((node->wid_caps & AC_WCAP_IN_AMP) ||
node              322 sound/pci/hda/hda_generic.c 				    (node->wid_caps & AC_WCAP_OUT_AMP)) {
node              324 sound/pci/hda/hda_generic.c 					spec->pcm_vol[n].node = node;
node              345 sound/pci/hda/hda_generic.c 	struct hda_gnode *node;
node              348 sound/pci/hda/hda_generic.c 	list_for_each_entry(node, &spec->nid_list, list) {
node              349 sound/pci/hda/hda_generic.c 		if (node->type != AC_WID_PIN)
node              352 sound/pci/hda/hda_generic.c 		if (! (node->pin_caps & AC_PINCAP_OUT))
node              354 sound/pci/hda/hda_generic.c 		if (defcfg_port_conn(node) == AC_JACK_PORT_NONE)
node              357 sound/pci/hda/hda_generic.c 			if (jack_type != defcfg_type(node))
node              359 sound/pci/hda/hda_generic.c 			if (node->wid_caps & AC_WCAP_DIGITAL)
node              363 sound/pci/hda/hda_generic.c 			if (! (node->pin_ctl & AC_PINCTL_OUT_EN))
node              367 sound/pci/hda/hda_generic.c 		err = parse_output_path(codec, spec, node, 0);
node              371 sound/pci/hda/hda_generic.c 			err = parse_output_path(codec, spec, node, 1);
node              377 sound/pci/hda/hda_generic.c 			unmute_output(codec, node);
node              379 sound/pci/hda/hda_generic.c 			snd_hda_codec_write_cache(codec, node->nid, 0,
node              382 sound/pci/hda/hda_generic.c 					    ((node->pin_caps & AC_PINCAP_HP_DRV) ?
node              384 sound/pci/hda/hda_generic.c 			return node;
node              397 sound/pci/hda/hda_generic.c 	struct hda_gnode *node;
node              403 sound/pci/hda/hda_generic.c 	node = parse_output_jack(codec, spec, AC_JACK_LINE_OUT);
node              404 sound/pci/hda/hda_generic.c 	if (node) /* found, remember the PIN node */
node              405 sound/pci/hda/hda_generic.c 		spec->out_pin_node[0] = node;
node              408 sound/pci/hda/hda_generic.c 		node = parse_output_jack(codec, spec, AC_JACK_SPEAKER);
node              409 sound/pci/hda/hda_generic.c 		if (node)
node              410 sound/pci/hda/hda_generic.c 			spec->out_pin_node[0] = node;
node              413 sound/pci/hda/hda_generic.c 	node = parse_output_jack(codec, spec, AC_JACK_HP_OUT);
node              414 sound/pci/hda/hda_generic.c 	if (node) {
node              416 sound/pci/hda/hda_generic.c 			spec->out_pin_node[0] = node;
node              418 sound/pci/hda/hda_generic.c 			spec->out_pin_node[1] = node;
node              467 sound/pci/hda/hda_generic.c 	unsigned int location = defcfg_location(node);
node              468 sound/pci/hda/hda_generic.c 	switch (defcfg_type(node)) {
node              485 sound/pci/hda/hda_generic.c 		    (node->pin_caps &
node              512 sound/pci/hda/hda_generic.c 	if (node->checked)
node              515 sound/pci/hda/hda_generic.c 	node->checked = 1;
node              516 sound/pci/hda/hda_generic.c 	if (node->type != AC_WID_PIN) {
node              517 sound/pci/hda/hda_generic.c 		for (i = 0; i < node->nconns; i++) {
node              519 sound/pci/hda/hda_generic.c 			child = hda_get_node(spec, node->conn_list[i]);
node              529 sound/pci/hda/hda_generic.c 				if (node->nconns > 1)
node              530 sound/pci/hda/hda_generic.c 					select_input_connection(codec, node, i);
node              531 sound/pci/hda/hda_generic.c 				unmute_input(codec, node, i);
node              532 sound/pci/hda/hda_generic.c 				unmute_output(codec, node);
node              540 sound/pci/hda/hda_generic.c 	if (! (node->pin_caps & AC_PINCAP_IN))
node              543 sound/pci/hda/hda_generic.c 	if (defcfg_port_conn(node) == AC_JACK_PORT_NONE)
node              546 sound/pci/hda/hda_generic.c 	if (node->wid_caps & AC_WCAP_DIGITAL)
node              556 sound/pci/hda/hda_generic.c 	type = get_input_type(node, &pinctl);
node              559 sound/pci/hda/hda_generic.c 		if (! (node->pin_ctl & AC_PINCTL_IN_EN))
node              568 sound/pci/hda/hda_generic.c 	unmute_input(codec, node, 0); /* index = 0? */
node              570 sound/pci/hda/hda_generic.c 	snd_hda_codec_write_cache(codec, node->nid, 0,
node              605 sound/pci/hda/hda_generic.c 	struct hda_gnode *node;
node              621 sound/pci/hda/hda_generic.c 		node = hda_get_node(spec, adc_node->conn_list[i]);
node              622 sound/pci/hda/hda_generic.c 		if (node && node->type == AC_WID_PIN) {
node              623 sound/pci/hda/hda_generic.c 			err = parse_adc_sub_nodes(codec, spec, node);
node              632 sound/pci/hda/hda_generic.c 		node = hda_get_node(spec, adc_node->conn_list[i]);
node              633 sound/pci/hda/hda_generic.c 		if (node && node->type != AC_WID_PIN) {
node              634 sound/pci/hda/hda_generic.c 			err = parse_adc_sub_nodes(codec, spec, node);
node              660 sound/pci/hda/hda_generic.c 	struct hda_gnode *node;
node              668 sound/pci/hda/hda_generic.c 	list_for_each_entry(node, &spec->nid_list, list) {
node              669 sound/pci/hda/hda_generic.c 		if (node->wid_caps & AC_WCAP_DIGITAL)
node              671 sound/pci/hda/hda_generic.c 		if (node->type == AC_WID_AUD_IN) {
node              672 sound/pci/hda/hda_generic.c 			err = parse_input_path(codec, node);
node              720 sound/pci/hda/hda_generic.c 	if ((node->wid_caps & AC_WCAP_IN_AMP) &&
node              721 sound/pci/hda/hda_generic.c 	    (node->amp_in_caps & AC_AMPCAP_MUTE)) {
node              722 sound/pci/hda/hda_generic.c 		knew = (struct snd_kcontrol_new)HDA_CODEC_MUTE(name, node->nid, index, HDA_INPUT);
node              724 sound/pci/hda/hda_generic.c 			add_input_loopback(codec, node->nid, HDA_INPUT, index);
node              725 sound/pci/hda/hda_generic.c 		snd_printdd("[%s] NID=0x%x, DIR=IN, IDX=0x%x\n", name, node->nid, index);
node              729 sound/pci/hda/hda_generic.c 	} else if ((node->wid_caps & AC_WCAP_OUT_AMP) &&
node              730 sound/pci/hda/hda_generic.c 		   (node->amp_out_caps & AC_AMPCAP_MUTE)) {
node              731 sound/pci/hda/hda_generic.c 		knew = (struct snd_kcontrol_new)HDA_CODEC_MUTE(name, node->nid, 0, HDA_OUTPUT);
node              733 sound/pci/hda/hda_generic.c 			add_input_loopback(codec, node->nid, HDA_OUTPUT, 0);
node              734 sound/pci/hda/hda_generic.c 		snd_printdd("[%s] NID=0x%x, DIR=OUT\n", name, node->nid);
node              744 sound/pci/hda/hda_generic.c 	if ((node->wid_caps & AC_WCAP_IN_AMP) &&
node              745 sound/pci/hda/hda_generic.c 	    (node->amp_in_caps & AC_AMPCAP_NUM_STEPS)) {
node              746 sound/pci/hda/hda_generic.c 		knew = (struct snd_kcontrol_new)HDA_CODEC_VOLUME(name, node->nid, index, HDA_INPUT);
node              747 sound/pci/hda/hda_generic.c 		snd_printdd("[%s] NID=0x%x, DIR=IN, IDX=0x%x\n", name, node->nid, index);
node              751 sound/pci/hda/hda_generic.c 	} else if ((node->wid_caps & AC_WCAP_OUT_AMP) &&
node              752 sound/pci/hda/hda_generic.c 		   (node->amp_out_caps & AC_AMPCAP_NUM_STEPS)) {
node              753 sound/pci/hda/hda_generic.c 		knew = (struct snd_kcontrol_new)HDA_CODEC_VOLUME(name, node->nid, 0, HDA_OUTPUT);
node              754 sound/pci/hda/hda_generic.c 		snd_printdd("[%s] NID=0x%x, DIR=OUT\n", name, node->nid);
node              790 sound/pci/hda/hda_generic.c 		err = create_mixer(codec, spec->pcm_vol[i].node,
node              807 sound/pci/hda/hda_generic.c 		return create_mixer(codec, spec->pcm_vol[0].node,
node              893 sound/pci/hda/hda_generic.c 	if (node->checked)
node              896 sound/pci/hda/hda_generic.c 	node->checked = 1;
node              897 sound/pci/hda/hda_generic.c 	if (node == dest_node) {
node              902 sound/pci/hda/hda_generic.c 	for (i = 0; i < node->nconns; i++) {
node              903 sound/pci/hda/hda_generic.c 		struct hda_gnode *child = hda_get_node(spec, node->conn_list[i]);
node              911 sound/pci/hda/hda_generic.c 				err = create_mixer(codec, node, i, type,
node              921 sound/pci/hda/hda_generic.c 			if (node->nconns > 1)
node              922 sound/pci/hda/hda_generic.c 				select_input_connection(codec, node, i);
node              923 sound/pci/hda/hda_generic.c 			unmute_input(codec, node, i);
node              924 sound/pci/hda/hda_generic.c 			unmute_output(codec, node);
node              937 sound/pci/hda/hda_generic.c 	struct hda_gnode *node;
node              944 sound/pci/hda/hda_generic.c 	list_for_each_entry(node, &spec->nid_list, list) {
node              945 sound/pci/hda/hda_generic.c 		if (node->type != AC_WID_PIN)
node              948 sound/pci/hda/hda_generic.c 		if (! (node->pin_caps & AC_PINCAP_IN))
node              950 sound/pci/hda/hda_generic.c 		type = get_input_type(node, NULL);
node              957 sound/pci/hda/hda_generic.c 						  node, type);
node              235 sound/pcmcia/pdaudiocf/pdaudiocf.c 	link->dev_node = &pdacf->node;
node              120 sound/pcmcia/pdaudiocf/pdaudiocf.h 	dev_node_t node;
node              244 sound/pcmcia/vx/vxpocket.c 	link->dev_node = &vxp->node;
node               46 sound/pcmcia/vx/vxpocket.h 	dev_node_t node;
node              832 sound/ppc/pmac.c 		ppc_md.feature_call(PMAC_FTR_SOUND_CHIP_ENABLE, chip->node, 0, enable);
node              848 sound/ppc/pmac.c 	if (chip->node)
node              879 sound/ppc/pmac.c 	if (chip->node) {
node              891 sound/ppc/pmac.c 	of_node_put(chip->node);
node              916 sound/ppc/pmac.c 	for (mio = chip->node->parent; mio; mio = mio->parent) {
node              968 sound/ppc/pmac.c 	chip->node = of_find_node_by_name(NULL, "awacs");
node              969 sound/ppc/pmac.c 	sound = of_node_get(chip->node);
node              975 sound/ppc/pmac.c 	if (!chip->node)
node              976 sound/ppc/pmac.c 		chip->node = of_find_node_by_name(NULL, "davbus");
node              981 sound/ppc/pmac.c 	if (! chip->node) {
node              982 sound/ppc/pmac.c 		chip->node = of_find_node_by_name(NULL, "i2s-a");
node              983 sound/ppc/pmac.c 		if (chip->node && chip->node->parent &&
node              984 sound/ppc/pmac.c 		    chip->node->parent->parent) {
node              985 sound/ppc/pmac.c 			if (of_device_is_compatible(chip->node->parent->parent,
node              990 sound/ppc/pmac.c 	if (! chip->node)
node              995 sound/ppc/pmac.c 		while (sound && sound->parent != chip->node)
node              999 sound/ppc/pmac.c 		of_node_put(chip->node);
node             1000 sound/ppc/pmac.c 		chip->node = NULL;
node             1014 sound/ppc/pmac.c 		of_node_put(chip->node);
node             1015 sound/ppc/pmac.c 		chip->node = NULL;
node             1060 sound/ppc/pmac.c 	macio = macio_find(chip->node, macio_unknown);
node             1212 sound/ppc/pmac.c 	np = chip->node;
node             1322 sound/ppc/pmac.c 		for (mio = chip->node->parent; mio; mio = mio->parent) {
node               95 sound/ppc/pmac.h 	struct device_node *node;
node             1096 sound/ppc/tumbler.c 	struct device_node *node;
node             1102 sound/ppc/tumbler.c 		node = find_compatible_audio_device(device);
node             1104 sound/ppc/tumbler.c 		node = find_audio_device(device);
node             1105 sound/ppc/tumbler.c 	if (! node) {
node             1111 sound/ppc/tumbler.c 	base = of_get_property(node, "AAPL,address", NULL);
node             1113 sound/ppc/tumbler.c 		base = of_get_property(node, "reg", NULL);
node             1117 sound/ppc/tumbler.c 			of_node_put(node);
node             1128 sound/ppc/tumbler.c 	base = of_get_property(node, "audio-gpio-active-state", NULL);
node             1143 sound/ppc/tumbler.c 			prop = of_get_property(node, platform, NULL);
node             1159 sound/ppc/tumbler.c 	ret = irq_of_parse_and_map(node, 0);
node             1160 sound/ppc/tumbler.c 	of_node_put(node);
node             1360 sound/ppc/tumbler.c 	for (np = chip->node->child; np; np = np->sibling) {
node              724 sound/soc/fsl/mpc5200_psc_i2s.c 	prop = of_get_property(op->node, "cell-index", &size);
node              730 sound/soc/fsl/mpc5200_psc_i2s.c 	irq = irq_of_parse_and_map(op->node, 0);
node              731 sound/soc/fsl/mpc5200_psc_i2s.c 	if (of_address_to_resource(op->node, 0, &res)) {
node              787 sound/soc/fsl/mpc5200_psc_i2s.c 	if (of_get_property(op->node, "fsl,cellslave", NULL))
node              795 sound/soc/fsl/mpc5200_psc_i2s.c 	if (!of_get_property(op->node, "codec-handle", NULL))
node              828 sound/soc/fsl/mpc5200_psc_i2s.c 	of_snd_soc_register_platform(&psc_i2s_pcm_soc_platform, op->node,
node              215 sound/soc/fsl/mpc8610_hpcd.c 	struct device_node *np = ofdev->node;
node              108 sound/soc/fsl/soc-of-simple.c 	pr_info("registering ASoC codec driver: %s\n", node->full_name);
node              111 sound/soc/fsl/soc-of-simple.c 	of_soc = of_snd_soc_get_device(node);
node              120 sound/soc/fsl/soc-of-simple.c 	of_soc->dai_link.name = (char *)node->name;
node              121 sound/soc/fsl/soc-of-simple.c 	of_soc->dai_link.stream_name = (char *)node->name;
node              142 sound/soc/fsl/soc-of-simple.c 	pr_info("registering ASoC platform driver: %s\n", node->full_name);
node              144 sound/soc/fsl/soc-of-simple.c 	handle = of_get_property(node, "codec-handle", &len);
node              159 sound/soc/fsl/soc-of-simple.c 	of_soc->platform_node = node;
node             2078 sound/sparc/cs4231.c 	if (!strcmp(op->node->parent->name, "ebus"))
node             2082 sound/sparc/cs4231.c 	if (!strcmp(op->node->parent->name, "sbus") ||
node             2083 sound/sparc/cs4231.c 	    !strcmp(op->node->parent->name, "sbi"))
node             2650 sound/sparc/dbri.c 	       dbri->irq, op->node->name[9], dbri->mm.version);