h 328 arch/x86/kernel/amd_iommu_init.c u8 *p = (void *)h, *end = (void *)h; h 331 arch/x86/kernel/amd_iommu_init.c p += sizeof(*h); h 332 arch/x86/kernel/amd_iommu_init.c end += h->length; h 334 arch/x86/kernel/amd_iommu_init.c find_last_devid_on_pci(PCI_BUS(h->devid), h 335 arch/x86/kernel/amd_iommu_init.c PCI_SLOT(h->devid), h 336 arch/x86/kernel/amd_iommu_init.c PCI_FUNC(h->devid), h 337 arch/x86/kernel/amd_iommu_init.c h->cap_ptr); h 369 arch/x86/kernel/amd_iommu_init.c struct ivhd_header *h; h 385 arch/x86/kernel/amd_iommu_init.c h = (struct ivhd_header *)p; h 386 arch/x86/kernel/amd_iommu_init.c switch (h->type) { h 388 arch/x86/kernel/amd_iommu_init.c find_last_devid_from_ivhd(h); h 393 arch/x86/kernel/amd_iommu_init.c p += h->length; h 559 arch/x86/kernel/amd_iommu_init.c u8 *p = (u8 *)h; h 570 arch/x86/kernel/amd_iommu_init.c h->flags & IVHD_FLAG_HT_TUN_EN ? h 574 arch/x86/kernel/amd_iommu_init.c h->flags & IVHD_FLAG_PASSPW_EN ? h 578 arch/x86/kernel/amd_iommu_init.c h->flags & IVHD_FLAG_RESPASSPW_EN ? h 582 arch/x86/kernel/amd_iommu_init.c h->flags & IVHD_FLAG_ISOC_EN ? h 595 arch/x86/kernel/amd_iommu_init.c end += h->length; h 700 arch/x86/kernel/amd_iommu_init.c iommu->dev = pci_get_bus_and_slot(PCI_BUS(h->devid), h->devid & 0xff); h 704 arch/x86/kernel/amd_iommu_init.c iommu->cap_ptr = h->cap_ptr; h 705 arch/x86/kernel/amd_iommu_init.c iommu->pci_seg = h->pci_seg; h 706 arch/x86/kernel/amd_iommu_init.c iommu->mmio_phys = h->mmio_phys; h 707 arch/x86/kernel/amd_iommu_init.c iommu->mmio_base = iommu_map_mmio_space(h->mmio_phys); h 723 arch/x86/kernel/amd_iommu_init.c init_iommu_from_acpi(iommu, h); h 736 arch/x86/kernel/amd_iommu_init.c struct ivhd_header *h; h 744 arch/x86/kernel/amd_iommu_init.c h = (struct ivhd_header *)p; h 750 arch/x86/kernel/amd_iommu_init.c ret = init_iommu_one(iommu, h); h 757 arch/x86/kernel/amd_iommu_init.c p += h->length; h 827 arch/x86/kernel/apic_32.c unsigned int l, h; h 829 arch/x86/kernel/apic_32.c rdmsr(MSR_IA32_APICBASE, l, h); h 831 arch/x86/kernel/apic_32.c wrmsr(MSR_IA32_APICBASE, l, h); h 1185 arch/x86/kernel/apic_32.c u32 h, l, features; h 1221 arch/x86/kernel/apic_32.c rdmsr(MSR_IA32_APICBASE, l, h); h 1227 arch/x86/kernel/apic_32.c wrmsr(MSR_IA32_APICBASE, l, h); h 1244 arch/x86/kernel/apic_32.c rdmsr(MSR_IA32_APICBASE, l, h); h 1632 arch/x86/kernel/apic_32.c unsigned int l, h; h 1655 arch/x86/kernel/apic_32.c rdmsr(MSR_IA32_APICBASE, l, h); h 1658 arch/x86/kernel/apic_32.c wrmsr(MSR_IA32_APICBASE, l, h); h 712 arch/x86/kernel/apic_64.c unsigned int l, h; h 714 arch/x86/kernel/apic_64.c rdmsr(MSR_IA32_APICBASE, l, h); h 716 arch/x86/kernel/apic_64.c wrmsr(MSR_IA32_APICBASE, l, h); h 1575 arch/x86/kernel/apic_64.c unsigned int l, h; h 1598 arch/x86/kernel/apic_64.c rdmsr(MSR_IA32_APICBASE, l, h); h 1601 arch/x86/kernel/apic_64.c wrmsr(MSR_IA32_APICBASE, l, h); h 56 arch/x86/kernel/cpu/amd.c u32 l, h; h 103 arch/x86/kernel/cpu/amd.c rdmsr(MSR_K6_WHCR, l, h); h 109 arch/x86/kernel/cpu/amd.c wrmsr(MSR_K6_WHCR, l, h); h 124 arch/x86/kernel/cpu/amd.c rdmsr(MSR_K6_WHCR, l, h); h 130 arch/x86/kernel/cpu/amd.c wrmsr(MSR_K6_WHCR, l, h); h 148 arch/x86/kernel/cpu/amd.c u32 l, h; h 158 arch/x86/kernel/cpu/amd.c rdmsr(MSR_K7_HWCR, l, h); h 160 arch/x86/kernel/cpu/amd.c wrmsr(MSR_K7_HWCR, l, h); h 171 arch/x86/kernel/cpu/amd.c rdmsr(MSR_K7_CLK_CTL, l, h); h 175 arch/x86/kernel/cpu/amd.c wrmsr(MSR_K7_CLK_CTL, (l & 0x000fffff)|0x20000000, h); h 154 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c u32 h; h 158 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c rdmsr(cmd->addr.msr.reg, cmd->val, h); h 296 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c u32 h; h 298 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c h = max_t(u32, aperf_cur.split.hi, mperf_cur.split.hi); h 299 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c shift_count = fls(h); h 59 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c u32 l, h; h 64 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c rdmsr_on_cpu(cpu, MSR_IA32_THERM_STATUS, &l, &h); h 72 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c rdmsr_on_cpu(cpu, MSR_IA32_THERM_CONTROL, &l, &h); h 75 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c wrmsr_on_cpu(cpu, MSR_IA32_THERM_CONTROL, l & ~(1<<4), h); h 86 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c wrmsr_on_cpu(cpu, MSR_IA32_THERM_CONTROL, l, h); h 247 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c u32 l, h; h 249 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c rdmsr_on_cpu(cpu, MSR_IA32_THERM_CONTROL, &l, &h); h 324 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c unsigned l, h; h 333 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c rdmsr(MSR_IA32_PERF_STATUS, l, h); h 343 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c rdmsr(MSR_IA32_PERF_CTL, l, h); h 356 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c unsigned l, h; h 391 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c rdmsr(MSR_IA32_MISC_ENABLE, l, h); h 396 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c wrmsr(MSR_IA32_MISC_ENABLE, l, h); h 399 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c rdmsr(MSR_IA32_MISC_ENABLE, l, h); h 473 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c unsigned int msr, oldmsr = 0, h = 0, cpu = policy->cpu; h 537 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c rdmsr(MSR_IA32_PERF_CTL, oldmsr, h); h 564 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c wrmsr(MSR_IA32_PERF_CTL, oldmsr, h); h 591 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c wrmsr(MSR_IA32_PERF_CTL, oldmsr, h); h 75 arch/x86/kernel/cpu/mcheck/k7.c u32 l, h; h 85 arch/x86/kernel/cpu/mcheck/k7.c rdmsr(MSR_IA32_MCG_CAP, l, h); h 35 arch/x86/kernel/cpu/mcheck/mce_intel_64.c u32 l, h; h 49 arch/x86/kernel/cpu/mcheck/mce_intel_64.c rdmsr(MSR_IA32_MISC_ENABLE, l, h); h 50 arch/x86/kernel/cpu/mcheck/mce_intel_64.c h = apic_read(APIC_LVTTHMR); h 51 arch/x86/kernel/cpu/mcheck/mce_intel_64.c if ((l & (1 << 3)) && (h & APIC_DM_SMI)) { h 60 arch/x86/kernel/cpu/mcheck/mce_intel_64.c if (h & APIC_VECTOR_MASK) { h 63 arch/x86/kernel/cpu/mcheck/mce_intel_64.c "installed\n", cpu, (h & APIC_VECTOR_MASK)); h 67 arch/x86/kernel/cpu/mcheck/mce_intel_64.c h = THERMAL_APIC_VECTOR; h 68 arch/x86/kernel/cpu/mcheck/mce_intel_64.c h |= (APIC_DM_FIXED | APIC_LVT_MASKED); h 69 arch/x86/kernel/cpu/mcheck/mce_intel_64.c apic_write(APIC_LVTTHMR, h); h 71 arch/x86/kernel/cpu/mcheck/mce_intel_64.c rdmsr(MSR_IA32_THERM_INTERRUPT, l, h); h 72 arch/x86/kernel/cpu/mcheck/mce_intel_64.c wrmsr(MSR_IA32_THERM_INTERRUPT, l | 0x03, h); h 74 arch/x86/kernel/cpu/mcheck/mce_intel_64.c rdmsr(MSR_IA32_MISC_ENABLE, l, h); h 75 arch/x86/kernel/cpu/mcheck/mce_intel_64.c wrmsr(MSR_IA32_MISC_ENABLE, l | (1 << 3), h); h 71 arch/x86/kernel/cpu/mcheck/p4.c u32 l, h; h 86 arch/x86/kernel/cpu/mcheck/p4.c rdmsr(MSR_IA32_MISC_ENABLE, l, h); h 87 arch/x86/kernel/cpu/mcheck/p4.c h = apic_read(APIC_LVTTHMR); h 88 arch/x86/kernel/cpu/mcheck/p4.c if ((l & (1<<3)) && (h & APIC_DM_SMI)) { h 95 arch/x86/kernel/cpu/mcheck/p4.c if (h & APIC_VECTOR_MASK) { h 98 arch/x86/kernel/cpu/mcheck/p4.c cpu, (h & APIC_VECTOR_MASK)); h 103 arch/x86/kernel/cpu/mcheck/p4.c h = THERMAL_APIC_VECTOR; /* our delivery vector */ h 104 arch/x86/kernel/cpu/mcheck/p4.c h |= (APIC_DM_FIXED | APIC_LVT_MASKED); /* we'll mask till we're ready */ h 105 arch/x86/kernel/cpu/mcheck/p4.c apic_write(APIC_LVTTHMR, h); h 107 arch/x86/kernel/cpu/mcheck/p4.c rdmsr(MSR_IA32_THERM_INTERRUPT, l, h); h 108 arch/x86/kernel/cpu/mcheck/p4.c wrmsr(MSR_IA32_THERM_INTERRUPT, l | 0x03 , h); h 113 arch/x86/kernel/cpu/mcheck/p4.c rdmsr(MSR_IA32_MISC_ENABLE, l, h); h 114 arch/x86/kernel/cpu/mcheck/p4.c wrmsr(MSR_IA32_MISC_ENABLE, l | (1<<3), h); h 130 arch/x86/kernel/cpu/mcheck/p4.c u32 h; h 132 arch/x86/kernel/cpu/mcheck/p4.c rdmsr(MSR_IA32_MCG_EAX, r->eax, h); h 133 arch/x86/kernel/cpu/mcheck/p4.c rdmsr(MSR_IA32_MCG_EBX, r->ebx, h); h 134 arch/x86/kernel/cpu/mcheck/p4.c rdmsr(MSR_IA32_MCG_ECX, r->ecx, h); h 135 arch/x86/kernel/cpu/mcheck/p4.c rdmsr(MSR_IA32_MCG_EDX, r->edx, h); h 136 arch/x86/kernel/cpu/mcheck/p4.c rdmsr(MSR_IA32_MCG_ESI, r->esi, h); h 137 arch/x86/kernel/cpu/mcheck/p4.c rdmsr(MSR_IA32_MCG_EDI, r->edi, h); h 138 arch/x86/kernel/cpu/mcheck/p4.c rdmsr(MSR_IA32_MCG_EBP, r->ebp, h); h 139 arch/x86/kernel/cpu/mcheck/p4.c rdmsr(MSR_IA32_MCG_ESP, r->esp, h); h 140 arch/x86/kernel/cpu/mcheck/p4.c rdmsr(MSR_IA32_MCG_EFLAGS, r->eflags, h); h 141 arch/x86/kernel/cpu/mcheck/p4.c rdmsr(MSR_IA32_MCG_EIP, r->eip, h); h 223 arch/x86/kernel/cpu/mcheck/p4.c u32 l, h; h 230 arch/x86/kernel/cpu/mcheck/p4.c rdmsr(MSR_IA32_MCG_CAP, l, h); h 245 arch/x86/kernel/cpu/mcheck/p4.c rdmsr(MSR_IA32_MCG_CAP, l, h); h 33 arch/x86/kernel/cpu/mcheck/p5.c u32 l, h; h 46 arch/x86/kernel/cpu/mcheck/p5.c rdmsr(MSR_IA32_P5_MC_ADDR, l, h); h 47 arch/x86/kernel/cpu/mcheck/p5.c rdmsr(MSR_IA32_P5_MC_TYPE, l, h); h 87 arch/x86/kernel/cpu/mcheck/p6.c u32 l, h; h 103 arch/x86/kernel/cpu/mcheck/p6.c rdmsr(MSR_IA32_MCG_CAP, l, h); h 1518 arch/x86/kernel/cpu/mtrr/main.c u32 l, h; h 1525 arch/x86/kernel/cpu/mtrr/main.c if (rdmsr_safe(MSR_K8_SYSCFG, &l, &h) < 0) h 565 arch/x86/kernel/vmi_32.c unsigned l, h; h 566 arch/x86/kernel/vmi_32.c rdmsr(MSR_EFER, l, h); h 567 arch/x86/kernel/vmi_32.c ap.efer = (unsigned long long) h << 32 | l; h 8 arch/x86/lib/msr-on-cpu.c u32 l, h; h 16 arch/x86/lib/msr-on-cpu.c rdmsr(rv->msr_no, rv->l, rv->h); h 23 arch/x86/lib/msr-on-cpu.c wrmsr(rv->msr_no, rv->l, rv->h); h 34 arch/x86/lib/msr-on-cpu.c *h = rv.h; h 46 arch/x86/lib/msr-on-cpu.c rv.h = h; h 58 arch/x86/lib/msr-on-cpu.c rv->err = rdmsr_safe(rv->msr_no, &rv->l, &rv->h); h 65 arch/x86/lib/msr-on-cpu.c rv->err = wrmsr_safe(rv->msr_no, rv->l, rv->h); h 76 arch/x86/lib/msr-on-cpu.c *h = rv.h; h 88 arch/x86/lib/msr-on-cpu.c rv.h = h; h 19 arch/x86/math-emu/reg_constant.c #define MAKE_REG(s, e, l, h) { l, h, \ h 263 arch/x86/mm/hugetlbpage.c struct hstate *h = hstate_file(file); h 276 arch/x86/mm/hugetlbpage.c addr = ALIGN(start_addr, huge_page_size(h)); h 298 arch/x86/mm/hugetlbpage.c addr = ALIGN(vma->vm_end, huge_page_size(h)); h 306 arch/x86/mm/hugetlbpage.c struct hstate *h = hstate_file(file); h 327 arch/x86/mm/hugetlbpage.c addr = (mm->free_area_cache - len) & huge_page_mask(h); h 358 arch/x86/mm/hugetlbpage.c addr = (vma->vm_start - len) & huge_page_mask(h); h 396 arch/x86/mm/hugetlbpage.c struct hstate *h = hstate_file(file); h 400 arch/x86/mm/hugetlbpage.c if (len & ~huge_page_mask(h)) h 412 arch/x86/mm/hugetlbpage.c addr = ALIGN(addr, huge_page_size(h)); h 598 arch/x86/mm/init_32.c unsigned int v[4], l, h; h 604 arch/x86/mm/init_32.c rdmsr(MSR_EFER, l, h); h 606 arch/x86/mm/init_32.c wrmsr(MSR_EFER, l, h); h 30 arch/x86/oprofile/op_model_amd.c #define CTR_READ(l, h, msrs, c) do {rdmsr(msrs->counters[(c)].addr, (l), (h)); } while (0) h 35 arch/x86/oprofile/op_model_amd.c #define CTRL_READ(l, h, msrs, c) do {rdmsr(msrs->controls[(c)].addr, (l), (h)); } while (0) h 36 arch/x86/oprofile/op_model_amd.c #define CTRL_WRITE(l, h, msrs, c) do {wrmsr(msrs->controls[(c)].addr, (l), (h)); } while (0) h 47 arch/x86/oprofile/op_model_amd.c #define CTRL_SET_HOST_ONLY(val, h) (val |= ((h & 1) << 9)) h 48 arch/x86/oprofile/op_model_amd.c #define CTRL_SET_GUEST_ONLY(val, h) (val |= ((h & 1) << 8)) h 371 arch/x86/oprofile/op_model_p4.c #define CTR_READ(l, h, i) do {rdmsr(p4_counters[(i)].counter_address, (l), (h)); } while (0) h 26 arch/x86/oprofile/op_model_ppro.c #define CTR_READ(l, h, msrs, c) do {rdmsr(msrs->counters[(c)].addr, (l), (h)); } while (0) h 32 arch/x86/oprofile/op_model_ppro.c #define CTRL_READ(l, h, msrs, c) do {rdmsr((msrs->controls[(c)].addr), (l), (h)); } while (0) h 33 arch/x86/oprofile/op_model_ppro.c #define CTRL_WRITE(l, h, msrs, c) do {wrmsr((msrs->controls[(c)].addr), (l), (h)); } while (0) h 810 arch/x86/pci/irq.c struct irq_router_handler *h; h 836 arch/x86/pci/irq.c for (h = pirq_routers; h->vendor; h++) { h 838 arch/x86/pci/irq.c if (rt->rtr_vendor == h->vendor && h 839 arch/x86/pci/irq.c h->probe(r, pirq_router_dev, rt->rtr_device)) h 842 arch/x86/pci/irq.c if (pirq_router_dev->vendor == h->vendor && h 843 arch/x86/pci/irq.c h->probe(r, pirq_router_dev, pirq_router_dev->device)) h 50 arch/x86/xen/time.c u32 h, l; h 59 arch/x86/xen/time.c h = p32[1]; h 63 arch/x86/xen/time.c } while (p32[1] != h); h 65 arch/x86/xen/time.c ret = (((u64)h) << 32) | l; h 61 crypto/sha256_generic.c u32 a, b, c, d, e, f, g, h, t1, t2; h 75 crypto/sha256_generic.c e=state[4]; f=state[5]; g=state[6]; h=state[7]; h 78 crypto/sha256_generic.c t1 = h + e1(e) + Ch(e,f,g) + 0x428a2f98 + W[ 0]; h 79 crypto/sha256_generic.c t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; h 81 crypto/sha256_generic.c t2 = e0(h) + Maj(h,a,b); c+=t1; g=t1+t2; h 83 crypto/sha256_generic.c t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; h 85 crypto/sha256_generic.c t2 = e0(f) + Maj(f,g,h); a+=t1; e=t1+t2; h 87 crypto/sha256_generic.c t2 = e0(e) + Maj(e,f,g); h+=t1; d=t1+t2; h 88 crypto/sha256_generic.c t1 = c + e1(h) + Ch(h,a,b) + 0x59f111f1 + W[ 5]; h 90 crypto/sha256_generic.c t1 = b + e1(g) + Ch(g,h,a) + 0x923f82a4 + W[ 6]; h 92 crypto/sha256_generic.c t1 = a + e1(f) + Ch(f,g,h) + 0xab1c5ed5 + W[ 7]; h 95 crypto/sha256_generic.c t1 = h + e1(e) + Ch(e,f,g) + 0xd807aa98 + W[ 8]; h 96 crypto/sha256_generic.c t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; h 98 crypto/sha256_generic.c t2 = e0(h) + Maj(h,a,b); c+=t1; g=t1+t2; h 100 crypto/sha256_generic.c t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; h 102 crypto/sha256_generic.c t2 = e0(f) + Maj(f,g,h); a+=t1; e=t1+t2; h 104 crypto/sha256_generic.c t2 = e0(e) + Maj(e,f,g); h+=t1; d=t1+t2; h 105 crypto/sha256_generic.c t1 = c + e1(h) + Ch(h,a,b) + 0x80deb1fe + W[13]; h 107 crypto/sha256_generic.c t1 = b + e1(g) + Ch(g,h,a) + 0x9bdc06a7 + W[14]; h 109 crypto/sha256_generic.c t1 = a + e1(f) + Ch(f,g,h) + 0xc19bf174 + W[15]; h 112 crypto/sha256_generic.c t1 = h + e1(e) + Ch(e,f,g) + 0xe49b69c1 + W[16]; h 113 crypto/sha256_generic.c t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; h 115 crypto/sha256_generic.c t2 = e0(h) + Maj(h,a,b); c+=t1; g=t1+t2; h 117 crypto/sha256_generic.c t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; h 119 crypto/sha256_generic.c t2 = e0(f) + Maj(f,g,h); a+=t1; e=t1+t2; h 121 crypto/sha256_generic.c t2 = e0(e) + Maj(e,f,g); h+=t1; d=t1+t2; h 122 crypto/sha256_generic.c t1 = c + e1(h) + Ch(h,a,b) + 0x4a7484aa + W[21]; h 124 crypto/sha256_generic.c t1 = b + e1(g) + Ch(g,h,a) + 0x5cb0a9dc + W[22]; h 126 crypto/sha256_generic.c t1 = a + e1(f) + Ch(f,g,h) + 0x76f988da + W[23]; h 129 crypto/sha256_generic.c t1 = h + e1(e) + Ch(e,f,g) + 0x983e5152 + W[24]; h 130 crypto/sha256_generic.c t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; h 132 crypto/sha256_generic.c t2 = e0(h) + Maj(h,a,b); c+=t1; g=t1+t2; h 134 crypto/sha256_generic.c t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; h 136 crypto/sha256_generic.c t2 = e0(f) + Maj(f,g,h); a+=t1; e=t1+t2; h 138 crypto/sha256_generic.c t2 = e0(e) + Maj(e,f,g); h+=t1; d=t1+t2; h 139 crypto/sha256_generic.c t1 = c + e1(h) + Ch(h,a,b) + 0xd5a79147 + W[29]; h 141 crypto/sha256_generic.c t1 = b + e1(g) + Ch(g,h,a) + 0x06ca6351 + W[30]; h 143 crypto/sha256_generic.c t1 = a + e1(f) + Ch(f,g,h) + 0x14292967 + W[31]; h 146 crypto/sha256_generic.c t1 = h + e1(e) + Ch(e,f,g) + 0x27b70a85 + W[32]; h 147 crypto/sha256_generic.c t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; h 149 crypto/sha256_generic.c t2 = e0(h) + Maj(h,a,b); c+=t1; g=t1+t2; h 151 crypto/sha256_generic.c t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; h 153 crypto/sha256_generic.c t2 = e0(f) + Maj(f,g,h); a+=t1; e=t1+t2; h 155 crypto/sha256_generic.c t2 = e0(e) + Maj(e,f,g); h+=t1; d=t1+t2; h 156 crypto/sha256_generic.c t1 = c + e1(h) + Ch(h,a,b) + 0x766a0abb + W[37]; h 158 crypto/sha256_generic.c t1 = b + e1(g) + Ch(g,h,a) + 0x81c2c92e + W[38]; h 160 crypto/sha256_generic.c t1 = a + e1(f) + Ch(f,g,h) + 0x92722c85 + W[39]; h 163 crypto/sha256_generic.c t1 = h + e1(e) + Ch(e,f,g) + 0xa2bfe8a1 + W[40]; h 164 crypto/sha256_generic.c t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; h 166 crypto/sha256_generic.c t2 = e0(h) + Maj(h,a,b); c+=t1; g=t1+t2; h 168 crypto/sha256_generic.c t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; h 170 crypto/sha256_generic.c t2 = e0(f) + Maj(f,g,h); a+=t1; e=t1+t2; h 172 crypto/sha256_generic.c t2 = e0(e) + Maj(e,f,g); h+=t1; d=t1+t2; h 173 crypto/sha256_generic.c t1 = c + e1(h) + Ch(h,a,b) + 0xd6990624 + W[45]; h 175 crypto/sha256_generic.c t1 = b + e1(g) + Ch(g,h,a) + 0xf40e3585 + W[46]; h 177 crypto/sha256_generic.c t1 = a + e1(f) + Ch(f,g,h) + 0x106aa070 + W[47]; h 180 crypto/sha256_generic.c t1 = h + e1(e) + Ch(e,f,g) + 0x19a4c116 + W[48]; h 181 crypto/sha256_generic.c t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; h 183 crypto/sha256_generic.c t2 = e0(h) + Maj(h,a,b); c+=t1; g=t1+t2; h 185 crypto/sha256_generic.c t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; h 187 crypto/sha256_generic.c t2 = e0(f) + Maj(f,g,h); a+=t1; e=t1+t2; h 189 crypto/sha256_generic.c t2 = e0(e) + Maj(e,f,g); h+=t1; d=t1+t2; h 190 crypto/sha256_generic.c t1 = c + e1(h) + Ch(h,a,b) + 0x4ed8aa4a + W[53]; h 192 crypto/sha256_generic.c t1 = b + e1(g) + Ch(g,h,a) + 0x5b9cca4f + W[54]; h 194 crypto/sha256_generic.c t1 = a + e1(f) + Ch(f,g,h) + 0x682e6ff3 + W[55]; h 197 crypto/sha256_generic.c t1 = h + e1(e) + Ch(e,f,g) + 0x748f82ee + W[56]; h 198 crypto/sha256_generic.c t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; h 200 crypto/sha256_generic.c t2 = e0(h) + Maj(h,a,b); c+=t1; g=t1+t2; h 202 crypto/sha256_generic.c t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; h 204 crypto/sha256_generic.c t2 = e0(f) + Maj(f,g,h); a+=t1; e=t1+t2; h 206 crypto/sha256_generic.c t2 = e0(e) + Maj(e,f,g); h+=t1; d=t1+t2; h 207 crypto/sha256_generic.c t1 = c + e1(h) + Ch(h,a,b) + 0xa4506ceb + W[61]; h 209 crypto/sha256_generic.c t1 = b + e1(g) + Ch(g,h,a) + 0xbef9a3f7 + W[62]; h 211 crypto/sha256_generic.c t1 = a + e1(f) + Ch(f,g,h) + 0xc67178f2 + W[63]; h 215 crypto/sha256_generic.c state[4] += e; state[5] += f; state[6] += g; state[7] += h; h 218 crypto/sha256_generic.c a = b = c = d = e = f = g = h = t1 = t2 = 0; h 94 crypto/sha512_generic.c u64 a, b, c, d, e, f, g, h, t1, t2; h 108 crypto/sha512_generic.c e=state[4]; f=state[5]; g=state[6]; h=state[7]; h 112 crypto/sha512_generic.c t1 = h + e1(e) + Ch(e,f,g) + sha512_K[i ] + W[i ]; h 113 crypto/sha512_generic.c t2 = e0(a) + Maj(a,b,c); d+=t1; h=t1+t2; h 115 crypto/sha512_generic.c t2 = e0(h) + Maj(h,a,b); c+=t1; g=t1+t2; h 117 crypto/sha512_generic.c t2 = e0(g) + Maj(g,h,a); b+=t1; f=t1+t2; h 119 crypto/sha512_generic.c t2 = e0(f) + Maj(f,g,h); a+=t1; e=t1+t2; h 121 crypto/sha512_generic.c t2 = e0(e) + Maj(e,f,g); h+=t1; d=t1+t2; h 122 crypto/sha512_generic.c t1 = c + e1(h) + Ch(h,a,b) + sha512_K[i+5] + W[i+5]; h 124 crypto/sha512_generic.c t1 = b + e1(g) + Ch(g,h,a) + sha512_K[i+6] + W[i+6]; h 126 crypto/sha512_generic.c t1 = a + e1(f) + Ch(f,g,h) + sha512_K[i+7] + W[i+7]; h 131 crypto/sha512_generic.c state[4] += e; state[5] += f; state[6] += g; state[7] += h; h 134 crypto/sha512_generic.c a = b = c = d = e = f = g = h = t1 = t2 = 0; h 25 fs/adfs/dir_fplus.c struct adfs_bigdirheader *h; h 44 fs/adfs/dir_fplus.c h = (struct adfs_bigdirheader *)dir->bh[0]->b_data; h 45 fs/adfs/dir_fplus.c size = le32_to_cpu(h->bigdirsize); h 51 fs/adfs/dir_fplus.c if (h->bigdirversion[0] != 0 || h->bigdirversion[1] != 0 || h 52 fs/adfs/dir_fplus.c h->bigdirversion[2] != 0 || size & 2047 || h 53 fs/adfs/dir_fplus.c h->bigdirstartname != cpu_to_le32(BIGDIRSTARTNAME)) h 73 fs/adfs/dir_fplus.c t->bigdirendmasseq != h->startmasseq || h 77 fs/adfs/dir_fplus.c dir->parent_id = le32_to_cpu(h->bigdirparent); h 90 fs/adfs/dir_fplus.c struct adfs_bigdirheader *h = (struct adfs_bigdirheader *)dir->bh[0]->b_data; h 93 fs/adfs/dir_fplus.c if (fpos <= le32_to_cpu(h->bigdirentries)) { h 127 fs/adfs/dir_fplus.c struct adfs_bigdirheader *h = (struct adfs_bigdirheader *)dir->bh[0]->b_data; h 132 fs/adfs/dir_fplus.c if (dir->pos >= le32_to_cpu(h->bigdirentries)) h 136 fs/adfs/dir_fplus.c offset += ((le32_to_cpu(h->bigdirnamelen) + 4) & ~3); h 149 fs/adfs/dir_fplus.c offset += ((le32_to_cpu(h->bigdirnamelen) + 4) & ~3); h 150 fs/adfs/dir_fplus.c offset += le32_to_cpu(h->bigdirentries) * sizeof(struct adfs_bigdirentry); h 67 fs/autofs/autofs_i.h struct autofs_dir_ent *h[AUTOFS_HASH_SIZE]; h 107 fs/autofs/dirhash.c memset(&dh->h, 0, AUTOFS_HASH_SIZE*sizeof(struct autofs_dir_ent *)); h 118 fs/autofs/dirhash.c for ( dhn = dh->h[(unsigned) name->hash % AUTOFS_HASH_SIZE] ; dhn ; dhn = dhn->next ) { h 139 fs/autofs/dirhash.c dhnp = &dh->h[(unsigned) ent->hash % AUTOFS_HASH_SIZE]; h 190 fs/autofs/dirhash.c ent = dh->h[bucket]; h 224 fs/autofs/dirhash.c for ( ent = dh->h[i] ; ent ; ent = ent->next ) { h 241 fs/autofs/dirhash.c for ( ent = sbi->dirhash.h[i] ; ent ; ent = nent ) { h 1457 fs/compat.c unsigned long h, l; h 1458 fs/compat.c if (__get_user(l, ufdset) || __get_user(h, ufdset+1)) h 1461 fs/compat.c *fdset++ = h << 32 | l; h 1489 fs/compat.c unsigned long h, l; h 1491 fs/compat.c h = l >> 32; h 1492 fs/compat.c if (__put_user(l, ufdset) || __put_user(h, ufdset+1)) h 179 fs/compat_ioctl.c err |= put_user(kevent.u.size.h, &up->u.size.h); h 279 fs/ext3/namei.c struct dx_hash_info h = *hinfo; h 291 fs/ext3/namei.c ext3fs_dirhash(de->name, de->name_len, &h); h 292 fs/ext3/namei.c printk(":%x.%u ", h.hash, h 716 fs/ext3/namei.c struct dx_hash_info h = *hinfo; h 721 fs/ext3/namei.c ext3fs_dirhash(de->name, de->name_len, &h); h 723 fs/ext3/namei.c map_tail->hash = h.hash; h 2203 fs/ext4/mballoc.c struct ext4_mb_history h; h 2211 fs/ext4/mballoc.c h.op = ac->ac_op; h 2212 fs/ext4/mballoc.c h.pid = current->pid; h 2213 fs/ext4/mballoc.c h.ino = ac->ac_inode ? ac->ac_inode->i_ino : 0; h 2214 fs/ext4/mballoc.c h.orig = ac->ac_o_ex; h 2215 fs/ext4/mballoc.c h.result = ac->ac_b_ex; h 2216 fs/ext4/mballoc.c h.flags = ac->ac_flags; h 2217 fs/ext4/mballoc.c h.found = ac->ac_found; h 2218 fs/ext4/mballoc.c h.groups = ac->ac_groups_scanned; h 2219 fs/ext4/mballoc.c h.cr = ac->ac_criteria; h 2220 fs/ext4/mballoc.c h.tail = ac->ac_tail; h 2221 fs/ext4/mballoc.c h.buddy = ac->ac_buddy; h 2222 fs/ext4/mballoc.c h.merged = 0; h 2226 fs/ext4/mballoc.c h.merged = 1; h 2227 fs/ext4/mballoc.c h.goal = ac->ac_g_ex; h 2228 fs/ext4/mballoc.c h.result = ac->ac_f_ex; h 2232 fs/ext4/mballoc.c memcpy(sbi->s_mb_history + sbi->s_mb_history_cur, &h, sizeof(h)); h 282 fs/ext4/namei.c struct dx_hash_info h = *hinfo; h 294 fs/ext4/namei.c ext4fs_dirhash(de->name, de->name_len, &h); h 295 fs/ext4/namei.c printk(":%x.%u ", h.hash, h 720 fs/ext4/namei.c struct dx_hash_info h = *hinfo; h 725 fs/ext4/namei.c ext4fs_dirhash(de->name, de->name_len, &h); h 727 fs/ext4/namei.c map_tail->hash = h.hash; h 90 fs/fuse/dev.c req->in.h.uid = current->fsuid; h 91 fs/fuse/dev.c req->in.h.gid = current->fsgid; h 92 fs/fuse/dev.c req->in.h.pid = current->pid; h 235 fs/fuse/dev.c req->in.h.unique = fuse_get_unique(fc); h 236 fs/fuse/dev.c req->in.h.len = sizeof(struct fuse_in_header) + h 353 fs/fuse/dev.c req->out.h.error = -EINTR; h 388 fs/fuse/dev.c req->out.h.error = -ENOTCONN; h 390 fs/fuse/dev.c req->out.h.error = -ECONNREFUSED; h 424 fs/fuse/dev.c req->out.h.error = -ENOTCONN; h 709 fs/fuse/dev.c arg.unique = req->in.h.unique; h 772 fs/fuse/dev.c reqsize = in->h.len; h 775 fs/fuse/dev.c req->out.h.error = -EIO; h 777 fs/fuse/dev.c if (in->h.opcode == FUSE_SETXATTR) h 778 fs/fuse/dev.c req->out.h.error = -E2BIG; h 784 fs/fuse/dev.c err = fuse_copy_one(&cs, &in->h, sizeof(in->h)); h 796 fs/fuse/dev.c req->out.h.error = -EIO; h 824 fs/fuse/dev.c if (req->in.h.unique == unique || req->intr_unique == unique) h 835 fs/fuse/dev.c if (out->h.error) h 918 fs/fuse/dev.c req->out.h = oh; h 932 fs/fuse/dev.c req->out.h.error = -EIO; h 973 fs/fuse/dev.c req->out.h.error = -ECONNABORTED; h 999 fs/fuse/dev.c req->out.h.error = -ECONNABORTED; h 120 fs/fuse/dir.c req->in.h.opcode = FUSE_LOOKUP; h 121 fs/fuse/dir.c req->in.h.nodeid = nodeid; h 194 fs/fuse/dir.c err = req->out.h.error; h 287 fs/fuse/dir.c err = req->out.h.error; h 420 fs/fuse/dir.c req->in.h.opcode = FUSE_CREATE; h 421 fs/fuse/dir.c req->in.h.nodeid = get_node_id(dir); h 436 fs/fuse/dir.c err = req->out.h.error; h 498 fs/fuse/dir.c req->in.h.nodeid = get_node_id(dir); h 506 fs/fuse/dir.c err = req->out.h.error; h 563 fs/fuse/dir.c req->in.h.opcode = FUSE_MKNOD; h 594 fs/fuse/dir.c req->in.h.opcode = FUSE_MKDIR; h 612 fs/fuse/dir.c req->in.h.opcode = FUSE_SYMLINK; h 629 fs/fuse/dir.c req->in.h.opcode = FUSE_UNLINK; h 630 fs/fuse/dir.c req->in.h.nodeid = get_node_id(dir); h 635 fs/fuse/dir.c err = req->out.h.error; h 660 fs/fuse/dir.c req->in.h.opcode = FUSE_RMDIR; h 661 fs/fuse/dir.c req->in.h.nodeid = get_node_id(dir); h 666 fs/fuse/dir.c err = req->out.h.error; h 689 fs/fuse/dir.c req->in.h.opcode = FUSE_RENAME; h 690 fs/fuse/dir.c req->in.h.nodeid = get_node_id(olddir); h 699 fs/fuse/dir.c err = req->out.h.error; h 739 fs/fuse/dir.c req->in.h.opcode = FUSE_LINK; h 803 fs/fuse/dir.c req->in.h.opcode = FUSE_GETATTR; h 804 fs/fuse/dir.c req->in.h.nodeid = get_node_id(inode); h 815 fs/fuse/dir.c err = req->out.h.error; h 902 fs/fuse/dir.c req->in.h.opcode = FUSE_ACCESS; h 903 fs/fuse/dir.c req->in.h.nodeid = get_node_id(inode); h 908 fs/fuse/dir.c err = req->out.h.error; h 1031 fs/fuse/dir.c err = req->out.h.error; h 1057 fs/fuse/dir.c req->in.h.opcode = FUSE_READLINK; h 1058 fs/fuse/dir.c req->in.h.nodeid = get_node_id(inode); h 1064 fs/fuse/dir.c if (req->out.h.error) { h 1066 fs/fuse/dir.c link = ERR_PTR(req->out.h.error); h 1258 fs/fuse/dir.c req->in.h.opcode = FUSE_SETATTR; h 1259 fs/fuse/dir.c req->in.h.nodeid = get_node_id(inode); h 1270 fs/fuse/dir.c err = req->out.h.error; h 1354 fs/fuse/dir.c req->in.h.opcode = FUSE_SETXATTR; h 1355 fs/fuse/dir.c req->in.h.nodeid = get_node_id(inode); h 1364 fs/fuse/dir.c err = req->out.h.error; h 1392 fs/fuse/dir.c req->in.h.opcode = FUSE_GETXATTR; h 1393 fs/fuse/dir.c req->in.h.nodeid = get_node_id(inode); h 1410 fs/fuse/dir.c ret = req->out.h.error; h 1444 fs/fuse/dir.c req->in.h.opcode = FUSE_LISTXATTR; h 1445 fs/fuse/dir.c req->in.h.nodeid = get_node_id(inode); h 1460 fs/fuse/dir.c ret = req->out.h.error; h 1487 fs/fuse/dir.c req->in.h.opcode = FUSE_REMOVEXATTR; h 1488 fs/fuse/dir.c req->in.h.nodeid = get_node_id(inode); h 1493 fs/fuse/dir.c err = req->out.h.error; h 34 fs/fuse/file.c req->in.h.opcode = isdir ? FUSE_OPENDIR : FUSE_OPEN; h 35 fs/fuse/file.c req->in.h.nodeid = get_node_id(inode); h 43 fs/fuse/file.c err = req->out.h.error; h 145 fs/fuse/file.c req->in.h.opcode = opcode; h 146 fs/fuse/file.c req->in.h.nodeid = nodeid; h 275 fs/fuse/file.c req->in.h.opcode = FUSE_FLUSH; h 276 fs/fuse/file.c req->in.h.nodeid = get_node_id(inode); h 282 fs/fuse/file.c err = req->out.h.error; h 340 fs/fuse/file.c req->in.h.opcode = isdir ? FUSE_FSYNCDIR : FUSE_FSYNC; h 341 fs/fuse/file.c req->in.h.nodeid = get_node_id(inode); h 346 fs/fuse/file.c err = req->out.h.error; h 373 fs/fuse/file.c req->in.h.opcode = opcode; h 374 fs/fuse/file.c req->in.h.nodeid = get_node_id(inode); h 448 fs/fuse/file.c err = req->out.h.error; h 477 fs/fuse/file.c if (!req->out.h.error && num_read < count) { h 486 fs/fuse/file.c if (!req->out.h.error) h 611 fs/fuse/file.c req->in.h.opcode = FUSE_WRITE; h 612 fs/fuse/file.c req->in.h.nodeid = get_node_id(inode); h 691 fs/fuse/file.c err = req->out.h.error; h 738 fs/fuse/file.c if (!req->out.h.error && !offset && count >= PAGE_CACHE_SIZE) h 845 fs/fuse/file.c err = req->out.h.error; h 987 fs/fuse/file.c if (req->out.h.error) { h 989 fs/fuse/file.c res = req->out.h.error; h 1114 fs/fuse/file.c mapping_set_error(inode->i_mapping, req->out.h.error); h 1303 fs/fuse/file.c req->in.h.opcode = opcode; h 1304 fs/fuse/file.c req->in.h.nodeid = get_node_id(inode); h 1327 fs/fuse/file.c err = req->out.h.error; h 1359 fs/fuse/file.c err = req->out.h.error; h 1426 fs/fuse/file.c req->in.h.opcode = FUSE_BMAP; h 1427 fs/fuse/file.c req->in.h.nodeid = get_node_id(inode); h 1435 fs/fuse/file.c err = req->out.h.error; h 119 fs/fuse/fuse_i.h struct fuse_in_header h; h 140 fs/fuse/fuse_i.h struct fuse_out_header h; h 92 fs/fuse/inode.c req->in.h.opcode = FUSE_FORGET; h 93 fs/fuse/inode.c req->in.h.nodeid = nodeid; h 270 fs/fuse/inode.c req->in.h.opcode = FUSE_DESTROY; h 331 fs/fuse/inode.c req->in.h.opcode = FUSE_STATFS; h 332 fs/fuse/inode.c req->in.h.nodeid = get_node_id(dentry->d_inode); h 338 fs/fuse/inode.c err = req->out.h.error; h 740 fs/fuse/inode.c if (req->out.h.error || arg->major != FUSE_KERNEL_VERSION) h 785 fs/fuse/inode.c req->in.h.opcode = FUSE_INIT; h 80 fs/gfs2/dir.c #define gfs2_disk_hash2offset(h) (((u64)(h)) >> 1) h 456 fs/gfs2/dir.c const struct gfs2_meta_header *h = buf; h 461 fs/gfs2/dir.c switch(be32_to_cpu(h->mh_type)) { h 474 fs/gfs2/dir.c be32_to_cpu(h->mh_type)); h 544 fs/gfs2/dir.c struct gfs2_meta_header *h = (struct gfs2_meta_header *)bh->b_data; h 546 fs/gfs2/dir.c if (be32_to_cpu(h->mh_type) == GFS2_METATYPE_LF) { h 133 fs/gfs2/glock.c unsigned int h; h 135 fs/gfs2/glock.c h = jhash(&name->ln_number, sizeof(u64), 0); h 136 fs/gfs2/glock.c h = jhash(&name->ln_type, sizeof(unsigned int), h); h 137 fs/gfs2/glock.c h = jhash(&sdp, sizeof(struct gfs2_sbd *), h); h 138 fs/gfs2/glock.c h &= GFS2_GL_HASH_MASK; h 140 fs/gfs2/glock.c return h; h 215 fs/gfs2/glock.c struct hlist_node *h; h 217 fs/gfs2/glock.c hlist_for_each_entry(gl, h, &gl_hash_table[hash].hb_list, gl_list) { h 1719 fs/gfs2/glock.c struct hlist_node *h; h 1727 fs/gfs2/glock.c hlist_for_each_entry(gl, h, &gl_hash_table[x].hb_list, gl_list) { h 94 fs/hfs/hfs.h __be16 h; h 207 fs/hfsplus/hfsplus_raw.h __be16 h; h 235 fs/hpfs/dnode.c int h; h 291 fs/hpfs/dnode.c h = ((char *)dnode_last_de(nd) - (char *)nd) / 2 + 10; h 302 fs/hpfs/dnode.c for (de = dnode_first_de(nd); (char *)de_next_de(de) - (char *)nd < h; de = de_next_de(de)) { h 196 fs/hpfs/ea.c unsigned char h[4]; h 347 fs/hpfs/ea.c h[0] = 0; h 348 fs/hpfs/ea.c h[1] = strlen(key); h 349 fs/hpfs/ea.c h[2] = size & 0xff; h 350 fs/hpfs/ea.c h[3] = size >> 8; h 351 fs/hpfs/ea.c if (hpfs_ea_write(s, fnode->ea_secno, fnode->ea_anode, fnode->ea_size_l, 4, h)) goto bail; h 352 fs/hpfs/ea.c if (hpfs_ea_write(s, fnode->ea_secno, fnode->ea_anode, fnode->ea_size_l + 4, h[1] + 1, key)) goto bail; h 353 fs/hpfs/ea.c if (hpfs_ea_write(s, fnode->ea_secno, fnode->ea_anode, fnode->ea_size_l + 5 + h[1], size, data)) goto bail; h 85 fs/hugetlbfs/inode.c struct hstate *h = hstate_file(file); h 98 fs/hugetlbfs/inode.c if (vma->vm_pgoff & ~(huge_page_mask(h) >> PAGE_SHIFT)) h 110 fs/hugetlbfs/inode.c vma->vm_pgoff >> huge_page_order(h), h 111 fs/hugetlbfs/inode.c len >> huge_page_shift(h), vma)) h 136 fs/hugetlbfs/inode.c struct hstate *h = hstate_file(file); h 138 fs/hugetlbfs/inode.c if (len & ~huge_page_mask(h)) h 150 fs/hugetlbfs/inode.c addr = ALIGN(addr, huge_page_size(h)); h 163 fs/hugetlbfs/inode.c addr = ALIGN(start_addr, huge_page_size(h)); h 181 fs/hugetlbfs/inode.c addr = ALIGN(vma->vm_end, huge_page_size(h)); h 232 fs/hugetlbfs/inode.c struct hstate *h = hstate_file(filp); h 235 fs/hugetlbfs/inode.c unsigned long index = *ppos >> huge_page_shift(h); h 236 fs/hugetlbfs/inode.c unsigned long offset = *ppos & ~huge_page_mask(h); h 251 fs/hugetlbfs/inode.c end_index = (isize - 1) >> huge_page_shift(h); h 257 fs/hugetlbfs/inode.c nr = huge_page_size(h); h 261 fs/hugetlbfs/inode.c nr = ((isize - 1) & ~huge_page_mask(h)) + 1; h 295 fs/hugetlbfs/inode.c index += offset >> huge_page_shift(h); h 296 fs/hugetlbfs/inode.c offset &= ~huge_page_mask(h); h 306 fs/hugetlbfs/inode.c *ppos = ((loff_t)index << huge_page_shift(h)) + offset; h 347 fs/hugetlbfs/inode.c struct hstate *h = hstate_inode(inode); h 349 fs/hugetlbfs/inode.c const pgoff_t start = lstart >> huge_page_shift(h); h 458 fs/hugetlbfs/inode.c struct hstate *h = hstate_inode(inode); h 460 fs/hugetlbfs/inode.c BUG_ON(offset & ~huge_page_mask(h)); h 475 fs/hugetlbfs/inode.c struct hstate *h = hstate_inode(inode); h 487 fs/hugetlbfs/inode.c if (!(attr->ia_size & ~huge_page_mask(h))) h 621 fs/hugetlbfs/inode.c struct hstate *h = hstate_inode(dentry->d_inode); h 624 fs/hugetlbfs/inode.c buf->f_bsize = huge_page_size(h); h 827 fs/hugetlbfs/inode.c struct hstate *h = pconfig->hstate; h 829 fs/hugetlbfs/inode.c size <<= huge_page_shift(h); h 830 fs/hugetlbfs/inode.c size *= h->max_huge_pages; h 833 fs/hugetlbfs/inode.c pconfig->nr_blocks = (size >> huge_page_shift(h)); h 3898 fs/jfs/jfs_dtree.c struct dtslot *h, *t; h 3917 fs/jfs/jfs_dtree.c h = &p->slot[fsi]; h 3918 fs/jfs/jfs_dtree.c p->header.freelist = h->next; h 3930 fs/jfs/jfs_dtree.c lh = (struct ldtentry *) h; h 3931 fs/jfs/jfs_dtree.c lh->next = h->next; h 3945 fs/jfs/jfs_dtree.c ih = (struct idtentry *) h; h 3946 fs/jfs/jfs_dtree.c ih->next = h->next; h 3960 fs/jfs/jfs_dtree.c t = h; h 4003 fs/jfs/jfs_dtree.c if (h == t) { h 4064 fs/jfs/jfs_dtree.c struct dtslot *h, *s, *d; h 4115 fs/jfs/jfs_dtree.c h = d = &dp->slot[dsi]; h 4122 fs/jfs/jfs_dtree.c dlh = (struct ldtentry *) h; h 4143 fs/jfs/jfs_dtree.c dih = (struct idtentry *) h; h 4211 fs/jfs/jfs_dtree.c if (h == d) { h 584 fs/jfs/jfs_logmgr.c lspn = le32_to_cpu(lp->h.page); h 634 fs/jfs/jfs_logmgr.c lp->h.eor = lp->t.eor = cpu_to_le16(bp->l_ceor); h 655 fs/jfs/jfs_logmgr.c lp->h.page = lp->t.page = cpu_to_le32(lspn + 1); h 656 fs/jfs/jfs_logmgr.c lp->h.eor = lp->t.eor = cpu_to_le16(LOGPHDRSIZE); h 795 fs/jfs/jfs_logmgr.c lp->h.eor = lp->t.eor = cpu_to_le16(bp->l_ceor); h 803 fs/jfs/jfs_logmgr.c lp->h.eor = lp->t.eor = cpu_to_le16(bp->l_ceor); h 887 fs/jfs/jfs_logmgr.c lp->h.eor = lp->t.eor = cpu_to_le16(bp->l_eor); h 1370 fs/jfs/jfs_logmgr.c le16_to_cpu(lp->h.eor)); h 1399 fs/jfs/jfs_logmgr.c lp->h.eor = lp->t.eor = cpu_to_le16(bp->l_eor); h 1688 fs/jfs/jfs_logmgr.c lp->h.eor = lp->t.eor = cpu_to_le16(bp->l_eor); h 2465 fs/jfs/jfs_logmgr.c lp->h.page = lp->t.page = cpu_to_le32(npages - 3); h 2466 fs/jfs/jfs_logmgr.c lp->h.eor = lp->t.eor = cpu_to_le16(LOGPHDRSIZE + LOGRDSIZE); h 2485 fs/jfs/jfs_logmgr.c lp->h.page = lp->t.page = cpu_to_le32(lspn); h 2486 fs/jfs/jfs_logmgr.c lp->h.eor = lp->t.eor = cpu_to_le16(LOGPHDRSIZE); h 127 fs/jfs/jfs_logmgr.h } h; h 257 fs/ncpfs/sock.c h->conn_low = server->connection; h 258 fs/ncpfs/sock.c h->conn_high = server->connection >> 8; h 259 fs/ncpfs/sock.c h->sequence = ++server->sequence; h 265 fs/ncpfs/sock.c struct ncp_request_header* h; h 269 fs/ncpfs/sock.c h = req->tx_iov[1].iov_base; h 270 fs/ncpfs/sock.c ncp_init_header(server, req, h); h 294 fs/ncpfs/sock.c struct ncp_request_header* h; h 297 fs/ncpfs/sock.c h = req->tx_iov[1].iov_base; h 298 fs/ncpfs/sock.c ncp_init_header(server, req, h); h 796 fs/ncpfs/sock.c struct ncp_request_header *h; h 800 fs/ncpfs/sock.c h = (struct ncp_request_header *) (server->packet); h 802 fs/ncpfs/sock.c *(__u16 *) & (h->data[0]) = htons(server->current_size - sizeof(*h) - 2); h 804 fs/ncpfs/sock.c h->type = NCP_REQUEST; h 809 fs/ncpfs/sock.c h->task = 2; /* (current->pid) & 0xff; */ h 810 fs/ncpfs/sock.c h->function = function; h 832 fs/ncpfs/sock.c struct ncp_request_header *h; h 838 fs/ncpfs/sock.c h = (struct ncp_request_header *) (server->packet); h 839 fs/ncpfs/sock.c h->type = NCP_ALLOC_SLOT_REQUEST; h 840 fs/ncpfs/sock.c h->task = 2; /* see above */ h 841 fs/ncpfs/sock.c h->function = 0; h 843 fs/ncpfs/sock.c result = ncp_do_request(server, sizeof(*h), server->packet, server->packet_size); h 846 fs/ncpfs/sock.c server->connection = h->conn_low + (h->conn_high * 256); h 854 fs/ncpfs/sock.c struct ncp_request_header *h; h 856 fs/ncpfs/sock.c h = (struct ncp_request_header *) (server->packet); h 857 fs/ncpfs/sock.c h->type = NCP_DEALLOC_SLOT_REQUEST; h 858 fs/ncpfs/sock.c h->task = 2; /* see above */ h 859 fs/ncpfs/sock.c h->function = 0; h 861 fs/ncpfs/sock.c return ncp_do_request(server, sizeof(*h), server->packet, server->packet_size); h 158 fs/nfs/idmap.c return &h->h_entries[fnvhash32(name, len) % IDMAP_HASH_SZ]; h 164 fs/nfs/idmap.c struct idmap_hashent *he = idmap_name_hash(h, name, len); h 176 fs/nfs/idmap.c return &h->h_entries[fnvhash32(&id, sizeof(id)) % IDMAP_HASH_SZ]; h 182 fs/nfs/idmap.c struct idmap_hashent *he = idmap_id_hash(h, id); h 198 fs/nfs/idmap.c return idmap_name_hash(h, name, len); h 204 fs/nfs/idmap.c return idmap_id_hash(h, id); h 250 fs/nfs/idmap.c he = idmap_lookup_name(h, name, namelen); h 260 fs/nfs/idmap.c im->im_type = h->h_type; h 311 fs/nfs/idmap.c he = idmap_lookup_id(h, id); h 319 fs/nfs/idmap.c im->im_type = h->h_type; h 382 fs/nfs/idmap.c struct idmap_hashtable *h; h 411 fs/nfs/idmap.c h = &idmap->idmap_user_hash; h 414 fs/nfs/idmap.c h = &idmap->idmap_group_hash; h 431 fs/nfs/idmap.c he = idmap_alloc_id(h, im_in.im_id); h 442 fs/nfs/idmap.c he = idmap_alloc_name(h, im_in.im_name, namelen_in); h 64 fs/nfsd/export.c struct svc_expkey *key = container_of(ref, struct svc_expkey, h.ref); h 66 fs/nfsd/export.c if (test_bit(CACHE_VALID, &key->h.flags) && h 67 fs/nfsd/export.c !test_bit(CACHE_NEGATIVE, &key->h.flags)) h 78 fs/nfsd/export.c struct svc_expkey *ek = container_of(h, struct svc_expkey, h); h 138 fs/nfsd/export.c key.h.flags = 0; h 139 fs/nfsd/export.c key.h.expiry_time = get_expiry(&mesg); h 140 fs/nfsd/export.c if (key.h.expiry_time == 0) h 159 fs/nfsd/export.c set_bit(CACHE_NEGATIVE, &key.h.flags); h 162 fs/nfsd/export.c cache_put(&ek->h, &svc_expkey_cache); h 175 fs/nfsd/export.c cache_put(&ek->h, &svc_expkey_cache); h 195 fs/nfsd/export.c if (h ==NULL) { h 199 fs/nfsd/export.c ek = container_of(h, struct svc_expkey, h); h 204 fs/nfsd/export.c if (test_bit(CACHE_VALID, &h->flags) && h 205 fs/nfsd/export.c !test_bit(CACHE_NEGATIVE, &h->flags)) { h 215 fs/nfsd/export.c struct svc_expkey *orig = container_of(a, struct svc_expkey, h); h 216 fs/nfsd/export.c struct svc_expkey *new = container_of(b, struct svc_expkey, h); h 228 fs/nfsd/export.c struct svc_expkey *new = container_of(cnew, struct svc_expkey, h); h 229 fs/nfsd/export.c struct svc_expkey *item = container_of(citem, struct svc_expkey, h); h 241 fs/nfsd/export.c struct svc_expkey *new = container_of(cnew, struct svc_expkey, h); h 242 fs/nfsd/export.c struct svc_expkey *item = container_of(citem, struct svc_expkey, h); h 252 fs/nfsd/export.c return &i->h; h 284 fs/nfsd/export.c ch = sunrpc_cache_lookup(&svc_expkey_cache, &item->h, h 287 fs/nfsd/export.c return container_of(ch, struct svc_expkey, h); h 304 fs/nfsd/export.c ch = sunrpc_cache_update(&svc_expkey_cache, &new->h, h 305 fs/nfsd/export.c &old->h, hash); h 307 fs/nfsd/export.c return container_of(ch, struct svc_expkey, h); h 332 fs/nfsd/export.c struct svc_export *exp = container_of(ref, struct svc_export, h.ref); h 345 fs/nfsd/export.c struct svc_export *exp = container_of(h, struct svc_export, h); h 545 fs/nfsd/export.c exp.h.flags = 0; h 556 fs/nfsd/export.c exp.h.expiry_time = get_expiry(&mesg); h 557 fs/nfsd/export.c if (exp.h.expiry_time == 0) h 564 fs/nfsd/export.c set_bit(CACHE_NEGATIVE, &exp.h.flags); h 648 fs/nfsd/export.c if (h ==NULL) { h 652 fs/nfsd/export.c exp = container_of(h, struct svc_export, h); h 657 fs/nfsd/export.c if (test_bit(CACHE_VALID, &h->flags) && h 658 fs/nfsd/export.c !test_bit(CACHE_NEGATIVE, &h->flags)) { h 677 fs/nfsd/export.c struct svc_export *orig = container_of(a, struct svc_export, h); h 678 fs/nfsd/export.c struct svc_export *new = container_of(b, struct svc_export, h); h 686 fs/nfsd/export.c struct svc_export *new = container_of(cnew, struct svc_export, h); h 687 fs/nfsd/export.c struct svc_export *item = container_of(citem, struct svc_export, h); h 701 fs/nfsd/export.c struct svc_export *new = container_of(cnew, struct svc_export, h); h 702 fs/nfsd/export.c struct svc_export *item = container_of(citem, struct svc_export, h); h 729 fs/nfsd/export.c return &i->h; h 758 fs/nfsd/export.c ch = sunrpc_cache_lookup(&svc_export_cache, &exp->h, h 761 fs/nfsd/export.c return container_of(ch, struct svc_export, h); h 775 fs/nfsd/export.c ch = sunrpc_cache_update(&svc_export_cache, &new->h, h 776 fs/nfsd/export.c &old->h, h 779 fs/nfsd/export.c return container_of(ch, struct svc_export, h); h 801 fs/nfsd/export.c err = cache_check(&svc_expkey_cache, &ek->h, reqp); h 816 fs/nfsd/export.c key.h.expiry_time = NEVER; h 817 fs/nfsd/export.c key.h.flags = 0; h 823 fs/nfsd/export.c cache_put(&ek->h, &svc_expkey_cache); h 875 fs/nfsd/export.c err = cache_check(&svc_export_cache, &exp->h, reqp); h 947 fs/nfsd/export.c ek->h.expiry_time = get_seconds()-1; h 948 fs/nfsd/export.c cache_put(&ek->h, &svc_expkey_cache); h 985 fs/nfsd/export.c ek->h.expiry_time = get_seconds()-1; h 986 fs/nfsd/export.c cache_put(&ek->h, &svc_expkey_cache); h 1061 fs/nfsd/export.c new.h.expiry_time = NEVER; h 1062 fs/nfsd/export.c new.h.flags = 0; h 1092 fs/nfsd/export.c cache_put(&fsid_key->h, &svc_expkey_cache); h 1109 fs/nfsd/export.c unexp->h.expiry_time = get_seconds()-1; h 1219 fs/nfsd/export.c cache_put(&ek->h, &svc_expkey_cache); h 1520 fs/nfsd/export.c struct svc_export *exp = container_of(cp, struct svc_export, h); h 1528 fs/nfsd/export.c cache_get(&exp->h); h 1529 fs/nfsd/export.c if (cache_check(&svc_export_cache, &exp->h, NULL)) h 1531 fs/nfsd/export.c cache_put(&exp->h, &svc_export_cache); h 69 fs/nfsd/nfs4idmap.c struct cache_head h; h 85 fs/nfsd/nfs4idmap.c struct ent *new = container_of(cnew, struct ent, h); h 86 fs/nfsd/nfs4idmap.c struct ent *itm = container_of(citm, struct ent, h); h 98 fs/nfsd/nfs4idmap.c struct ent *map = container_of(ref, struct ent, h.ref); h 107 fs/nfsd/nfs4idmap.c return &e->h; h 137 fs/nfsd/nfs4idmap.c struct ent *ent = container_of(ch, struct ent, h); h 151 fs/nfsd/nfs4idmap.c struct ent *a = container_of(ca, struct ent, h); h 152 fs/nfsd/nfs4idmap.c struct ent *b = container_of(cb, struct ent, h); h 163 fs/nfsd/nfs4idmap.c if (h == NULL) { h 167 fs/nfsd/nfs4idmap.c ent = container_of(h, struct ent, h); h 171 fs/nfsd/nfs4idmap.c if (test_bit(CACHE_VALID, &h->flags)) h 242 fs/nfsd/nfs4idmap.c ent.h.expiry_time = get_expiry(&buf); h 243 fs/nfsd/nfs4idmap.c if (ent.h.expiry_time == 0) h 257 fs/nfsd/nfs4idmap.c set_bit(CACHE_NEGATIVE, &ent.h.flags); h 267 fs/nfsd/nfs4idmap.c cache_put(&res->h, &idtoname_cache); h 281 fs/nfsd/nfs4idmap.c &item->h, h 284 fs/nfsd/nfs4idmap.c return container_of(ch, struct ent, h); h 293 fs/nfsd/nfs4idmap.c &new->h, &old->h, h 296 fs/nfsd/nfs4idmap.c return container_of(ch, struct ent, h); h 318 fs/nfsd/nfs4idmap.c struct ent *ent = container_of(ch, struct ent, h); h 330 fs/nfsd/nfs4idmap.c struct ent *a = container_of(ca, struct ent, h); h 331 fs/nfsd/nfs4idmap.c struct ent *b = container_of(cb, struct ent, h); h 342 fs/nfsd/nfs4idmap.c if (h == NULL) { h 346 fs/nfsd/nfs4idmap.c ent = container_of(h, struct ent, h); h 350 fs/nfsd/nfs4idmap.c if (test_bit(CACHE_VALID, &h->flags)) h 411 fs/nfsd/nfs4idmap.c ent.h.expiry_time = get_expiry(&buf); h 412 fs/nfsd/nfs4idmap.c if (ent.h.expiry_time == 0) h 420 fs/nfsd/nfs4idmap.c set_bit(CACHE_NEGATIVE, &ent.h.flags); h 430 fs/nfsd/nfs4idmap.c cache_put(&res->h, &nametoid_cache); h 443 fs/nfsd/nfs4idmap.c &item->h, h 446 fs/nfsd/nfs4idmap.c return container_of(ch, struct ent, h); h 455 fs/nfsd/nfs4idmap.c &new->h, &old->h, h 458 fs/nfsd/nfs4idmap.c return container_of(ch, struct ent, h); h 541 fs/nfsd/nfs4idmap.c return cache_check(detail, &(*item)->h, &mdr->req); h 555 fs/nfsd/nfs4idmap.c if (!test_bit(CACHE_VALID, &(*item)->h.flags) h 556 fs/nfsd/nfs4idmap.c || (*item)->h.expiry_time < get_seconds() h 557 fs/nfsd/nfs4idmap.c || detail->flush_time > (*item)->h.last_refresh) h 560 fs/nfsd/nfs4idmap.c if (test_bit(CACHE_NEGATIVE, &(*item)->h.flags)) h 564 fs/nfsd/nfs4idmap.c cache_put(&(*item)->h, detail); h 587 fs/nfsd/nfs4idmap.c test_bit(CACHE_VALID, &(*item)->h.flags), 1 * HZ); h 623 fs/nfsd/nfs4idmap.c cache_put(&item->h, &nametoid_cache); h 645 fs/nfsd/nfs4idmap.c cache_put(&item->h, &idtoname_cache); h 60 fs/nfsd/nfs4proc.c cache_get(&src->fh_export->h); h 498 fs/nfsd/nfsfh.c cache_get(&exp->h); h 592 fs/nfsd/nfsfh.c cache_put(&exp->h, &svc_export_cache); h 19 fs/nls/nls_euc-jp.c #define IS_SJIS_JISX0208(h, l) ((((0x81 <= (h)) && ((h) <= 0x9F)) \ h 20 fs/nls/nls_euc-jp.c || ((0xE0 <= (h)) && ((h) <= 0xEA))) \ h 23 fs/nls/nls_euc-jp.c #define IS_SJIS_UDC_LOW(h, l) (((0xF0 <= (h)) && ((h) <= 0xF4)) \ h 25 fs/nls/nls_euc-jp.c #define IS_SJIS_UDC_HI(h, l) (((0xF5 <= (h)) && ((h) <= 0xF9)) \ h 27 fs/nls/nls_euc-jp.c #define IS_SJIS_IBM(h, l) (((0xFA <= (h)) && ((h) <= 0xFC)) \ h 29 fs/nls/nls_euc-jp.c #define IS_SJIS_NECIBM(h, l) (((0xED <= (h)) && ((h) <= 0xEE)) \ h 44 fs/nls/nls_euc-jp.c #define IS_EUC_JISX0208(h, l) (IS_EUC_BYTE(h) && IS_EUC_BYTE(l)) h 45 fs/nls/nls_euc-jp.c #define IS_EUC_JISX0201KANA(h, l) (((h) == SS2) && (0xA1 <= (l) && (l) <= 0xDF)) h 46 fs/nls/nls_euc-jp.c #define IS_EUC_UDC_LOW(h, l) (((0xF5 <= (h)) && ((h) <= 0xFE)) \ h 48 fs/nls/nls_euc-jp.c #define IS_EUC_UDC_HI(h, l) IS_EUC_UDC_LOW(h, l) /* G3 block */ h 142 fs/nls/nls_euc-jp.c (((h) == 0xA2 && (l) == 0xCC) || ((h) == 0xA2 && (l) == 0xE8)) h 77 fs/partitions/ldm.c int h; h 80 fs/partitions/ldm.c if ((x = src[0] - '0') <= '9'-'0') h = x; h 81 fs/partitions/ldm.c else if ((x = src[0] - 'a') <= 'f'-'a') h = x+10; h 82 fs/partitions/ldm.c else if ((x = src[0] - 'A') <= 'F'-'A') h = x+10; h 84 fs/partitions/ldm.c h <<= 4; h 87 fs/partitions/ldm.c if ((x = src[1] - '0') <= '9'-'0') return h | x; h 88 fs/partitions/ldm.c if ((x = src[1] - 'a') <= 'f'-'a') return h | (x+10); h 89 fs/partitions/ldm.c if ((x = src[1] - 'A') <= 'F'-'A') return h | (x+10); h 82 fs/proc/proc_sysctl.c struct ctl_table_header *h = NULL; h 100 fs/proc/proc_sysctl.c for (h = sysctl_head_next(NULL); h; h = sysctl_head_next(h)) { h 101 fs/proc/proc_sysctl.c if (h->attached_to != table) h 103 fs/proc/proc_sysctl.c p = find_in_table(h->attached_by, name); h 113 fs/proc/proc_sysctl.c inode = proc_sys_make_inode(dir->i_sb, h ? h : head, p); h 114 fs/proc/proc_sysctl.c if (h) h 115 fs/proc/proc_sysctl.c sysctl_head_finish(h); h 246 fs/proc/proc_sysctl.c struct ctl_table_header *h = NULL; h 280 fs/proc/proc_sysctl.c for (h = sysctl_head_next(NULL); h; h = sysctl_head_next(h)) { h 281 fs/proc/proc_sysctl.c if (h->attached_to != table) h 283 fs/proc/proc_sysctl.c ret = scan(h, h->attached_by, &pos, filp, dirent, filldir); h 285 fs/proc/proc_sysctl.c sysctl_head_finish(h); h 1828 fs/reiserfs/do_balan.c int Sh_position = PATH_H_POSITION(tb->tb_path, h + 1); h 1830 fs/reiserfs/do_balan.c RFALSE(PATH_H_PPARENT(tb->tb_path, h) == NULL || tb->FL[h] == NULL, h 1832 fs/reiserfs/do_balan.c h, tb->FL[h], h, PATH_H_PPARENT(tb->tb_path, h)); h 1835 fs/reiserfs/do_balan.c return B_NR_ITEMS(tb->FL[h]); h 1842 fs/reiserfs/do_balan.c int Sh_position = PATH_H_POSITION(tb->tb_path, h + 1); h 1844 fs/reiserfs/do_balan.c RFALSE(PATH_H_PPARENT(tb->tb_path, h) == NULL || tb->FR[h] == NULL, h 1846 fs/reiserfs/do_balan.c h, PATH_H_PPARENT(tb->tb_path, h), h, tb->FR[h]); h 1848 fs/reiserfs/do_balan.c if (Sh_position == B_NR_ITEMS(PATH_H_PPARENT(tb->tb_path, h))) h 1983 fs/reiserfs/do_balan.c int h; h 1986 fs/reiserfs/do_balan.c for (h = 1; tb->insert_size[h]; h++) { h 1987 fs/reiserfs/do_balan.c check_internal_node(tb->tb_sb, PATH_H_PBUFFER(tb->tb_path, h), h 1989 fs/reiserfs/do_balan.c if (tb->lnum[h]) h 1990 fs/reiserfs/do_balan.c check_internal_node(tb->tb_sb, tb->L[h], "BAD L"); h 1991 fs/reiserfs/do_balan.c if (tb->rnum[h]) h 1992 fs/reiserfs/do_balan.c check_internal_node(tb->tb_sb, tb->R[h], "BAD R"); h 2090 fs/reiserfs/do_balan.c h; /* level of the tree being processed */ h 2131 fs/reiserfs/do_balan.c for (h = 1; h < MAX_HEIGHT && tb->insert_size[h]; h++) h 2133 fs/reiserfs/do_balan.c balance_internal(tb, h, child_pos, insert_key, insert_ptr); h 84 fs/reiserfs/fix_node.c Sh = PATH_H_PBUFFER(tb->tb_path, h); h 88 fs/reiserfs/fix_node.c MAX_CHILD_SIZE(Sh) - B_FREE_SPACE(Sh) + tb->insert_size[h]; h 91 fs/reiserfs/fix_node.c if (h) { h 212 fs/reiserfs/fix_node.c if (h > 0) { h 213 fs/reiserfs/fix_node.c tb->lnum[h] = cur_free / (DC_SIZE + KEY_SIZE); h 221 fs/reiserfs/fix_node.c tb->lnum[h] = 0; h 292 fs/reiserfs/fix_node.c if (h > 0) { h 293 fs/reiserfs/fix_node.c tb->rnum[h] = cur_free / (DC_SIZE + KEY_SIZE); h 301 fs/reiserfs/fix_node.c tb->rnum[h] = 0; h 318 fs/reiserfs/fix_node.c tb->rnum[h] = vn->vn_nr_item; h 395 fs/reiserfs/fix_node.c RFALSE(tb->insert_size[h] < 0 || (mode != M_INSERT && mode != M_PASTE), h 398 fs/reiserfs/fix_node.c max_node_size = MAX_CHILD_SIZE(PATH_H_PBUFFER(tb->tb_path, h)); h 406 fs/reiserfs/fix_node.c if (h > 0) { h 591 fs/reiserfs/fix_node.c tb->lnum[h] = lnum; h 592 fs/reiserfs/fix_node.c tb->rnum[h] = rnum; h 593 fs/reiserfs/fix_node.c tb->blknum[h] = blk_num; h 595 fs/reiserfs/fix_node.c if (h == 0) { /* only for leaf level */ h 605 fs/reiserfs/fix_node.c PROC_INFO_ADD(tb->tb_sb, lnum[h], lnum); h 606 fs/reiserfs/fix_node.c PROC_INFO_ADD(tb->tb_sb, rnum[h], rnum); h 608 fs/reiserfs/fix_node.c PROC_INFO_ADD(tb->tb_sb, lbytes[h], lb); h 609 fs/reiserfs/fix_node.c PROC_INFO_ADD(tb->tb_sb, rbytes[h], rb); h 714 fs/reiserfs/fix_node.c if (h)\ h 721 fs/reiserfs/fix_node.c set_parameters (tb, h, to_l, 0, lnver, NULL, -1, -1);\ h 726 fs/reiserfs/fix_node.c set_parameters (tb, h, lpar, 0, lnver, snum012+lset,\ h 729 fs/reiserfs/fix_node.c set_parameters (tb, h, lpar - (tb->lbytes!=-1), 0, lnver, snum012+lset,\ h 734 fs/reiserfs/fix_node.c if (h)\ h 740 fs/reiserfs/fix_node.c set_parameters (tb, h, 0, to_r, rnver, NULL, -1, -1);\ h 745 fs/reiserfs/fix_node.c set_parameters (tb, h, 0, rpar, rnver, snum012+rset,\ h 748 fs/reiserfs/fix_node.c set_parameters (tb, h, 0, rpar - (tb->rbytes!=-1), rnver, snum012+rset,\ h 860 fs/reiserfs/fix_node.c if ((f = PATH_H_PPARENT(tb->tb_path, h)) == NULL || h 861 fs/reiserfs/fix_node.c (l = tb->FL[h]) == NULL) h 865 fs/reiserfs/fix_node.c order = PATH_H_B_ITEM_ORDER(tb->tb_path, h) - 1; h 882 fs/reiserfs/fix_node.c if ((f = PATH_H_PPARENT(tb->tb_path, h)) == NULL || h 883 fs/reiserfs/fix_node.c (r = tb->FR[h]) == NULL) h 887 fs/reiserfs/fix_node.c order = PATH_H_B_ITEM_ORDER(tb->tb_path, h) + 1; h 1166 fs/reiserfs/fix_node.c struct buffer_head *Sh = PATH_H_PBUFFER(tb->tb_path, h); h 1167 fs/reiserfs/fix_node.c int levbytes = tb->insert_size[h]; h 1172 fs/reiserfs/fix_node.c if (tb->CFR[h]) h 1173 fs/reiserfs/fix_node.c r_key = B_N_PDELIM_KEY(tb->CFR[h], tb->rkey[h]); h 1178 fs/reiserfs/fix_node.c ((!h h 1181 fs/reiserfs/fix_node.c ((!h && r_key h 1183 fs/reiserfs/fix_node.c + ((h) ? KEY_SIZE : 0)) { h 1186 fs/reiserfs/fix_node.c if (!h) h 1190 fs/reiserfs/fix_node.c set_parameters(tb, h, 0, 0, 1, NULL, -1, -1); h 1194 fs/reiserfs/fix_node.c PROC_INFO_INC(tb->tb_sb, can_node_be_removed[h]); h 1252 fs/reiserfs/fix_node.c Sh = PATH_H_PBUFFER(tb->tb_path, h); h 1253 fs/reiserfs/fix_node.c levbytes = tb->insert_size[h]; h 1257 fs/reiserfs/fix_node.c if (!h) h 1260 fs/reiserfs/fix_node.c switch (n_ret_value = get_empty_nodes(tb, h)) { h 1262 fs/reiserfs/fix_node.c set_parameters(tb, h, 0, 0, 1, NULL, -1, -1); h 1274 fs/reiserfs/fix_node.c if ((n_ret_value = get_parents(tb, h)) != CARRY_ON) /* get parents of S[h] neighbors. */ h 1280 fs/reiserfs/fix_node.c rfree = get_rfree(tb, h); h 1281 fs/reiserfs/fix_node.c lfree = get_lfree(tb, h); h 1283 fs/reiserfs/fix_node.c if (can_node_be_removed(vn->vn_mode, lfree, sfree, rfree, tb, h) == h 1288 fs/reiserfs/fix_node.c create_virtual_node(tb, h); h 1295 fs/reiserfs/fix_node.c check_left(tb, h, lfree); h 1302 fs/reiserfs/fix_node.c check_right(tb, h, rfree); h 1306 fs/reiserfs/fix_node.c if (h && (tb->rnum[h] + tb->lnum[h] >= vn->vn_nr_item + 1)) { h 1316 fs/reiserfs/fix_node.c ((MAX_NR_KEY(Sh) << 1) + 2 - tb->lnum[h] - tb->rnum[h] + h 1318 fs/reiserfs/fix_node.c tb->rnum[h]); h 1319 fs/reiserfs/fix_node.c set_parameters(tb, h, vn->vn_nr_item + 1 - to_r, to_r, 0, NULL, h 1325 fs/reiserfs/fix_node.c RFALSE(h && h 1326 fs/reiserfs/fix_node.c (tb->lnum[h] >= vn->vn_nr_item + 1 || h 1327 fs/reiserfs/fix_node.c tb->rnum[h] >= vn->vn_nr_item + 1), h 1329 fs/reiserfs/fix_node.c RFALSE(!h && ((tb->lnum[h] >= vn->vn_nr_item && (tb->lbytes == -1)) || h 1330 fs/reiserfs/fix_node.c (tb->rnum[h] >= vn->vn_nr_item && (tb->rbytes == -1))), h 1335 fs/reiserfs/fix_node.c if (!h && is_leaf_removable(tb)) h 1343 fs/reiserfs/fix_node.c if (!h) h 1345 fs/reiserfs/fix_node.c set_parameters(tb, h, 0, 0, 1, NULL, -1, -1); h 1372 fs/reiserfs/fix_node.c lpar = tb->lnum[h]; h 1373 fs/reiserfs/fix_node.c rpar = tb->rnum[h]; h 1380 fs/reiserfs/fix_node.c nver = get_num_ver(vn->vn_mode, tb, h, h 1381 fs/reiserfs/fix_node.c 0, -1, h ? vn->vn_nr_item : 0, -1, h 1384 fs/reiserfs/fix_node.c if (!h) { h 1388 fs/reiserfs/fix_node.c nver1 = get_num_ver(vn->vn_mode, tb, h, h 1402 fs/reiserfs/fix_node.c lnver = get_num_ver(vn->vn_mode, tb, h, h 1403 fs/reiserfs/fix_node.c lpar - ((h || tb->lbytes == -1) ? 0 : 1), h 1404 fs/reiserfs/fix_node.c -1, h ? vn->vn_nr_item : 0, -1, h 1406 fs/reiserfs/fix_node.c if (!h) { h 1409 fs/reiserfs/fix_node.c lnver1 = get_num_ver(vn->vn_mode, tb, h, h 1425 fs/reiserfs/fix_node.c rnver = get_num_ver(vn->vn_mode, tb, h, h 1427 fs/reiserfs/fix_node.c h ? (vn->vn_nr_item - rpar) : (rpar - h 1433 fs/reiserfs/fix_node.c if (!h) { h 1436 fs/reiserfs/fix_node.c rnver1 = get_num_ver(vn->vn_mode, tb, h, h 1453 fs/reiserfs/fix_node.c lrnver = get_num_ver(vn->vn_mode, tb, h, h 1454 fs/reiserfs/fix_node.c lpar - ((h || tb->lbytes == -1) ? 0 : 1), h 1456 fs/reiserfs/fix_node.c h ? (vn->vn_nr_item - rpar) : (rpar - h 1462 fs/reiserfs/fix_node.c if (!h) { h 1465 fs/reiserfs/fix_node.c lrnver1 = get_num_ver(vn->vn_mode, tb, h, h 1484 fs/reiserfs/fix_node.c RFALSE(h && h 1485 fs/reiserfs/fix_node.c (tb->lnum[h] != 1 || h 1486 fs/reiserfs/fix_node.c tb->rnum[h] != 1 || h 1488 fs/reiserfs/fix_node.c || h != 1), "vs-8230: bad h"); h 1490 fs/reiserfs/fix_node.c set_parameters(tb, h, tb->lnum[h], tb->rnum[h], h 1494 fs/reiserfs/fix_node.c set_parameters(tb, h, h 1495 fs/reiserfs/fix_node.c tb->lnum[h] - h 1497 fs/reiserfs/fix_node.c tb->rnum[h] - h 1506 fs/reiserfs/fix_node.c set_parameters(tb, h, 0, 0, nver, snum012 + nset, -1, h 1528 fs/reiserfs/fix_node.c if (is_left_neighbor_in_cache(tb, h)) { h 1565 fs/reiserfs/fix_node.c Sh = PATH_H_PBUFFER(tb->tb_path, h); h 1566 fs/reiserfs/fix_node.c Fh = PATH_H_PPARENT(tb->tb_path, h); h 1573 fs/reiserfs/fix_node.c create_virtual_node(tb, h); h 1577 fs/reiserfs/fix_node.c set_parameters(tb, h, 0, 0, 1, NULL, -1, -1); h 1583 fs/reiserfs/fix_node.c set_parameters(tb, h, 0, 0, 0, NULL, -1, -1); h 1587 fs/reiserfs/fix_node.c if ((n_ret_value = get_parents(tb, h)) != CARRY_ON) h 1591 fs/reiserfs/fix_node.c rfree = get_rfree(tb, h); h 1592 fs/reiserfs/fix_node.c lfree = get_lfree(tb, h); h 1595 fs/reiserfs/fix_node.c check_left(tb, h, lfree); h 1596 fs/reiserfs/fix_node.c check_right(tb, h, rfree); h 1602 fs/reiserfs/fix_node.c if (tb->lnum[h] >= vn->vn_nr_item + 1) { h 1610 fs/reiserfs/fix_node.c h)) == h 1611 fs/reiserfs/fix_node.c 0) ? B_NR_ITEMS(tb->FL[h]) : n - 1; h 1612 fs/reiserfs/fix_node.c n = dc_size(B_N_CHILD(tb->FL[h], order_L)) / h 1614 fs/reiserfs/fix_node.c set_parameters(tb, h, -n - 1, 0, 0, NULL, -1, h 1619 fs/reiserfs/fix_node.c if (tb->rnum[h] >= vn->vn_nr_item + 1) { h 1627 fs/reiserfs/fix_node.c h)) == h 1629 fs/reiserfs/fix_node.c n = dc_size(B_N_CHILD(tb->FR[h], order_R)) / h 1631 fs/reiserfs/fix_node.c set_parameters(tb, h, 0, -n - 1, 0, NULL, -1, h 1637 fs/reiserfs/fix_node.c if (tb->rnum[h] + tb->lnum[h] >= vn->vn_nr_item + 1) { h 1642 fs/reiserfs/fix_node.c ((MAX_NR_KEY(Sh) << 1) + 2 - tb->lnum[h] - h 1643 fs/reiserfs/fix_node.c tb->rnum[h] + vn->vn_nr_item + 1) / 2 - h 1644 fs/reiserfs/fix_node.c (MAX_NR_KEY(Sh) + 1 - tb->rnum[h]); h 1645 fs/reiserfs/fix_node.c set_parameters(tb, h, vn->vn_nr_item + 1 - to_r, to_r, h 1651 fs/reiserfs/fix_node.c set_parameters(tb, h, 0, 0, 1, NULL, -1, -1); h 1657 fs/reiserfs/fix_node.c if (tb->lnum[h] >= vn->vn_nr_item + 1) h 1658 fs/reiserfs/fix_node.c if (is_left_neighbor_in_cache(tb, h) h 1659 fs/reiserfs/fix_node.c || tb->rnum[h] < vn->vn_nr_item + 1 || !tb->FR[h]) { h 1666 fs/reiserfs/fix_node.c h)) == h 1667 fs/reiserfs/fix_node.c 0) ? B_NR_ITEMS(tb->FL[h]) : n - 1; h 1668 fs/reiserfs/fix_node.c n = dc_size(B_N_CHILD(tb->FL[h], order_L)) / (DC_SIZE + h 1670 fs/reiserfs/fix_node.c set_parameters(tb, h, -n - 1, 0, 0, NULL, -1, -1); h 1675 fs/reiserfs/fix_node.c if (tb->rnum[h] >= vn->vn_nr_item + 1) { h 1682 fs/reiserfs/fix_node.c h)) == B_NR_ITEMS(Fh)) ? 0 : (n + 1); h 1683 fs/reiserfs/fix_node.c n = dc_size(B_N_CHILD(tb->FR[h], order_R)) / (DC_SIZE + h 1685 fs/reiserfs/fix_node.c set_parameters(tb, h, 0, -n - 1, 0, NULL, -1, -1); h 1690 fs/reiserfs/fix_node.c if (tb->rnum[h] + tb->lnum[h] >= vn->vn_nr_item + 1) { h 1694 fs/reiserfs/fix_node.c ((MAX_NR_KEY(Sh) << 1) + 2 - tb->lnum[h] - tb->rnum[h] + h 1696 fs/reiserfs/fix_node.c tb->rnum[h]); h 1697 fs/reiserfs/fix_node.c set_parameters(tb, h, vn->vn_nr_item + 1 - to_r, to_r, 0, NULL, h 1703 fs/reiserfs/fix_node.c RFALSE(!tb->FL[h] && !tb->FR[h], "vs-8235: trying to borrow for root"); h 1706 fs/reiserfs/fix_node.c if (is_left_neighbor_in_cache(tb, h) || !tb->FR[h]) { h 1710 fs/reiserfs/fix_node.c (MAX_NR_KEY(Sh) + 1 - tb->lnum[h] + vn->vn_nr_item + h 1712 fs/reiserfs/fix_node.c set_parameters(tb, h, -from_l, 0, 1, NULL, -1, -1); h 1716 fs/reiserfs/fix_node.c set_parameters(tb, h, 0, h 1717 fs/reiserfs/fix_node.c -((MAX_NR_KEY(Sh) + 1 - tb->rnum[h] + vn->vn_nr_item + h 1754 fs/reiserfs/fix_node.c levbytes = tb->insert_size[h]; h 1763 fs/reiserfs/fix_node.c set_parameters(tb, h, 0, 0, 1, NULL, -1, -1); h 1767 fs/reiserfs/fix_node.c if ((n_ret_value = get_parents(tb, h)) != CARRY_ON) h 1771 fs/reiserfs/fix_node.c rfree = get_rfree(tb, h); h 1772 fs/reiserfs/fix_node.c lfree = get_lfree(tb, h); h 1774 fs/reiserfs/fix_node.c create_virtual_node(tb, h); h 1784 fs/reiserfs/fix_node.c check_left(tb, h, lfree); h 1785 fs/reiserfs/fix_node.c check_right(tb, h, rfree); h 1789 fs/reiserfs/fix_node.c if (is_left_neighbor_in_cache(tb, h) || ((tb->rnum[0] - ((tb->rbytes == -1) ? 0 : 1)) < vn->vn_nr_item) || /* S can not be merged with R */ h 1790 fs/reiserfs/fix_node.c !tb->FR[h]) { h 1792 fs/reiserfs/fix_node.c RFALSE(!tb->FL[h], h 1796 fs/reiserfs/fix_node.c set_parameters(tb, h, -1, 0, 0, NULL, -1, -1); h 1802 fs/reiserfs/fix_node.c set_parameters(tb, h, 0, -1, 0, NULL, -1, -1); h 1812 fs/reiserfs/fix_node.c set_parameters(tb, h, 0, 0, 1, NULL, -1, -1); h 1831 fs/reiserfs/fix_node.c RFALSE(!(PATH_H_PBUFFER(tb->tb_path, h)), h 1834 fs/reiserfs/fix_node.c if (h) h 1835 fs/reiserfs/fix_node.c return dc_check_balance_internal(tb, h); h 1837 fs/reiserfs/fix_node.c return dc_check_balance_leaf(tb, h); h 1878 fs/reiserfs/fix_node.c if (tb->insert_size[h] > 0) h 1880 fs/reiserfs/fix_node.c return ip_check_balance(tb, h); h 1883 fs/reiserfs/fix_node.c return dc_check_balance(tb, h); h 37 fs/reiserfs/ibalance.c src_bi->bi_bh = PATH_H_PBUFFER(tb->tb_path, h); h 38 fs/reiserfs/ibalance.c src_bi->bi_parent = PATH_H_PPARENT(tb->tb_path, h); h 39 fs/reiserfs/ibalance.c src_bi->bi_position = PATH_H_POSITION(tb->tb_path, h + 1); h 41 fs/reiserfs/ibalance.c dest_bi->bi_bh = tb->L[h]; h 42 fs/reiserfs/ibalance.c dest_bi->bi_parent = tb->FL[h]; h 43 fs/reiserfs/ibalance.c dest_bi->bi_position = get_left_neighbor_position(tb, h); h 44 fs/reiserfs/ibalance.c *d_key = tb->lkey[h]; h 45 fs/reiserfs/ibalance.c *cf = tb->CFL[h]; h 49 fs/reiserfs/ibalance.c src_bi->bi_bh = tb->L[h]; h 50 fs/reiserfs/ibalance.c src_bi->bi_parent = tb->FL[h]; h 51 fs/reiserfs/ibalance.c src_bi->bi_position = get_left_neighbor_position(tb, h); h 53 fs/reiserfs/ibalance.c dest_bi->bi_bh = PATH_H_PBUFFER(tb->tb_path, h); h 54 fs/reiserfs/ibalance.c dest_bi->bi_parent = PATH_H_PPARENT(tb->tb_path, h); h 55 fs/reiserfs/ibalance.c dest_bi->bi_position = PATH_H_POSITION(tb->tb_path, h + 1); /* dest position is analog of dest->b_item_order */ h 56 fs/reiserfs/ibalance.c *d_key = tb->lkey[h]; h 57 fs/reiserfs/ibalance.c *cf = tb->CFL[h]; h 62 fs/reiserfs/ibalance.c src_bi->bi_bh = tb->R[h]; h 63 fs/reiserfs/ibalance.c src_bi->bi_parent = tb->FR[h]; h 64 fs/reiserfs/ibalance.c src_bi->bi_position = get_right_neighbor_position(tb, h); h 66 fs/reiserfs/ibalance.c dest_bi->bi_bh = PATH_H_PBUFFER(tb->tb_path, h); h 67 fs/reiserfs/ibalance.c dest_bi->bi_parent = PATH_H_PPARENT(tb->tb_path, h); h 68 fs/reiserfs/ibalance.c dest_bi->bi_position = PATH_H_POSITION(tb->tb_path, h + 1); h 69 fs/reiserfs/ibalance.c *d_key = tb->rkey[h]; h 70 fs/reiserfs/ibalance.c *cf = tb->CFR[h]; h 75 fs/reiserfs/ibalance.c src_bi->bi_bh = PATH_H_PBUFFER(tb->tb_path, h); h 76 fs/reiserfs/ibalance.c src_bi->bi_parent = PATH_H_PPARENT(tb->tb_path, h); h 77 fs/reiserfs/ibalance.c src_bi->bi_position = PATH_H_POSITION(tb->tb_path, h + 1); h 79 fs/reiserfs/ibalance.c dest_bi->bi_bh = tb->R[h]; h 80 fs/reiserfs/ibalance.c dest_bi->bi_parent = tb->FR[h]; h 81 fs/reiserfs/ibalance.c dest_bi->bi_position = get_right_neighbor_position(tb, h); h 82 fs/reiserfs/ibalance.c *d_key = tb->rkey[h]; h 83 fs/reiserfs/ibalance.c *cf = tb->CFR[h]; h 88 fs/reiserfs/ibalance.c dest_bi->bi_bh = tb->L[h]; h 89 fs/reiserfs/ibalance.c dest_bi->bi_parent = tb->FL[h]; h 90 fs/reiserfs/ibalance.c dest_bi->bi_position = get_left_neighbor_position(tb, h); h 95 fs/reiserfs/ibalance.c dest_bi->bi_bh = PATH_H_PBUFFER(tb->tb_path, h); h 96 fs/reiserfs/ibalance.c dest_bi->bi_parent = PATH_H_PPARENT(tb->tb_path, h); h 97 fs/reiserfs/ibalance.c dest_bi->bi_position = PATH_H_POSITION(tb->tb_path, h + 1); h 102 fs/reiserfs/ibalance.c dest_bi->bi_bh = tb->R[h]; h 103 fs/reiserfs/ibalance.c dest_bi->bi_parent = tb->FR[h]; h 104 fs/reiserfs/ibalance.c dest_bi->bi_position = get_right_neighbor_position(tb, h); h 470 fs/reiserfs/ibalance.c internal_define_dest_src_infos(mode, tb, h, &dest_bi, &src_bi, h 507 fs/reiserfs/ibalance.c internal_define_dest_src_infos(INTERNAL_SHIFT_FROM_S_TO_L, tb, h, h 535 fs/reiserfs/ibalance.c internal_define_dest_src_infos(mode, tb, h, &dest_bi, &src_bi, h 544 fs/reiserfs/ibalance.c RFALSE(src_bi.bi_bh != PATH_H_PBUFFER(tb->tb_path, h) /*tb->S[h] */ || h 545 fs/reiserfs/ibalance.c dest_bi.bi_bh != tb->R[h], h 547 fs/reiserfs/ibalance.c src_bi.bi_bh, PATH_H_PBUFFER(tb->tb_path, h)); h 549 fs/reiserfs/ibalance.c if (tb->CFL[h]) h 550 fs/reiserfs/ibalance.c replace_key(tb, cf, d_key_position, tb->CFL[h], h 551 fs/reiserfs/ibalance.c tb->lkey[h]); h 574 fs/reiserfs/ibalance.c internal_define_dest_src_infos(INTERNAL_SHIFT_FROM_S_TO_R, tb, h, h 594 fs/reiserfs/ibalance.c struct buffer_head *tbSh = PATH_H_PBUFFER(tb->tb_path, h); h 597 fs/reiserfs/ibalance.c insert_num = tb->insert_size[h] / ((int)(DC_SIZE + KEY_SIZE)); h 602 fs/reiserfs/ibalance.c bi.bi_parent = PATH_H_PPARENT(tb->tb_path, h); h 603 fs/reiserfs/ibalance.c bi.bi_position = PATH_H_POSITION(tb->tb_path, h + 1); h 607 fs/reiserfs/ibalance.c RFALSE(tb->blknum[h] > 1, h 608 fs/reiserfs/ibalance.c "tb->blknum[%d]=%d when insert_size < 0", h, tb->blknum[h]); h 612 fs/reiserfs/ibalance.c if (tb->lnum[h] == 0 && tb->rnum[h] == 0) { h 613 fs/reiserfs/ibalance.c if (tb->blknum[h] == 0) { h 625 fs/reiserfs/ibalance.c if (!tb->L[h - 1] || !B_NR_ITEMS(tb->L[h - 1])) h 626 fs/reiserfs/ibalance.c new_root = tb->R[h - 1]; h 628 fs/reiserfs/ibalance.c new_root = tb->L[h - 1]; h 639 fs/reiserfs/ibalance.c if (h > 1) h 651 fs/reiserfs/ibalance.c if (tb->L[h] && tb->lnum[h] == -B_NR_ITEMS(tb->L[h]) - 1) { /* join S[h] with L[h] */ h 653 fs/reiserfs/ibalance.c RFALSE(tb->rnum[h] != 0, h 655 fs/reiserfs/ibalance.c h, tb->rnum[h]); h 657 fs/reiserfs/ibalance.c internal_shift_left(INTERNAL_SHIFT_FROM_S_TO_L, tb, h, n + 1); h 663 fs/reiserfs/ibalance.c if (tb->R[h] && tb->rnum[h] == -B_NR_ITEMS(tb->R[h]) - 1) { /* join S[h] with R[h] */ h 664 fs/reiserfs/ibalance.c RFALSE(tb->lnum[h] != 0, h 666 fs/reiserfs/ibalance.c h, tb->lnum[h]); h 668 fs/reiserfs/ibalance.c internal_shift_right(INTERNAL_SHIFT_FROM_S_TO_R, tb, h, n + 1); h 674 fs/reiserfs/ibalance.c if (tb->lnum[h] < 0) { /* borrow from left neighbor L[h] */ h 675 fs/reiserfs/ibalance.c RFALSE(tb->rnum[h] != 0, h 676 fs/reiserfs/ibalance.c "wrong tb->rnum[%d]==%d when borrow from L[h]", h, h 677 fs/reiserfs/ibalance.c tb->rnum[h]); h 679 fs/reiserfs/ibalance.c internal_shift_right(INTERNAL_SHIFT_FROM_L_TO_S, tb, h, h 680 fs/reiserfs/ibalance.c -tb->lnum[h]); h 684 fs/reiserfs/ibalance.c if (tb->rnum[h] < 0) { /* borrow from right neighbor R[h] */ h 685 fs/reiserfs/ibalance.c RFALSE(tb->lnum[h] != 0, h 687 fs/reiserfs/ibalance.c h, tb->lnum[h]); h 688 fs/reiserfs/ibalance.c internal_shift_left(INTERNAL_SHIFT_FROM_R_TO_S, tb, h, -tb->rnum[h]); /*tb->S[h], tb->CFR[h], tb->rkey[h], tb->R[h], -tb->rnum[h]); */ h 692 fs/reiserfs/ibalance.c if (tb->lnum[h] > 0) { /* split S[h] into two parts and put them into neighbors */ h 693 fs/reiserfs/ibalance.c RFALSE(tb->rnum[h] == 0 || tb->lnum[h] + tb->rnum[h] != n + 1, h 695 fs/reiserfs/ibalance.c h, tb->lnum[h], h, tb->rnum[h], n); h 697 fs/reiserfs/ibalance.c internal_shift_left(INTERNAL_SHIFT_FROM_S_TO_L, tb, h, tb->lnum[h]); /*tb->L[h], tb->CFL[h], tb->lkey[h], tb->S[h], tb->lnum[h]); */ h 698 fs/reiserfs/ibalance.c internal_shift_right(INTERNAL_SHIFT_FROM_S_TO_R, tb, h, h 699 fs/reiserfs/ibalance.c tb->rnum[h]); h 707 fs/reiserfs/ibalance.c h, tb->lnum[h], h, tb->rnum[h]); h 713 fs/reiserfs/ibalance.c RFALSE(tb->L[h] == NULL || tb->CFL[h] == NULL, h 715 fs/reiserfs/ibalance.c tb->L[h], tb->CFL[h]); h 717 fs/reiserfs/ibalance.c if (B_NR_ITEMS(PATH_H_PBUFFER(tb->tb_path, h)) == 0) h 720 fs/reiserfs/ibalance.c memcpy(B_N_PDELIM_KEY(tb->CFL[h], tb->lkey[h]), key, KEY_SIZE); h 722 fs/reiserfs/ibalance.c do_balance_mark_internal_dirty(tb, tb->CFL[h], 0); h 728 fs/reiserfs/ibalance.c RFALSE(tb->R[h] == NULL || tb->CFR[h] == NULL, h 730 fs/reiserfs/ibalance.c tb->R[h], tb->CFR[h]); h 731 fs/reiserfs/ibalance.c RFALSE(B_NR_ITEMS(tb->R[h]) == 0, h 733 fs/reiserfs/ibalance.c B_NR_ITEMS(tb->R[h])); h 735 fs/reiserfs/ibalance.c memcpy(B_N_PDELIM_KEY(tb->CFR[h], tb->rkey[h]), key, KEY_SIZE); h 737 fs/reiserfs/ibalance.c do_balance_mark_internal_dirty(tb, tb->CFR[h], 0); h 760 fs/reiserfs/ibalance.c struct buffer_head *tbSh = PATH_H_PBUFFER(tb->tb_path, h); h 769 fs/reiserfs/ibalance.c RFALSE(h < 1, "h (%d) can not be < 1 on internal level", h); h 771 fs/reiserfs/ibalance.c PROC_INFO_INC(tb->tb_sb, balance_at[h]); h 775 fs/reiserfs/ibalance.c h + 1) /*tb->S[h]->b_item_order */ : 0; h 779 fs/reiserfs/ibalance.c insert_num = tb->insert_size[h] / ((int)(KEY_SIZE + DC_SIZE)); h 785 fs/reiserfs/ibalance.c RFALSE(h > 1 && (insert_num > 1 || insert_num < -1), h 787 fs/reiserfs/ibalance.c insert_num, h); h 791 fs/reiserfs/ibalance.c balance_internal_when_delete(tb, h, child_pos); h 796 fs/reiserfs/ibalance.c if (tb->lnum[h] > 0) { h 800 fs/reiserfs/ibalance.c n = B_NR_ITEMS(tb->L[h]); /* number of items in L[h] */ h 801 fs/reiserfs/ibalance.c if (tb->lnum[h] <= child_pos) { h 803 fs/reiserfs/ibalance.c internal_shift_left(INTERNAL_SHIFT_FROM_S_TO_L, tb, h, h 804 fs/reiserfs/ibalance.c tb->lnum[h]); h 806 fs/reiserfs/ibalance.c child_pos -= tb->lnum[h]; h 807 fs/reiserfs/ibalance.c } else if (tb->lnum[h] > child_pos + insert_num) { h 809 fs/reiserfs/ibalance.c internal_shift_left(INTERNAL_SHIFT_FROM_S_TO_L, tb, h, h 810 fs/reiserfs/ibalance.c tb->lnum[h] - insert_num); h 816 fs/reiserfs/ibalance.c bi.bi_bh = tb->L[h]; h 817 fs/reiserfs/ibalance.c bi.bi_parent = tb->FL[h]; h 818 fs/reiserfs/ibalance.c bi.bi_position = get_left_neighbor_position(tb, h); h 830 fs/reiserfs/ibalance.c internal_shift1_left(tb, h, child_pos + 1); h 832 fs/reiserfs/ibalance.c k = tb->lnum[h] - child_pos - 1; h 834 fs/reiserfs/ibalance.c bi.bi_bh = tb->L[h]; h 835 fs/reiserfs/ibalance.c bi.bi_parent = tb->FL[h]; h 836 fs/reiserfs/ibalance.c bi.bi_position = get_left_neighbor_position(tb, h); h 842 fs/reiserfs/ibalance.c replace_lkey(tb, h, insert_key + k); h 861 fs/reiserfs/ibalance.c if (tb->rnum[h] > 0) { h 865 fs/reiserfs/ibalance.c if (n - tb->rnum[h] >= child_pos) h 868 fs/reiserfs/ibalance.c internal_shift_right(INTERNAL_SHIFT_FROM_S_TO_R, tb, h, h 869 fs/reiserfs/ibalance.c tb->rnum[h]); h 870 fs/reiserfs/ibalance.c else if (n + insert_num - tb->rnum[h] < child_pos) { h 874 fs/reiserfs/ibalance.c internal_shift_right(INTERNAL_SHIFT_FROM_S_TO_R, tb, h, h 875 fs/reiserfs/ibalance.c tb->rnum[h] - insert_num); h 879 fs/reiserfs/ibalance.c bi.bi_bh = tb->R[h]; h 880 fs/reiserfs/ibalance.c bi.bi_parent = tb->FR[h]; h 881 fs/reiserfs/ibalance.c bi.bi_position = get_right_neighbor_position(tb, h); h 885 fs/reiserfs/ibalance.c tb->rnum[h] - 1, h 893 fs/reiserfs/ibalance.c internal_shift1_right(tb, h, n - child_pos + 1); h 895 fs/reiserfs/ibalance.c k = tb->rnum[h] - n + child_pos - 1; h 897 fs/reiserfs/ibalance.c bi.bi_bh = tb->R[h]; h 898 fs/reiserfs/ibalance.c bi.bi_parent = tb->FR[h]; h 899 fs/reiserfs/ibalance.c bi.bi_position = get_right_neighbor_position(tb, h); h 905 fs/reiserfs/ibalance.c replace_rkey(tb, h, insert_key + insert_num - k - 1); h 908 fs/reiserfs/ibalance.c dc = B_N_CHILD(tb->R[h], 0); h 918 fs/reiserfs/ibalance.c do_balance_mark_internal_dirty(tb, tb->R[h], 0); h 925 fs/reiserfs/ibalance.c RFALSE(tb->blknum[h] > 2, "blknum can not be > 2 for internal level"); h 926 fs/reiserfs/ibalance.c RFALSE(tb->blknum[h] < 0, "blknum can not be < 0"); h 928 fs/reiserfs/ibalance.c if (!tb->blknum[h]) { /* node S[h] is empty now */ h 939 fs/reiserfs/ibalance.c struct buffer_head *tbSh_1 = PATH_H_PBUFFER(tb->tb_path, h - 1); h 942 fs/reiserfs/ibalance.c if (tb->blknum[h] != 1) h 948 fs/reiserfs/ibalance.c set_blkh_level(blkh, h + 1); h 957 fs/reiserfs/ibalance.c tb->insert_size[h] -= DC_SIZE; h 976 fs/reiserfs/ibalance.c if (tb->blknum[h] == 2) { h 983 fs/reiserfs/ibalance.c set_blkh_level(B_BLK_HEAD(S_new), h + 1); h 991 fs/reiserfs/ibalance.c src_bi.bi_parent = PATH_H_PPARENT(tb->tb_path, h); h 992 fs/reiserfs/ibalance.c src_bi.bi_position = PATH_H_POSITION(tb->tb_path, h + 1); h 1077 fs/reiserfs/ibalance.c bi.bi_parent = PATH_H_PPARENT(tb->tb_path, h); h 1078 fs/reiserfs/ibalance.c bi.bi_position = PATH_H_POSITION(tb->tb_path, h + 1); h 57 fs/reiserfs/journal.c #define JOURNAL_LIST_ENTRY(h) (list_entry((h), struct reiserfs_journal_list, \ h 59 fs/reiserfs/journal.c #define JOURNAL_WORK_ENTRY(h) (list_entry((h), struct reiserfs_journal_list, \ h 587 fs/reiserfs/prints.c int h = 0; h 603 fs/reiserfs/prints.c for (h = 0; h < ARRAY_SIZE(tb->insert_size); h++) { h 604 fs/reiserfs/prints.c if (PATH_H_PATH_OFFSET(tb->tb_path, h) <= h 607 fs/reiserfs/prints.c h) > ILLEGAL_PATH_ELEMENT_OFFSET) { h 608 fs/reiserfs/prints.c tbSh = PATH_H_PBUFFER(tb->tb_path, h); h 609 fs/reiserfs/prints.c tbFh = PATH_H_PPARENT(tb->tb_path, h); h 616 fs/reiserfs/prints.c h, h 619 fs/reiserfs/prints.c (tb->L[h]) ? (long long)(tb->L[h]->b_blocknr) : (-1LL), h 620 fs/reiserfs/prints.c (tb->L[h]) ? atomic_read(&(tb->L[h]->b_count)) : -1, h 621 fs/reiserfs/prints.c (tb->R[h]) ? (long long)(tb->R[h]->b_blocknr) : (-1LL), h 622 fs/reiserfs/prints.c (tb->R[h]) ? atomic_read(&(tb->R[h]->b_count)) : -1, h 624 fs/reiserfs/prints.c (tb->FL[h]) ? (long long)(tb->FL[h]-> h 626 fs/reiserfs/prints.c (tb->FR[h]) ? (long long)(tb->FR[h]-> h 628 fs/reiserfs/prints.c (tb->CFL[h]) ? (long long)(tb->CFL[h]-> h 630 fs/reiserfs/prints.c (tb->CFR[h]) ? (long long)(tb->CFR[h]-> h 644 fs/reiserfs/prints.c h = 0; h 646 fs/reiserfs/prints.c h++; h 649 fs/reiserfs/prints.c h, tb->insert_size[h], tb->lnum[h], tb->rnum[h], h 650 fs/reiserfs/prints.c tb->blknum[h]); h 651 fs/reiserfs/prints.c } while (tb->insert_size[h]); h 658 fs/reiserfs/prints.c h = 0; h 122 fs/ubifs/log.c long long h, t; h 124 fs/ubifs/log.c h = (long long)c->lhead_lnum * c->leb_size + c->lhead_offs; h 127 fs/ubifs/log.c if (h >= t) h 128 fs/ubifs/log.c return c->log_bytes - h + t; h 130 fs/ubifs/log.c return t - h; h 1413 fs/ubifs/lpt.c int err, i, h, iip, shft; h 1425 fs/ubifs/lpt.c for (h = 1; h < c->lpt_hght; h++) { h 1547 fs/ubifs/lpt.c int err, i, h, iip, shft; h 1562 fs/ubifs/lpt.c for (h = 1; h < c->lpt_hght; h++) { h 1884 fs/ubifs/lpt.c int err = 0, i, h, iip, shft; h 1916 fs/ubifs/lpt.c for (h = 1; h < c->lpt_hght; h++) { h 1919 fs/ubifs/lpt.c nnode = scan_get_nnode(c, path + h, nnode, iip); h 1927 fs/ubifs/lpt.c pnode = scan_get_pnode(c, path + h, nnode, iip); h 1939 fs/ubifs/lpt.c ret = scan_cb(c, lprops, path[h].in_tree, data); h 1946 fs/ubifs/lpt.c for (h = 1; h < c->lpt_hght; h++) { h 1950 fs/ubifs/lpt.c if (path[h].in_tree) h 1957 fs/ubifs/lpt.c memcpy(nnode, &path[h].nnode, sz); h 1960 fs/ubifs/lpt.c path[h].ptr.nnode = nnode; h 1961 fs/ubifs/lpt.c path[h].in_tree = 1; h 1962 fs/ubifs/lpt.c path[h + 1].cnode.parent = nnode; h 1964 fs/ubifs/lpt.c if (path[h].in_tree) h 1975 fs/ubifs/lpt.c memcpy(pnode, &path[h].pnode, sz); h 1978 fs/ubifs/lpt.c path[h].ptr.pnode = pnode; h 1979 fs/ubifs/lpt.c path[h].in_tree = 1; h 2017 fs/ubifs/lpt.c h -= 1; h 2018 fs/ubifs/lpt.c ubifs_assert(h >= 0); h 2019 fs/ubifs/lpt.c nnode = path[h].ptr.nnode; h 2027 fs/ubifs/lpt.c h += 1; h 2028 fs/ubifs/lpt.c for (; h < c->lpt_hght; h++) { h 2029 fs/ubifs/lpt.c nnode = scan_get_nnode(c, path + h, nnode, iip); h 2036 fs/ubifs/lpt.c pnode = scan_get_pnode(c, path + h, nnode, iip); h 567 fs/ubifs/lpt_commit.c int err, h, iip, shft; h 578 fs/ubifs/lpt_commit.c for (h = 1; h < c->lpt_hght; h++) { h 1306 fs/ubifs/lpt_commit.c int h, i, found; h 1312 fs/ubifs/lpt_commit.c for (h = 1; h < c->lpt_hght; h++) { h 1318 fs/ubifs/lpt_commit.c *hght = h; h 1341 fs/ubifs/lpt_commit.c int iip, h, i, found; h 1359 fs/ubifs/lpt_commit.c for (h = *hght + 1; h < c->lpt_hght; h++) { h 1365 fs/ubifs/lpt_commit.c *hght = h; h 913 fs/xfs/quota/xfs_dquot.c xfs_dqhash_t *h; h 923 fs/xfs/quota/xfs_dquot.c h = XFS_DQ_HASH(mp, id, type); h 949 fs/xfs/quota/xfs_dquot.c XFS_DQ_HASH_LOCK(h); h 955 fs/xfs/quota/xfs_dquot.c if (xfs_qm_dqlookup(mp, id, h, O_dqpp) == 0) { h 964 fs/xfs/quota/xfs_dquot.c XFS_DQ_HASH_UNLOCK(h); h 983 fs/xfs/quota/xfs_dquot.c version = h->qh_version; h 984 fs/xfs/quota/xfs_dquot.c XFS_DQ_HASH_UNLOCK(h); h 1049 fs/xfs/quota/xfs_dquot.c XFS_DQ_HASH_LOCK(h); h 1050 fs/xfs/quota/xfs_dquot.c if (version != h->qh_version) { h 1059 fs/xfs/quota/xfs_dquot.c if (xfs_qm_dqlookup(mp, id, h, &tmpdqp) == 0) { h 1065 fs/xfs/quota/xfs_dquot.c XFS_DQ_HASH_UNLOCK(h); h 1076 fs/xfs/quota/xfs_dquot.c ASSERT(XFS_DQ_IS_HASH_LOCKED(h)); h 1077 fs/xfs/quota/xfs_dquot.c dqp->q_hash = h; h 1078 fs/xfs/quota/xfs_dquot.c XQM_HASHLIST_INSERT(h, dqp); h 1095 fs/xfs/quota/xfs_dquot.c XFS_DQ_HASH_UNLOCK(h); h 1170 fs/xfs/quota/xfs_qm_syscalls.c if (((d) = (h)->qh_next)) h 1173 fs/xfs/quota/xfs_qm_syscalls.c (dqp)->HL_PREVP = &((h)->qh_next); h 1174 fs/xfs/quota/xfs_qm_syscalls.c (h)->qh_next = (xfs_dquot_t *)dqp; h 1175 fs/xfs/quota/xfs_qm_syscalls.c (h)->qh_version++; h 1176 fs/xfs/quota/xfs_qm_syscalls.c (h)->qh_nelems++; h 1285 fs/xfs/quota/xfs_qm_syscalls.c xfs_dqhash_t *h; h 1287 fs/xfs/quota/xfs_qm_syscalls.c h = DQTEST_HASH(mp, id, type); h 1288 fs/xfs/quota/xfs_qm_syscalls.c for (d = (xfs_dqtest_t *) h->qh_next; d != NULL; h 1300 fs/xfs/quota/xfs_qm_syscalls.c d->q_hash = h; h 1301 fs/xfs/quota/xfs_qm_syscalls.c xfs_qm_hashinsert(h, d); h 49 fs/xfs/quota/xfs_quota_priv.h #define XQMLCK(h) (mutex_lock(&((h)->qh_lock))) h 50 fs/xfs/quota/xfs_quota_priv.h #define XQMUNLCK(h) (mutex_unlock(&((h)->qh_lock))) h 55 fs/xfs/quota/xfs_quota_priv.h if (mutex_trylock(&h->qh_lock)) { h 56 fs/xfs/quota/xfs_quota_priv.h mutex_unlock(&h->qh_lock); h 63 fs/xfs/quota/xfs_quota_priv.h #define XFS_DQ_HASH_LOCK(h) XQMLCK(h) h 64 fs/xfs/quota/xfs_quota_priv.h #define XFS_DQ_HASH_UNLOCK(h) XQMUNLCK(h) h 65 fs/xfs/quota/xfs_quota_priv.h #define XFS_DQ_IS_HASH_LOCKED(h) XQMISLCKD(h) h 113 fs/xfs/quota/xfs_quota_priv.h (h)->qh_version++; \ h 114 fs/xfs/quota/xfs_quota_priv.h (h)->qh_nelems--; \ h 120 fs/xfs/quota/xfs_quota_priv.h if (((d) = (h)->qh_next)) \ h 123 fs/xfs/quota/xfs_quota_priv.h (dqp)->PVP = &((h)->qh_next); \ h 124 fs/xfs/quota/xfs_quota_priv.h (h)->qh_next = dqp; \ h 125 fs/xfs/quota/xfs_quota_priv.h (h)->qh_version++; \ h 126 fs/xfs/quota/xfs_quota_priv.h (h)->qh_nelems++; \ h 137 fs/xfs/quota/xfs_quota_priv.h _LIST_INSERT(h, dqp, HL_PREVP, HL_NEXT) h 140 fs/xfs/quota/xfs_quota_priv.h xfs_qm_freelist_append(h, dqp) h 143 fs/xfs/quota/xfs_quota_priv.h _LIST_INSERT(h, dqp, MPL_PREVP, MPL_NEXT) h 146 fs/xfs/quota/xfs_quota_priv.h _LIST_REMOVE(h, dqp, HL_PREVP, HL_NEXT) h 150 fs/xfs/quota/xfs_quota_priv.h { _LIST_REMOVE(h, dqp, MPL_PREVP, MPL_NEXT); \ h 317 include/asm-cris/arch-v32/hwregs/extmem_defs.h unsigned int h : 32; h 13 include/asm-cris/dma-mapping.h #define dma_alloc_noncoherent(d, s, h, f) dma_alloc_coherent(d, s, h, f) h 14 include/asm-cris/dma-mapping.h #define dma_free_noncoherent(d, s, v, h) dma_free_coherent(d, s, v, h) h 283 include/asm-frv/bitops.h struct { u32 h, l; }; h 304 include/asm-frv/bitops.h : "0r"(_.h), "r"(_.l) h 365 include/asm-frv/bitops.h struct { u32 h, l; }; h 380 include/asm-frv/bitops.h : "0r"(_.h), "r"(_.l) h 10 include/asm-frv/dma-mapping.h #define dma_alloc_noncoherent(d, s, h, f) dma_alloc_coherent(d, s, h, f) h 11 include/asm-frv/dma-mapping.h #define dma_free_noncoherent(d, s, v, h) dma_free_coherent(d, s, v, h) h 20 include/asm-generic/bitops/fls64.h __u32 h = x >> 32; h 21 include/asm-generic/bitops/fls64.h if (h) h 22 include/asm-generic/bitops/fls64.h return fls(h) + 32; h 19 include/asm-generic/dma-mapping-broken.h #define dma_alloc_noncoherent(d, s, h, f) dma_alloc_coherent(d, s, h, f) h 20 include/asm-generic/dma-mapping-broken.h #define dma_free_noncoherent(d, s, v, h) dma_free_coherent(d, s, v, h) h 267 include/asm-generic/dma-mapping.h #define dma_alloc_noncoherent(d, s, h, f) dma_alloc_coherent(d, s, h, f) h 268 include/asm-generic/dma-mapping.h #define dma_free_noncoherent(d, s, v, h) dma_free_coherent(d, s, v, h) h 195 include/asm-generic/rtc.h struct rtc_time h; h 197 include/asm-generic/rtc.h get_rtc_time(&h); h 198 include/asm-generic/rtc.h return h.tm_sec; h 53 include/asm-m68k/rtc.h struct rtc_time h; h 55 include/asm-m68k/rtc.h get_rtc_time(&h); h 56 include/asm-m68k/rtc.h return h.tm_sec; h 26 include/asm-mn10300/dma-mapping.h #define dma_alloc_noncoherent(d, s, h, f) dma_alloc_coherent((d), (s), (h), (f)) h 27 include/asm-mn10300/dma-mapping.h #define dma_free_noncoherent(d, s, v, h) dma_free_coherent((d), (s), (v), (h)) h 77 include/asm-parisc/pci.h #define HBA_PORT_BASE(h) ((h) << HBA_PORT_SPACE_BITS) h 115 include/asm-parisc/rtc.h struct rtc_time h; h 117 include/asm-parisc/rtc.h get_rtc_time(&h); h 118 include/asm-parisc/rtc.h return h.tm_sec; h 16 include/asm-parisc/system.h unsigned int h:1; h 95 include/asm-um/dma-mapping.h #define dma_alloc_noncoherent(d, s, h, f) dma_alloc_coherent(d, s, h, f) h 96 include/asm-um/dma-mapping.h #define dma_free_noncoherent(d, s, v, h) dma_free_coherent(d, s, v, h) h 85 include/asm-x86/dma-mapping.h #define dma_alloc_noncoherent(d, s, h, f) dma_alloc_coherent(d, s, h, f) h 86 include/asm-x86/dma-mapping.h #define dma_free_noncoherent(d, s, v, h) dma_free_coherent(d, s, v, h) h 20 include/asm-x86/hugetlb.h struct hstate *h = hstate_file(file); h 21 include/asm-x86/hugetlb.h if (len & ~huge_page_mask(h)) h 23 include/asm-x86/hugetlb.h if (addr & ~huge_page_mask(h)) h 225 include/asm-x86/msr.h rdmsr(msr_no, *l, *h); h 230 include/asm-x86/msr.h wrmsr(msr_no, l, h); h 236 include/asm-x86/msr.h return rdmsr_safe(msr_no, l, h); h 240 include/asm-x86/msr.h return wrmsr_safe(msr_no, l, h); h 27 include/asm-xtensa/dma-mapping.h #define dma_alloc_noncoherent(d, s, h, f) dma_alloc_coherent(d, s, h, f) h 28 include/asm-xtensa/dma-mapping.h #define dma_free_noncoherent(d, s, v, h) dma_free_coherent(d, s, v, h) h 117 include/drm/i810_drm.h unsigned int h; h 135 include/drm/i810_drm.h unsigned int h; h 167 include/drm/i830_drm.h unsigned int h; h 57 include/drm/i915_drm.h unsigned int h; h 220 include/linux/aio.h return list_entry(h, struct kiocb, ki_list); h 237 include/linux/amba/clcd.h #define CHECK(e,l,h) (var->e < l || var->e > h) h 193 include/linux/byteorder/swab.h __u32 h = x >> 32; h 195 include/linux/byteorder/swab.h return (((__u64)__swab32(l)) << 32) | ((__u64)(__swab32(h))); h 65 include/linux/dvb/video.h int h; h 231 include/linux/hugetlb.h return (unsigned long)PAGE_SIZE << h->order; h 236 include/linux/hugetlb.h return h->mask; h 241 include/linux/hugetlb.h return h->order; h 246 include/linux/hugetlb.h return h->order + PAGE_SHIFT; h 251 include/linux/hugetlb.h return 1 << h->order; h 256 include/linux/hugetlb.h return huge_page_size(h) / 512; h 553 include/linux/list.h h->next = NULL; h 554 include/linux/list.h h->pprev = NULL; h 559 include/linux/list.h return !h->pprev; h 564 include/linux/list.h return !h->first; h 593 include/linux/list.h struct hlist_node *first = h->first; h 597 include/linux/list.h h->first = n; h 598 include/linux/list.h n->pprev = &h->first; h 84 include/linux/nfsd/export.h struct cache_head h; h 103 include/linux/nfsd/export.h struct cache_head h; h 143 include/linux/nfsd/export.h cache_put(&exp->h, &svc_export_cache); h 148 include/linux/nfsd/export.h cache_get(&exp->h); h 176 include/linux/pkt_cls.h #define TC_U32_HTID(h) ((h)&0xFFF00000) h 177 include/linux/pkt_cls.h #define TC_U32_USERHTID(h) (TC_U32_HTID(h)>>20) h 178 include/linux/pkt_cls.h #define TC_U32_HASH(h) (((h)>>12)&0xFF) h 179 include/linux/pkt_cls.h #define TC_U32_NODE(h) ((h)&0xFFF) h 180 include/linux/pkt_cls.h #define TC_U32_KEY(h) ((h)&0xFFFFF) h 68 include/linux/pkt_sched.h #define TC_H_MAJ(h) ((h)&TC_H_MAJ_MASK) h 69 include/linux/pkt_sched.h #define TC_H_MIN(h) ((h)&TC_H_MIN_MASK) h 357 include/linux/quota.h # /* nodep */ include <sys/cdefs.h> h 305 include/linux/rculist.h struct hlist_node *first = h->first; h 308 include/linux/rculist.h n->pprev = &h->first; h 309 include/linux/rculist.h rcu_assign_pointer(h->first, n); h 1212 include/linux/reiserfs_fs.h #define PATH_H_PBUFFER(p_s_path, h) PATH_OFFSET_PBUFFER (p_s_path, p_s_path->path_length - (h)) /* tb->S[h] */ h 1213 include/linux/reiserfs_fs.h #define PATH_H_PPARENT(path, h) PATH_H_PBUFFER (path, (h) + 1) /* tb->F[h] or tb->S[0]->b_parent */ h 1214 include/linux/reiserfs_fs.h #define PATH_H_POSITION(path, h) PATH_OFFSET_POSITION (path, path->path_length - (h)) h 1215 include/linux/reiserfs_fs.h #define PATH_H_B_ITEM_ORDER(path, h) PATH_H_POSITION(path, h + 1) /* tb->S[h]->b_item_order */ h 13 include/linux/romfs_fs.h #define __mkw(h,l) (((h)&0x00ff)<< 8|((l)&0x00ff)) h 14 include/linux/romfs_fs.h #define __mkl(h,l) (((h)&0xffff)<<16|((l)&0xffff)) h 74 include/linux/sunrpc/cache.h struct cache_head *h, h 81 include/linux/sunrpc/cache.h struct cache_head *h); h 142 include/linux/sunrpc/cache.h kref_get(&h->ref); h 143 include/linux/sunrpc/cache.h return h; h 149 include/linux/sunrpc/cache.h if (atomic_read(&h->ref.refcount) <= 2 && h 150 include/linux/sunrpc/cache.h h->expiry_time < cd->nextcheck) h 151 include/linux/sunrpc/cache.h cd->nextcheck = h->expiry_time; h 152 include/linux/sunrpc/cache.h kref_put(&h->ref, cd->cache_put); h 163 include/linux/sunrpc/cache.h return (h->expiry_time != 0 && test_bit(CACHE_VALID, &h->flags)); h 75 include/linux/swab.h __u32 h = val >> 32; h 77 include/linux/swab.h return (((__u64)___swab32(l)) << 32) | ((__u64)(___swab32(h))); h 903 include/net/bluetooth/hci.h #define hci_handle_pack(h, f) (__u16) ((h & 0x0fff)|(f << 12)) h 904 include/net/bluetooth/hci.h #define hci_handle(h) (h & 0x0fff) h 905 include/net/bluetooth/hci.h #define hci_flags(h) (h >> 12) h 246 include/net/bluetooth/hci_core.h struct hci_conn_hash *h = &hdev->conn_hash; h 247 include/net/bluetooth/hci_core.h INIT_LIST_HEAD(&h->list); h 248 include/net/bluetooth/hci_core.h spin_lock_init(&h->lock); h 249 include/net/bluetooth/hci_core.h h->acl_num = 0; h 250 include/net/bluetooth/hci_core.h h->sco_num = 0; h 255 include/net/bluetooth/hci_core.h struct hci_conn_hash *h = &hdev->conn_hash; h 256 include/net/bluetooth/hci_core.h list_add(&c->list, &h->list); h 258 include/net/bluetooth/hci_core.h h->acl_num++; h 260 include/net/bluetooth/hci_core.h h->sco_num++; h 265 include/net/bluetooth/hci_core.h struct hci_conn_hash *h = &hdev->conn_hash; h 268 include/net/bluetooth/hci_core.h h->acl_num--; h 270 include/net/bluetooth/hci_core.h h->sco_num--; h 276 include/net/bluetooth/hci_core.h struct hci_conn_hash *h = &hdev->conn_hash; h 280 include/net/bluetooth/hci_core.h list_for_each(p, &h->list) { h 291 include/net/bluetooth/hci_core.h struct hci_conn_hash *h = &hdev->conn_hash; h 295 include/net/bluetooth/hci_core.h list_for_each(p, &h->list) { h 306 include/net/bluetooth/hci_core.h struct hci_conn_hash *h = &hdev->conn_hash; h 310 include/net/bluetooth/hci_core.h list_for_each(p, &h->list) { h 146 include/net/netfilter/nf_conntrack_tuple.h ((enum ip_conntrack_dir)(h)->tuple.dst.dir) h 44 include/net/raw.h struct raw_hashinfo *h; h 276 include/net/sch_generic.h unsigned int h; h 278 include/net/sch_generic.h h = qdisc_class_hash(id, hash->hashmask); h 279 include/net/sch_generic.h hlist_for_each_entry(cl, n, &hash->hash[h], hnode) { h 636 include/net/sctp/sctp.h int h = (lport << 16) + rport; h 637 include/net/sctp/sctp.h h ^= h>>8; h 638 include/net/sctp/sctp.h return (h & (sctp_assoc_hashsize - 1)); h 647 include/net/sctp/sctp.h int h = (lport << 16) + rport; h 648 include/net/sctp/sctp.h h ^= vtag; h 649 include/net/sctp/sctp.h return (h & (sctp_assoc_hashsize-1)); h 602 include/net/sock.h } h; h 98 include/scsi/scsi_transport_spi.h #define spi_signalling(h) (((struct spi_host_attrs *)(h)->shost_data)->signalling) h 228 include/video/pm3fb.h #define PM3VideoOverlayHeight_HEIGHT(h) (((h) & 0xfff) << 0) h 655 include/video/pm3fb.h #define PM3FBWriteMode_StripeHeight(h) (((h) & 0x7) << 9) h 685 include/video/pm3fb.h #define PM3LBDestReadMode_StripeHeight(h) (((h) & 0x7) << 5) h 706 include/video/pm3fb.h #define PM3LBSourceReadMode_StripeHeight(h) (((h) & 0x7) << 5) h 726 include/video/pm3fb.h #define PM3LBWriteMode_StripeHeight(h) (((h) & 0x7) << 6) h 974 include/video/pm3fb.h #define PM3Render2D_Height(h) (((h) & 0x0fff) << 16) h 981 include/video/pm3fb.h #define PM3Render2DGlyph_Height(h) (((h) & 0x7f) << 7) h 123 include/xen/interface/vcpu.h GUEST_HANDLE(vcpu_runstate_info) h; h 242 ipc/msg.c tmp = h->next; h 243 ipc/msg.c while (tmp != h) { h 565 ipc/shm.c struct hstate *h = hstate_file(shp->shm_file); h 566 ipc/shm.c *rss += pages_per_huge_page(h) * mapping->nrpages; h 1013 kernel/auditfilter.c int h = audit_hash_ino((u32)ino); h 1015 kernel/auditfilter.c list_add_rcu(&nentry->list, &audit_inode_hash[h]); h 1108 kernel/auditfilter.c int h; h 1112 kernel/auditfilter.c for (h = 0; h < AUDIT_INODE_BUCKETS; h++) { h 1113 kernel/auditfilter.c list = &audit_inode_hash[h]; h 1270 kernel/auditfilter.c int h, err; h 1281 kernel/auditfilter.c h = audit_hash_ino(inode_f->val); h 1282 kernel/auditfilter.c list = &audit_inode_hash[h]; h 1311 kernel/auditfilter.c h = audit_hash_ino((u32)watch->ino); h 1312 kernel/auditfilter.c list = &audit_inode_hash[h]; h 1356 kernel/auditfilter.c int h, ret = 0; h 1367 kernel/auditfilter.c h = audit_hash_ino(inode_f->val); h 1368 kernel/auditfilter.c list = &audit_inode_hash[h]; h 703 kernel/auditsc.c int h = audit_hash_ino((u32)n->ino); h 704 kernel/auditsc.c struct list_head *list = &audit_inode_hash[h]; h 186 kernel/softirq.c struct softirq_action *h; h 204 kernel/softirq.c h = softirq_vec; h 208 kernel/softirq.c h->action(h); h 211 kernel/softirq.c h++; h 77 kernel/user.c struct hlist_node *h; h 79 kernel/user.c hlist_for_each_entry(user, h, hashent, uidhash_node) { h 336 lib/inflate.c int h; /* table level */ h 442 lib/inflate.c h = -1; /* no tables yet--level -1 */ h 462 lib/inflate.c h++; h 487 lib/inflate.c if (h) h 496 lib/inflate.c u[h] = ++q; /* table starts after link */ h 500 lib/inflate.c if (h) h 502 lib/inflate.c x[h] = i; /* save pattern for backing up */ h 507 lib/inflate.c u[h-1][j] = r; /* connect to last table */ h 541 lib/inflate.c while ((i & ((1 << w) - 1)) != x[h]) h 543 lib/inflate.c h--; /* don't need to update q */ h 1086 lib/inflate.c unsigned h; /* maximum struct huft's malloc'ed */ h 1095 lib/inflate.c h = 0; h 1104 lib/inflate.c if (hufts > h) h 1105 lib/inflate.c h = hufts; h 1122 lib/inflate.c fprintf(stderr, "<%u> ", h); h 142 lib/zlib_deflate/deflate.c #define UPDATE_HASH(s,h,c) (h = (((h)<<s->hash_shift) ^ (c)) & s->hash_mask) h 418 lib/zlib_deflate/deftree.c int h; /* heap index */ h 432 lib/zlib_deflate/deftree.c for (h = s->heap_max+1; h < HEAP_SIZE; h++) { h 433 lib/zlib_deflate/deftree.c n = s->heap[h]; h 474 lib/zlib_deflate/deftree.c m = s->heap[--h]; h 44 mm/hugetlb.c for ((h) = hstates; (h) < &hstates[max_hstate]; (h)++) h 217 mm/hugetlb.c return ((address - vma->vm_start) >> huge_page_shift(h)) + h 218 mm/hugetlb.c (vma->vm_pgoff >> huge_page_order(h)); h 328 mm/hugetlb.c h->resv_huge_pages--; h 334 mm/hugetlb.c h->resv_huge_pages--; h 372 mm/hugetlb.c struct hstate *h = hstate_vma(vma); h 375 mm/hugetlb.c for (i = 0; i < pages_per_huge_page(h); i++) { h 384 mm/hugetlb.c list_add(&page->lru, &h->hugepage_freelists[nid]); h 385 mm/hugetlb.c h->free_huge_pages++; h 386 mm/hugetlb.c h->free_huge_pages_node[nid]++; h 395 mm/hugetlb.c if (!list_empty(&h->hugepage_freelists[nid])) { h 396 mm/hugetlb.c page = list_entry(h->hugepage_freelists[nid].next, h 399 mm/hugetlb.c h->free_huge_pages--; h 400 mm/hugetlb.c h->free_huge_pages_node[nid]--; h 426 mm/hugetlb.c h->free_huge_pages - h->resv_huge_pages == 0) h 430 mm/hugetlb.c if (avoid_reserve && h->free_huge_pages - h->resv_huge_pages == 0) h 437 mm/hugetlb.c !list_empty(&h->hugepage_freelists[nid])) { h 438 mm/hugetlb.c page = list_entry(h->hugepage_freelists[nid].next, h 441 mm/hugetlb.c h->free_huge_pages--; h 442 mm/hugetlb.c h->free_huge_pages_node[nid]--; h 445 mm/hugetlb.c decrement_hugepage_resv_vma(h, vma); h 458 mm/hugetlb.c h->nr_huge_pages--; h 459 mm/hugetlb.c h->nr_huge_pages_node[page_to_nid(page)]--; h 460 mm/hugetlb.c for (i = 0; i < pages_per_huge_page(h); i++) { h 468 mm/hugetlb.c __free_pages(page, huge_page_order(h)); h 473 mm/hugetlb.c struct hstate *h; h 475 mm/hugetlb.c for_each_hstate(h) { h 476 mm/hugetlb.c if (huge_page_size(h) == size) h 477 mm/hugetlb.c return h; h 488 mm/hugetlb.c struct hstate *h = page_hstate(page); h 498 mm/hugetlb.c if (h->surplus_huge_pages_node[nid] && huge_page_order(h) < MAX_ORDER) { h 499 mm/hugetlb.c update_and_free_page(h, page); h 500 mm/hugetlb.c h->surplus_huge_pages--; h 501 mm/hugetlb.c h->surplus_huge_pages_node[nid]--; h 503 mm/hugetlb.c enqueue_huge_page(h, page); h 528 mm/hugetlb.c if (delta < 0 && !h->surplus_huge_pages_node[nid]) h 531 mm/hugetlb.c if (delta > 0 && h->surplus_huge_pages_node[nid] >= h 532 mm/hugetlb.c h->nr_huge_pages_node[nid]) h 535 mm/hugetlb.c h->surplus_huge_pages += delta; h 536 mm/hugetlb.c h->surplus_huge_pages_node[nid] += delta; h 549 mm/hugetlb.c h->nr_huge_pages++; h 550 mm/hugetlb.c h->nr_huge_pages_node[nid]++; h 559 mm/hugetlb.c if (h->order >= MAX_ORDER) h 565 mm/hugetlb.c huge_page_order(h)); h 568 mm/hugetlb.c __free_pages(page, huge_page_order(h)); h 571 mm/hugetlb.c prep_new_huge_page(h, page, nid); h 591 mm/hugetlb.c next_nid = next_node(h->hugetlb_next_nid, node_online_map); h 594 mm/hugetlb.c h->hugetlb_next_nid = next_nid; h 605 mm/hugetlb.c start_nid = h->hugetlb_next_nid; h 608 mm/hugetlb.c page = alloc_fresh_huge_page_node(h, h->hugetlb_next_nid); h 611 mm/hugetlb.c next_nid = hstate_next_node(h); h 612 mm/hugetlb.c } while (!page && h->hugetlb_next_nid != start_nid); h 628 mm/hugetlb.c if (h->order >= MAX_ORDER) h 655 mm/hugetlb.c if (h->surplus_huge_pages >= h->nr_overcommit_huge_pages) { h 659 mm/hugetlb.c h->nr_huge_pages++; h 660 mm/hugetlb.c h->surplus_huge_pages++; h 666 mm/hugetlb.c huge_page_order(h)); h 669 mm/hugetlb.c __free_pages(page, huge_page_order(h)); h 686 mm/hugetlb.c h->nr_huge_pages_node[nid]++; h 687 mm/hugetlb.c h->surplus_huge_pages_node[nid]++; h 690 mm/hugetlb.c h->nr_huge_pages--; h 691 mm/hugetlb.c h->surplus_huge_pages--; h 710 mm/hugetlb.c needed = (h->resv_huge_pages + delta) - h->free_huge_pages; h 712 mm/hugetlb.c h->resv_huge_pages += delta; h 723 mm/hugetlb.c page = alloc_buddy_huge_page(h, NULL, 0); h 744 mm/hugetlb.c needed = (h->resv_huge_pages + delta) - h 745 mm/hugetlb.c (h->free_huge_pages + allocated); h 758 mm/hugetlb.c h->resv_huge_pages += delta; h 766 mm/hugetlb.c enqueue_huge_page(h, page); h 810 mm/hugetlb.c h->resv_huge_pages -= unused_resv_pages; h 813 mm/hugetlb.c if (h->order >= MAX_ORDER) h 816 mm/hugetlb.c nr_pages = min(unused_resv_pages, h->surplus_huge_pages); h 823 mm/hugetlb.c if (!h->surplus_huge_pages_node[nid]) h 826 mm/hugetlb.c if (!list_empty(&h->hugepage_freelists[nid])) { h 827 mm/hugetlb.c page = list_entry(h->hugepage_freelists[nid].next, h 830 mm/hugetlb.c update_and_free_page(h, page); h 831 mm/hugetlb.c h->free_huge_pages--; h 832 mm/hugetlb.c h->free_huge_pages_node[nid]--; h 833 mm/hugetlb.c h->surplus_huge_pages--; h 834 mm/hugetlb.c h->surplus_huge_pages_node[nid]--; h 857 mm/hugetlb.c pgoff_t idx = vma_hugecache_offset(h, vma, addr); h 866 mm/hugetlb.c pgoff_t idx = vma_hugecache_offset(h, vma, addr); h 882 mm/hugetlb.c pgoff_t idx = vma_hugecache_offset(h, vma, addr); h 886 mm/hugetlb.c pgoff_t idx = vma_hugecache_offset(h, vma, addr); h 897 mm/hugetlb.c struct hstate *h = hstate_vma(vma); h 910 mm/hugetlb.c chg = vma_needs_reservation(h, vma, addr); h 918 mm/hugetlb.c page = dequeue_huge_page_vma(h, vma, addr, avoid_reserve); h 922 mm/hugetlb.c page = alloc_buddy_huge_page(h, vma, addr); h 932 mm/hugetlb.c vma_commit_reservation(h, vma, addr); h 946 mm/hugetlb.c NODE_DATA(h->hugetlb_next_nid), h 947 mm/hugetlb.c huge_page_size(h), huge_page_size(h), 0); h 959 mm/hugetlb.c hstate_next_node(h); h 965 mm/hugetlb.c BUG_ON((unsigned long)virt_to_phys(m) & (huge_page_size(h) - 1)); h 968 mm/hugetlb.c m->hstate = h; h 979 mm/hugetlb.c struct hstate *h = m->hstate; h 982 mm/hugetlb.c prep_compound_page(page, h->order); h 983 mm/hugetlb.c prep_new_huge_page(h, page, page_to_nid(page)); h 991 mm/hugetlb.c for (i = 0; i < h->max_huge_pages; ++i) { h 992 mm/hugetlb.c if (h->order >= MAX_ORDER) { h 993 mm/hugetlb.c if (!alloc_bootmem_huge_page(h)) h 995 mm/hugetlb.c } else if (!alloc_fresh_huge_page(h)) h 998 mm/hugetlb.c h->max_huge_pages = i; h 1003 mm/hugetlb.c struct hstate *h; h 1005 mm/hugetlb.c for_each_hstate(h) { h 1007 mm/hugetlb.c if (h->order < MAX_ORDER) h 1008 mm/hugetlb.c hugetlb_hstate_alloc_pages(h); h 1025 mm/hugetlb.c struct hstate *h; h 1027 mm/hugetlb.c for_each_hstate(h) { h 1031 mm/hugetlb.c memfmt(buf, huge_page_size(h)), h 1032 mm/hugetlb.c h->free_huge_pages); h 1041 mm/hugetlb.c if (h->order >= MAX_ORDER) h 1046 mm/hugetlb.c struct list_head *freel = &h->hugepage_freelists[i]; h 1048 mm/hugetlb.c if (count >= h->nr_huge_pages) h 1053 mm/hugetlb.c update_and_free_page(h, page); h 1054 mm/hugetlb.c h->free_huge_pages--; h 1055 mm/hugetlb.c h->free_huge_pages_node[page_to_nid(page)]--; h 1065 mm/hugetlb.c #define persistent_huge_pages(h) (h->nr_huge_pages - h->surplus_huge_pages) h 1070 mm/hugetlb.c if (h->order >= MAX_ORDER) h 1071 mm/hugetlb.c return h->max_huge_pages; h 1085 mm/hugetlb.c while (h->surplus_huge_pages && count > persistent_huge_pages(h)) { h 1086 mm/hugetlb.c if (!adjust_pool_surplus(h, -1)) h 1090 mm/hugetlb.c while (count > persistent_huge_pages(h)) { h 1097 mm/hugetlb.c ret = alloc_fresh_huge_page(h); h 1119 mm/hugetlb.c min_count = h->resv_huge_pages + h->nr_huge_pages - h->free_huge_pages; h 1121 mm/hugetlb.c try_to_free_low(h, min_count); h 1122 mm/hugetlb.c while (min_count < persistent_huge_pages(h)) { h 1123 mm/hugetlb.c struct page *page = dequeue_huge_page(h); h 1126 mm/hugetlb.c update_and_free_page(h, page); h 1128 mm/hugetlb.c while (count < persistent_huge_pages(h)) { h 1129 mm/hugetlb.c if (!adjust_pool_surplus(h, 1)) h 1133 mm/hugetlb.c ret = persistent_huge_pages(h); h 1161 mm/hugetlb.c struct hstate *h = kobj_to_hstate(kobj); h 1162 mm/hugetlb.c return sprintf(buf, "%lu\n", h->nr_huge_pages); h 1169 mm/hugetlb.c struct hstate *h = kobj_to_hstate(kobj); h 1175 mm/hugetlb.c h->max_huge_pages = set_max_huge_pages(h, input); h 1184 mm/hugetlb.c struct hstate *h = kobj_to_hstate(kobj); h 1185 mm/hugetlb.c return sprintf(buf, "%lu\n", h->nr_overcommit_huge_pages); h 1192 mm/hugetlb.c struct hstate *h = kobj_to_hstate(kobj); h 1199 mm/hugetlb.c h->nr_overcommit_huge_pages = input; h 1209 mm/hugetlb.c struct hstate *h = kobj_to_hstate(kobj); h 1210 mm/hugetlb.c return sprintf(buf, "%lu\n", h->free_huge_pages); h 1217 mm/hugetlb.c struct hstate *h = kobj_to_hstate(kobj); h 1218 mm/hugetlb.c return sprintf(buf, "%lu\n", h->resv_huge_pages); h 1225 mm/hugetlb.c struct hstate *h = kobj_to_hstate(kobj); h 1226 mm/hugetlb.c return sprintf(buf, "%lu\n", h->surplus_huge_pages); h 1247 mm/hugetlb.c hstate_kobjs[h - hstates] = kobject_create_and_add(h->name, h 1249 mm/hugetlb.c if (!hstate_kobjs[h - hstates]) h 1252 mm/hugetlb.c retval = sysfs_create_group(hstate_kobjs[h - hstates], h 1255 mm/hugetlb.c kobject_put(hstate_kobjs[h - hstates]); h 1262 mm/hugetlb.c struct hstate *h; h 1269 mm/hugetlb.c for_each_hstate(h) { h 1270 mm/hugetlb.c err = hugetlb_sysfs_add_hstate(h); h 1273 mm/hugetlb.c h->name); h 1279 mm/hugetlb.c struct hstate *h; h 1281 mm/hugetlb.c for_each_hstate(h) { h 1282 mm/hugetlb.c kobject_put(hstate_kobjs[h - hstates]); h 1322 mm/hugetlb.c struct hstate *h; h 1331 mm/hugetlb.c h = &hstates[max_hstate++]; h 1332 mm/hugetlb.c h->order = order; h 1333 mm/hugetlb.c h->mask = ~((1ULL << (order + PAGE_SHIFT)) - 1); h 1334 mm/hugetlb.c h->nr_huge_pages = 0; h 1335 mm/hugetlb.c h->free_huge_pages = 0; h 1337 mm/hugetlb.c INIT_LIST_HEAD(&h->hugepage_freelists[i]); h 1338 mm/hugetlb.c h->hugetlb_next_nid = first_node(node_online_map); h 1339 mm/hugetlb.c snprintf(h->name, HSTATE_NAME_LEN, "hugepages-%lukB", h 1340 mm/hugetlb.c huge_page_size(h)/1024); h 1342 mm/hugetlb.c parsed_hstate = h; h 1405 mm/hugetlb.c struct hstate *h = &default_hstate; h 1409 mm/hugetlb.c tmp = h->max_huge_pages; h 1416 mm/hugetlb.c h->max_huge_pages = set_max_huge_pages(h, tmp); h 1437 mm/hugetlb.c struct hstate *h = &default_hstate; h 1441 mm/hugetlb.c tmp = h->nr_overcommit_huge_pages; h 1449 mm/hugetlb.c h->nr_overcommit_huge_pages = tmp; h 1460 mm/hugetlb.c struct hstate *h = &default_hstate; h 1467 mm/hugetlb.c h->nr_huge_pages, h 1468 mm/hugetlb.c h->free_huge_pages, h 1469 mm/hugetlb.c h->resv_huge_pages, h 1470 mm/hugetlb.c h->surplus_huge_pages, h 1471 mm/hugetlb.c 1UL << (huge_page_order(h) + PAGE_SHIFT - 10)); h 1476 mm/hugetlb.c struct hstate *h = &default_hstate; h 1481 mm/hugetlb.c nid, h->nr_huge_pages_node[nid], h 1482 mm/hugetlb.c nid, h->free_huge_pages_node[nid], h 1483 mm/hugetlb.c nid, h->surplus_huge_pages_node[nid]); h 1489 mm/hugetlb.c struct hstate *h = &default_hstate; h 1490 mm/hugetlb.c return h->nr_huge_pages * pages_per_huge_page(h); h 1516 mm/hugetlb.c if (gather_surplus_pages(h, delta) < 0) h 1519 mm/hugetlb.c if (delta > cpuset_mems_nr(h->free_huge_pages_node)) { h 1520 mm/hugetlb.c return_unused_surplus_pages(h, delta); h 1527 mm/hugetlb.c return_unused_surplus_pages(h, (unsigned long) -delta); h 1552 mm/hugetlb.c struct hstate *h = hstate_vma(vma); h 1559 mm/hugetlb.c start = vma_hugecache_offset(h, vma, vma->vm_start); h 1560 mm/hugetlb.c end = vma_hugecache_offset(h, vma, vma->vm_end); h 1568 mm/hugetlb.c hugetlb_acct_memory(h, -reserve); h 1628 mm/hugetlb.c struct hstate *h = hstate_vma(vma); h 1629 mm/hugetlb.c unsigned long sz = huge_page_size(h); h 1673 mm/hugetlb.c struct hstate *h = hstate_vma(vma); h 1674 mm/hugetlb.c unsigned long sz = huge_page_size(h); h 1684 mm/hugetlb.c BUG_ON(start & ~huge_page_mask(h)); h 1685 mm/hugetlb.c BUG_ON(end & ~huge_page_mask(h)); h 1794 mm/hugetlb.c struct hstate *h = hstate_vma(vma); h 1855 mm/hugetlb.c ptep = huge_pte_offset(mm, address & huge_page_mask(h)); h 1877 mm/hugetlb.c idx = vma_hugecache_offset(h, vma, address); h 1885 mm/hugetlb.c struct hstate *h = hstate_vma(vma); h 1906 mm/hugetlb.c idx = vma_hugecache_offset(h, vma, address); h 1915 mm/hugetlb.c size = i_size_read(mapping->host) >> huge_page_shift(h); h 1923 mm/hugetlb.c clear_huge_page(page, address, huge_page_size(h)); h 1939 mm/hugetlb.c inode->i_blocks += blocks_per_huge_page(h); h 1952 mm/hugetlb.c if (vma_needs_reservation(h, vma, address) < 0) { h 1958 mm/hugetlb.c size = i_size_read(mapping->host) >> huge_page_shift(h); h 1996 mm/hugetlb.c struct hstate *h = hstate_vma(vma); h 1998 mm/hugetlb.c ptep = huge_pte_alloc(mm, address, huge_page_size(h)); h 2025 mm/hugetlb.c if (vma_needs_reservation(h, vma, address) < 0) { h 2031 mm/hugetlb.c pagecache_page = hugetlbfs_pagecache_page(h, h 2071 mm/hugetlb.c struct hstate *h = hstate_vma(vma); h 2083 mm/hugetlb.c pte = huge_pte_offset(mm, vaddr & huge_page_mask(h)); h 2101 mm/hugetlb.c pfn_offset = (vaddr & ~huge_page_mask(h)) >> PAGE_SHIFT; h 2117 mm/hugetlb.c pfn_offset < pages_per_huge_page(h)) { h 2139 mm/hugetlb.c struct hstate *h = hstate_vma(vma); h 2146 mm/hugetlb.c for (; address < end; address += huge_page_size(h)) { h 2169 mm/hugetlb.c struct hstate *h = hstate_inode(inode); h 2198 mm/hugetlb.c ret = hugetlb_acct_memory(h, chg); h 2210 mm/hugetlb.c struct hstate *h = hstate_inode(inode); h 2214 mm/hugetlb.c inode->i_blocks -= blocks_per_huge_page(h); h 2218 mm/hugetlb.c hugetlb_acct_memory(h, -(chg - freed)); h 2222 mm/mempolicy.c struct hstate *h = hstate_vma(vma); h 2223 mm/mempolicy.c unsigned long sz = huge_page_size(h); h 2227 mm/mempolicy.c addr & huge_page_mask(h)); h 1179 mm/slub.c page = container_of((struct list_head *)h, struct page, lru); h 2406 mm/slub.c struct page *page, *h; h 2409 mm/slub.c list_for_each_entry_safe(page, h, &n->partial, lru) { h 268 net/bluetooth/bnep/core.c struct bnep_ext_hdr *h; h 272 net/bluetooth/bnep/core.c h = (void *) skb->data; h 273 net/bluetooth/bnep/core.c if (!skb_pull(skb, sizeof(*h))) { h 278 net/bluetooth/bnep/core.c BT_DBG("type 0x%x len %d", h->type, h->len); h 280 net/bluetooth/bnep/core.c switch (h->type & BNEP_TYPE_MASK) { h 290 net/bluetooth/bnep/core.c if (!skb_pull(skb, h->len)) { h 294 net/bluetooth/bnep/core.c } while (!err && (h->type & BNEP_EXT_HEADER)); h 536 net/bluetooth/hci_conn.c struct hci_conn_hash *h = &hdev->conn_hash; h 541 net/bluetooth/hci_conn.c p = h->list.next; h 542 net/bluetooth/hci_conn.c while (p != &h->list) { h 975 net/bluetooth/hci_core.c struct hci_event_hdr *h = data; h 976 net/bluetooth/hci_core.c len = HCI_EVENT_HDR_SIZE + h->plen; h 983 net/bluetooth/hci_core.c struct hci_acl_hdr *h = data; h 984 net/bluetooth/hci_core.c len = HCI_ACL_HDR_SIZE + __le16_to_cpu(h->dlen); h 991 net/bluetooth/hci_core.c struct hci_sco_hdr *h = data; h 992 net/bluetooth/hci_core.c len = HCI_SCO_HDR_SIZE + h->dlen; h 1278 net/bluetooth/hci_core.c struct hci_conn_hash *h = &hdev->conn_hash; h 1285 net/bluetooth/hci_core.c list_for_each(p, &h->list) { h 1316 net/bluetooth/hci_core.c struct hci_conn_hash *h = &hdev->conn_hash; h 1323 net/bluetooth/hci_core.c list_for_each(p, &h->list) { h 89 net/bridge/br_fdb.c struct hlist_node *h; h 90 net/bridge/br_fdb.c hlist_for_each(h, &br->hash[i]) { h 93 net/bridge/br_fdb.c f = hlist_entry(h, struct net_bridge_fdb_entry, hlist); h 129 net/bridge/br_fdb.c struct hlist_node *h, *n; h 131 net/bridge/br_fdb.c hlist_for_each_entry_safe(f, h, n, &br->hash[i], hlist) { h 157 net/bridge/br_fdb.c struct hlist_node *h, *n; h 158 net/bridge/br_fdb.c hlist_for_each_entry_safe(f, h, n, &br->hash[i], hlist) { h 177 net/bridge/br_fdb.c struct hlist_node *h, *g; h 179 net/bridge/br_fdb.c hlist_for_each_safe(h, g, &br->hash[i]) { h 181 net/bridge/br_fdb.c = hlist_entry(h, struct net_bridge_fdb_entry, hlist); h 215 net/bridge/br_fdb.c struct hlist_node *h; h 218 net/bridge/br_fdb.c hlist_for_each_entry_rcu(fdb, h, &br->hash[br_mac_hash(addr)], hlist) { h 266 net/bridge/br_fdb.c struct hlist_node *h; h 273 net/bridge/br_fdb.c hlist_for_each_entry_rcu(f, h, &br->hash[i], hlist) { h 309 net/bridge/br_fdb.c struct hlist_node *h; h 312 net/bridge/br_fdb.c hlist_for_each_entry_rcu(fdb, h, head, hlist) { h 109 net/bridge/netfilter/ebtables.c if (FWINV2(ntohs(h->h_proto) >= 1536, EBT_IPROTO)) h 112 net/bridge/netfilter/ebtables.c FWINV2(e->ethproto != h->h_proto, EBT_IPROTO)) h 129 net/bridge/netfilter/ebtables.c verdict |= (h->h_source[i] ^ e->sourcemac[i]) & h 137 net/bridge/netfilter/ebtables.c verdict |= (h->h_dest[i] ^ e->destmac[i]) & h 309 net/bridge/netfilter/ebtables.c #define find_inlist_lock(h,n,p,e,m) find_inlist_lock_noload((h),(n),(e),(m)) h 591 net/core/neighbour.c u32 h; h 593 net/core/neighbour.c for (h = 0; h <= PNEIGH_HASHMASK; h++) { h 594 net/core/neighbour.c np = &tbl->phash_buckets[h]; h 2084 net/core/neighbour.c int rc, h, s_h = cb->args[1]; h 2088 net/core/neighbour.c for (h = 0; h <= tbl->hash_mask; h++) { h 2089 net/core/neighbour.c if (h < s_h) h 2091 net/core/neighbour.c if (h > s_h) h 2093 net/core/neighbour.c for (n = tbl->hash_buckets[h], idx = 0; n; n = n->next) { h 2113 net/core/neighbour.c cb->args[1] = h; h 171 net/dccp/ccids/lib/packet_history.c struct tfrc_rx_hist_entry *entry = tfrc_rx_hist_last_rcv(h); h 183 net/dccp/ccids/lib/packet_history.c if (dccp_delta_seqno(tfrc_rx_hist_loss_prev(h)->tfrchrx_seqno, seq) <= 0) h 186 net/dccp/ccids/lib/packet_history.c for (i = 1; i <= h->loss_count; i++) h 187 net/dccp/ccids/lib/packet_history.c if (tfrc_rx_hist_entry(h, i)->tfrchrx_seqno == seq) h 196 net/dccp/ccids/lib/packet_history.c const u8 idx_a = tfrc_rx_hist_index(h, a), h 197 net/dccp/ccids/lib/packet_history.c idx_b = tfrc_rx_hist_index(h, b); h 198 net/dccp/ccids/lib/packet_history.c struct tfrc_rx_hist_entry *tmp = h->ring[idx_a]; h 200 net/dccp/ccids/lib/packet_history.c h->ring[idx_a] = h->ring[idx_b]; h 201 net/dccp/ccids/lib/packet_history.c h->ring[idx_b] = tmp; h 215 net/dccp/ccids/lib/packet_history.c u64 s0 = tfrc_rx_hist_loss_prev(h)->tfrchrx_seqno, h 219 net/dccp/ccids/lib/packet_history.c h->loss_count = 1; h 220 net/dccp/ccids/lib/packet_history.c tfrc_rx_hist_entry_from_skb(tfrc_rx_hist_entry(h, 1), skb, n1); h 226 net/dccp/ccids/lib/packet_history.c u64 s0 = tfrc_rx_hist_loss_prev(h)->tfrchrx_seqno, h 227 net/dccp/ccids/lib/packet_history.c s1 = tfrc_rx_hist_entry(h, 1)->tfrchrx_seqno, h 231 net/dccp/ccids/lib/packet_history.c h->loss_count = 2; h 232 net/dccp/ccids/lib/packet_history.c tfrc_rx_hist_entry_from_skb(tfrc_rx_hist_entry(h, 2), skb, n2); h 239 net/dccp/ccids/lib/packet_history.c u64 n1 = tfrc_rx_hist_entry(h, 1)->tfrchrx_ndp; h 243 net/dccp/ccids/lib/packet_history.c h->loss_count = 0; h 244 net/dccp/ccids/lib/packet_history.c h->loss_start = tfrc_rx_hist_index(h, 1); h 247 net/dccp/ccids/lib/packet_history.c tfrc_rx_hist_entry_from_skb(tfrc_rx_hist_loss_prev(h), skb, n2); h 253 net/dccp/ccids/lib/packet_history.c tfrc_rx_hist_swap(h, 0, 3); h 254 net/dccp/ccids/lib/packet_history.c h->loss_start = tfrc_rx_hist_index(h, 3); h 255 net/dccp/ccids/lib/packet_history.c tfrc_rx_hist_entry_from_skb(tfrc_rx_hist_entry(h, 1), skb, n2); h 256 net/dccp/ccids/lib/packet_history.c h->loss_count = 2; h 263 net/dccp/ccids/lib/packet_history.c u64 s0 = tfrc_rx_hist_loss_prev(h)->tfrchrx_seqno, h 264 net/dccp/ccids/lib/packet_history.c s1 = tfrc_rx_hist_entry(h, 1)->tfrchrx_seqno, h 265 net/dccp/ccids/lib/packet_history.c s2 = tfrc_rx_hist_entry(h, 2)->tfrchrx_seqno, h 269 net/dccp/ccids/lib/packet_history.c h->loss_count = 3; h 270 net/dccp/ccids/lib/packet_history.c tfrc_rx_hist_entry_from_skb(tfrc_rx_hist_entry(h, 3), skb, n3); h 280 net/dccp/ccids/lib/packet_history.c tfrc_rx_hist_swap(h, 2, 3); h 281 net/dccp/ccids/lib/packet_history.c tfrc_rx_hist_entry_from_skb(tfrc_rx_hist_entry(h, 2), skb, n3); h 282 net/dccp/ccids/lib/packet_history.c h->loss_count = 3; h 289 net/dccp/ccids/lib/packet_history.c u64 n1 = tfrc_rx_hist_entry(h, 1)->tfrchrx_ndp; h 293 net/dccp/ccids/lib/packet_history.c u64 n2 = tfrc_rx_hist_entry(h, 2)->tfrchrx_ndp; h 297 net/dccp/ccids/lib/packet_history.c h->loss_start = tfrc_rx_hist_index(h, 2); h 298 net/dccp/ccids/lib/packet_history.c h->loss_count = 0; h 301 net/dccp/ccids/lib/packet_history.c h->loss_start = tfrc_rx_hist_index(h, 1); h 302 net/dccp/ccids/lib/packet_history.c h->loss_count = 1; h 306 net/dccp/ccids/lib/packet_history.c tfrc_rx_hist_entry_from_skb(tfrc_rx_hist_loss_prev(h), skb, n3); h 315 net/dccp/ccids/lib/packet_history.c tfrc_rx_hist_swap(h, 0, 3); h 316 net/dccp/ccids/lib/packet_history.c h->loss_start = tfrc_rx_hist_index(h, 3); h 317 net/dccp/ccids/lib/packet_history.c tfrc_rx_hist_entry_from_skb(tfrc_rx_hist_entry(h, 1), skb, n3); h 318 net/dccp/ccids/lib/packet_history.c h->loss_count = 3; h 332 net/dccp/ccids/lib/packet_history.c u64 s1 = tfrc_rx_hist_entry(h, 1)->tfrchrx_seqno, h 333 net/dccp/ccids/lib/packet_history.c s2 = tfrc_rx_hist_entry(h, 2)->tfrchrx_seqno, h 334 net/dccp/ccids/lib/packet_history.c s3 = tfrc_rx_hist_entry(h, 3)->tfrchrx_seqno; h 335 net/dccp/ccids/lib/packet_history.c u64 n2 = tfrc_rx_hist_entry(h, 2)->tfrchrx_ndp, h 336 net/dccp/ccids/lib/packet_history.c n3 = tfrc_rx_hist_entry(h, 3)->tfrchrx_ndp; h 342 net/dccp/ccids/lib/packet_history.c h->loss_start = tfrc_rx_hist_index(h, 3); h 343 net/dccp/ccids/lib/packet_history.c h->loss_count = 0; h 346 net/dccp/ccids/lib/packet_history.c h->loss_start = tfrc_rx_hist_index(h, 2); h 347 net/dccp/ccids/lib/packet_history.c h->loss_count = 1; h 351 net/dccp/ccids/lib/packet_history.c h->loss_start = tfrc_rx_hist_index(h, 1); h 352 net/dccp/ccids/lib/packet_history.c h->loss_count = 2; h 378 net/dccp/ccids/lib/packet_history.c if (h->loss_count == 0) { h 379 net/dccp/ccids/lib/packet_history.c __do_track_loss(h, skb, ndp); h 380 net/dccp/ccids/lib/packet_history.c } else if (h->loss_count == 1) { h 381 net/dccp/ccids/lib/packet_history.c __one_after_loss(h, skb, ndp); h 382 net/dccp/ccids/lib/packet_history.c } else if (h->loss_count != 2) { h 383 net/dccp/ccids/lib/packet_history.c DCCP_BUG("invalid loss_count %d", h->loss_count); h 384 net/dccp/ccids/lib/packet_history.c } else if (__two_after_loss(h, skb, ndp)) { h 388 net/dccp/ccids/lib/packet_history.c is_new_loss = tfrc_lh_interval_add(lh, h, calc_first_li, sk); h 389 net/dccp/ccids/lib/packet_history.c __three_after_loss(h); h 400 net/dccp/ccids/lib/packet_history.c h->ring[i] = kmem_cache_alloc(tfrc_rx_hist_slab, GFP_ATOMIC); h 401 net/dccp/ccids/lib/packet_history.c if (h->ring[i] == NULL) h 405 net/dccp/ccids/lib/packet_history.c h->loss_count = h->loss_start = 0; h 410 net/dccp/ccids/lib/packet_history.c kmem_cache_free(tfrc_rx_hist_slab, h->ring[i]); h 411 net/dccp/ccids/lib/packet_history.c h->ring[i] = NULL; h 422 net/dccp/ccids/lib/packet_history.c if (h->ring[i] != NULL) { h 423 net/dccp/ccids/lib/packet_history.c kmem_cache_free(tfrc_rx_hist_slab, h->ring[i]); h 424 net/dccp/ccids/lib/packet_history.c h->ring[i] = NULL; h 435 net/dccp/ccids/lib/packet_history.c return h->ring[0]; h 444 net/dccp/ccids/lib/packet_history.c return h->ring[h->rtt_sample_prev]; h 456 net/dccp/ccids/lib/packet_history.c tfrc_rx_hist_rtt_last_s(h)->tfrchrx_ccval); h 459 net/dccp/ccids/lib/packet_history.c if (h->rtt_sample_prev == 2) { /* previous candidate stored */ h 460 net/dccp/ccids/lib/packet_history.c sample = SUB16(tfrc_rx_hist_rtt_prev_s(h)->tfrchrx_ccval, h 461 net/dccp/ccids/lib/packet_history.c tfrc_rx_hist_rtt_last_s(h)->tfrchrx_ccval); h 464 net/dccp/ccids/lib/packet_history.c ktime_us_delta(tfrc_rx_hist_rtt_prev_s(h)->tfrchrx_tstamp, h 465 net/dccp/ccids/lib/packet_history.c tfrc_rx_hist_rtt_last_s(h)->tfrchrx_tstamp); h 474 net/dccp/ccids/lib/packet_history.c tfrc_rx_hist_rtt_prev_s(h)->tfrchrx_ccval, h 475 net/dccp/ccids/lib/packet_history.c tfrc_rx_hist_rtt_last_s(h)->tfrchrx_ccval); h 477 net/dccp/ccids/lib/packet_history.c h->rtt_sample_prev = 1; h 482 net/dccp/ccids/lib/packet_history.c sample = ktime_to_us(net_timedelta(tfrc_rx_hist_rtt_last_s(h)->tfrchrx_tstamp)); h 484 net/dccp/ccids/lib/packet_history.c h->rtt_sample_prev = 2; h 493 net/dccp/ccids/lib/packet_history.c h->rtt_sample_prev = 0; /* use current entry as next reference */ h 91 net/dccp/ccids/lib/packet_history.h return (h->loss_start + n) & TFRC_NDUPACK; h 100 net/dccp/ccids/lib/packet_history.h return h->ring[tfrc_rx_hist_index(h, h->loss_count)]; h 109 net/dccp/ccids/lib/packet_history.h return h->ring[tfrc_rx_hist_index(h, n)]; h 118 net/dccp/ccids/lib/packet_history.h return h->ring[h->loss_start]; h 124 net/dccp/ccids/lib/packet_history.h return h->loss_count > 0; h 943 net/dccp/ipv4.c .h.hashinfo = &dccp_hashinfo, h 1143 net/dccp/ipv6.c .h.hashinfo = &dccp_hashinfo, h 1606 net/decnet/dn_route.c int h, s_h; h 1619 net/decnet/dn_route.c for(h = 0; h <= dn_rt_hash_mask; h++) { h 1620 net/decnet/dn_route.c if (h < s_h) h 1622 net/decnet/dn_route.c if (h > s_h) h 1625 net/decnet/dn_route.c for(rt = rcu_dereference(dn_rt_hash_table[h].chain), idx = 0; h 1644 net/decnet/dn_route.c cb->args[0] = h; h 88 net/decnet/dn_table.c u16 h = dn_ntohs(key.datum)>>(16 - dz->dz_order); h 89 net/decnet/dn_table.c h ^= (h >> 10); h 90 net/decnet/dn_table.c h ^= (h >> 6); h 91 net/decnet/dn_table.c h &= DZ_HASHMASK(dz); h 92 net/decnet/dn_table.c return *(dn_fib_idx_t *)&h; h 418 net/decnet/dn_table.c int h, s_h; h 421 net/decnet/dn_table.c for(h = 0; h < dz->dz_divisor; h++) { h 422 net/decnet/dn_table.c if (h < s_h) h 424 net/decnet/dn_table.c if (h > s_h) h 426 net/decnet/dn_table.c if (dz->dz_hash == NULL || dz->dz_hash[h] == NULL) h 428 net/decnet/dn_table.c if (dn_hash_dump_bucket(skb, cb, tb, dz, dz->dz_hash[h]) < 0) { h 429 net/decnet/dn_table.c cb->args[3] = h; h 433 net/decnet/dn_table.c cb->args[3] = h; h 467 net/decnet/dn_table.c unsigned int h, s_h; h 483 net/decnet/dn_table.c for (h = s_h; h < DN_FIB_TABLE_HASHSZ; h++, s_h = 0) { h 485 net/decnet/dn_table.c hlist_for_each_entry(tb, node, &dn_fib_table_hash[h], hlist) { h 500 net/decnet/dn_table.c cb->args[0] = h; h 815 net/decnet/dn_table.c unsigned int h; h 823 net/decnet/dn_table.c h = n & (DN_FIB_TABLE_HASHSZ - 1); h 825 net/decnet/dn_table.c hlist_for_each_entry_rcu(t, node, &dn_fib_table_hash[h], hlist) { h 852 net/decnet/dn_table.c hlist_add_head_rcu(&t->hlist, &dn_fib_table_hash[h]); h 872 net/decnet/dn_table.c unsigned int h; h 874 net/decnet/dn_table.c for (h = 0; h < DN_FIB_TABLE_HASHSZ; h++) { h 875 net/decnet/dn_table.c hlist_for_each_entry(tb, node, &dn_fib_table_hash[h], hlist) h 895 net/decnet/dn_table.c unsigned int h; h 898 net/decnet/dn_table.c for (h = 0; h < DN_FIB_TABLE_HASHSZ; h++) { h 899 net/decnet/dn_table.c hlist_for_each_entry_safe(t, node, next, &dn_fib_table_hash[h], h 77 net/ipv4/fib_frontend.c unsigned int h; h 88 net/ipv4/fib_frontend.c h = id & (FIB_TABLE_HASHSZ - 1); h 89 net/ipv4/fib_frontend.c hlist_add_head_rcu(&tb->tb_hlist, &net->ipv4.fib_table_hash[h]); h 98 net/ipv4/fib_frontend.c unsigned int h; h 102 net/ipv4/fib_frontend.c h = id & (FIB_TABLE_HASHSZ - 1); h 105 net/ipv4/fib_frontend.c head = &net->ipv4.fib_table_hash[h]; h 138 net/ipv4/fib_frontend.c unsigned int h; h 140 net/ipv4/fib_frontend.c for (h = 0; h < FIB_TABLE_HASHSZ; h++) { h 141 net/ipv4/fib_frontend.c head = &net->ipv4.fib_table_hash[h]; h 628 net/ipv4/fib_frontend.c unsigned int h, s_h; h 642 net/ipv4/fib_frontend.c for (h = s_h; h < FIB_TABLE_HASHSZ; h++, s_e = 0) { h 644 net/ipv4/fib_frontend.c head = &net->ipv4.fib_table_hash[h]; h 660 net/ipv4/fib_frontend.c cb->args[0] = h; h 81 net/ipv4/fib_hash.c u32 h = ntohl(key)>>(32 - fz->fz_order); h 82 net/ipv4/fib_hash.c h ^= (h>>20); h 83 net/ipv4/fib_hash.c h ^= (h>>10); h 84 net/ipv4/fib_hash.c h ^= (h>>5); h 85 net/ipv4/fib_hash.c h &= FZ_HASHMASK(fz); h 86 net/ipv4/fib_hash.c return h; h 728 net/ipv4/fib_hash.c int h, s_h; h 733 net/ipv4/fib_hash.c for (h = s_h; h < fz->fz_divisor; h++) { h 734 net/ipv4/fib_hash.c if (hlist_empty(&fz->fz_hash[h])) h 736 net/ipv4/fib_hash.c if (fn_hash_dump_bucket(skb, cb, tb, fz, &fz->fz_hash[h]) < 0) { h 737 net/ipv4/fib_hash.c cb->args[3] = h; h 743 net/ipv4/fib_hash.c cb->args[3] = h; h 2220 net/ipv4/fib_trie.c unsigned int h; h 2227 net/ipv4/fib_trie.c for (h = 0; h < FIB_TABLE_HASHSZ; h++) { h 2228 net/ipv4/fib_trie.c struct hlist_head *head = &net->ipv4.fib_table_hash[h]; h 2270 net/ipv4/fib_trie.c unsigned int h; h 2272 net/ipv4/fib_trie.c for (h = 0; h < FIB_TABLE_HASHSZ; h++) { h 2273 net/ipv4/fib_trie.c struct hlist_head *head = &net->ipv4.fib_table_hash[h]; h 2306 net/ipv4/fib_trie.c unsigned int h; h 2316 net/ipv4/fib_trie.c h = tb->tb_id & (FIB_TABLE_HASHSZ - 1); h 2325 net/ipv4/fib_trie.c while (++h < FIB_TABLE_HASHSZ) { h 2326 net/ipv4/fib_trie.c struct hlist_head *head = &net->ipv4.fib_table_hash[h]; h 92 net/ipv4/inet_connection_sock.c struct inet_hashinfo *hashinfo = sk->sk_prot->h.hashinfo; h 408 net/ipv4/inet_connection_sock.c const u32 h = inet_synq_hash(inet_rsk(req)->rmt_addr, inet_rsk(req)->rmt_port, h 411 net/ipv4/inet_connection_sock.c reqsk_queue_hash_req(&icsk->icsk_accept_queue, h, req, timeout); h 886 net/ipv4/inet_diag.c const __u16 type = h->idiag_type; h 895 net/ipv4/inet_diag.c inet_diag_table[type] = h; h 906 net/ipv4/inet_diag.c const __u16 type = h->idiag_type; h 72 net/ipv4/inet_hashtables.c struct inet_hashinfo *hashinfo = sk->sk_prot->h.hashinfo; h 98 net/ipv4/inet_hashtables.c struct inet_hashinfo *table = sk->sk_prot->h.hashinfo; h 340 net/ipv4/inet_hashtables.c struct inet_hashinfo *hashinfo = sk->sk_prot->h.hashinfo; h 361 net/ipv4/inet_hashtables.c struct inet_hashinfo *hashinfo = sk->sk_prot->h.hashinfo; h 394 net/ipv4/inet_hashtables.c struct inet_hashinfo *hashinfo = sk->sk_prot->h.hashinfo; h 420 net/ipv4/inet_timewait_sock.c int h; h 423 net/ipv4/inet_timewait_sock.c for (h = 0; h < (hashinfo->ehash_size); h++) { h 425 net/ipv4/inet_timewait_sock.c inet_ehash_bucket(hashinfo, h); h 426 net/ipv4/inet_timewait_sock.c rwlock_t *lock = inet_ehash_lockp(hashinfo, h); h 55 net/ipv4/ip_fragment.c struct inet_skb_parm h; h 239 net/ipv4/ip_gre.c unsigned h = HASH(key); h 246 net/ipv4/ip_gre.c h ^= HASH(remote); h 249 net/ipv4/ip_gre.c return &ign->tunnels[prio][h]; h 472 net/ipv4/ip_gre.c u8 *h; h 486 net/ipv4/ip_gre.c h = skb->data; h 487 net/ipv4/ip_gre.c flags = *(__be16*)h; h 511 net/ipv4/ip_gre.c key = *(__be32*)(h + offset); h 515 net/ipv4/ip_gre.c seqno = ntohl(*(__be32*)(h + offset)); h 520 net/ipv4/ip_gre.c gre_proto = *(__be16 *)(h + 2); h 537 net/ipv4/ip_gre.c if ((*(h + offset) & 0xF0) != 0x40) h 1238 net/ipv4/ip_gre.c int h; h 1239 net/ipv4/ip_gre.c for (h = 0; h < HASH_SIZE; h++) { h 1241 net/ipv4/ip_gre.c while ((t = ign->tunnels[prio][h]) != NULL) h 711 net/ipv4/ipconfig.c struct iphdr *h; h 724 net/ipv4/ipconfig.c h = ip_hdr(skb); h 725 net/ipv4/ipconfig.c h->version = 4; h 726 net/ipv4/ipconfig.c h->ihl = 5; h 727 net/ipv4/ipconfig.c h->tot_len = htons(sizeof(struct bootp_pkt)); h 728 net/ipv4/ipconfig.c h->frag_off = htons(IP_DF); h 729 net/ipv4/ipconfig.c h->ttl = 64; h 730 net/ipv4/ipconfig.c h->protocol = IPPROTO_UDP; h 731 net/ipv4/ipconfig.c h->daddr = htonl(INADDR_BROADCAST); h 732 net/ipv4/ipconfig.c h->check = ip_fast_csum((unsigned char *) h, h->ihl); h 851 net/ipv4/ipconfig.c struct iphdr *h; h 871 net/ipv4/ipconfig.c h = &b->iph; h 873 net/ipv4/ipconfig.c if (h->ihl != 5 || h->version != 4 || h->protocol != IPPROTO_UDP) h 877 net/ipv4/ipconfig.c if (h->frag_off & htons(IP_OFFSET | IP_MF)) { h 884 net/ipv4/ipconfig.c if (skb->len < ntohs(h->tot_len)) h 887 net/ipv4/ipconfig.c if (ip_fast_csum((char *) h, h->ihl)) h 893 net/ipv4/ipconfig.c if (ntohs(h->tot_len) < ntohs(b->udph.len) + sizeof(struct iphdr)) h 909 net/ipv4/ipconfig.c h = &b->iph; h 170 net/ipv4/ipip.c unsigned h = 0; h 175 net/ipv4/ipip.c h ^= HASH(remote); h 179 net/ipv4/ipip.c h ^= HASH(local); h 181 net/ipv4/ipip.c return &ipn->tunnels[prio][h]; h 760 net/ipv4/ipip.c int h; h 761 net/ipv4/ipip.c for (h = 0; h < HASH_SIZE; h++) { h 763 net/ipv4/ipip.c while ((t = ipn->tunnels[prio][h]) != NULL) h 525 net/ipv4/netfilter/arp_tables.c unsigned int h; h 541 net/ipv4/netfilter/arp_tables.c for (h = 0; h < NF_ARP_NUMHOOKS; h++) { h 542 net/ipv4/netfilter/arp_tables.c if ((unsigned char *)e - base == hook_entries[h]) h 543 net/ipv4/netfilter/arp_tables.c newinfo->hook_entry[h] = hook_entries[h]; h 544 net/ipv4/netfilter/arp_tables.c if ((unsigned char *)e - base == underflows[h]) h 545 net/ipv4/netfilter/arp_tables.c newinfo->underflow[h] = underflows[h]; h 1204 net/ipv4/netfilter/arp_tables.c int ret, off, h; h 1248 net/ipv4/netfilter/arp_tables.c for (h = 0; h < NF_ARP_NUMHOOKS; h++) { h 1249 net/ipv4/netfilter/arp_tables.c if ((unsigned char *)e - base == hook_entries[h]) h 1250 net/ipv4/netfilter/arp_tables.c newinfo->hook_entry[h] = hook_entries[h]; h 1251 net/ipv4/netfilter/arp_tables.c if ((unsigned char *)e - base == underflows[h]) h 1252 net/ipv4/netfilter/arp_tables.c newinfo->underflow[h] = underflows[h]; h 1277 net/ipv4/netfilter/arp_tables.c int ret, h; h 1294 net/ipv4/netfilter/arp_tables.c for (h = 0; h < NF_ARP_NUMHOOKS; h++) { h 1295 net/ipv4/netfilter/arp_tables.c if ((unsigned char *)de - base < newinfo->hook_entry[h]) h 1296 net/ipv4/netfilter/arp_tables.c newinfo->hook_entry[h] -= origsize - *size; h 1297 net/ipv4/netfilter/arp_tables.c if ((unsigned char *)de - base < newinfo->underflow[h]) h 1298 net/ipv4/netfilter/arp_tables.c newinfo->underflow[h] -= origsize - *size; h 738 net/ipv4/netfilter/ip_tables.c unsigned int h; h 754 net/ipv4/netfilter/ip_tables.c for (h = 0; h < NF_INET_NUMHOOKS; h++) { h 755 net/ipv4/netfilter/ip_tables.c if ((unsigned char *)e - base == hook_entries[h]) h 756 net/ipv4/netfilter/ip_tables.c newinfo->hook_entry[h] = hook_entries[h]; h 757 net/ipv4/netfilter/ip_tables.c if ((unsigned char *)e - base == underflows[h]) h 758 net/ipv4/netfilter/ip_tables.c newinfo->underflow[h] = underflows[h]; h 1546 net/ipv4/netfilter/ip_tables.c int ret, off, h; h 1595 net/ipv4/netfilter/ip_tables.c for (h = 0; h < NF_INET_NUMHOOKS; h++) { h 1596 net/ipv4/netfilter/ip_tables.c if ((unsigned char *)e - base == hook_entries[h]) h 1597 net/ipv4/netfilter/ip_tables.c newinfo->hook_entry[h] = hook_entries[h]; h 1598 net/ipv4/netfilter/ip_tables.c if ((unsigned char *)e - base == underflows[h]) h 1599 net/ipv4/netfilter/ip_tables.c newinfo->underflow[h] = underflows[h]; h 1625 net/ipv4/netfilter/ip_tables.c int ret, h; h 1646 net/ipv4/netfilter/ip_tables.c for (h = 0; h < NF_INET_NUMHOOKS; h++) { h 1647 net/ipv4/netfilter/ip_tables.c if ((unsigned char *)de - base < newinfo->hook_entry[h]) h 1648 net/ipv4/netfilter/ip_tables.c newinfo->hook_entry[h] -= origsize - *size; h 1649 net/ipv4/netfilter/ip_tables.c if ((unsigned char *)de - base < newinfo->underflow[h]) h 1650 net/ipv4/netfilter/ip_tables.c newinfo->underflow[h] -= origsize - *size; h 252 net/ipv4/netfilter/nf_conntrack_l3proto_ipv4.c const struct nf_conntrack_tuple_hash *h; h 275 net/ipv4/netfilter/nf_conntrack_l3proto_ipv4.c h = nf_conntrack_find_get(sock_net(sk), &tuple); h 276 net/ipv4/netfilter/nf_conntrack_l3proto_ipv4.c if (h) { h 278 net/ipv4/netfilter/nf_conntrack_l3proto_ipv4.c struct nf_conn *ct = nf_ct_tuplehash_to_ctrack(h); h 132 net/ipv4/netfilter/nf_conntrack_proto_icmp.c const struct nf_conntrack_tuple_hash *h; h 158 net/ipv4/netfilter/nf_conntrack_proto_icmp.c h = nf_conntrack_find_get(net, &innertuple); h 159 net/ipv4/netfilter/nf_conntrack_proto_icmp.c if (!h) { h 164 net/ipv4/netfilter/nf_conntrack_proto_icmp.c if (NF_CT_DIRECTION(h) == IP_CT_DIR_REPLY) h 168 net/ipv4/netfilter/nf_conntrack_proto_icmp.c skb->nfct = &nf_ct_tuplehash_to_ctrack(h)->ct_general; h 150 net/ipv4/netfilter/nf_nat_core.c unsigned int h = hash_by_src(tuple); h 156 net/ipv4/netfilter/nf_nat_core.c hlist_for_each_entry_rcu(nat, n, &net->ipv4.nat_bysource[h], bysource) { h 87 net/ipv4/raw.c struct raw_hashinfo *h = sk->sk_prot->h.raw_hash; h 90 net/ipv4/raw.c head = &h->ht[inet_sk(sk)->num & (RAW_HTABLE_SIZE - 1)]; h 92 net/ipv4/raw.c write_lock_bh(&h->lock); h 95 net/ipv4/raw.c write_unlock_bh(&h->lock); h 101 net/ipv4/raw.c struct raw_hashinfo *h = sk->sk_prot->h.raw_hash; h 103 net/ipv4/raw.c write_lock_bh(&h->lock); h 106 net/ipv4/raw.c write_unlock_bh(&h->lock); h 843 net/ipv4/raw.c .h.raw_hash = &raw_v4_hashinfo, h 860 net/ipv4/raw.c sk_for_each(sk, node, &state->h->ht[state->bucket]) h 880 net/ipv4/raw.c sk = sk_head(&state->h->ht[state->bucket]); h 900 net/ipv4/raw.c read_lock(&state->h->lock); h 922 net/ipv4/raw.c read_unlock(&state->h->lock); h 972 net/ipv4/raw.c i->h = h; h 2844 net/ipv4/route.c int h, s_h; h 2854 net/ipv4/route.c for (h = s_h; h <= rt_hash_mask; h++, s_idx = 0) { h 2855 net/ipv4/route.c if (!rt_hash_table[h].chain) h 2858 net/ipv4/route.c for (rt = rcu_dereference(rt_hash_table[h].chain), idx = 0; rt; h 2878 net/ipv4/route.c cb->args[0] = h; h 2383 net/ipv4/tcp_ipv4.c .h.hashinfo = &tcp_hashinfo, h 157 net/ipv4/udp.c struct hlist_head *udptable = sk->sk_prot->h.udp_hash; h 1491 net/ipv4/udp.c .h.udp_hash = udp_hash, h 53 net/ipv4/udplite.c .h.udp_hash = udplite_hash, h 132 net/ipv6/addrlabel.c ip6addrlbl_free(container_of(h, struct ip6addrlbl_entry, rcu)); h 700 net/ipv6/exthdrs.c struct ipv6_opt_hdr *h = (struct ipv6_opt_hdr *)skb_push(skb, ipv6_optlen(opt)); h 702 net/ipv6/exthdrs.c memcpy(h, opt, ipv6_optlen(opt)); h 703 net/ipv6/exthdrs.c h->nexthdr = *proto; h 118 net/ipv6/inet6_connection_sock.c const u32 h = inet6_synq_hash(&inet6_rsk(req)->rmt_addr, h 122 net/ipv6/inet6_connection_sock.c reqsk_queue_hash_req(&icsk->icsk_accept_queue, h, req, timeout); h 27 net/ipv6/inet6_hashtables.c struct inet_hashinfo *hashinfo = sk->sk_prot->h.hashinfo; h 175 net/ipv6/ip6_fib.c unsigned int h; h 183 net/ipv6/ip6_fib.c h = tb->tb6_id & (FIB_TABLE_HASHSZ - 1); h 189 net/ipv6/ip6_fib.c hlist_add_head_rcu(&tb->tb6_hlist, &net->ipv6.fib_table_hash[h]); h 230 net/ipv6/ip6_fib.c unsigned int h; h 234 net/ipv6/ip6_fib.c h = id & (FIB_TABLE_HASHSZ - 1); h 236 net/ipv6/ip6_fib.c head = &net->ipv6.fib_table_hash[h]; h 348 net/ipv6/ip6_fib.c unsigned int h, s_h; h 384 net/ipv6/ip6_fib.c for (h = s_h; h < FIB_TABLE_HASHSZ; h++, s_e = 0) { h 386 net/ipv6/ip6_fib.c head = &net->ipv6.fib_table_hash[h]; h 399 net/ipv6/ip6_fib.c cb->args[0] = h; h 1367 net/ipv6/ip6_fib.c unsigned int h; h 1370 net/ipv6/ip6_fib.c for (h = 0; h < FIB_TABLE_HASHSZ; h++) { h 1371 net/ipv6/ip6_fib.c head = &net->ipv6.fib_table_hash[h]; h 169 net/ipv6/ip6_tunnel.c unsigned h = 0; h 174 net/ipv6/ip6_tunnel.c h = HASH(remote) ^ HASH(local); h 176 net/ipv6/ip6_tunnel.c return &ip6n->tnls[prio][h]; h 1396 net/ipv6/ip6_tunnel.c int h; h 1399 net/ipv6/ip6_tunnel.c for (h = 0; h < HASH_SIZE; h++) { h 1400 net/ipv6/ip6_tunnel.c while ((t = ip6n->tnls_r_l[h]) != NULL) h 763 net/ipv6/netfilter/ip6_tables.c unsigned int h; h 779 net/ipv6/netfilter/ip6_tables.c for (h = 0; h < NF_INET_NUMHOOKS; h++) { h 780 net/ipv6/netfilter/ip6_tables.c if ((unsigned char *)e - base == hook_entries[h]) h 781 net/ipv6/netfilter/ip6_tables.c newinfo->hook_entry[h] = hook_entries[h]; h 782 net/ipv6/netfilter/ip6_tables.c if ((unsigned char *)e - base == underflows[h]) h 783 net/ipv6/netfilter/ip6_tables.c newinfo->underflow[h] = underflows[h]; h 1573 net/ipv6/netfilter/ip6_tables.c int ret, off, h; h 1622 net/ipv6/netfilter/ip6_tables.c for (h = 0; h < NF_INET_NUMHOOKS; h++) { h 1623 net/ipv6/netfilter/ip6_tables.c if ((unsigned char *)e - base == hook_entries[h]) h 1624 net/ipv6/netfilter/ip6_tables.c newinfo->hook_entry[h] = hook_entries[h]; h 1625 net/ipv6/netfilter/ip6_tables.c if ((unsigned char *)e - base == underflows[h]) h 1626 net/ipv6/netfilter/ip6_tables.c newinfo->underflow[h] = underflows[h]; h 1652 net/ipv6/netfilter/ip6_tables.c int ret, h; h 1673 net/ipv6/netfilter/ip6_tables.c for (h = 0; h < NF_INET_NUMHOOKS; h++) { h 1674 net/ipv6/netfilter/ip6_tables.c if ((unsigned char *)de - base < newinfo->hook_entry[h]) h 1675 net/ipv6/netfilter/ip6_tables.c newinfo->hook_entry[h] -= origsize - *size; h 1676 net/ipv6/netfilter/ip6_tables.c if ((unsigned char *)de - base < newinfo->underflow[h]) h 1677 net/ipv6/netfilter/ip6_tables.c newinfo->underflow[h] -= origsize - *size; h 132 net/ipv6/netfilter/nf_conntrack_proto_icmpv6.c const struct nf_conntrack_tuple_hash *h; h 160 net/ipv6/netfilter/nf_conntrack_proto_icmpv6.c h = nf_conntrack_find_get(net, &intuple); h 161 net/ipv6/netfilter/nf_conntrack_proto_icmpv6.c if (!h) { h 165 net/ipv6/netfilter/nf_conntrack_proto_icmpv6.c if (NF_CT_DIRECTION(h) == IP_CT_DIR_REPLY) h 170 net/ipv6/netfilter/nf_conntrack_proto_icmpv6.c skb->nfct = &nf_ct_tuplehash_to_ctrack(h)->ct_general; h 54 net/ipv6/netfilter/nf_conntrack_reasm.c struct inet6_skb_parm h; h 1211 net/ipv6/raw.c .h.raw_hash = &raw_v6_hashinfo, h 59 net/ipv6/reassembly.c struct inet6_skb_parm h; h 113 net/ipv6/sit.c unsigned h = 0; h 118 net/ipv6/sit.c h ^= HASH(remote); h 122 net/ipv6/sit.c h ^= HASH(local); h 124 net/ipv6/sit.c return &sitn->tunnels[prio][h]; h 994 net/ipv6/sit.c int h; h 995 net/ipv6/sit.c for (h = 0; h < HASH_SIZE; h++) { h 997 net/ipv6/sit.c while ((t = sitn->tunnels[prio][h]) != NULL) h 2050 net/ipv6/tcp_ipv6.c .h.hashinfo = &tcp_hashinfo, h 1053 net/ipv6/udp.c .h.udp_hash = udp_hash, h 52 net/ipv6/udplite.c .h.udp_hash = udplite_hash, h 63 net/ipv6/xfrm6_tunnel.c unsigned h; h 65 net/ipv6/xfrm6_tunnel.c h = (__force u32)(addr->a6[0] ^ addr->a6[1] ^ addr->a6[2] ^ addr->a6[3]); h 66 net/ipv6/xfrm6_tunnel.c h ^= h >> 16; h 67 net/ipv6/xfrm6_tunnel.c h ^= h >> 8; h 68 net/ipv6/xfrm6_tunnel.c h &= XFRM6_TUNNEL_SPI_BYADDR_HSIZE - 1; h 70 net/ipv6/xfrm6_tunnel.c return h; h 214 net/irda/irqueue.c __u32 h = 0; h 218 net/irda/irqueue.c h = (h<<4) + *name++; h 219 net/irda/irqueue.c if ((g = (h & 0xf0000000))) h 220 net/irda/irqueue.c h ^=g>>24; h 221 net/irda/irqueue.c h &=~g; h 223 net/irda/irqueue.c return h; h 465 net/mac80211/rx.c #define msh_h_get(h, l) ((struct ieee80211s_hdr *) ((u8 *)h + l)) h 258 net/netfilter/core.c int i, h; h 260 net/netfilter/core.c for (h = 0; h < NF_MAX_HOOKS; h++) h 261 net/netfilter/core.c INIT_LIST_HEAD(&nf_hooks[i][h]); h 65 net/netfilter/nf_conntrack_core.c u_int32_t h; h 72 net/netfilter/nf_conntrack_core.c h = jhash2((u32 *)tuple, n, h 76 net/netfilter/nf_conntrack_core.c return ((u64)h * size) >> 32; h 240 net/netfilter/nf_conntrack_core.c struct nf_conntrack_tuple_hash *h; h 248 net/netfilter/nf_conntrack_core.c hlist_for_each_entry_rcu(h, n, &net->ct.hash[hash], hnode) { h 249 net/netfilter/nf_conntrack_core.c if (nf_ct_tuple_equal(tuple, &h->tuple)) { h 252 net/netfilter/nf_conntrack_core.c return h; h 266 net/netfilter/nf_conntrack_core.c struct nf_conntrack_tuple_hash *h; h 270 net/netfilter/nf_conntrack_core.c h = __nf_conntrack_find(net, tuple); h 271 net/netfilter/nf_conntrack_core.c if (h) { h 272 net/netfilter/nf_conntrack_core.c ct = nf_ct_tuplehash_to_ctrack(h); h 274 net/netfilter/nf_conntrack_core.c h = NULL; h 278 net/netfilter/nf_conntrack_core.c return h; h 312 net/netfilter/nf_conntrack_core.c struct nf_conntrack_tuple_hash *h; h 347 net/netfilter/nf_conntrack_core.c hlist_for_each_entry(h, n, &net->ct.hash[hash], hnode) h 349 net/netfilter/nf_conntrack_core.c &h->tuple)) h 351 net/netfilter/nf_conntrack_core.c hlist_for_each_entry(h, n, &net->ct.hash[repl_hash], hnode) h 353 net/netfilter/nf_conntrack_core.c &h->tuple)) h 395 net/netfilter/nf_conntrack_core.c struct nf_conntrack_tuple_hash *h; h 403 net/netfilter/nf_conntrack_core.c hlist_for_each_entry_rcu(h, n, &net->ct.hash[hash], hnode) { h 404 net/netfilter/nf_conntrack_core.c if (nf_ct_tuplehash_to_ctrack(h) != ignored_conntrack && h 405 net/netfilter/nf_conntrack_core.c nf_ct_tuple_equal(tuple, &h->tuple)) { h 425 net/netfilter/nf_conntrack_core.c struct nf_conntrack_tuple_hash *h; h 433 net/netfilter/nf_conntrack_core.c hlist_for_each_entry_rcu(h, n, &net->ct.hash[hash], h 435 net/netfilter/nf_conntrack_core.c tmp = nf_ct_tuplehash_to_ctrack(h); h 623 net/netfilter/nf_conntrack_core.c struct nf_conntrack_tuple_hash *h; h 634 net/netfilter/nf_conntrack_core.c h = nf_conntrack_find_get(net, &tuple); h 635 net/netfilter/nf_conntrack_core.c if (!h) { h 636 net/netfilter/nf_conntrack_core.c h = init_conntrack(net, &tuple, l3proto, l4proto, skb, dataoff); h 637 net/netfilter/nf_conntrack_core.c if (!h) h 639 net/netfilter/nf_conntrack_core.c if (IS_ERR(h)) h 640 net/netfilter/nf_conntrack_core.c return (void *)h; h 642 net/netfilter/nf_conntrack_core.c ct = nf_ct_tuplehash_to_ctrack(h); h 645 net/netfilter/nf_conntrack_core.c if (NF_CT_DIRECTION(h) == IP_CT_DIR_REPLY) { h 949 net/netfilter/nf_conntrack_core.c struct nf_conntrack_tuple_hash *h; h 955 net/netfilter/nf_conntrack_core.c hlist_for_each_entry(h, n, &net->ct.hash[*bucket], hnode) { h 956 net/netfilter/nf_conntrack_core.c ct = nf_ct_tuplehash_to_ctrack(h); h 961 net/netfilter/nf_conntrack_core.c hlist_for_each_entry(h, n, &net->ct.unconfirmed, hnode) { h 962 net/netfilter/nf_conntrack_core.c ct = nf_ct_tuplehash_to_ctrack(h); h 1092 net/netfilter/nf_conntrack_core.c struct nf_conntrack_tuple_hash *h; h 1118 net/netfilter/nf_conntrack_core.c h = hlist_entry(init_net.ct.hash[i].first, h 1120 net/netfilter/nf_conntrack_core.c hlist_del_rcu(&h->hnode); h 1121 net/netfilter/nf_conntrack_core.c bucket = __hash_conntrack(&h->tuple, hashsize, rnd); h 1122 net/netfilter/nf_conntrack_core.c hlist_add_head(&h->hnode, &hash[bucket]); h 90 net/netfilter/nf_conntrack_expect.c unsigned int h; h 95 net/netfilter/nf_conntrack_expect.c h = nf_ct_expect_dst_hash(tuple); h 96 net/netfilter/nf_conntrack_expect.c hlist_for_each_entry_rcu(i, n, &net->ct.expect_hash[h], hnode) { h 127 net/netfilter/nf_conntrack_expect.c unsigned int h; h 132 net/netfilter/nf_conntrack_expect.c h = nf_ct_expect_dst_hash(tuple); h 133 net/netfilter/nf_conntrack_expect.c hlist_for_each_entry(i, n, &net->ct.expect_hash[h], hnode) { h 312 net/netfilter/nf_conntrack_expect.c unsigned int h = nf_ct_expect_dst_hash(&exp->tuple); h 319 net/netfilter/nf_conntrack_expect.c hlist_add_head_rcu(&exp->hnode, &net->ct.expect_hash[h]); h 373 net/netfilter/nf_conntrack_expect.c unsigned int h; h 383 net/netfilter/nf_conntrack_expect.c h = nf_ct_expect_dst_hash(&expect->tuple); h 384 net/netfilter/nf_conntrack_expect.c hlist_for_each_entry(i, n, &net->ct.expect_hash[h], hnode) { h 53 net/netfilter/nf_conntrack_helper.c unsigned int h; h 58 net/netfilter/nf_conntrack_helper.c h = helper_hash(tuple); h 59 net/netfilter/nf_conntrack_helper.c hlist_for_each_entry_rcu(helper, n, &nf_ct_helper_hash[h], hnode) { h 70 net/netfilter/nf_conntrack_helper.c struct nf_conntrack_helper *h; h 75 net/netfilter/nf_conntrack_helper.c hlist_for_each_entry_rcu(h, n, &nf_ct_helper_hash[i], hnode) { h 76 net/netfilter/nf_conntrack_helper.c if (!strcmp(h->name, name)) h 77 net/netfilter/nf_conntrack_helper.c return h; h 112 net/netfilter/nf_conntrack_helper.c unsigned int h = helper_hash(&me->tuple); h 118 net/netfilter/nf_conntrack_helper.c hlist_add_head_rcu(&me->hnode, &nf_ct_helper_hash[h]); h 129 net/netfilter/nf_conntrack_helper.c struct nf_conntrack_tuple_hash *h; h 148 net/netfilter/nf_conntrack_helper.c hlist_for_each_entry(h, n, &net->ct.unconfirmed, hnode) h 149 net/netfilter/nf_conntrack_helper.c unhelp(h, me); h 151 net/netfilter/nf_conntrack_helper.c hlist_for_each_entry(h, n, &net->ct.hash[i], hnode) h 152 net/netfilter/nf_conntrack_helper.c unhelp(h, me); h 543 net/netfilter/nf_conntrack_netlink.c struct nf_conntrack_tuple_hash *h; h 552 net/netfilter/nf_conntrack_netlink.c hlist_for_each_entry_rcu(h, n, &init_net.ct.hash[cb->args[0]], h 554 net/netfilter/nf_conntrack_netlink.c if (NF_CT_DIRECTION(h) != IP_CT_DIR_ORIGINAL) h 556 net/netfilter/nf_conntrack_netlink.c ct = nf_ct_tuplehash_to_ctrack(h); h 784 net/netfilter/nf_conntrack_netlink.c struct nf_conntrack_tuple_hash *h; h 804 net/netfilter/nf_conntrack_netlink.c h = nf_conntrack_find_get(&init_net, &tuple); h 805 net/netfilter/nf_conntrack_netlink.c if (!h) h 808 net/netfilter/nf_conntrack_netlink.c ct = nf_ct_tuplehash_to_ctrack(h); h 828 net/netfilter/nf_conntrack_netlink.c struct nf_conntrack_tuple_hash *h; h 850 net/netfilter/nf_conntrack_netlink.c h = nf_conntrack_find_get(&init_net, &tuple); h 851 net/netfilter/nf_conntrack_netlink.c if (!h) h 854 net/netfilter/nf_conntrack_netlink.c ct = nf_ct_tuplehash_to_ctrack(h); h 1197 net/netfilter/nf_conntrack_netlink.c struct nf_conntrack_tuple_hash *h = NULL; h 1216 net/netfilter/nf_conntrack_netlink.c h = __nf_conntrack_find(&init_net, &otuple); h 1218 net/netfilter/nf_conntrack_netlink.c h = __nf_conntrack_find(&init_net, &rtuple); h 1220 net/netfilter/nf_conntrack_netlink.c if (h == NULL) { h 1270 net/netfilter/nf_conntrack_netlink.c err = ctnetlink_change_conntrack(nf_ct_tuplehash_to_ctrack(h), h 1573 net/netfilter/nf_conntrack_netlink.c struct nf_conntrack_helper *h; h 1610 net/netfilter/nf_conntrack_netlink.c h = __nf_conntrack_helper_find_byname(name); h 1611 net/netfilter/nf_conntrack_netlink.c if (!h) { h 1620 net/netfilter/nf_conntrack_netlink.c if (m_help->helper == h h 1656 net/netfilter/nf_conntrack_netlink.c struct nf_conntrack_tuple_hash *h = NULL; h 1674 net/netfilter/nf_conntrack_netlink.c h = nf_conntrack_find_get(&init_net, &master_tuple); h 1675 net/netfilter/nf_conntrack_netlink.c if (!h) h 1677 net/netfilter/nf_conntrack_netlink.c ct = nf_ct_tuplehash_to_ctrack(h); h 1704 net/netfilter/nf_conntrack_netlink.c nf_ct_put(nf_ct_tuplehash_to_ctrack(h)); h 141 net/netfilter/nf_conntrack_pptp.c const struct nf_conntrack_tuple_hash *h; h 148 net/netfilter/nf_conntrack_pptp.c h = nf_conntrack_find_get(net, t); h 149 net/netfilter/nf_conntrack_pptp.c if (h) { h 150 net/netfilter/nf_conntrack_pptp.c sibling = nf_ct_tuplehash_to_ctrack(h); h 857 net/netfilter/nfnetlink_log.c h = h->next; h 858 net/netfilter/nfnetlink_log.c while (!h) { h 862 net/netfilter/nfnetlink_log.c h = instance_table[st->bucket].first; h 864 net/netfilter/nfnetlink_log.c return h; h 93 net/netfilter/nfnetlink_queue.c unsigned int h; h 122 net/netfilter/nfnetlink_queue.c h = instance_hashfn(queue_num); h 123 net/netfilter/nfnetlink_queue.c hlist_add_head_rcu(&inst->hlist, &instance_table[h]); h 819 net/netfilter/nfnetlink_queue.c h = h->next; h 820 net/netfilter/nfnetlink_queue.c while (!h) { h 824 net/netfilter/nfnetlink_queue.c h = instance_table[st->bucket].first; h 826 net/netfilter/nfnetlink_queue.c return h; h 35 net/netfilter/xt_RATEEST.c unsigned int h; h 37 net/netfilter/xt_RATEEST.c h = xt_rateest_hash(est->name); h 38 net/netfilter/xt_RATEEST.c hlist_add_head(&est->list, &rateest_hash[h]); h 45 net/netfilter/xt_RATEEST.c unsigned int h; h 47 net/netfilter/xt_RATEEST.c h = xt_rateest_hash(name); h 49 net/netfilter/xt_RATEEST.c hlist_for_each_entry(est, n, &rateest_hash[h], list) { h 125 net/netfilter/xt_recent.c unsigned int h; h 128 net/netfilter/xt_recent.c h = recent_entry_hash4(addrp); h 130 net/netfilter/xt_recent.c h = recent_entry_hash6(addrp); h 132 net/netfilter/xt_recent.c list_for_each_entry(e, &table->iphash[h], list) h 78 net/netlabel/netlabel_addrlist.h while (i != h && !n->valid) { h 90 net/netlabel/netlabel_addrlist.h while (i != h && !n->valid) { h 136 net/netlabel/netlabel_addrlist.h while (i != h && !n->valid) { h 148 net/netlabel/netlabel_addrlist.h while (i != h && !n->valid) { h 215 net/packet/af_packet.c } h; h 220 net/packet/af_packet.c h.raw = po->pg_vec[pg_vec_pos] + (frame_offset * po->frame_size); h 223 net/packet/af_packet.c if (status != h.h1->tp_status ? TP_STATUS_USER : h 228 net/packet/af_packet.c if (status != h.h2->tp_status ? TP_STATUS_USER : h 233 net/packet/af_packet.c return h.raw; h 242 net/packet/af_packet.c } h; h 244 net/packet/af_packet.c h.raw = frame; h 247 net/packet/af_packet.c h.h1->tp_status = status; h 250 net/packet/af_packet.c h.h2->tp_status = status; h 599 net/packet/af_packet.c } h; h 668 net/packet/af_packet.c h.raw = packet_lookup_frame(po, po->head, TP_STATUS_KERNEL); h 669 net/packet/af_packet.c if (!h.raw) h 681 net/packet/af_packet.c skb_copy_bits(skb, 0, h.raw + macoff, snaplen); h 685 net/packet/af_packet.c h.h1->tp_len = skb->len; h 686 net/packet/af_packet.c h.h1->tp_snaplen = snaplen; h 687 net/packet/af_packet.c h.h1->tp_mac = macoff; h 688 net/packet/af_packet.c h.h1->tp_net = netoff; h 693 net/packet/af_packet.c h.h1->tp_sec = tv.tv_sec; h 694 net/packet/af_packet.c h.h1->tp_usec = tv.tv_usec; h 695 net/packet/af_packet.c hdrlen = sizeof(*h.h1); h 698 net/packet/af_packet.c h.h2->tp_len = skb->len; h 699 net/packet/af_packet.c h.h2->tp_snaplen = snaplen; h 700 net/packet/af_packet.c h.h2->tp_mac = macoff; h 701 net/packet/af_packet.c h.h2->tp_net = netoff; h 706 net/packet/af_packet.c h.h2->tp_sec = ts.tv_sec; h 707 net/packet/af_packet.c h.h2->tp_nsec = ts.tv_nsec; h 708 net/packet/af_packet.c h.h2->tp_vlan_tci = skb->vlan_tci; h 709 net/packet/af_packet.c hdrlen = sizeof(*h.h2); h 715 net/packet/af_packet.c sll = h.raw + TPACKET_ALIGN(hdrlen); h 726 net/packet/af_packet.c __packet_set_status(po, h.raw, status); h 731 net/packet/af_packet.c u8 *h_end = h.raw + macoff + snaplen - 1; h 733 net/packet/af_packet.c p_start = virt_to_page(h.raw); h 67 net/phonet/pep.c } *ph, h; h 70 net/phonet/pep.c ph = skb_header_pointer(skb, 0, 2, &h); h 30 net/sched/act_api.c unsigned int h = tcf_hash(p->tcfc_index, hinfo->hmask); h 33 net/sched/act_api.c for (p1p = &hinfo->htab[h]; *p1p; p1p = &(*p1p)->tcfc_next) { h 241 net/sched/act_api.c unsigned int h = tcf_hash(p->tcfc_index, hinfo->hmask); h 244 net/sched/act_api.c p->tcfc_next = hinfo->htab[h]; h 245 net/sched/act_api.c hinfo->htab[h] = p; h 587 net/sched/act_api.c struct tcf_act_hdr *h = a->priv; h 589 net/sched/act_api.c if (h == NULL) h 598 net/sched/act_api.c TCA_STATS, TCA_XSTATS, &h->tcf_lock, &d); h 603 net/sched/act_api.c &h->tcf_lock, &d); h 612 net/sched/act_api.c if (gnet_stats_copy_basic(&d, &h->tcf_bstats) < 0 || h 613 net/sched/act_api.c gnet_stats_copy_rate_est(&d, &h->tcf_rate_est) < 0 || h 614 net/sched/act_api.c gnet_stats_copy_queue(&d, &h->tcf_qstats) < 0) h 101 net/sched/act_police.c unsigned int h = tcf_hash(p->tcf_index, POL_TAB_MASK); h 104 net/sched/act_police.c for (p1p = &tcf_police_ht[h]; *p1p; p1p = &(*p1p)->tcfc_next) { h 132 net/sched/act_police.c unsigned h; h 232 net/sched/act_police.c h = tcf_hash(police->tcf_index, POL_TAB_MASK); h 234 net/sched/act_police.c police->tcf_next = tcf_police_ht[h]; h 235 net/sched/act_police.c tcf_police_ht[h] = &police->common; h 153 net/sched/cls_fw.c int h; h 158 net/sched/cls_fw.c for (h=0; h<HTSIZE; h++) { h 159 net/sched/cls_fw.c while ((f=head->ht[h]) != NULL) { h 160 net/sched/cls_fw.c head->ht[h] = f->next; h 305 net/sched/cls_fw.c int h; h 313 net/sched/cls_fw.c for (h = 0; h < HTSIZE; h++) { h 316 net/sched/cls_fw.c for (f = head->ht[h]; f; f = f->next) { h 89 net/sched/cls_route.c int h = route4_fastmap_hash(id, iif); h 90 net/sched/cls_route.c head->fastmap[h].id = id; h 91 net/sched/cls_route.c head->fastmap[h].iif = iif; h 92 net/sched/cls_route.c head->fastmap[h].filter = f; h 137 net/sched/cls_route.c u32 id, h; h 149 net/sched/cls_route.c h = route4_fastmap_hash(id, iif); h 150 net/sched/cls_route.c if (id == head->fastmap[h].id && h 151 net/sched/cls_route.c iif == head->fastmap[h].iif && h 152 net/sched/cls_route.c (f = head->fastmap[h].filter) != NULL) { h 160 net/sched/cls_route.c h = route4_hash_to(id); h 163 net/sched/cls_route.c if ((b = head->table[h]) != NULL) { h 176 net/sched/cls_route.c if (h < 256) { h 177 net/sched/cls_route.c h = 256; h 199 net/sched/cls_route.c u32 h = id&0xFF; h 201 net/sched/cls_route.c h += 256; h 202 net/sched/cls_route.c return h; h 291 net/sched/cls_route.c unsigned h = 0; h 298 net/sched/cls_route.c h = f->handle; h 301 net/sched/cls_route.c for (fp = &b->ht[from_hash(h>>16)]; *fp; fp = &(*fp)->next) { h 318 net/sched/cls_route.c head->table[to_hash(h)] = NULL; h 435 net/sched/cls_route.c unsigned int h, th; h 482 net/sched/cls_route.c h = from_hash(f->handle >> 16); h 483 net/sched/cls_route.c for (fp = &f->bkt->ht[h]; (f1=*fp) != NULL; fp = &f1->next) h 493 net/sched/cls_route.c h = from_hash(old_handle >> 16); h 495 net/sched/cls_route.c for (fp = &b->ht[h]; *fp; fp = &(*fp)->next) { h 517 net/sched/cls_route.c unsigned h, h1; h 525 net/sched/cls_route.c for (h = 0; h <= 256; h++) { h 526 net/sched/cls_route.c struct route4_bucket *b = head->table[h]; h 105 net/sched/cls_rsvp.h unsigned h = (__force __u32)dst[RSVP_DST_LEN-1]; h 106 net/sched/cls_rsvp.h h ^= h>>16; h 107 net/sched/cls_rsvp.h h ^= h>>8; h 108 net/sched/cls_rsvp.h return (h ^ protocol ^ tunnelid) & 0xFF; h 113 net/sched/cls_rsvp.h unsigned h = (__force __u32)src[RSVP_DST_LEN-1]; h 114 net/sched/cls_rsvp.h h ^= h>>16; h 115 net/sched/cls_rsvp.h h ^= h>>8; h 116 net/sched/cls_rsvp.h h ^= h>>4; h 117 net/sched/cls_rsvp.h return h & 0xF; h 293 net/sched/cls_rsvp.h unsigned h = f->handle; h 298 net/sched/cls_rsvp.h for (fp = &s->ht[(h>>8)&0xFF]; *fp; fp = &(*fp)->next) { h 312 net/sched/cls_rsvp.h for (sp = &((struct rsvp_head*)tp->root)->ht[h&0xFF]; h 336 net/sched/cls_rsvp.h u32 h; h 339 net/sched/cls_rsvp.h h = data->hgenerator|salt; h 340 net/sched/cls_rsvp.h if (rsvp_get(tp, h) == 0) h 341 net/sched/cls_rsvp.h return h; h 560 net/sched/cls_rsvp.h unsigned h, h1; h 565 net/sched/cls_rsvp.h for (h = 0; h < 256; h++) { h 568 net/sched/cls_rsvp.h for (s = head->ht[h]; s; s = s->next) { h 49 net/sched/cls_tcindex.c struct tcindex_filter **h; /* imperfect hash; only used if !perfect; h 77 net/sched/cls_tcindex.c else if (p->h) { h 78 net/sched/cls_tcindex.c for (f = p->h[key % p->hash]; f; f = f->next) h 166 net/sched/cls_tcindex.c for (walk = p->h+i; *walk; walk = &(*walk)->next) h 245 net/sched/cls_tcindex.c } else if (cp.h && cp.hash != cp.alloc_hash) h 262 net/sched/cls_tcindex.c if (!cp.perfect && !cp.h) h 276 net/sched/cls_tcindex.c if (!cp.perfect && !cp.h) { h 283 net/sched/cls_tcindex.c cp.h = kcalloc(cp.hash, sizeof(f), GFP_KERNEL); h 284 net/sched/cls_tcindex.c if (!cp.h) h 321 net/sched/cls_tcindex.c for (fp = p->h+(handle % p->hash); *fp; fp = &(*fp)->next) h 333 net/sched/cls_tcindex.c kfree(cp.h); h 386 net/sched/cls_tcindex.c if (!p->h) h 389 net/sched/cls_tcindex.c for (f = p->h[i]; f; f = next) { h 422 net/sched/cls_tcindex.c kfree(p->h); h 438 net/sched/cls_tcindex.c pr_debug("p->perfect %p p->h %p\n", p->perfect, p->h); h 460 net/sched/cls_tcindex.c for (f = p->h[i]; !t->tcm_handle && f; h 91 net/sched/cls_u32.c unsigned h = ntohl(key & sel->hmask)>>fshift; h 93 net/sched/cls_u32.c return h; h 355 net/sched/cls_u32.c unsigned h; h 357 net/sched/cls_u32.c for (h=0; h<=ht->divisor; h++) { h 358 net/sched/cls_u32.c while ((n = ht->ht[h]) != NULL) { h 359 net/sched/cls_u32.c ht->ht[h] = n->next; h 658 net/sched/cls_u32.c unsigned h; h 673 net/sched/cls_u32.c for (h = 0; h <= ht->divisor; h++) { h 674 net/sched/cls_u32.c for (n = ht->ht[h]; n; n = n->next) { h 502 net/sched/sch_api.c struct hlist_head *h; h 505 net/sched/sch_api.c h = kmalloc(size, GFP_KERNEL); h 507 net/sched/sch_api.c h = (struct hlist_head *) h 510 net/sched/sch_api.c if (h != NULL) { h 512 net/sched/sch_api.c INIT_HLIST_HEAD(&h[i]); h 514 net/sched/sch_api.c return h; h 522 net/sched/sch_api.c kfree(h); h 524 net/sched/sch_api.c free_pages((unsigned long)h, get_order(size)); h 533 net/sched/sch_api.c unsigned int i, h; h 550 net/sched/sch_api.c h = qdisc_class_hash(cl->classid, nmask); h 551 net/sched/sch_api.c hlist_add_head(&cl->hnode, &nhash[h]); h 586 net/sched/sch_api.c unsigned int h; h 589 net/sched/sch_api.c h = qdisc_class_hash(cl->classid, clhash->hashmask); h 590 net/sched/sch_api.c hlist_add_head(&cl->hnode, &clhash->hash[h]); h 1076 net/sched/sch_cbq.c unsigned int h; h 1081 net/sched/sch_cbq.c for (h = 0; h < q->clhash.hashsize; h++) { h 1082 net/sched/sch_cbq.c hlist_for_each_entry(cl, n, &q->clhash.hash[h], common.hnode) { h 1102 net/sched/sch_cbq.c unsigned h; h 1119 net/sched/sch_cbq.c for (h = 0; h < q->clhash.hashsize; h++) { h 1123 net/sched/sch_cbq.c hlist_for_each_entry(c, n, &q->clhash.hash[h], h 1243 net/sched/sch_cbq.c unsigned h; h 1258 net/sched/sch_cbq.c for (h = 0; h < q->clhash.hashsize; h++) { h 1259 net/sched/sch_cbq.c hlist_for_each_entry(cl, n, &q->clhash.hash[h], common.hnode) { h 1729 net/sched/sch_cbq.c unsigned h; h 1739 net/sched/sch_cbq.c for (h = 0; h < q->clhash.hashsize; h++) { h 1740 net/sched/sch_cbq.c hlist_for_each_entry(cl, n, &q->clhash.hash[h], common.hnode) h 1743 net/sched/sch_cbq.c for (h = 0; h < q->clhash.hashsize; h++) { h 1744 net/sched/sch_cbq.c hlist_for_each_entry_safe(cl, n, next, &q->clhash.hash[h], h 2025 net/sched/sch_cbq.c unsigned h; h 2030 net/sched/sch_cbq.c for (h = 0; h < q->clhash.hashsize; h++) { h 2031 net/sched/sch_cbq.c hlist_for_each_entry(cl, n, &q->clhash.hash[h], common.hnode) { h 114 net/sched/sch_sfq.c return jhash_2words(h, h1, q->perturbation) & (SFQ_HASH_DIVISOR - 1); h 119 net/sched/sch_sfq.c u32 h, h2; h 125 net/sched/sch_sfq.c h = iph->daddr; h 140 net/sched/sch_sfq.c h = iph->daddr.s6_addr32[3]; h 152 net/sched/sch_sfq.c h = (unsigned long)skb->dst ^ skb->protocol; h 156 net/sched/sch_sfq.c return sfq_fold_hash(q, h, h2); h 72 net/sunrpc/auth_gss/svcauth_gss.c struct cache_head h; h 93 net/sunrpc/auth_gss/svcauth_gss.c struct rsi *rsii = container_of(ref, struct rsi, h.ref); h 106 net/sunrpc/auth_gss/svcauth_gss.c struct rsi *item = container_of(a, struct rsi, h); h 107 net/sunrpc/auth_gss/svcauth_gss.c struct rsi *tmp = container_of(b, struct rsi, h); h 128 net/sunrpc/auth_gss/svcauth_gss.c struct rsi *new = container_of(cnew, struct rsi, h); h 129 net/sunrpc/auth_gss/svcauth_gss.c struct rsi *item = container_of(citem, struct rsi, h); h 147 net/sunrpc/auth_gss/svcauth_gss.c struct rsi *new = container_of(cnew, struct rsi, h); h 148 net/sunrpc/auth_gss/svcauth_gss.c struct rsi *item = container_of(citem, struct rsi, h); h 168 net/sunrpc/auth_gss/svcauth_gss.c return &rsii->h; h 177 net/sunrpc/auth_gss/svcauth_gss.c struct rsi *rsii = container_of(h, struct rsi, h); h 218 net/sunrpc/auth_gss/svcauth_gss.c rsii.h.flags = 0; h 255 net/sunrpc/auth_gss/svcauth_gss.c rsii.h.expiry_time = expiry; h 261 net/sunrpc/auth_gss/svcauth_gss.c cache_put(&rsip->h, &rsi_cache); h 286 net/sunrpc/auth_gss/svcauth_gss.c ch = sunrpc_cache_lookup(&rsi_cache, &item->h, hash); h 288 net/sunrpc/auth_gss/svcauth_gss.c return container_of(ch, struct rsi, h); h 298 net/sunrpc/auth_gss/svcauth_gss.c ch = sunrpc_cache_update(&rsi_cache, &new->h, h 299 net/sunrpc/auth_gss/svcauth_gss.c &old->h, hash); h 301 net/sunrpc/auth_gss/svcauth_gss.c return container_of(ch, struct rsi, h); h 330 net/sunrpc/auth_gss/svcauth_gss.c struct cache_head h; h 353 net/sunrpc/auth_gss/svcauth_gss.c struct rsc *rsci = container_of(ref, struct rsc, h.ref); h 368 net/sunrpc/auth_gss/svcauth_gss.c struct rsc *new = container_of(a, struct rsc, h); h 369 net/sunrpc/auth_gss/svcauth_gss.c struct rsc *tmp = container_of(b, struct rsc, h); h 377 net/sunrpc/auth_gss/svcauth_gss.c struct rsc *new = container_of(cnew, struct rsc, h); h 378 net/sunrpc/auth_gss/svcauth_gss.c struct rsc *tmp = container_of(ctmp, struct rsc, h); h 391 net/sunrpc/auth_gss/svcauth_gss.c struct rsc *new = container_of(cnew, struct rsc, h); h 392 net/sunrpc/auth_gss/svcauth_gss.c struct rsc *tmp = container_of(ctmp, struct rsc, h); h 407 net/sunrpc/auth_gss/svcauth_gss.c return &rsci->h; h 431 net/sunrpc/auth_gss/svcauth_gss.c rsci.h.flags = 0; h 447 net/sunrpc/auth_gss/svcauth_gss.c set_bit(CACHE_NEGATIVE, &rsci.h.flags); h 490 net/sunrpc/auth_gss/svcauth_gss.c rsci.h.expiry_time = expiry; h 497 net/sunrpc/auth_gss/svcauth_gss.c cache_put(&rscp->h, &rsc_cache); h 521 net/sunrpc/auth_gss/svcauth_gss.c ch = sunrpc_cache_lookup(&rsc_cache, &item->h, hash); h 523 net/sunrpc/auth_gss/svcauth_gss.c return container_of(ch, struct rsc, h); h 533 net/sunrpc/auth_gss/svcauth_gss.c ch = sunrpc_cache_update(&rsc_cache, &new->h, h 534 net/sunrpc/auth_gss/svcauth_gss.c &old->h, hash); h 536 net/sunrpc/auth_gss/svcauth_gss.c return container_of(ch, struct rsc, h); h 555 net/sunrpc/auth_gss/svcauth_gss.c if (cache_check(&rsc_cache, &found->h, NULL)) h 725 net/sunrpc/auth_gss/svcauth_gss.c struct auth_domain h; h 744 net/sunrpc/auth_gss/svcauth_gss.c struct gss_domain *gd = container_of(dom, struct gss_domain, h); h 761 net/sunrpc/auth_gss/svcauth_gss.c kref_init(&new->h.ref); h 762 net/sunrpc/auth_gss/svcauth_gss.c new->h.name = kstrdup(name, GFP_KERNEL); h 763 net/sunrpc/auth_gss/svcauth_gss.c if (!new->h.name) h 765 net/sunrpc/auth_gss/svcauth_gss.c new->h.flavour = &svcauthops_gss; h 769 net/sunrpc/auth_gss/svcauth_gss.c test = auth_domain_lookup(name, &new->h); h 770 net/sunrpc/auth_gss/svcauth_gss.c if (test != &new->h) { /* Duplicate registration */ h 772 net/sunrpc/auth_gss/svcauth_gss.c kfree(new->h.name); h 956 net/sunrpc/auth_gss/svcauth_gss.c cache_put(&rsci->h, &rsc_cache); h 1007 net/sunrpc/auth_gss/svcauth_gss.c switch (cache_check(&rsi_cache, &rsip->h, &rqstp->rq_chandle)) { h 1033 net/sunrpc/auth_gss/svcauth_gss.c cache_put(&rsip->h, &rsi_cache); h 1129 net/sunrpc/auth_gss/svcauth_gss.c rsci->h.expiry_time = get_seconds(); h 1130 net/sunrpc/auth_gss/svcauth_gss.c set_bit(CACHE_NEGATIVE, &rsci->h.flags); h 1166 net/sunrpc/auth_gss/svcauth_gss.c cache_get(&rsci->h); h 1187 net/sunrpc/auth_gss/svcauth_gss.c cache_put(&rsci->h, &rsc_cache); h 1365 net/sunrpc/auth_gss/svcauth_gss.c cache_put(&gsd->rsci->h, &rsc_cache); h 1374 net/sunrpc/auth_gss/svcauth_gss.c struct gss_domain *gd = container_of(dom, struct gss_domain, h); h 43 net/sunrpc/cache.c h->next = NULL; h 44 net/sunrpc/cache.c h->flags = 0; h 45 net/sunrpc/cache.c kref_init(&h->ref); h 46 net/sunrpc/cache.c h->expiry_time = now + CACHE_NEW_EXPIRY; h 47 net/sunrpc/cache.c h->last_refresh = now; h 198 net/sunrpc/cache.c if (!test_bit(CACHE_VALID, &h->flags) || h 199 net/sunrpc/cache.c h->expiry_time < get_seconds()) h 201 net/sunrpc/cache.c else if (detail->flush_time > h->last_refresh) h 205 net/sunrpc/cache.c if (test_bit(CACHE_NEGATIVE, &h->flags)) h 211 net/sunrpc/cache.c refresh_age = (h->expiry_time - h->last_refresh); h 212 net/sunrpc/cache.c age = get_seconds() - h->last_refresh; h 220 net/sunrpc/cache.c if (!test_and_set_bit(CACHE_PENDING, &h->flags)) { h 221 net/sunrpc/cache.c switch (cache_make_upcall(detail, h)) { h 223 net/sunrpc/cache.c clear_bit(CACHE_PENDING, &h->flags); h 225 net/sunrpc/cache.c set_bit(CACHE_NEGATIVE, &h->flags); h 226 net/sunrpc/cache.c cache_fresh_unlocked(h, detail, h 227 net/sunrpc/cache.c cache_fresh_locked(h, get_seconds()+CACHE_NEW_EXPIRY)); h 233 net/sunrpc/cache.c clear_bit(CACHE_PENDING, &h->flags); h 234 net/sunrpc/cache.c cache_revisit_request(h); h 241 net/sunrpc/cache.c if (cache_defer_req(rqstp, h) != 0) h 245 net/sunrpc/cache.c cache_put(h, detail); h 1064 net/sunrpc/cache.c detail->cache_request(detail, h, &bp, &len); h 1072 net/sunrpc/cache.c crq->item = cache_get(h); h 28 net/sunrpc/svcauth_unix.c struct auth_domain h; h 43 net/sunrpc/svcauth_unix.c if (new && rv != &new->h) h 44 net/sunrpc/svcauth_unix.c auth_domain_put(&new->h); h 56 net/sunrpc/svcauth_unix.c kref_init(&new->h.ref); h 57 net/sunrpc/svcauth_unix.c new->h.name = kstrdup(name, GFP_KERNEL); h 58 net/sunrpc/svcauth_unix.c if (new->h.name == NULL) { h 62 net/sunrpc/svcauth_unix.c new->h.flavour = &svcauth_unix; h 64 net/sunrpc/svcauth_unix.c rv = auth_domain_lookup(name, &new->h); h 71 net/sunrpc/svcauth_unix.c struct unix_domain *ud = container_of(dom, struct unix_domain, h); h 87 net/sunrpc/svcauth_unix.c struct cache_head h; h 98 net/sunrpc/svcauth_unix.c struct ip_map *im = container_of(item, struct ip_map,h); h 102 net/sunrpc/svcauth_unix.c auth_domain_put(&im->m_client->h); h 126 net/sunrpc/svcauth_unix.c struct ip_map *orig = container_of(corig, struct ip_map, h); h 127 net/sunrpc/svcauth_unix.c struct ip_map *new = container_of(cnew, struct ip_map, h); h 133 net/sunrpc/svcauth_unix.c struct ip_map *new = container_of(cnew, struct ip_map, h); h 134 net/sunrpc/svcauth_unix.c struct ip_map *item = container_of(citem, struct ip_map, h); h 141 net/sunrpc/svcauth_unix.c struct ip_map *new = container_of(cnew, struct ip_map, h); h 142 net/sunrpc/svcauth_unix.c struct ip_map *item = container_of(citem, struct ip_map, h); h 144 net/sunrpc/svcauth_unix.c kref_get(&item->m_client->h.ref); h 152 net/sunrpc/svcauth_unix.c return &i->h; h 162 net/sunrpc/svcauth_unix.c struct ip_map *im = container_of(h, struct ip_map, h); h 248 net/sunrpc/svcauth_unix.c container_of(dom, struct unix_domain, h), h 268 net/sunrpc/svcauth_unix.c if (h == NULL) { h 272 net/sunrpc/svcauth_unix.c im = container_of(h, struct ip_map, h); h 276 net/sunrpc/svcauth_unix.c if (test_bit(CACHE_VALID, &h->flags) && h 277 net/sunrpc/svcauth_unix.c !test_bit(CACHE_NEGATIVE, &h->flags)) h 278 net/sunrpc/svcauth_unix.c dom = im->m_client->h.name; h 318 net/sunrpc/svcauth_unix.c ch = sunrpc_cache_lookup(&ip_map_cache, &ip.h, h 323 net/sunrpc/svcauth_unix.c return container_of(ch, struct ip_map, h); h 334 net/sunrpc/svcauth_unix.c ip.h.flags = 0; h 336 net/sunrpc/svcauth_unix.c set_bit(CACHE_NEGATIVE, &ip.h.flags); h 345 net/sunrpc/svcauth_unix.c ip.h.expiry_time = expiry; h 347 net/sunrpc/svcauth_unix.c &ip.h, &ipm->h, h 363 net/sunrpc/svcauth_unix.c udom = container_of(dom, struct unix_domain, h); h 379 net/sunrpc/svcauth_unix.c udom = container_of(dom, struct unix_domain, h); h 394 net/sunrpc/svcauth_unix.c if (cache_check(&ip_map_cache, &ipm->h, NULL)) h 398 net/sunrpc/svcauth_unix.c if (test_and_set_bit(CACHE_NEGATIVE, &ipm->h.flags) == 0) h 399 net/sunrpc/svcauth_unix.c auth_domain_put(&ipm->m_client->h); h 402 net/sunrpc/svcauth_unix.c rv = &ipm->m_client->h; h 405 net/sunrpc/svcauth_unix.c cache_put(&ipm->h, &ip_map_cache); h 426 net/sunrpc/svcauth_unix.c if (!cache_valid(&ipm->h)) { h 434 net/sunrpc/svcauth_unix.c cache_put(&ipm->h, &ip_map_cache); h 437 net/sunrpc/svcauth_unix.c cache_get(&ipm->h); h 459 net/sunrpc/svcauth_unix.c cache_put(&ipm->h, &ip_map_cache); h 466 net/sunrpc/svcauth_unix.c cache_put(&ipm->h, &ip_map_cache); h 479 net/sunrpc/svcauth_unix.c struct cache_head h; h 488 net/sunrpc/svcauth_unix.c struct unix_gid *ug = container_of(item, struct unix_gid, h); h 497 net/sunrpc/svcauth_unix.c struct unix_gid *orig = container_of(corig, struct unix_gid, h); h 498 net/sunrpc/svcauth_unix.c struct unix_gid *new = container_of(cnew, struct unix_gid, h); h 503 net/sunrpc/svcauth_unix.c struct unix_gid *new = container_of(cnew, struct unix_gid, h); h 504 net/sunrpc/svcauth_unix.c struct unix_gid *item = container_of(citem, struct unix_gid, h); h 509 net/sunrpc/svcauth_unix.c struct unix_gid *new = container_of(cnew, struct unix_gid, h); h 510 net/sunrpc/svcauth_unix.c struct unix_gid *item = container_of(citem, struct unix_gid, h); h 519 net/sunrpc/svcauth_unix.c return &g->h; h 529 net/sunrpc/svcauth_unix.c struct unix_gid *ug = container_of(h, struct unix_gid, h); h 584 net/sunrpc/svcauth_unix.c ug.h.flags = 0; h 585 net/sunrpc/svcauth_unix.c ug.h.expiry_time = expiry; h 587 net/sunrpc/svcauth_unix.c &ug.h, &ugp->h, h 611 net/sunrpc/svcauth_unix.c if (h == NULL) { h 615 net/sunrpc/svcauth_unix.c ug = container_of(h, struct unix_gid, h); h 616 net/sunrpc/svcauth_unix.c if (test_bit(CACHE_VALID, &h->flags) && h 617 net/sunrpc/svcauth_unix.c !test_bit(CACHE_NEGATIVE, &h->flags)) h 650 net/sunrpc/svcauth_unix.c ch = sunrpc_cache_lookup(&unix_gid_cache, &ug.h, h 653 net/sunrpc/svcauth_unix.c return container_of(ch, struct unix_gid, h); h 664 net/sunrpc/svcauth_unix.c switch (cache_check(&unix_gid_cache, &ug->h, &rqstp->rq_chandle)) { h 710 net/sunrpc/svcauth_unix.c switch (cache_check(&ip_map_cache, &ipm->h, &rqstp->rq_chandle)) { h 719 net/sunrpc/svcauth_unix.c rqstp->rq_client = &ipm->m_client->h; h 32 net/xfrm/xfrm_hash.h unsigned int h = family ^ reqid; h 35 net/xfrm/xfrm_hash.h h ^= __xfrm4_daddr_saddr_hash(daddr, saddr); h 38 net/xfrm/xfrm_hash.h h ^= __xfrm6_daddr_saddr_hash(daddr, saddr); h 41 net/xfrm/xfrm_hash.h return (h ^ (h >> 16)) & hmask; h 49 net/xfrm/xfrm_hash.h unsigned int h = family; h 52 net/xfrm/xfrm_hash.h h ^= __xfrm4_daddr_saddr_hash(daddr, saddr); h 55 net/xfrm/xfrm_hash.h h ^= __xfrm6_daddr_saddr_hash(daddr, saddr); h 58 net/xfrm/xfrm_hash.h return (h ^ (h >> 16)) & hmask; h 65 net/xfrm/xfrm_hash.h unsigned int h = (__force u32)spi ^ proto; h 68 net/xfrm/xfrm_hash.h h ^= __xfrm4_addr_hash(daddr); h 71 net/xfrm/xfrm_hash.h h ^= __xfrm6_addr_hash(daddr); h 74 net/xfrm/xfrm_hash.h return (h ^ (h >> 10) ^ (h >> 20)) & hmask; h 86 net/xfrm/xfrm_hash.h unsigned int h = 0; h 94 net/xfrm/xfrm_hash.h h = __xfrm4_daddr_saddr_hash(daddr, saddr); h 102 net/xfrm/xfrm_hash.h h = __xfrm6_daddr_saddr_hash(daddr, saddr); h 105 net/xfrm/xfrm_hash.h h ^= (h >> 16); h 106 net/xfrm/xfrm_hash.h return h & hmask; h 111 net/xfrm/xfrm_hash.h unsigned int h = 0; h 115 net/xfrm/xfrm_hash.h h = __xfrm4_daddr_saddr_hash(daddr, saddr); h 119 net/xfrm/xfrm_hash.h h = __xfrm6_daddr_saddr_hash(daddr, saddr); h 122 net/xfrm/xfrm_hash.h h ^= (h >> 16); h 123 net/xfrm/xfrm_hash.h return h & hmask; h 369 net/xfrm/xfrm_policy.c unsigned int h; h 371 net/xfrm/xfrm_policy.c h = __addr_hash(&pol->selector.daddr, &pol->selector.saddr, h 375 net/xfrm/xfrm_policy.c hlist_add_head(&pol->bydst, ndsttable+h); h 376 net/xfrm/xfrm_policy.c h0 = h; h 378 net/xfrm/xfrm_policy.c if (h != h0) h 399 net/xfrm/xfrm_policy.c unsigned int h; h 401 net/xfrm/xfrm_policy.c h = __idx_hash(pol->index, nhashmask); h 402 net/xfrm/xfrm_policy.c hlist_add_head(&pol->byidx, nidxtable+h); h 103 net/xfrm/xfrm_state.c unsigned int h; h 105 net/xfrm/xfrm_state.c h = __xfrm_dst_hash(&x->id.daddr, &x->props.saddr, h 108 net/xfrm/xfrm_state.c hlist_add_head(&x->bydst, ndsttable+h); h 110 net/xfrm/xfrm_state.c h = __xfrm_src_hash(&x->id.daddr, &x->props.saddr, h 113 net/xfrm/xfrm_state.c hlist_add_head(&x->bysrc, nsrctable+h); h 116 net/xfrm/xfrm_state.c h = __xfrm_spi_hash(&x->id.daddr, x->id.spi, h 119 net/xfrm/xfrm_state.c hlist_add_head(&x->byspi, nspitable+h); h 686 net/xfrm/xfrm_state.c unsigned int h = xfrm_spi_hash(daddr, spi, proto, family); h 690 net/xfrm/xfrm_state.c hlist_for_each_entry(x, entry, xfrm_state_byspi+h, byspi) { h 718 net/xfrm/xfrm_state.c unsigned int h = xfrm_src_hash(daddr, saddr, family); h 722 net/xfrm/xfrm_state.c hlist_for_each_entry(x, entry, xfrm_state_bysrc+h, bysrc) { h 777 net/xfrm/xfrm_state.c unsigned int h; h 787 net/xfrm/xfrm_state.c h = xfrm_dst_hash(daddr, saddr, tmpl->reqid, family); h 788 net/xfrm/xfrm_state.c hlist_for_each_entry(x, entry, xfrm_state_bydst+h, bydst) { h 858 net/xfrm/xfrm_state.c hlist_add_head(&x->bydst, xfrm_state_bydst+h); h 859 net/xfrm/xfrm_state.c h = xfrm_src_hash(daddr, saddr, family); h 860 net/xfrm/xfrm_state.c hlist_add_head(&x->bysrc, xfrm_state_bysrc+h); h 862 net/xfrm/xfrm_state.c h = xfrm_spi_hash(&x->id.daddr, x->id.spi, x->id.proto, family); h 863 net/xfrm/xfrm_state.c hlist_add_head(&x->byspi, xfrm_state_byspi+h); h 892 net/xfrm/xfrm_state.c unsigned int h; h 897 net/xfrm/xfrm_state.c h = xfrm_dst_hash(daddr, saddr, reqid, family); h 898 net/xfrm/xfrm_state.c hlist_for_each_entry(x, entry, xfrm_state_bydst+h, bydst) { h 922 net/xfrm/xfrm_state.c unsigned int h; h 928 net/xfrm/xfrm_state.c h = xfrm_dst_hash(&x->id.daddr, &x->props.saddr, h 930 net/xfrm/xfrm_state.c hlist_add_head(&x->bydst, xfrm_state_bydst+h); h 932 net/xfrm/xfrm_state.c h = xfrm_src_hash(&x->id.daddr, &x->props.saddr, x->props.family); h 933 net/xfrm/xfrm_state.c hlist_add_head(&x->bysrc, xfrm_state_bysrc+h); h 936 net/xfrm/xfrm_state.c h = xfrm_spi_hash(&x->id.daddr, x->id.spi, x->id.proto, h 939 net/xfrm/xfrm_state.c hlist_add_head(&x->byspi, xfrm_state_byspi+h); h 960 net/xfrm/xfrm_state.c unsigned int h; h 962 net/xfrm/xfrm_state.c h = xfrm_dst_hash(&xnew->id.daddr, &xnew->props.saddr, reqid, family); h 963 net/xfrm/xfrm_state.c hlist_for_each_entry(x, entry, xfrm_state_bydst+h, bydst) { h 984 net/xfrm/xfrm_state.c unsigned int h = xfrm_dst_hash(daddr, saddr, reqid, family); h 988 net/xfrm/xfrm_state.c hlist_for_each_entry(x, entry, xfrm_state_bydst+h, bydst) { h 1056 net/xfrm/xfrm_state.c hlist_add_head(&x->bydst, xfrm_state_bydst+h); h 1057 net/xfrm/xfrm_state.c h = xfrm_src_hash(daddr, saddr, family); h 1058 net/xfrm/xfrm_state.c hlist_add_head(&x->bysrc, xfrm_state_bysrc+h); h 1204 net/xfrm/xfrm_state.c unsigned int h; h 1209 net/xfrm/xfrm_state.c h = xfrm_dst_hash(&m->old_daddr, &m->old_saddr, h 1211 net/xfrm/xfrm_state.c hlist_for_each_entry(x, entry, xfrm_state_bydst+h, bydst) { h 1226 net/xfrm/xfrm_state.c h = xfrm_src_hash(&m->old_daddr, &m->old_saddr, h 1228 net/xfrm/xfrm_state.c hlist_for_each_entry(x, entry, xfrm_state_bysrc+h, bysrc) { h 1497 net/xfrm/xfrm_state.c unsigned int h; h 1522 net/xfrm/xfrm_state.c for (h=0; h<high-low+1; h++) { h 1534 net/xfrm/xfrm_state.c h = xfrm_spi_hash(&x->id.daddr, x->id.spi, x->id.proto, x->props.family); h 1535 net/xfrm/xfrm_state.c hlist_add_head(&x->byspi, xfrm_state_byspi+h); h 149 scripts/genksyms/genksyms.c unsigned long h = crc32(name) % HASH_BUCKETS; h 152 scripts/genksyms/genksyms.c for (sym = symtab[h]; sym; sym = sym->hash_next) h 163 scripts/genksyms/genksyms.c unsigned long h = crc32(name) % HASH_BUCKETS; h 166 scripts/genksyms/genksyms.c for (sym = symtab[h]; sym; sym = sym->hash_next) { h 183 scripts/genksyms/genksyms.c sym->hash_next = symtab[h]; h 184 scripts/genksyms/genksyms.c symtab[h] = sym; h 572 scripts/kconfig/gconf.c gint w, h; h 575 scripts/kconfig/gconf.c gtk_window_get_default_size(GTK_WINDOW(main_wnd), &w, &h); h 577 scripts/kconfig/gconf.c gdk_window_get_size(widget->window, &w, &h); h 579 scripts/kconfig/gconf.c if (h == old_h) h 581 scripts/kconfig/gconf.c old_h = h; h 583 scripts/kconfig/gconf.c gtk_paned_set_position(GTK_PANED(vpaned), 2 * h / 3); h 832 scripts/kconfig/gconf.c gint w, h; h 835 scripts/kconfig/gconf.c gtk_window_get_default_size(GTK_WINDOW(main_wnd), &w, &h); h 63 scripts/kconfig/lxdialog/util.c dlg.dialog.hl = (h); \ h 49 security/selinux/ss/avtab.c newnode->next = h->htable[hvalue]; h 50 security/selinux/ss/avtab.c h->htable[hvalue] = newnode; h 53 security/selinux/ss/avtab.c h->nel++; h 63 security/selinux/ss/avtab.c if (!h || !h->htable) h 66 security/selinux/ss/avtab.c hvalue = avtab_hash(key, h->mask); h 67 security/selinux/ss/avtab.c for (prev = NULL, cur = h->htable[hvalue]; h 86 security/selinux/ss/avtab.c newnode = avtab_insert_node(h, hvalue, prev, cur, key, datum); h 104 security/selinux/ss/avtab.c if (!h || !h->htable) h 106 security/selinux/ss/avtab.c hvalue = avtab_hash(key, h->mask); h 107 security/selinux/ss/avtab.c for (prev = NULL, cur = h->htable[hvalue]; h 125 security/selinux/ss/avtab.c return avtab_insert_node(h, hvalue, prev, cur, key, datum); h 134 security/selinux/ss/avtab.c if (!h || !h->htable) h 137 security/selinux/ss/avtab.c hvalue = avtab_hash(key, h->mask); h 138 security/selinux/ss/avtab.c for (cur = h->htable[hvalue]; cur; cur = cur->next) { h 169 security/selinux/ss/avtab.c if (!h || !h->htable) h 172 security/selinux/ss/avtab.c hvalue = avtab_hash(key, h->mask); h 173 security/selinux/ss/avtab.c for (cur = h->htable[hvalue]; cur; cur = cur->next) { h 227 security/selinux/ss/avtab.c if (!h || !h->htable) h 230 security/selinux/ss/avtab.c for (i = 0; i < h->nslot; i++) { h 231 security/selinux/ss/avtab.c cur = h->htable[i]; h 237 security/selinux/ss/avtab.c h->htable[i] = NULL; h 239 security/selinux/ss/avtab.c kfree(h->htable); h 240 security/selinux/ss/avtab.c h->htable = NULL; h 241 security/selinux/ss/avtab.c h->nslot = 0; h 242 security/selinux/ss/avtab.c h->mask = 0; h 247 security/selinux/ss/avtab.c h->htable = NULL; h 248 security/selinux/ss/avtab.c h->nel = 0; h 273 security/selinux/ss/avtab.c h->htable = kcalloc(nslot, sizeof(*(h->htable)), GFP_KERNEL); h 274 security/selinux/ss/avtab.c if (!h->htable) h 278 security/selinux/ss/avtab.c h->nel = 0; h 279 security/selinux/ss/avtab.c h->nslot = nslot; h 280 security/selinux/ss/avtab.c h->mask = mask; h 282 security/selinux/ss/avtab.c h->nslot, nrules); h 295 security/selinux/ss/avtab.c for (i = 0; i < h->nslot; i++) { h 296 security/selinux/ss/avtab.c cur = h->htable[i]; h 313 security/selinux/ss/avtab.c tag, h->nel, slots_used, h->nslot, max_chain_len, h 243 security/selinux/ss/conditional.c if (hashtab_insert(h, key, booldatum)) h 43 security/selinux/ss/hashtab.c if (!h || h->nel == HASHTAB_MAX_NODES) h 46 security/selinux/ss/hashtab.c hvalue = h->hash_value(h, key); h 48 security/selinux/ss/hashtab.c cur = h->htable[hvalue]; h 49 security/selinux/ss/hashtab.c while (cur && h->keycmp(h, key, cur->key) > 0) { h 54 security/selinux/ss/hashtab.c if (cur && (h->keycmp(h, key, cur->key) == 0)) h 66 security/selinux/ss/hashtab.c newnode->next = h->htable[hvalue]; h 67 security/selinux/ss/hashtab.c h->htable[hvalue] = newnode; h 70 security/selinux/ss/hashtab.c h->nel++; h 79 security/selinux/ss/hashtab.c if (!h) h 82 security/selinux/ss/hashtab.c hvalue = h->hash_value(h, key); h 83 security/selinux/ss/hashtab.c cur = h->htable[hvalue]; h 84 security/selinux/ss/hashtab.c while (cur && h->keycmp(h, key, cur->key) > 0) h 87 security/selinux/ss/hashtab.c if (cur == NULL || (h->keycmp(h, key, cur->key) != 0)) h 98 security/selinux/ss/hashtab.c if (!h) h 101 security/selinux/ss/hashtab.c for (i = 0; i < h->size; i++) { h 102 security/selinux/ss/hashtab.c cur = h->htable[i]; h 108 security/selinux/ss/hashtab.c h->htable[i] = NULL; h 111 security/selinux/ss/hashtab.c kfree(h->htable); h 112 security/selinux/ss/hashtab.c h->htable = NULL; h 114 security/selinux/ss/hashtab.c kfree(h); h 125 security/selinux/ss/hashtab.c if (!h) h 128 security/selinux/ss/hashtab.c for (i = 0; i < h->size; i++) { h 129 security/selinux/ss/hashtab.c cur = h->htable[i]; h 148 security/selinux/ss/hashtab.c for (slots_used = max_chain_len = i = 0; i < h->size; i++) { h 149 security/selinux/ss/hashtab.c cur = h->htable[i]; h 25 security/selinux/ss/hashtab.h u32 (*hash_value)(struct hashtab *h, const void *key); h 27 security/selinux/ss/hashtab.h int (*keycmp)(struct hashtab *h, const void *key1, const void *key2); h 402 security/selinux/ss/policydb.c struct hashtab *h = s[i].table; h 405 security/selinux/ss/policydb.c hashtab_stat(h, &info); h 407 security/selinux/ss/policydb.c "longest chain length %d\n", symtab_name[i], h->nel, h 408 security/selinux/ss/policydb.c info.slots_used, h->size, info.max_chain_len); h 959 security/selinux/ss/policydb.c rc = hashtab_insert(h, key, perdatum); h 1012 security/selinux/ss/policydb.c rc = hashtab_insert(h, key, comdatum); h 1185 security/selinux/ss/policydb.c rc = hashtab_insert(h, key, cladatum); h 1252 security/selinux/ss/policydb.c rc = hashtab_insert(h, key, role); h 1308 security/selinux/ss/policydb.c rc = hashtab_insert(h, key, typdatum); h 1398 security/selinux/ss/policydb.c rc = hashtab_insert(h, key, usrdatum); h 1449 security/selinux/ss/policydb.c rc = hashtab_insert(h, key, levdatum); h 1491 security/selinux/ss/policydb.c rc = hashtab_insert(h, key, catdatum); h 1502 security/selinux/ss/policydb.c static int (*read_f[SYM_NUM]) (struct policydb *p, struct hashtab *h, void *fp) = h 216 security/selinux/ss/sidtab.c cur = h->htable[i]; h 231 security/selinux/ss/sidtab.c "chain length %d\n", tag, h->nel, slots_used, SIDTAB_SIZE, h 23 security/selinux/ss/symtab.c return val & (h->size - 1); h 55 sound/drivers/vx/vx_mixer.c u16 h; h 67 sound/drivers/vx/vx_mixer.c u16 h; h 110 sound/parisc/harmony.c return __raw_readl(h->iobase + r); h 116 sound/parisc/harmony.c __raw_writel(v, h->iobase + r); h 122 sound/parisc/harmony.c while (harmony_read(h, HARMONY_CNTL) & HARMONY_CNTL_C) ; h 128 sound/parisc/harmony.c harmony_write(h, HARMONY_RESET, 1); h 130 sound/parisc/harmony.c harmony_write(h, HARMONY_RESET, 0); h 137 sound/parisc/harmony.c harmony_wait_for_control(h); h 138 sound/parisc/harmony.c dstatus = harmony_read(h, HARMONY_DSTATUS); h 140 sound/parisc/harmony.c harmony_write(h, HARMONY_DSTATUS, dstatus); h 147 sound/parisc/harmony.c harmony_wait_for_control(h); h 148 sound/parisc/harmony.c dstatus = harmony_read(h, HARMONY_DSTATUS); h 150 sound/parisc/harmony.c harmony_write(h, HARMONY_DSTATUS, dstatus); h 158 sound/parisc/harmony.c spin_lock_irqsave(&h->mixer_lock, flags); h 159 sound/parisc/harmony.c harmony_wait_for_control(h); h 160 sound/parisc/harmony.c harmony_write(h, HARMONY_GAINCTL, HARMONY_GAIN_SILENCE); h 161 sound/parisc/harmony.c spin_unlock_irqrestore(&h->mixer_lock, flags); h 169 sound/parisc/harmony.c spin_lock_irqsave(&h->mixer_lock, flags); h 170 sound/parisc/harmony.c harmony_wait_for_control(h); h 171 sound/parisc/harmony.c harmony_write(h, HARMONY_GAINCTL, h->st.gain); h 172 sound/parisc/harmony.c spin_unlock_irqrestore(&h->mixer_lock, flags); h 181 sound/parisc/harmony.c spin_lock_irqsave(&h->lock, flags); h 184 sound/parisc/harmony.c (h->st.format << 6) | h 185 sound/parisc/harmony.c (h->st.stereo << 5) | h 186 sound/parisc/harmony.c (h->st.rate)); h 188 sound/parisc/harmony.c harmony_wait_for_control(h); h 189 sound/parisc/harmony.c harmony_write(h, HARMONY_CNTL, ctrl); h 191 sound/parisc/harmony.c spin_unlock_irqrestore(&h->lock, flags); h 198 sound/parisc/harmony.c struct snd_harmony *h = dev; h 200 sound/parisc/harmony.c spin_lock(&h->lock); h 201 sound/parisc/harmony.c harmony_disable_interrupts(h); h 202 sound/parisc/harmony.c harmony_wait_for_control(h); h 203 sound/parisc/harmony.c dstatus = harmony_read(h, HARMONY_DSTATUS); h 204 sound/parisc/harmony.c spin_unlock(&h->lock); h 207 sound/parisc/harmony.c if (h->psubs && h->st.playing) { h 208 sound/parisc/harmony.c spin_lock(&h->lock); h 209 sound/parisc/harmony.c h->pbuf.buf += h->pbuf.count; /* PAGE_SIZE */ h 210 sound/parisc/harmony.c h->pbuf.buf %= h->pbuf.size; /* MAX_BUFS*PAGE_SIZE */ h 212 sound/parisc/harmony.c harmony_write(h, HARMONY_PNXTADD, h 213 sound/parisc/harmony.c h->pbuf.addr + h->pbuf.buf); h 214 sound/parisc/harmony.c h->stats.play_intr++; h 215 sound/parisc/harmony.c spin_unlock(&h->lock); h 216 sound/parisc/harmony.c snd_pcm_period_elapsed(h->psubs); h 218 sound/parisc/harmony.c spin_lock(&h->lock); h 219 sound/parisc/harmony.c harmony_write(h, HARMONY_PNXTADD, h->sdma.addr); h 220 sound/parisc/harmony.c h->stats.silence_intr++; h 221 sound/parisc/harmony.c spin_unlock(&h->lock); h 226 sound/parisc/harmony.c if (h->csubs && h->st.capturing) { h 227 sound/parisc/harmony.c spin_lock(&h->lock); h 228 sound/parisc/harmony.c h->cbuf.buf += h->cbuf.count; h 229 sound/parisc/harmony.c h->cbuf.buf %= h->cbuf.size; h 231 sound/parisc/harmony.c harmony_write(h, HARMONY_RNXTADD, h 232 sound/parisc/harmony.c h->cbuf.addr + h->cbuf.buf); h 233 sound/parisc/harmony.c h->stats.rec_intr++; h 234 sound/parisc/harmony.c spin_unlock(&h->lock); h 235 sound/parisc/harmony.c snd_pcm_period_elapsed(h->csubs); h 237 sound/parisc/harmony.c spin_lock(&h->lock); h 238 sound/parisc/harmony.c harmony_write(h, HARMONY_RNXTADD, h->gdma.addr); h 239 sound/parisc/harmony.c h->stats.graveyard_intr++; h 240 sound/parisc/harmony.c spin_unlock(&h->lock); h 244 sound/parisc/harmony.c spin_lock(&h->lock); h 245 sound/parisc/harmony.c harmony_enable_interrupts(h); h 246 sound/parisc/harmony.c spin_unlock(&h->lock); h 308 sound/parisc/harmony.c struct snd_harmony *h = snd_pcm_substream_chip(ss); h 310 sound/parisc/harmony.c if (h->st.capturing) h 313 sound/parisc/harmony.c spin_lock(&h->lock); h 316 sound/parisc/harmony.c h->st.playing = 1; h 317 sound/parisc/harmony.c harmony_write(h, HARMONY_PNXTADD, h->pbuf.addr); h 318 sound/parisc/harmony.c harmony_write(h, HARMONY_RNXTADD, h->gdma.addr); h 319 sound/parisc/harmony.c harmony_unmute(h); h 320 sound/parisc/harmony.c harmony_enable_interrupts(h); h 323 sound/parisc/harmony.c h->st.playing = 0; h 324 sound/parisc/harmony.c harmony_mute(h); h 325 sound/parisc/harmony.c harmony_write(h, HARMONY_PNXTADD, h->sdma.addr); h 326 sound/parisc/harmony.c harmony_disable_interrupts(h); h 332 sound/parisc/harmony.c spin_unlock(&h->lock); h 336 sound/parisc/harmony.c spin_unlock(&h->lock); h 344 sound/parisc/harmony.c struct snd_harmony *h = snd_pcm_substream_chip(ss); h 346 sound/parisc/harmony.c if (h->st.playing) h 349 sound/parisc/harmony.c spin_lock(&h->lock); h 352 sound/parisc/harmony.c h->st.capturing = 1; h 353 sound/parisc/harmony.c harmony_write(h, HARMONY_PNXTADD, h->sdma.addr); h 354 sound/parisc/harmony.c harmony_write(h, HARMONY_RNXTADD, h->cbuf.addr); h 355 sound/parisc/harmony.c harmony_unmute(h); h 356 sound/parisc/harmony.c harmony_enable_interrupts(h); h 359 sound/parisc/harmony.c h->st.capturing = 0; h 360 sound/parisc/harmony.c harmony_mute(h); h 361 sound/parisc/harmony.c harmony_write(h, HARMONY_RNXTADD, h->gdma.addr); h 362 sound/parisc/harmony.c harmony_disable_interrupts(h); h 368 sound/parisc/harmony.c spin_unlock(&h->lock); h 372 sound/parisc/harmony.c spin_unlock(&h->lock); h 380 sound/parisc/harmony.c int o = h->st.format; h 399 sound/parisc/harmony.c snd_pcm_format_set_silence(fmt, h->sdma.area, SILENCE_BUFSZ / h 410 sound/parisc/harmony.c struct snd_harmony *h = snd_pcm_substream_chip(ss); h 413 sound/parisc/harmony.c if (h->st.capturing) h 416 sound/parisc/harmony.c h->pbuf.size = snd_pcm_lib_buffer_bytes(ss); h 417 sound/parisc/harmony.c h->pbuf.count = snd_pcm_lib_period_bytes(ss); h 418 sound/parisc/harmony.c if (h->pbuf.buf >= h->pbuf.size) h 419 sound/parisc/harmony.c h->pbuf.buf = 0; h 420 sound/parisc/harmony.c h->st.playing = 0; h 422 sound/parisc/harmony.c h->st.rate = snd_harmony_rate_bits(rt->rate); h 423 sound/parisc/harmony.c h->st.format = snd_harmony_set_data_format(h, rt->format, 0); h 426 sound/parisc/harmony.c h->st.stereo = HARMONY_SS_STEREO; h 428 sound/parisc/harmony.c h->st.stereo = HARMONY_SS_MONO; h 430 sound/parisc/harmony.c harmony_set_control(h); h 432 sound/parisc/harmony.c h->pbuf.addr = rt->dma_addr; h 440 sound/parisc/harmony.c struct snd_harmony *h = snd_pcm_substream_chip(ss); h 443 sound/parisc/harmony.c if (h->st.playing) h 446 sound/parisc/harmony.c h->cbuf.size = snd_pcm_lib_buffer_bytes(ss); h 447 sound/parisc/harmony.c h->cbuf.count = snd_pcm_lib_period_bytes(ss); h 448 sound/parisc/harmony.c if (h->cbuf.buf >= h->cbuf.size) h 449 sound/parisc/harmony.c h->cbuf.buf = 0; h 450 sound/parisc/harmony.c h->st.capturing = 0; h 452 sound/parisc/harmony.c h->st.rate = snd_harmony_rate_bits(rt->rate); h 453 sound/parisc/harmony.c h->st.format = snd_harmony_set_data_format(h, rt->format, 0); h 456 sound/parisc/harmony.c h->st.stereo = HARMONY_SS_STEREO; h 458 sound/parisc/harmony.c h->st.stereo = HARMONY_SS_MONO; h 460 sound/parisc/harmony.c harmony_set_control(h); h 462 sound/parisc/harmony.c h->cbuf.addr = rt->dma_addr; h 471 sound/parisc/harmony.c struct snd_harmony *h = snd_pcm_substream_chip(ss); h 475 sound/parisc/harmony.c if (!(h->st.playing) || (h->psubs == NULL)) h 478 sound/parisc/harmony.c if ((h->pbuf.addr == 0) || (h->pbuf.size == 0)) h 481 sound/parisc/harmony.c pcuradd = harmony_read(h, HARMONY_PCURADD); h 482 sound/parisc/harmony.c played = pcuradd - h->pbuf.addr; h 486 sound/parisc/harmony.c pcuradd, h->pbuf.addr, played); h 489 sound/parisc/harmony.c if (pcuradd > h->pbuf.addr + h->pbuf.size) { h 500 sound/parisc/harmony.c struct snd_harmony *h = snd_pcm_substream_chip(ss); h 504 sound/parisc/harmony.c if (!(h->st.capturing) || (h->csubs == NULL)) h 507 sound/parisc/harmony.c if ((h->cbuf.addr == 0) || (h->cbuf.size == 0)) h 510 sound/parisc/harmony.c rcuradd = harmony_read(h, HARMONY_RCURADD); h 511 sound/parisc/harmony.c caught = rcuradd - h->cbuf.addr; h 515 sound/parisc/harmony.c rcuradd, h->cbuf.addr, caught); h 518 sound/parisc/harmony.c if (rcuradd > h->cbuf.addr + h->cbuf.size) { h 528 sound/parisc/harmony.c struct snd_harmony *h = snd_pcm_substream_chip(ss); h 532 sound/parisc/harmony.c h->psubs = ss; h 547 sound/parisc/harmony.c struct snd_harmony *h = snd_pcm_substream_chip(ss); h 551 sound/parisc/harmony.c h->csubs = ss; h 566 sound/parisc/harmony.c struct snd_harmony *h = snd_pcm_substream_chip(ss); h 567 sound/parisc/harmony.c h->psubs = NULL; h 574 sound/parisc/harmony.c struct snd_harmony *h = snd_pcm_substream_chip(ss); h 575 sound/parisc/harmony.c h->csubs = NULL; h 584 sound/parisc/harmony.c struct snd_harmony *h = snd_pcm_substream_chip(ss); h 587 sound/parisc/harmony.c if (err > 0 && h->dma.type == SNDRV_DMA_TYPE_CONTINUOUS) h 627 sound/parisc/harmony.c harmony_disable_interrupts(h); h 629 sound/parisc/harmony.c err = snd_pcm_new(h->card, "harmony", 0, 1, 1, &pcm); h 638 sound/parisc/harmony.c pcm->private_data = h; h 641 sound/parisc/harmony.c h->pcm = pcm; h 643 sound/parisc/harmony.c h->psubs = NULL; h 644 sound/parisc/harmony.c h->csubs = NULL; h 647 sound/parisc/harmony.c h->dma.type = SNDRV_DMA_TYPE_DEV; h 648 sound/parisc/harmony.c h->dma.dev = &h->dev->dev; h 649 sound/parisc/harmony.c err = snd_dma_alloc_pages(h->dma.type, h 650 sound/parisc/harmony.c h->dma.dev, h 652 sound/parisc/harmony.c &h->gdma); h 659 sound/parisc/harmony.c err = snd_dma_alloc_pages(h->dma.type, h 660 sound/parisc/harmony.c h->dma.dev, h 662 sound/parisc/harmony.c &h->sdma); h 669 sound/parisc/harmony.c err = snd_pcm_lib_preallocate_pages_for_all(pcm, h->dma.type, h 670 sound/parisc/harmony.c h->dma.dev, h 678 sound/parisc/harmony.c h->st.format = snd_harmony_set_data_format(h, h 687 sound/parisc/harmony.c harmony_wait_for_control(h); h 688 sound/parisc/harmony.c harmony_write(h, HARMONY_GAINCTL, h->st.gain); h 712 sound/parisc/harmony.c struct snd_harmony *h = snd_kcontrol_chip(kc); h 719 sound/parisc/harmony.c spin_lock_irq(&h->mixer_lock); h 721 sound/parisc/harmony.c left = (h->st.gain >> shift_left) & mask; h 722 sound/parisc/harmony.c right = (h->st.gain >> shift_right) & mask; h 732 sound/parisc/harmony.c spin_unlock_irq(&h->mixer_lock); h 741 sound/parisc/harmony.c struct snd_harmony *h = snd_kcontrol_chip(kc); h 747 sound/parisc/harmony.c int old_gain = h->st.gain; h 749 sound/parisc/harmony.c spin_lock_irq(&h->mixer_lock); h 754 sound/parisc/harmony.c h->st.gain &= ~( (mask << shift_left ) ); h 755 sound/parisc/harmony.c h->st.gain |= (left << shift_left); h 761 sound/parisc/harmony.c h->st.gain &= ~( (mask << shift_right) ); h 762 sound/parisc/harmony.c h->st.gain |= (right << shift_right); h 765 sound/parisc/harmony.c snd_harmony_set_new_gain(h); h 767 sound/parisc/harmony.c spin_unlock_irq(&h->mixer_lock); h 769 sound/parisc/harmony.c return h->st.gain != old_gain; h 791 sound/parisc/harmony.c struct snd_harmony *h = snd_kcontrol_chip(kc); h 794 sound/parisc/harmony.c spin_lock_irq(&h->mixer_lock); h 796 sound/parisc/harmony.c value = (h->st.gain >> HARMONY_GAIN_IS_SHIFT) & 1; h 799 sound/parisc/harmony.c spin_unlock_irq(&h->mixer_lock); h 808 sound/parisc/harmony.c struct snd_harmony *h = snd_kcontrol_chip(kc); h 810 sound/parisc/harmony.c int old_gain = h->st.gain; h 812 sound/parisc/harmony.c spin_lock_irq(&h->mixer_lock); h 815 sound/parisc/harmony.c h->st.gain &= ~HARMONY_GAIN_IS_MASK; h 816 sound/parisc/harmony.c h->st.gain |= value << HARMONY_GAIN_IS_SHIFT; h 818 sound/parisc/harmony.c snd_harmony_set_new_gain(h); h 820 sound/parisc/harmony.c spin_unlock_irq(&h->mixer_lock); h 822 sound/parisc/harmony.c return h->st.gain != old_gain; h 859 sound/parisc/harmony.c harmony_mute(h); h 860 sound/parisc/harmony.c harmony_reset(h); h 861 sound/parisc/harmony.c h->st.gain = HARMONY_GAIN_DEFAULT; h 862 sound/parisc/harmony.c harmony_unmute(h); h 868 sound/parisc/harmony.c struct snd_card *card = h->card; h 871 sound/parisc/harmony.c if (snd_BUG_ON(!h)) h 877 sound/parisc/harmony.c snd_ctl_new1(&snd_harmony_controls[idx], h)); h 882 sound/parisc/harmony.c snd_harmony_mixer_reset(h); h 890 sound/parisc/harmony.c if (h->gdma.addr) h 891 sound/parisc/harmony.c snd_dma_free_pages(&h->gdma); h 892 sound/parisc/harmony.c if (h->sdma.addr) h 893 sound/parisc/harmony.c snd_dma_free_pages(&h->sdma); h 895 sound/parisc/harmony.c if (h->irq >= 0) h 896 sound/parisc/harmony.c free_irq(h->irq, h); h 898 sound/parisc/harmony.c if (h->iobase) h 899 sound/parisc/harmony.c iounmap(h->iobase); h 901 sound/parisc/harmony.c parisc_set_drvdata(h->dev, NULL); h 903 sound/parisc/harmony.c kfree(h); h 910 sound/parisc/harmony.c struct snd_harmony *h = dev->device_data; h 911 sound/parisc/harmony.c return snd_harmony_free(h); h 920 sound/parisc/harmony.c struct snd_harmony *h; h 927 sound/parisc/harmony.c h = kzalloc(sizeof(*h), GFP_KERNEL); h 928 sound/parisc/harmony.c if (h == NULL) h 931 sound/parisc/harmony.c h->hpa = padev->hpa.start; h 932 sound/parisc/harmony.c h->card = card; h 933 sound/parisc/harmony.c h->dev = padev; h 934 sound/parisc/harmony.c h->irq = -1; h 935 sound/parisc/harmony.c h->iobase = ioremap_nocache(padev->hpa.start, HARMONY_SIZE); h 936 sound/parisc/harmony.c if (h->iobase == NULL) { h 944 sound/parisc/harmony.c "harmony", h); h 950 sound/parisc/harmony.c h->irq = padev->irq; h 952 sound/parisc/harmony.c spin_lock_init(&h->mixer_lock); h 953 sound/parisc/harmony.c spin_lock_init(&h->lock); h 956 sound/parisc/harmony.c h, &ops)) < 0) { h 962 sound/parisc/harmony.c *rchip = h; h 967 sound/parisc/harmony.c snd_harmony_free(h); h 976 sound/parisc/harmony.c struct snd_harmony *h; h 982 sound/parisc/harmony.c err = snd_harmony_create(card, padev, &h); h 986 sound/parisc/harmony.c err = snd_harmony_pcm_init(h); h 990 sound/parisc/harmony.c err = snd_harmony_mixer_init(h); h 997 sound/parisc/harmony.c card->shortname, h->hpa, h->irq); h 263 sound/pci/riptide/riptide.c #define SEND_LSEL(p,b,c,d,e,f,g,h) sendcmd(p,PARM,LSEL|BYTE1(b)|BYTE2(c)|BYTE3(d),BYTE0(e)|BYTE1(f)|BYTE2(g)|BYTE3(h),RET(0)) /* select paths for internal connections */