cpu 254 arch/x86/boot/boot.h extern struct cpu_features cpu; cpu 31 arch/x86/boot/cpucheck.c struct cpu_features cpu; cpu 115 arch/x86/boot/cpucheck.c set_bit(X86_FEATURE_FPU, cpu.flags); cpu 129 arch/x86/boot/cpucheck.c "=c" (cpu.flags[4]), cpu 130 arch/x86/boot/cpucheck.c "=d" (cpu.flags[0]) cpu 133 arch/x86/boot/cpucheck.c cpu.level = (tfms >> 8) & 15; cpu 134 arch/x86/boot/cpucheck.c cpu.model = (tfms >> 4) & 15; cpu 135 arch/x86/boot/cpucheck.c if (cpu.level >= 6) cpu 136 arch/x86/boot/cpucheck.c cpu.model += ((tfms >> 16) & 0xf) << 4; cpu 149 arch/x86/boot/cpucheck.c "=c" (cpu.flags[6]), cpu 150 arch/x86/boot/cpucheck.c "=d" (cpu.flags[1]) cpu 164 arch/x86/boot/cpucheck.c err_flags[i] = req_flags[i] & ~cpu.flags[i]; cpu 184 arch/x86/boot/cpucheck.c memset(&cpu.flags, 0, sizeof cpu.flags); cpu 185 arch/x86/boot/cpucheck.c cpu.level = 3; cpu 188 arch/x86/boot/cpucheck.c cpu.level = 4; cpu 193 arch/x86/boot/cpucheck.c if (test_bit(X86_FEATURE_LM, cpu.flags)) cpu 194 arch/x86/boot/cpucheck.c cpu.level = 64; cpu 214 arch/x86/boot/cpucheck.c is_centaur() && cpu.model >= 6) { cpu 225 arch/x86/boot/cpucheck.c set_bit(X86_FEATURE_CX8, cpu.flags); cpu 237 arch/x86/boot/cpucheck.c : "+a" (level), "=d" (cpu.flags[0]) cpu 247 arch/x86/boot/cpucheck.c *cpu_level_ptr = cpu.level; cpu 251 arch/x86/boot/cpucheck.c return (cpu.level < req_level || err) ? -1 : 0; cpu 78 arch/x86/boot/main.c if (cpu.level < 6) cpu 542 arch/x86/kernel/acpi/boot.c int cpu; cpu 584 arch/x86/kernel/acpi/boot.c cpu = first_cpu(new_map); cpu 586 arch/x86/kernel/acpi/boot.c *pcpu = cpu; cpu 599 arch/x86/kernel/acpi/boot.c per_cpu(x86_cpu_to_apicid, cpu) = -1; cpu 600 arch/x86/kernel/acpi/boot.c cpu_clear(cpu, cpu_present_map); cpu 30 arch/x86/kernel/acpi/cstate.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu 73 arch/x86/kernel/acpi/cstate.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu 88 arch/x86/kernel/acpi/cstate.c percpu_entry = per_cpu_ptr(cpu_cstate_entry, cpu); cpu 94 arch/x86/kernel/acpi/cstate.c retval = set_cpus_allowed_ptr(current, &cpumask_of_cpu(cpu)); cpu 138 arch/x86/kernel/acpi/cstate.c unsigned int cpu = smp_processor_id(); cpu 141 arch/x86/kernel/acpi/cstate.c percpu_entry = per_cpu_ptr(cpu_cstate_entry, cpu); cpu 665 arch/x86/kernel/apic_32.c int cpu = smp_processor_id(); cpu 666 arch/x86/kernel/apic_32.c struct clock_event_device *evt = &per_cpu(lapic_events, cpu); cpu 681 arch/x86/kernel/apic_32.c "Spurious LAPIC timer interrupt on cpu %d\n", cpu); cpu 693 arch/x86/kernel/apic_32.c per_cpu(irq_stat, cpu).apic_timer_irqs++; cpu 1492 arch/x86/kernel/apic_32.c int cpu; cpu 1514 arch/x86/kernel/apic_32.c cpu = first_cpu(tmp_map); cpu 1523 arch/x86/kernel/apic_32.c cpu = 0; cpu 1556 arch/x86/kernel/apic_32.c cpu_to_apicid[cpu] = apicid; cpu 1557 arch/x86/kernel/apic_32.c bios_cpu_apicid[cpu] = apicid; cpu 1559 arch/x86/kernel/apic_32.c per_cpu(x86_cpu_to_apicid, cpu) = apicid; cpu 1560 arch/x86/kernel/apic_32.c per_cpu(x86_bios_cpu_apicid, cpu) = apicid; cpu 1564 arch/x86/kernel/apic_32.c cpu_set(cpu, cpu_possible_map); cpu 1565 arch/x86/kernel/apic_32.c cpu_set(cpu, cpu_present_map); cpu 548 arch/x86/kernel/apic_64.c int cpu = smp_processor_id(); cpu 549 arch/x86/kernel/apic_64.c struct clock_event_device *evt = &per_cpu(lapic_events, cpu); cpu 564 arch/x86/kernel/apic_64.c "Spurious LAPIC timer interrupt on cpu %d\n", cpu); cpu 576 arch/x86/kernel/apic_64.c per_cpu(irq_stat, cpu).apic_timer_irqs++; cpu 1430 arch/x86/kernel/apic_64.c int cpu; cpu 1452 arch/x86/kernel/apic_64.c cpu = first_cpu(tmp_map); cpu 1461 arch/x86/kernel/apic_64.c cpu = 0; cpu 1494 arch/x86/kernel/apic_64.c cpu_to_apicid[cpu] = apicid; cpu 1495 arch/x86/kernel/apic_64.c bios_cpu_apicid[cpu] = apicid; cpu 1497 arch/x86/kernel/apic_64.c per_cpu(x86_cpu_to_apicid, cpu) = apicid; cpu 1498 arch/x86/kernel/apic_64.c per_cpu(x86_bios_cpu_apicid, cpu) = apicid; cpu 1502 arch/x86/kernel/apic_64.c cpu_set(cpu, cpu_possible_map); cpu 1503 arch/x86/kernel/apic_64.c cpu_set(cpu, cpu_present_map); cpu 601 arch/x86/kernel/apm_32.c int cpu; cpu 607 arch/x86/kernel/apm_32.c cpu = get_cpu(); cpu 608 arch/x86/kernel/apm_32.c gdt = get_cpu_gdt_table(cpu); cpu 644 arch/x86/kernel/apm_32.c int cpu; cpu 650 arch/x86/kernel/apm_32.c cpu = get_cpu(); cpu 651 arch/x86/kernel/apm_32.c gdt = get_cpu_gdt_table(cpu); cpu 61 arch/x86/kernel/asm-offsets_32.c OFFSET(TI_cpu, thread_info, cpu); cpu 223 arch/x86/kernel/cpu/amd.c int cpu = smp_processor_id(); cpu 250 arch/x86/kernel/cpu/amd.c numa_set_node(cpu, node); cpu 252 arch/x86/kernel/cpu/amd.c printk(KERN_INFO "CPU %d/%x -> Node %d\n", cpu, apicid, node); cpu 864 arch/x86/kernel/cpu/common.c struct x8664_pda *pda = cpu_pda(cpu); cpu 874 arch/x86/kernel/cpu/common.c pda->cpunumber = cpu; cpu 881 arch/x86/kernel/cpu/common.c if (cpu == 0) { cpu 892 arch/x86/kernel/cpu/common.c cpu); cpu 896 arch/x86/kernel/cpu/common.c if (pda->nodenumber == 0 && cpu_to_node(cpu) != NUMA_NO_NODE) cpu 897 arch/x86/kernel/cpu/common.c pda->nodenumber = cpu_to_node(cpu); cpu 956 arch/x86/kernel/cpu/common.c int cpu = stack_smp_processor_id(); cpu 957 arch/x86/kernel/cpu/common.c struct tss_struct *t = &per_cpu(init_tss, cpu); cpu 958 arch/x86/kernel/cpu/common.c struct orig_ist *orig_ist = &per_cpu(orig_ist, cpu); cpu 965 arch/x86/kernel/cpu/common.c if (cpu != 0) cpu 966 arch/x86/kernel/cpu/common.c pda_init(cpu); cpu 972 arch/x86/kernel/cpu/common.c if (cpu_test_and_set(cpu, cpu_initialized)) cpu 973 arch/x86/kernel/cpu/common.c panic("CPU#%d already initialized!\n", cpu); cpu 975 arch/x86/kernel/cpu/common.c printk(KERN_INFO "Initializing CPU#%d\n", cpu); cpu 995 arch/x86/kernel/cpu/common.c if (cpu != 0 && x2apic) cpu 1007 arch/x86/kernel/cpu/common.c if (cpu) { cpu 1011 arch/x86/kernel/cpu/common.c "stack %ld %d\n", v, cpu); cpu 1034 arch/x86/kernel/cpu/common.c set_tss_desc(cpu, t); cpu 1076 arch/x86/kernel/cpu/common.c int cpu = smp_processor_id(); cpu 1078 arch/x86/kernel/cpu/common.c struct tss_struct *t = &per_cpu(init_tss, cpu); cpu 1081 arch/x86/kernel/cpu/common.c if (cpu_test_and_set(cpu, cpu_initialized)) { cpu 1082 arch/x86/kernel/cpu/common.c printk(KERN_WARNING "CPU#%d already initialized!\n", cpu); cpu 1086 arch/x86/kernel/cpu/common.c printk(KERN_INFO "Initializing CPU#%d\n", cpu); cpu 1104 arch/x86/kernel/cpu/common.c set_tss_desc(cpu, t); cpu 1110 arch/x86/kernel/cpu/common.c __set_tss_desc(cpu, GDT_ENTRY_DOUBLEFAULT_TSS, &doublefault_tss); cpu 81 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c struct cpuinfo_x86 *cpu = &cpu_data(cpuid); cpu 83 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c if (cpu->x86_vendor != X86_VENDOR_INTEL || cpu 84 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c !cpu_has(cpu, X86_FEATURE_EST)) cpu 275 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c set_cpus_allowed_ptr(current, &cpumask_of_cpu(cpu)); cpu 276 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c if (get_cpu() != cpu) { cpu 330 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c retval = per_cpu(drv_data, policy->cpu)->max_freq * perf_percent / 100; cpu 335 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c dprintk("cpu %d: performance percent %d\n", cpu, perf_percent); cpu 341 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c struct acpi_cpufreq_data *data = per_cpu(drv_data, cpu); cpu 345 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c dprintk("get_cur_freq_on_cpu (%d)\n", cpu); cpu 353 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c freq = extract_freq(get_cur_val(&cpumask_of_cpu(cpu)), data); cpu 385 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c struct acpi_cpufreq_data *data = per_cpu(drv_data, policy->cpu); cpu 395 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c dprintk("acpi_cpufreq_target %d (%d)\n", target_freq, policy->cpu); cpu 451 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c cpu_set(policy->cpu, cmd.mask); cpu 456 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c freqs.cpu = i; cpu 465 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c policy->cpu); cpu 471 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c freqs.cpu = i; cpu 481 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c struct acpi_cpufreq_data *data = per_cpu(drv_data, policy->cpu); cpu 572 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c unsigned int cpu = policy->cpu; cpu 575 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c struct cpuinfo_x86 *c = &cpu_data(policy->cpu); cpu 584 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c data->acpi_data = percpu_ptr(acpi_perf_data, cpu); cpu 585 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c per_cpu(drv_data, cpu) = data; cpu 590 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c result = acpi_processor_register_performance(data->acpi_data, cpu); cpu 611 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c policy->cpus = per_cpu(cpu_core_map, cpu); cpu 634 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c if (!check_est_cpu(cpu)) { cpu 685 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c policy->cur = acpi_cpufreq_guess_freq(data, policy->cpu); cpu 689 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c policy->cur = get_cur_freq_on_cpu(cpu); cpu 706 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c dprintk("CPU%u - ACPI performance management activated.\n", cpu); cpu 714 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c cpufreq_frequency_table_get_attr(data->freq_table, policy->cpu); cpu 727 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c acpi_processor_unregister_performance(perf, cpu); cpu 730 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c per_cpu(drv_data, cpu) = NULL; cpu 737 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c struct acpi_cpufreq_data *data = per_cpu(drv_data, policy->cpu); cpu 742 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c cpufreq_frequency_table_put_attr(policy->cpu); cpu 743 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c per_cpu(drv_data, policy->cpu) = NULL; cpu 745 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c policy->cpu); cpu 754 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c struct acpi_cpufreq_data *data = per_cpu(drv_data, policy->cpu); cpu 238 arch/x86/kernel/cpu/cpufreq/cpufreq-nforce2.c if (cpu) cpu 263 arch/x86/kernel/cpu/cpufreq/cpufreq-nforce2.c freqs.old = nforce2_get(policy->cpu); cpu 265 arch/x86/kernel/cpu/cpufreq/cpufreq-nforce2.c freqs.cpu = 0; /* Only one CPU on nForce2 platforms */ cpu 318 arch/x86/kernel/cpu/cpufreq/cpufreq-nforce2.c if (policy->cpu != 0) cpu 365 arch/x86/kernel/cpu/cpufreq/cpufreq-nforce2.c policy->cur = nforce2_get(policy->cpu); cpu 41 arch/x86/kernel/cpu/cpufreq/e_powersaver.c if (cpu) cpu 43 arch/x86/kernel/cpu/cpufreq/e_powersaver.c centaur = eps_cpu[cpu]; cpu 61 arch/x86/kernel/cpu/cpufreq/e_powersaver.c freqs.old = eps_get(cpu); cpu 63 arch/x86/kernel/cpu/cpufreq/e_powersaver.c freqs.cpu = cpu; cpu 121 arch/x86/kernel/cpu/cpufreq/e_powersaver.c unsigned int cpu = policy->cpu; cpu 125 arch/x86/kernel/cpu/cpufreq/e_powersaver.c if (unlikely(eps_cpu[cpu] == NULL)) cpu 127 arch/x86/kernel/cpu/cpufreq/e_powersaver.c centaur = eps_cpu[cpu]; cpu 130 arch/x86/kernel/cpu/cpufreq/e_powersaver.c &eps_cpu[cpu]->freq_table[0], cpu 139 arch/x86/kernel/cpu/cpufreq/e_powersaver.c ret = eps_set_state(centaur, cpu, dest_state); cpu 148 arch/x86/kernel/cpu/cpufreq/e_powersaver.c &eps_cpu[policy->cpu]->freq_table[0]); cpu 168 arch/x86/kernel/cpu/cpufreq/e_powersaver.c if (policy->cpu != 0) cpu 296 arch/x86/kernel/cpu/cpufreq/e_powersaver.c cpufreq_frequency_table_get_attr(¢aur->freq_table[0], policy->cpu); cpu 302 arch/x86/kernel/cpu/cpufreq/e_powersaver.c unsigned int cpu = policy->cpu; cpu 306 arch/x86/kernel/cpu/cpufreq/e_powersaver.c if (eps_cpu[cpu] == NULL) cpu 308 arch/x86/kernel/cpu/cpufreq/e_powersaver.c centaur = eps_cpu[cpu]; cpu 313 arch/x86/kernel/cpu/cpufreq/e_powersaver.c eps_set_state(centaur, cpu, hi & 0xffff); cpu 315 arch/x86/kernel/cpu/cpufreq/e_powersaver.c cpufreq_frequency_table_put_attr(policy->cpu); cpu 316 arch/x86/kernel/cpu/cpufreq/e_powersaver.c kfree(eps_cpu[cpu]); cpu 317 arch/x86/kernel/cpu/cpufreq/e_powersaver.c eps_cpu[cpu] = NULL; cpu 126 arch/x86/kernel/cpu/cpufreq/elanfreq.c freqs.cpu = 0; /* elanfreq.c is UP only driver */ cpu 229 arch/x86/kernel/cpu/cpufreq/elanfreq.c cpufreq_frequency_table_get_attr(elanfreq_table, policy->cpu); cpu 236 arch/x86/kernel/cpu/cpufreq/elanfreq.c cpufreq_frequency_table_put_attr(policy->cpu); cpu 259 arch/x86/kernel/cpu/cpufreq/gx-suspmod.c freqs.cpu = 0; cpu 336 arch/x86/kernel/cpu/cpufreq/gx-suspmod.c policy->cpu = 0; cpu 376 arch/x86/kernel/cpu/cpufreq/gx-suspmod.c policy->cpu = 0; cpu 397 arch/x86/kernel/cpu/cpufreq/gx-suspmod.c if (!policy || policy->cpu != 0) cpu 415 arch/x86/kernel/cpu/cpufreq/gx-suspmod.c policy->cpu = 0; cpu 273 arch/x86/kernel/cpu/cpufreq/longhaul.c freqs.cpu = 0; /* longhaul.c is UP only driver */ cpu 683 arch/x86/kernel/cpu/cpufreq/longhaul.c if (cpu) cpu 932 arch/x86/kernel/cpu/cpufreq/longhaul.c cpufreq_frequency_table_get_attr(longhaul_table, policy->cpu); cpu 939 arch/x86/kernel/cpu/cpufreq/longhaul.c cpufreq_frequency_table_put_attr(policy->cpu); cpu 63 arch/x86/kernel/cpu/cpufreq/longrun.c policy->cpu = 0; cpu 133 arch/x86/kernel/cpu/cpufreq/longrun.c policy->cpu = 0; cpu 149 arch/x86/kernel/cpu/cpufreq/longrun.c if (cpu) cpu 265 arch/x86/kernel/cpu/cpufreq/longrun.c if (policy->cpu != 0) cpu 61 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c if (!cpu_online(cpu) || (newstate > DC_DISABLE) || (newstate == DC_RESV)) cpu 64 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c rdmsr_on_cpu(cpu, MSR_IA32_THERM_STATUS, &l, &h); cpu 67 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c dprintk("CPU#%d currently thermal throttled\n", cpu); cpu 69 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c if (has_N44_O17_errata[cpu] && (newstate == DC_25PT || newstate == DC_DFLT)) cpu 72 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c rdmsr_on_cpu(cpu, MSR_IA32_THERM_CONTROL, &l, &h); cpu 74 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c dprintk("CPU#%d disabling modulation\n", cpu); cpu 75 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c wrmsr_on_cpu(cpu, MSR_IA32_THERM_CONTROL, l & ~(1<<4), h); cpu 78 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c cpu, ((125 * newstate) / 10)); cpu 86 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c wrmsr_on_cpu(cpu, MSR_IA32_THERM_CONTROL, l, h); cpu 118 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c freqs.old = cpufreq_p4_get(policy->cpu); cpu 126 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c freqs.cpu = i; cpu 138 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c freqs.cpu = i; cpu 198 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c struct cpuinfo_x86 *c = &cpu_data(policy->cpu); cpu 203 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c policy->cpus = per_cpu(cpu_sibling_map, policy->cpu); cpu 213 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c has_N44_O17_errata[policy->cpu] = 1; cpu 224 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c if ((i<2) && (has_N44_O17_errata[policy->cpu])) cpu 229 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c cpufreq_frequency_table_get_attr(p4clockmod_table, policy->cpu); cpu 241 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c cpufreq_frequency_table_put_attr(policy->cpu); cpu 249 arch/x86/kernel/cpu/cpufreq/p4-clockmod.c rdmsr_on_cpu(cpu, MSR_IA32_THERM_CONTROL, &l, &h); cpu 82 arch/x86/kernel/cpu/cpufreq/powernow-k6.c freqs.cpu = 0; /* powernow-k6.c is UP only driver */ cpu 146 arch/x86/kernel/cpu/cpufreq/powernow-k6.c if (policy->cpu != 0) cpu 169 arch/x86/kernel/cpu/cpufreq/powernow-k6.c cpufreq_frequency_table_get_attr(clock_ratio, policy->cpu); cpu 182 arch/x86/kernel/cpu/cpufreq/powernow-k6.c cpufreq_frequency_table_put_attr(policy->cpu); cpu 258 arch/x86/kernel/cpu/cpufreq/powernow-k7.c freqs.cpu = 0; cpu 559 arch/x86/kernel/cpu/cpufreq/powernow-k7.c if (cpu) cpu 598 arch/x86/kernel/cpu/cpufreq/powernow-k7.c if (policy->cpu != 0) cpu 644 arch/x86/kernel/cpu/cpufreq/powernow-k7.c cpufreq_frequency_table_get_attr(powernow_table, policy->cpu); cpu 650 arch/x86/kernel/cpu/cpufreq/powernow-k7.c cpufreq_frequency_table_put_attr(policy->cpu); cpu 483 arch/x86/kernel/cpu/cpufreq/powernow-k8.c set_cpus_allowed_ptr(current, &cpumask_of_cpu(cpu)); cpu 485 arch/x86/kernel/cpu/cpufreq/powernow-k8.c if (smp_processor_id() != cpu) { cpu 486 arch/x86/kernel/cpu/cpufreq/powernow-k8.c printk(KERN_ERR PFX "limiting to cpu %u failed\n", cpu); cpu 642 arch/x86/kernel/cpu/cpufreq/powernow-k8.c if (first_cpu(per_cpu(cpu_core_map, data->cpu)) == data->cpu) cpu 756 arch/x86/kernel/cpu/cpufreq/powernow-k8.c if (acpi_processor_register_performance(&data->acpi_data, data->cpu)) { cpu 796 arch/x86/kernel/cpu/cpufreq/powernow-k8.c if (first_cpu(per_cpu(cpu_core_map, data->cpu)) == data->cpu) cpu 809 arch/x86/kernel/cpu/cpufreq/powernow-k8.c acpi_processor_unregister_performance(&data->acpi_data, data->cpu); cpu 916 arch/x86/kernel/cpu/cpufreq/powernow-k8.c acpi_processor_unregister_performance(&data->acpi_data, data->cpu); cpu 967 arch/x86/kernel/cpu/cpufreq/powernow-k8.c freqs.cpu = i; cpu 975 arch/x86/kernel/cpu/cpufreq/powernow-k8.c freqs.cpu = i; cpu 998 arch/x86/kernel/cpu/cpufreq/powernow-k8.c freqs.cpu = i; cpu 1006 arch/x86/kernel/cpu/cpufreq/powernow-k8.c freqs.cpu = i; cpu 1016 arch/x86/kernel/cpu/cpufreq/powernow-k8.c struct powernow_k8_data *data = per_cpu(powernow_data, pol->cpu); cpu 1030 arch/x86/kernel/cpu/cpufreq/powernow-k8.c set_cpus_allowed_ptr(current, &cpumask_of_cpu(pol->cpu)); cpu 1032 arch/x86/kernel/cpu/cpufreq/powernow-k8.c if (smp_processor_id() != pol->cpu) { cpu 1033 arch/x86/kernel/cpu/cpufreq/powernow-k8.c printk(KERN_ERR PFX "limiting to cpu %u failed\n", pol->cpu); cpu 1043 arch/x86/kernel/cpu/cpufreq/powernow-k8.c pol->cpu, targfreq, pol->min, pol->max, relation); cpu 1092 arch/x86/kernel/cpu/cpufreq/powernow-k8.c struct powernow_k8_data *data = per_cpu(powernow_data, pol->cpu); cpu 1107 arch/x86/kernel/cpu/cpufreq/powernow-k8.c if (!cpu_online(pol->cpu)) cpu 1110 arch/x86/kernel/cpu/cpufreq/powernow-k8.c if (!check_supported_cpu(pol->cpu)) cpu 1119 arch/x86/kernel/cpu/cpufreq/powernow-k8.c data->cpu = pol->cpu; cpu 1141 arch/x86/kernel/cpu/cpufreq/powernow-k8.c if (pol->cpu != 0) { cpu 1156 arch/x86/kernel/cpu/cpufreq/powernow-k8.c set_cpus_allowed_ptr(current, &cpumask_of_cpu(pol->cpu)); cpu 1158 arch/x86/kernel/cpu/cpufreq/powernow-k8.c if (smp_processor_id() != pol->cpu) { cpu 1159 arch/x86/kernel/cpu/cpufreq/powernow-k8.c printk(KERN_ERR PFX "limiting to cpu %u failed\n", pol->cpu); cpu 1178 arch/x86/kernel/cpu/cpufreq/powernow-k8.c pol->cpus = cpumask_of_cpu(pol->cpu); cpu 1180 arch/x86/kernel/cpu/cpufreq/powernow-k8.c pol->cpus = per_cpu(cpu_core_map, pol->cpu); cpu 1203 arch/x86/kernel/cpu/cpufreq/powernow-k8.c cpufreq_frequency_table_get_attr(data->powernow_table, pol->cpu); cpu 1211 arch/x86/kernel/cpu/cpufreq/powernow-k8.c per_cpu(powernow_data, pol->cpu) = data; cpu 1225 arch/x86/kernel/cpu/cpufreq/powernow-k8.c struct powernow_k8_data *data = per_cpu(powernow_data, pol->cpu); cpu 1232 arch/x86/kernel/cpu/cpufreq/powernow-k8.c cpufreq_frequency_table_put_attr(pol->cpu); cpu 1247 arch/x86/kernel/cpu/cpufreq/powernow-k8.c first = first_cpu(per_cpu(cpu_core_map, cpu)); cpu 1253 arch/x86/kernel/cpu/cpufreq/powernow-k8.c set_cpus_allowed_ptr(current, &cpumask_of_cpu(cpu)); cpu 1254 arch/x86/kernel/cpu/cpufreq/powernow-k8.c if (smp_processor_id() != cpu) { cpu 1256 arch/x86/kernel/cpu/cpufreq/powernow-k8.c "limiting to CPU %d failed in powernowk8_get\n", cpu); cpu 9 arch/x86/kernel/cpu/cpufreq/powernow-k8.h unsigned int cpu; cpu 62 arch/x86/kernel/cpu/cpufreq/sc520_freq.c freqs.cpu = 0; /* AMD Elan is UP */ cpu 121 arch/x86/kernel/cpu/cpufreq/sc520_freq.c cpufreq_frequency_table_get_attr(sc520_freq_table, policy->cpu); cpu 129 arch/x86/kernel/cpu/cpufreq/sc520_freq.c cpufreq_frequency_table_put_attr(policy->cpu); cpu 235 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c struct cpuinfo_x86 *cpu = &cpu_data(policy->cpu); cpu 239 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c if (centrino_verify_cpu_id(cpu, model->cpu_id) && cpu 241 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c strcmp(cpu->x86_model_id, model->model_name) == 0)) cpu 248 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c cpu->x86_model_id); cpu 255 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c cpu->x86_model_id); cpu 260 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c per_cpu(centrino_model, policy->cpu) = model; cpu 295 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c if ((per_cpu(centrino_cpu, cpu) == &cpu_ids[CPU_BANIAS]) || cpu 296 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c (per_cpu(centrino_cpu, cpu) == &cpu_ids[CPU_DOTHAN_A1]) || cpu 297 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c (per_cpu(centrino_cpu, cpu) == &cpu_ids[CPU_DOTHAN_B0])) { cpu 302 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c if ((!per_cpu(centrino_model, cpu)) || cpu 303 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c (!per_cpu(centrino_model, cpu)->op_points)) cpu 308 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c per_cpu(centrino_model, cpu)->op_points[i].frequency cpu 311 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c if (msr == per_cpu(centrino_model, cpu)->op_points[i].index) cpu 312 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c return per_cpu(centrino_model, cpu)-> cpu 316 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c return per_cpu(centrino_model, cpu)->op_points[i-1].frequency; cpu 329 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c set_cpus_allowed_ptr(current, &cpumask_of_cpu(cpu)); cpu 330 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c if (smp_processor_id() != cpu) cpu 334 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c clock_freq = extract_clock(l, cpu, 0); cpu 344 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c clock_freq = extract_clock(l, cpu, 1); cpu 354 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c struct cpuinfo_x86 *cpu = &cpu_data(policy->cpu); cpu 361 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c if (cpu->x86_vendor != X86_VENDOR_INTEL || cpu 362 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c !cpu_has(cpu, X86_FEATURE_EST)) cpu 365 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c if (cpu_has(cpu, X86_FEATURE_CONSTANT_TSC)) cpu 368 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c if (policy->cpu != 0) cpu 372 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c if (centrino_verify_cpu_id(cpu, &cpu_ids[i])) cpu 376 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c per_cpu(centrino_cpu, policy->cpu) = &cpu_ids[i]; cpu 378 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c if (!per_cpu(centrino_cpu, policy->cpu)) { cpu 407 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c freq = get_cur_freq(policy->cpu); cpu 415 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c per_cpu(centrino_model, policy->cpu)->op_points); cpu 420 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c per_cpu(centrino_model, policy->cpu)->op_points, policy->cpu); cpu 427 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c unsigned int cpu = policy->cpu; cpu 429 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c if (!per_cpu(centrino_model, cpu)) cpu 432 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c cpufreq_frequency_table_put_attr(cpu); cpu 434 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c per_cpu(centrino_model, cpu) = NULL; cpu 449 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c per_cpu(centrino_model, policy->cpu)->op_points); cpu 473 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c unsigned int msr, oldmsr = 0, h = 0, cpu = policy->cpu; cpu 486 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c if (unlikely(per_cpu(centrino_model, cpu) == NULL)) { cpu 492 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c per_cpu(centrino_model, cpu)->op_points, cpu 534 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c msr = per_cpu(centrino_model, cpu)->op_points[newstate].index; cpu 545 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c freqs.old = extract_clock(oldmsr, cpu, 0); cpu 546 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c freqs.new = extract_clock(msr, cpu, 0); cpu 552 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c freqs.cpu = k; cpu 575 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c freqs.cpu = k; cpu 598 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c freqs.cpu = j; cpu 649 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c struct cpuinfo_x86 *cpu = &cpu_data(0); cpu 651 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c if (!cpu_has(cpu, X86_FEATURE_EST)) cpu 247 arch/x86/kernel/cpu/cpufreq/speedstep-ich.c return _speedstep_get(&cpumask_of_cpu(cpu)); cpu 272 arch/x86/kernel/cpu/cpufreq/speedstep-ich.c freqs.cpu = policy->cpu; cpu 283 arch/x86/kernel/cpu/cpufreq/speedstep-ich.c freqs.cpu = i; cpu 296 arch/x86/kernel/cpu/cpufreq/speedstep-ich.c freqs.cpu = i; cpu 325 arch/x86/kernel/cpu/cpufreq/speedstep-ich.c policy->cpus = per_cpu(cpu_sibling_map, policy->cpu); cpu 357 arch/x86/kernel/cpu/cpufreq/speedstep-ich.c cpufreq_frequency_table_get_attr(speedstep_freqs, policy->cpu); cpu 365 arch/x86/kernel/cpu/cpufreq/speedstep-ich.c cpufreq_frequency_table_put_attr(policy->cpu); cpu 233 arch/x86/kernel/cpu/cpufreq/speedstep-smi.c freqs.cpu = 0; /* speedstep.c is UP only driver */ cpu 265 arch/x86/kernel/cpu/cpufreq/speedstep-smi.c if (policy->cpu != 0) cpu 309 arch/x86/kernel/cpu/cpufreq/speedstep-smi.c cpufreq_frequency_table_get_attr(speedstep_freqs, policy->cpu); cpu 316 arch/x86/kernel/cpu/cpufreq/speedstep-smi.c cpufreq_frequency_table_put_attr(policy->cpu); cpu 322 arch/x86/kernel/cpu/cpufreq/speedstep-smi.c if (cpu) cpu 167 arch/x86/kernel/cpu/intel.c int cpu = smp_processor_id(); cpu 175 arch/x86/kernel/cpu/intel.c numa_set_node(cpu, node); cpu 177 arch/x86/kernel/cpu/intel.c printk(KERN_INFO "CPU %d/%x -> Node %d\n", cpu, apicid, node); cpu 324 arch/x86/kernel/cpu/intel_cacheinfo.c unsigned int cpu = c->cpu_index; cpu 446 arch/x86/kernel/cpu/intel_cacheinfo.c per_cpu(cpu_llc_id, cpu) = l2_id; cpu 453 arch/x86/kernel/cpu/intel_cacheinfo.c per_cpu(cpu_llc_id, cpu) = l3_id; cpu 488 arch/x86/kernel/cpu/intel_cacheinfo.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu 490 arch/x86/kernel/cpu/intel_cacheinfo.c this_leaf = CPUID4_INFO_IDX(cpu, index); cpu 494 arch/x86/kernel/cpu/intel_cacheinfo.c cpu_set(cpu, this_leaf->shared_cpu_map); cpu 502 arch/x86/kernel/cpu/intel_cacheinfo.c if (i != cpu && per_cpu(cpuid4_info, i)) { cpu 504 arch/x86/kernel/cpu/intel_cacheinfo.c cpu_set(cpu, sibling_leaf->shared_cpu_map); cpu 515 arch/x86/kernel/cpu/intel_cacheinfo.c this_leaf = CPUID4_INFO_IDX(cpu, index); cpu 518 arch/x86/kernel/cpu/intel_cacheinfo.c cpu_clear(cpu, sibling_leaf->shared_cpu_map); cpu 531 arch/x86/kernel/cpu/intel_cacheinfo.c cache_remove_shared_cpu_map(cpu, i); cpu 533 arch/x86/kernel/cpu/intel_cacheinfo.c kfree(per_cpu(cpuid4_info, cpu)); cpu 534 arch/x86/kernel/cpu/intel_cacheinfo.c per_cpu(cpuid4_info, cpu) = NULL; cpu 547 arch/x86/kernel/cpu/intel_cacheinfo.c per_cpu(cpuid4_info, cpu) = kzalloc( cpu 549 arch/x86/kernel/cpu/intel_cacheinfo.c if (per_cpu(cpuid4_info, cpu) == NULL) cpu 553 arch/x86/kernel/cpu/intel_cacheinfo.c retval = set_cpus_allowed_ptr(current, &cpumask_of_cpu(cpu)); cpu 559 arch/x86/kernel/cpu/intel_cacheinfo.c this_leaf = CPUID4_INFO_IDX(cpu, j); cpu 565 arch/x86/kernel/cpu/intel_cacheinfo.c cache_remove_shared_cpu_map(cpu, i); cpu 568 arch/x86/kernel/cpu/intel_cacheinfo.c cache_shared_cpu_map_setup(cpu, j); cpu 574 arch/x86/kernel/cpu/intel_cacheinfo.c kfree(per_cpu(cpuid4_info, cpu)); cpu 575 arch/x86/kernel/cpu/intel_cacheinfo.c per_cpu(cpuid4_info, cpu) = NULL; cpu 593 arch/x86/kernel/cpu/intel_cacheinfo.c unsigned int cpu; cpu 800 arch/x86/kernel/cpu/intel_cacheinfo.c fattr->show(CPUID4_INFO_IDX(this_leaf->cpu, this_leaf->index), cpu 814 arch/x86/kernel/cpu/intel_cacheinfo.c fattr->store(CPUID4_INFO_IDX(this_leaf->cpu, this_leaf->index), cpu 836 arch/x86/kernel/cpu/intel_cacheinfo.c kfree(per_cpu(cache_kobject, cpu)); cpu 837 arch/x86/kernel/cpu/intel_cacheinfo.c kfree(per_cpu(index_kobject, cpu)); cpu 838 arch/x86/kernel/cpu/intel_cacheinfo.c per_cpu(cache_kobject, cpu) = NULL; cpu 839 arch/x86/kernel/cpu/intel_cacheinfo.c per_cpu(index_kobject, cpu) = NULL; cpu 840 arch/x86/kernel/cpu/intel_cacheinfo.c free_cache_attributes(cpu); cpu 850 arch/x86/kernel/cpu/intel_cacheinfo.c err = detect_cache_attributes(cpu); cpu 855 arch/x86/kernel/cpu/intel_cacheinfo.c per_cpu(cache_kobject, cpu) = cpu 857 arch/x86/kernel/cpu/intel_cacheinfo.c if (unlikely(per_cpu(cache_kobject, cpu) == NULL)) cpu 860 arch/x86/kernel/cpu/intel_cacheinfo.c per_cpu(index_kobject, cpu) = kzalloc( cpu 862 arch/x86/kernel/cpu/intel_cacheinfo.c if (unlikely(per_cpu(index_kobject, cpu) == NULL)) cpu 868 arch/x86/kernel/cpu/intel_cacheinfo.c cpuid4_cache_sysfs_exit(cpu); cpu 877 arch/x86/kernel/cpu/intel_cacheinfo.c unsigned int cpu = sys_dev->id; cpu 882 arch/x86/kernel/cpu/intel_cacheinfo.c retval = cpuid4_cache_sysfs_init(cpu); cpu 886 arch/x86/kernel/cpu/intel_cacheinfo.c retval = kobject_init_and_add(per_cpu(cache_kobject, cpu), cpu 890 arch/x86/kernel/cpu/intel_cacheinfo.c cpuid4_cache_sysfs_exit(cpu); cpu 895 arch/x86/kernel/cpu/intel_cacheinfo.c this_object = INDEX_KOBJECT_PTR(cpu,i); cpu 896 arch/x86/kernel/cpu/intel_cacheinfo.c this_object->cpu = cpu; cpu 900 arch/x86/kernel/cpu/intel_cacheinfo.c per_cpu(cache_kobject, cpu), cpu 904 arch/x86/kernel/cpu/intel_cacheinfo.c kobject_put(&(INDEX_KOBJECT_PTR(cpu,j)->kobj)); cpu 906 arch/x86/kernel/cpu/intel_cacheinfo.c kobject_put(per_cpu(cache_kobject, cpu)); cpu 907 arch/x86/kernel/cpu/intel_cacheinfo.c cpuid4_cache_sysfs_exit(cpu); cpu 912 arch/x86/kernel/cpu/intel_cacheinfo.c cpu_set(cpu, cache_dev_map); cpu 914 arch/x86/kernel/cpu/intel_cacheinfo.c kobject_uevent(per_cpu(cache_kobject, cpu), KOBJ_ADD); cpu 920 arch/x86/kernel/cpu/intel_cacheinfo.c unsigned int cpu = sys_dev->id; cpu 923 arch/x86/kernel/cpu/intel_cacheinfo.c if (per_cpu(cpuid4_info, cpu) == NULL) cpu 925 arch/x86/kernel/cpu/intel_cacheinfo.c if (!cpu_isset(cpu, cache_dev_map)) cpu 927 arch/x86/kernel/cpu/intel_cacheinfo.c cpu_clear(cpu, cache_dev_map); cpu 930 arch/x86/kernel/cpu/intel_cacheinfo.c kobject_put(&(INDEX_KOBJECT_PTR(cpu,i)->kobj)); cpu 931 arch/x86/kernel/cpu/intel_cacheinfo.c kobject_put(per_cpu(cache_kobject, cpu)); cpu 932 arch/x86/kernel/cpu/intel_cacheinfo.c cpuid4_cache_sysfs_exit(cpu); cpu 938 arch/x86/kernel/cpu/intel_cacheinfo.c unsigned int cpu = (unsigned long)hcpu; cpu 941 arch/x86/kernel/cpu/intel_cacheinfo.c sys_dev = get_cpu_sysdev(cpu); cpu 113 arch/x86/kernel/cpu/mcheck/mce_64.c m->cpu, m->mcgstatus, m->bank, m->status); cpu 203 arch/x86/kernel/cpu/mcheck/mce_64.c m.cpu = smp_processor_id(); cpu 340 arch/x86/kernel/cpu/mcheck/mce_64.c m.cpu = cpu; cpu 627 arch/x86/kernel/cpu/mcheck/mce_64.c mcelog.entry[i].tsc < cpu_tsc[mcelog.entry[i].cpu]) { cpu 762 arch/x86/kernel/cpu/mcheck/mce_64.c void (*threshold_cpu_callback)(unsigned long action, unsigned int cpu) __cpuinitdata; cpu 836 arch/x86/kernel/cpu/mcheck/mce_64.c memset(&per_cpu(device_mce, cpu).kobj, 0, sizeof(struct kobject)); cpu 837 arch/x86/kernel/cpu/mcheck/mce_64.c per_cpu(device_mce,cpu).id = cpu; cpu 838 arch/x86/kernel/cpu/mcheck/mce_64.c per_cpu(device_mce,cpu).cls = &mce_sysclass; cpu 840 arch/x86/kernel/cpu/mcheck/mce_64.c err = sysdev_register(&per_cpu(device_mce,cpu)); cpu 845 arch/x86/kernel/cpu/mcheck/mce_64.c err = sysdev_create_file(&per_cpu(device_mce,cpu), cpu 850 arch/x86/kernel/cpu/mcheck/mce_64.c cpu_set(cpu, mce_device_initialized); cpu 855 arch/x86/kernel/cpu/mcheck/mce_64.c sysdev_remove_file(&per_cpu(device_mce,cpu), cpu 858 arch/x86/kernel/cpu/mcheck/mce_64.c sysdev_unregister(&per_cpu(device_mce,cpu)); cpu 867 arch/x86/kernel/cpu/mcheck/mce_64.c if (!cpu_isset(cpu, mce_device_initialized)) cpu 871 arch/x86/kernel/cpu/mcheck/mce_64.c sysdev_remove_file(&per_cpu(device_mce,cpu), cpu 873 arch/x86/kernel/cpu/mcheck/mce_64.c sysdev_unregister(&per_cpu(device_mce,cpu)); cpu 874 arch/x86/kernel/cpu/mcheck/mce_64.c cpu_clear(cpu, mce_device_initialized); cpu 881 arch/x86/kernel/cpu/mcheck/mce_64.c unsigned int cpu = (unsigned long)hcpu; cpu 886 arch/x86/kernel/cpu/mcheck/mce_64.c mce_create_device(cpu); cpu 888 arch/x86/kernel/cpu/mcheck/mce_64.c threshold_cpu_callback(action, cpu); cpu 893 arch/x86/kernel/cpu/mcheck/mce_64.c threshold_cpu_callback(action, cpu); cpu 894 arch/x86/kernel/cpu/mcheck/mce_64.c mce_remove_device(cpu); cpu 53 arch/x86/kernel/cpu/mcheck/mce_amd_64.c unsigned int cpu; cpu 120 arch/x86/kernel/cpu/mcheck/mce_amd_64.c unsigned int cpu = smp_processor_id(); cpu 152 arch/x86/kernel/cpu/mcheck/mce_amd_64.c per_cpu(bank_map, cpu) |= (1 << bank); cpu 191 arch/x86/kernel/cpu/mcheck/mce_amd_64.c m.cpu = smp_processor_id(); cpu 195 arch/x86/kernel/cpu/mcheck/mce_amd_64.c if (!(per_cpu(bank_map, m.cpu) & (1 << bank))) cpu 259 arch/x86/kernel/cpu/mcheck/mce_amd_64.c cpu_set(cpu, *newmask); cpu 286 arch/x86/kernel/cpu/mcheck/mce_amd_64.c affinity_set(b->cpu, &oldmask, &newmask); cpu 309 arch/x86/kernel/cpu/mcheck/mce_amd_64.c affinity_set(b->cpu, &oldmask, &newmask); cpu 320 arch/x86/kernel/cpu/mcheck/mce_amd_64.c affinity_set(b->cpu, &oldmask, &newmask); cpu 331 arch/x86/kernel/cpu/mcheck/mce_amd_64.c affinity_set(b->cpu, &oldmask, &newmask); cpu 422 arch/x86/kernel/cpu/mcheck/mce_amd_64.c b->cpu = cpu; cpu 429 arch/x86/kernel/cpu/mcheck/mce_amd_64.c if (per_cpu(threshold_banks, cpu)[bank]->blocks) cpu 431 arch/x86/kernel/cpu/mcheck/mce_amd_64.c &per_cpu(threshold_banks, cpu)[bank]->blocks->miscj); cpu 433 arch/x86/kernel/cpu/mcheck/mce_amd_64.c per_cpu(threshold_banks, cpu)[bank]->blocks = b; cpu 436 arch/x86/kernel/cpu/mcheck/mce_amd_64.c per_cpu(threshold_banks, cpu)[bank]->kobj, cpu 449 arch/x86/kernel/cpu/mcheck/mce_amd_64.c err = allocate_threshold_blocks(cpu, bank, ++block, address); cpu 477 arch/x86/kernel/cpu/mcheck/mce_amd_64.c if (cpu_data(cpu).cpu_core_id && shared_bank[bank]) { /* symlink */ cpu 478 arch/x86/kernel/cpu/mcheck/mce_amd_64.c i = first_cpu(per_cpu(cpu_core_map, cpu)); cpu 485 arch/x86/kernel/cpu/mcheck/mce_amd_64.c if (per_cpu(threshold_banks, cpu)[bank]) cpu 493 arch/x86/kernel/cpu/mcheck/mce_amd_64.c err = sysfs_create_link(&per_cpu(device_mce, cpu).kobj, cpu 498 arch/x86/kernel/cpu/mcheck/mce_amd_64.c b->cpus = per_cpu(cpu_core_map, cpu); cpu 499 arch/x86/kernel/cpu/mcheck/mce_amd_64.c per_cpu(threshold_banks, cpu)[bank] = b; cpu 510 arch/x86/kernel/cpu/mcheck/mce_amd_64.c b->kobj = kobject_create_and_add(name, &per_cpu(device_mce, cpu).kobj); cpu 517 arch/x86/kernel/cpu/mcheck/mce_amd_64.c b->cpus = per_cpu(cpu_core_map, cpu); cpu 520 arch/x86/kernel/cpu/mcheck/mce_amd_64.c per_cpu(threshold_banks, cpu)[bank] = b; cpu 522 arch/x86/kernel/cpu/mcheck/mce_amd_64.c affinity_set(cpu, &oldmask, &newmask); cpu 523 arch/x86/kernel/cpu/mcheck/mce_amd_64.c err = allocate_threshold_blocks(cpu, bank, 0, cpu 531 arch/x86/kernel/cpu/mcheck/mce_amd_64.c if (i == cpu) cpu 545 arch/x86/kernel/cpu/mcheck/mce_amd_64.c per_cpu(threshold_banks, cpu)[bank] = NULL; cpu 558 arch/x86/kernel/cpu/mcheck/mce_amd_64.c if (!(per_cpu(bank_map, cpu) & (1 << bank))) cpu 560 arch/x86/kernel/cpu/mcheck/mce_amd_64.c err = threshold_create_bank(cpu, bank); cpu 579 arch/x86/kernel/cpu/mcheck/mce_amd_64.c struct threshold_bank *head = per_cpu(threshold_banks, cpu)[bank]; cpu 590 arch/x86/kernel/cpu/mcheck/mce_amd_64.c kfree(per_cpu(threshold_banks, cpu)[bank]->blocks); cpu 591 arch/x86/kernel/cpu/mcheck/mce_amd_64.c per_cpu(threshold_banks, cpu)[bank]->blocks = NULL; cpu 600 arch/x86/kernel/cpu/mcheck/mce_amd_64.c b = per_cpu(threshold_banks, cpu)[bank]; cpu 612 arch/x86/kernel/cpu/mcheck/mce_amd_64.c if (shared_bank[bank] && b->blocks->cpu != cpu) { cpu 613 arch/x86/kernel/cpu/mcheck/mce_amd_64.c sysfs_remove_link(&per_cpu(device_mce, cpu).kobj, name); cpu 614 arch/x86/kernel/cpu/mcheck/mce_amd_64.c per_cpu(threshold_banks, cpu)[bank] = NULL; cpu 621 arch/x86/kernel/cpu/mcheck/mce_amd_64.c if (i == cpu) cpu 628 arch/x86/kernel/cpu/mcheck/mce_amd_64.c deallocate_threshold_block(cpu, bank); cpu 634 arch/x86/kernel/cpu/mcheck/mce_amd_64.c per_cpu(threshold_banks, cpu)[bank] = NULL; cpu 642 arch/x86/kernel/cpu/mcheck/mce_amd_64.c if (!(per_cpu(bank_map, cpu) & (1 << bank))) cpu 644 arch/x86/kernel/cpu/mcheck/mce_amd_64.c threshold_remove_bank(cpu, bank); cpu 652 arch/x86/kernel/cpu/mcheck/mce_amd_64.c if (cpu >= NR_CPUS) cpu 658 arch/x86/kernel/cpu/mcheck/mce_amd_64.c threshold_create_device(cpu); cpu 662 arch/x86/kernel/cpu/mcheck/mce_amd_64.c threshold_remove_device(cpu); cpu 37 arch/x86/kernel/cpu/mcheck/mce_intel_64.c unsigned int cpu = smp_processor_id(); cpu 53 arch/x86/kernel/cpu/mcheck/mce_intel_64.c "CPU%d: Thermal monitoring handled by SMI\n", cpu); cpu 63 arch/x86/kernel/cpu/mcheck/mce_intel_64.c "installed\n", cpu, (h & APIC_VECTOR_MASK)); cpu 80 arch/x86/kernel/cpu/mcheck/mce_intel_64.c cpu, tm2 ? "TM2" : "TM1"); cpu 72 arch/x86/kernel/cpu/mcheck/p4.c unsigned int cpu = smp_processor_id(); cpu 90 arch/x86/kernel/cpu/mcheck/p4.c cpu); cpu 98 arch/x86/kernel/cpu/mcheck/p4.c cpu, (h & APIC_VECTOR_MASK)); cpu 118 arch/x86/kernel/cpu/mcheck/p4.c printk(KERN_INFO "CPU%d: Thermal monitoring enabled\n", cpu); cpu 41 arch/x86/kernel/cpu/mcheck/therm_throt.c unsigned int cpu = dev->id; \ cpu 45 arch/x86/kernel/cpu/mcheck/therm_throt.c if (cpu_online(cpu)) \ cpu 47 arch/x86/kernel/cpu/mcheck/therm_throt.c per_cpu(thermal_throttle_##name, cpu)); \ cpu 87 arch/x86/kernel/cpu/mcheck/therm_throt.c unsigned int cpu = smp_processor_id(); cpu 101 arch/x86/kernel/cpu/mcheck/therm_throt.c "cpu clock throttled (total events = %lu)\n", cpu, cpu 106 arch/x86/kernel/cpu/mcheck/therm_throt.c printk(KERN_CRIT "CPU%d: Temperature/speed normal\n", cpu); cpu 132 arch/x86/kernel/cpu/mcheck/therm_throt.c unsigned int cpu = (unsigned long)hcpu; cpu 136 arch/x86/kernel/cpu/mcheck/therm_throt.c sys_dev = get_cpu_sysdev(cpu); cpu 164 arch/x86/kernel/cpu/mcheck/therm_throt.c unsigned int cpu = 0; cpu 176 arch/x86/kernel/cpu/mcheck/therm_throt.c for_each_online_cpu(cpu) { cpu 177 arch/x86/kernel/cpu/mcheck/therm_throt.c err = thermal_throttle_add_dev(get_cpu_sysdev(cpu)); cpu 18 arch/x86/kernel/cpu/proc.c cpus_weight(per_cpu(cpu_core_map, cpu))); cpu 60 arch/x86/kernel/cpu/proc.c cpus_weight(per_cpu(cpu_core_map, cpu))); cpu 83 arch/x86/kernel/cpu/proc.c unsigned int cpu = 0; cpu 87 arch/x86/kernel/cpu/proc.c cpu = c->cpu_index; cpu 94 arch/x86/kernel/cpu/proc.c cpu, cpu 106 arch/x86/kernel/cpu/proc.c unsigned int freq = cpufreq_quick_get(cpu); cpu 118 arch/x86/kernel/cpu/proc.c show_cpuinfo_core(m, c, cpu); cpu 89 arch/x86/kernel/cpuid.c int cpu = iminor(file->f_path.dentry->d_inode); cpu 100 arch/x86/kernel/cpuid.c err = smp_call_function_single(cpu, cpuid_smp_cpuid, &cmd, 1); cpu 117 arch/x86/kernel/cpuid.c unsigned int cpu; cpu 123 arch/x86/kernel/cpuid.c cpu = iminor(file->f_path.dentry->d_inode); cpu 124 arch/x86/kernel/cpuid.c if (cpu >= NR_CPUS || !cpu_online(cpu)) { cpu 128 arch/x86/kernel/cpuid.c c = &cpu_data(cpu); cpu 150 arch/x86/kernel/cpuid.c dev = device_create_drvdata(cpuid_class, NULL, MKDEV(CPUID_MAJOR, cpu), cpu 151 arch/x86/kernel/cpuid.c NULL, "cpu%d", cpu); cpu 157 arch/x86/kernel/cpuid.c device_destroy(cpuid_class, MKDEV(CPUID_MAJOR, cpu)); cpu 164 arch/x86/kernel/cpuid.c unsigned int cpu = (unsigned long)hcpu; cpu 169 arch/x86/kernel/cpuid.c err = cpuid_device_create(cpu); cpu 174 arch/x86/kernel/cpuid.c cpuid_device_destroy(cpu); cpu 225 arch/x86/kernel/cpuid.c int cpu = 0; cpu 227 arch/x86/kernel/cpuid.c for_each_online_cpu(cpu) cpu 228 arch/x86/kernel/cpuid.c cpuid_device_destroy(cpu); cpu 45 arch/x86/kernel/crash.c int cpu; cpu 51 arch/x86/kernel/crash.c cpu = raw_smp_processor_id(); cpu 57 arch/x86/kernel/crash.c if (cpu == crashing_cpu) cpu 67 arch/x86/kernel/crash.c crash_save_cpu(regs, cpu); cpu 54 arch/x86/kernel/dumpstack_64.c unsigned long end = per_cpu(orig_ist, cpu).ist[k]; cpu 164 arch/x86/kernel/dumpstack_64.c const unsigned cpu = get_cpu(); cpu 165 arch/x86/kernel/dumpstack_64.c unsigned long *irqstack_end = (unsigned long *)cpu_pda(cpu)->irqstackptr; cpu 200 arch/x86/kernel/dumpstack_64.c estack_end = in_exception_stack(cpu, (unsigned long)stack, cpu 306 arch/x86/kernel/dumpstack_64.c const int cpu = smp_processor_id(); cpu 308 arch/x86/kernel/dumpstack_64.c (unsigned long *) (cpu_pda(cpu)->irqstackptr); cpu 310 arch/x86/kernel/dumpstack_64.c (unsigned long *) (cpu_pda(cpu)->irqstackptr - IRQSTACKSIZE); cpu 375 arch/x86/kernel/dumpstack_64.c const int cpu = smp_processor_id(); cpu 376 arch/x86/kernel/dumpstack_64.c struct task_struct *cur = cpu_pda(cpu)->pcurrent; cpu 379 arch/x86/kernel/dumpstack_64.c printk("CPU %d ", cpu); cpu 437 arch/x86/kernel/dumpstack_64.c int cpu; cpu 444 arch/x86/kernel/dumpstack_64.c cpu = smp_processor_id(); cpu 446 arch/x86/kernel/dumpstack_64.c if (cpu == die_owner) cpu 452 arch/x86/kernel/dumpstack_64.c die_owner = cpu; cpu 336 arch/x86/kernel/es7000_32.c psaival = (0x1000000 | vect | cpu); cpu 196 arch/x86/kernel/genapic_flat_64.c return cpumask_of_cpu(cpu); cpu 219 arch/x86/kernel/genapic_flat_64.c int cpu; cpu 225 arch/x86/kernel/genapic_flat_64.c cpu = first_cpu(cpumask); cpu 226 arch/x86/kernel/genapic_flat_64.c if ((unsigned)cpu < nr_cpu_ids) cpu 227 arch/x86/kernel/genapic_flat_64.c return per_cpu(x86_cpu_to_apicid, cpu); cpu 36 arch/x86/kernel/genx2apic_cluster.c cpu_set(cpu, domain); cpu 94 arch/x86/kernel/genx2apic_cluster.c int cpu; cpu 100 arch/x86/kernel/genx2apic_cluster.c cpu = first_cpu(cpumask); cpu 101 arch/x86/kernel/genx2apic_cluster.c if ((unsigned)cpu < NR_CPUS) cpu 102 arch/x86/kernel/genx2apic_cluster.c return per_cpu(x86_cpu_to_logical_apicid, cpu); cpu 135 arch/x86/kernel/genx2apic_cluster.c int cpu = smp_processor_id(); cpu 137 arch/x86/kernel/genx2apic_cluster.c per_cpu(x86_cpu_to_logical_apicid, cpu) = apic_read(APIC_LDR); cpu 40 arch/x86/kernel/genx2apic_phys.c cpu_set(cpu, domain); cpu 92 arch/x86/kernel/genx2apic_phys.c int cpu; cpu 98 arch/x86/kernel/genx2apic_phys.c cpu = first_cpu(cpumask); cpu 99 arch/x86/kernel/genx2apic_phys.c if ((unsigned)cpu < NR_CPUS) cpu 100 arch/x86/kernel/genx2apic_phys.c return per_cpu(x86_cpu_to_apicid, cpu); cpu 87 arch/x86/kernel/genx2apic_uv_x.c cpu_set(cpu, domain); cpu 117 arch/x86/kernel/genx2apic_uv_x.c apicid = per_cpu(x86_cpu_to_apicid, cpu); cpu 129 arch/x86/kernel/genx2apic_uv_x.c unsigned int cpu; cpu 131 arch/x86/kernel/genx2apic_uv_x.c for_each_possible_cpu(cpu) cpu 132 arch/x86/kernel/genx2apic_uv_x.c if (cpu_isset(cpu, mask)) cpu 133 arch/x86/kernel/genx2apic_uv_x.c uv_send_IPI_one(cpu, vector); cpu 162 arch/x86/kernel/genx2apic_uv_x.c int cpu; cpu 168 arch/x86/kernel/genx2apic_uv_x.c cpu = first_cpu(cpumask); cpu 169 arch/x86/kernel/genx2apic_uv_x.c if ((unsigned)cpu < nr_cpu_ids) cpu 170 arch/x86/kernel/genx2apic_uv_x.c return per_cpu(x86_cpu_to_apicid, cpu); cpu 366 arch/x86/kernel/genx2apic_uv_x.c int bytes, nid, cpu, lcpu, pnode, blade, i, j, m_val, n_val; cpu 417 arch/x86/kernel/genx2apic_uv_x.c for_each_present_cpu(cpu) { cpu 418 arch/x86/kernel/genx2apic_uv_x.c nid = cpu_to_node(cpu); cpu 419 arch/x86/kernel/genx2apic_uv_x.c pnode = uv_apicid_to_pnode(per_cpu(x86_cpu_to_apicid, cpu)); cpu 424 arch/x86/kernel/genx2apic_uv_x.c uv_cpu_hub_info(cpu)->lowmem_remap_base = lowmem_redir_base; cpu 425 arch/x86/kernel/genx2apic_uv_x.c uv_cpu_hub_info(cpu)->lowmem_remap_top = cpu 427 arch/x86/kernel/genx2apic_uv_x.c uv_cpu_hub_info(cpu)->m_val = m_val; cpu 428 arch/x86/kernel/genx2apic_uv_x.c uv_cpu_hub_info(cpu)->n_val = m_val; cpu 429 arch/x86/kernel/genx2apic_uv_x.c uv_cpu_hub_info(cpu)->numa_blade_id = blade; cpu 430 arch/x86/kernel/genx2apic_uv_x.c uv_cpu_hub_info(cpu)->blade_processor_id = lcpu; cpu 431 arch/x86/kernel/genx2apic_uv_x.c uv_cpu_hub_info(cpu)->pnode = pnode; cpu 432 arch/x86/kernel/genx2apic_uv_x.c uv_cpu_hub_info(cpu)->pnode_mask = (1 << n_val) - 1; cpu 433 arch/x86/kernel/genx2apic_uv_x.c uv_cpu_hub_info(cpu)->gpa_mask = (1 << (m_val + n_val)) - 1; cpu 434 arch/x86/kernel/genx2apic_uv_x.c uv_cpu_hub_info(cpu)->gnode_upper = gnode_upper; cpu 435 arch/x86/kernel/genx2apic_uv_x.c uv_cpu_hub_info(cpu)->global_mmr_base = mmr_base; cpu 436 arch/x86/kernel/genx2apic_uv_x.c uv_cpu_hub_info(cpu)->coherency_domain_number = 0;/* ZZZ */ cpu 438 arch/x86/kernel/genx2apic_uv_x.c uv_cpu_to_blade[cpu] = blade; cpu 443 arch/x86/kernel/genx2apic_uv_x.c cpu, per_cpu(x86_cpu_to_apicid, cpu), pnode, nid, cpu 393 arch/x86/kernel/io_apic_32.c #define CPU_IRQ(cpu) (irq_cpu_data[cpu].irq) cpu 394 arch/x86/kernel/io_apic_32.c #define LAST_CPU_IRQ(cpu, irq) (irq_cpu_data[cpu].last_irq[irq]) cpu 395 arch/x86/kernel/io_apic_32.c #define IRQ_DELTA(cpu, irq) (irq_cpu_data[cpu].irq_delta[irq]) cpu 398 arch/x86/kernel/io_apic_32.c (idle_cpu(cpu) && ((now) - per_cpu(irq_stat, (cpu)).idle_timestamp > 1)) cpu 400 arch/x86/kernel/io_apic_32.c #define IRQ_ALLOWED(cpu, allowed_mask) cpu_isset(cpu, allowed_mask) cpu 417 arch/x86/kernel/io_apic_32.c int cpu = curr_cpu; cpu 422 arch/x86/kernel/io_apic_32.c if (unlikely(cpu == curr_cpu)) cpu 426 arch/x86/kernel/io_apic_32.c cpu++; cpu 427 arch/x86/kernel/io_apic_32.c if (cpu >= NR_CPUS) cpu 428 arch/x86/kernel/io_apic_32.c cpu = 0; cpu 430 arch/x86/kernel/io_apic_32.c cpu--; cpu 431 arch/x86/kernel/io_apic_32.c if (cpu == -1) cpu 432 arch/x86/kernel/io_apic_32.c cpu = NR_CPUS-1; cpu 434 arch/x86/kernel/io_apic_32.c } while (!cpu_online(cpu) || !IRQ_ALLOWED(cpu, allowed_mask) || cpu 435 arch/x86/kernel/io_apic_32.c (search_idle && !IDLE_ENOUGH(cpu, now))); cpu 437 arch/x86/kernel/io_apic_32.c return cpu; cpu 450 arch/x86/kernel/io_apic_32.c new_cpu = move(cpu, allowed_mask, now, 1); cpu 451 arch/x86/kernel/io_apic_32.c if (cpu != new_cpu) cpu 801 arch/x86/kernel/io_apic_64.c int cpu; cpu 821 arch/x86/kernel/io_apic_64.c for_each_cpu_mask_nr(cpu, mask) { cpu 826 arch/x86/kernel/io_apic_64.c domain = vector_allocation_domain(cpu); cpu 876 arch/x86/kernel/io_apic_64.c int cpu, vector; cpu 884 arch/x86/kernel/io_apic_64.c for_each_cpu_mask_nr(cpu, mask) cpu 885 arch/x86/kernel/io_apic_64.c per_cpu(vector_irq, cpu)[vector] = -1; cpu 899 arch/x86/kernel/io_apic_64.c if (!cpu_isset(cpu, irq_cfg[irq].domain)) cpu 902 arch/x86/kernel/io_apic_64.c per_cpu(vector_irq, cpu)[vector] = irq; cpu 906 arch/x86/kernel/io_apic_64.c irq = per_cpu(vector_irq, cpu)[vector]; cpu 909 arch/x86/kernel/io_apic_64.c if (!cpu_isset(cpu, irq_cfg[irq].domain)) cpu 910 arch/x86/kernel/io_apic_64.c per_cpu(vector_irq, cpu)[vector] = -1; cpu 151 arch/x86/kernel/irq_32.c if (hardirq_ctx[cpu]) cpu 154 arch/x86/kernel/irq_32.c irqctx = (union irq_ctx*) &hardirq_stack[cpu*THREAD_SIZE]; cpu 157 arch/x86/kernel/irq_32.c irqctx->tinfo.cpu = cpu; cpu 161 arch/x86/kernel/irq_32.c hardirq_ctx[cpu] = irqctx; cpu 163 arch/x86/kernel/irq_32.c irqctx = (union irq_ctx*) &softirq_stack[cpu*THREAD_SIZE]; cpu 166 arch/x86/kernel/irq_32.c irqctx->tinfo.cpu = cpu; cpu 170 arch/x86/kernel/irq_32.c softirq_ctx[cpu] = irqctx; cpu 173 arch/x86/kernel/irq_32.c cpu,hardirq_ctx[cpu],softirq_ctx[cpu]); cpu 178 arch/x86/kernel/irq_32.c hardirq_ctx[cpu] = NULL; cpu 362 arch/x86/kernel/irq_32.c u64 sum = nmi_count(cpu); cpu 365 arch/x86/kernel/irq_32.c sum += per_cpu(irq_stat, cpu).apic_timer_irqs; cpu 368 arch/x86/kernel/irq_32.c sum += per_cpu(irq_stat, cpu).irq_resched_count; cpu 369 arch/x86/kernel/irq_32.c sum += per_cpu(irq_stat, cpu).irq_call_count; cpu 370 arch/x86/kernel/irq_32.c sum += per_cpu(irq_stat, cpu).irq_tlb_count; cpu 373 arch/x86/kernel/irq_32.c sum += per_cpu(irq_stat, cpu).irq_thermal_count; cpu 376 arch/x86/kernel/irq_32.c sum += per_cpu(irq_stat, cpu).irq_spurious_count; cpu 162 arch/x86/kernel/irq_64.c u64 sum = cpu_pda(cpu)->__nmi_count; cpu 164 arch/x86/kernel/irq_64.c sum += cpu_pda(cpu)->apic_timer_irqs; cpu 166 arch/x86/kernel/irq_64.c sum += cpu_pda(cpu)->irq_resched_count; cpu 167 arch/x86/kernel/irq_64.c sum += cpu_pda(cpu)->irq_call_count; cpu 168 arch/x86/kernel/irq_64.c sum += cpu_pda(cpu)->irq_tlb_count; cpu 171 arch/x86/kernel/irq_64.c sum += cpu_pda(cpu)->irq_thermal_count; cpu 172 arch/x86/kernel/irq_64.c sum += cpu_pda(cpu)->irq_threshold_count; cpu 174 arch/x86/kernel/irq_64.c sum += cpu_pda(cpu)->irq_spurious_count; cpu 93 arch/x86/kernel/kvmclock.c int cpu = smp_processor_id(); cpu 95 arch/x86/kernel/kvmclock.c low = (int)__pa(&per_cpu(hv_clock, cpu)) | 1; cpu 96 arch/x86/kernel/kvmclock.c high = ((u64)__pa(&per_cpu(hv_clock, cpu)) >> 32); cpu 98 arch/x86/kernel/kvmclock.c cpu, high, low, txt); cpu 95 arch/x86/kernel/microcode_amd.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu 101 arch/x86/kernel/microcode_amd.c cpu); cpu 136 arch/x86/kernel/microcode_amd.c "not found in equivalent cpu table \n", cpu); cpu 144 arch/x86/kernel/microcode_amd.c cpu, mc_header->processor_rev_id[0], cpu 152 arch/x86/kernel/microcode_amd.c cpu, mc_header->processor_rev_id[1], cpu 165 arch/x86/kernel/microcode_amd.c printk(KERN_ERR "microcode: CPU%d NB mismatch \n", cpu); cpu 179 arch/x86/kernel/microcode_amd.c printk(KERN_ERR "microcode: CPU%d SB mismatch \n", cpu); cpu 203 arch/x86/kernel/microcode_amd.c BUG_ON(cpu_num != cpu); cpu 327 arch/x86/kernel/microcode_amd.c struct ucode_cpu_info *uci = ucode_cpu_info + cpu; cpu 351 arch/x86/kernel/microcode_amd.c if (get_matching_microcode(cpu, mc, new_rev)) { cpu 370 arch/x86/kernel/microcode_amd.c cpu, new_rev, uci->cpu_sig.rev); cpu 393 arch/x86/kernel/microcode_amd.c BUG_ON(cpu != raw_smp_processor_id()); cpu 401 arch/x86/kernel/microcode_amd.c ret = generic_load_microcode(cpu, (void*)firmware->data, firmware->size, cpu 418 arch/x86/kernel/microcode_amd.c struct ucode_cpu_info *uci = ucode_cpu_info + cpu; cpu 115 arch/x86/kernel/microcode_core.c int cpu; cpu 119 arch/x86/kernel/microcode_core.c for_each_online_cpu(cpu) { cpu 120 arch/x86/kernel/microcode_core.c struct ucode_cpu_info *uci = ucode_cpu_info + cpu; cpu 125 arch/x86/kernel/microcode_core.c set_cpus_allowed_ptr(current, &cpumask_of_cpu(cpu)); cpu 126 arch/x86/kernel/microcode_core.c error = microcode_ops->request_microcode_user(cpu, buf, size); cpu 130 arch/x86/kernel/microcode_core.c microcode_ops->apply_microcode(cpu); cpu 216 arch/x86/kernel/microcode_core.c int cpu = dev->id; cpu 224 arch/x86/kernel/microcode_core.c if (cpu_online(cpu)) { cpu 225 arch/x86/kernel/microcode_core.c set_cpus_allowed_ptr(current, &cpumask_of_cpu(cpu)); cpu 228 arch/x86/kernel/microcode_core.c err = microcode_ops->request_microcode_fw(cpu, cpu 231 arch/x86/kernel/microcode_core.c microcode_ops->apply_microcode(cpu); cpu 277 arch/x86/kernel/microcode_core.c struct ucode_cpu_info *uci = ucode_cpu_info + cpu; cpu 280 arch/x86/kernel/microcode_core.c microcode_ops->microcode_fini_cpu(cpu); cpu 287 arch/x86/kernel/microcode_core.c struct ucode_cpu_info *uci = ucode_cpu_info + cpu; cpu 290 arch/x86/kernel/microcode_core.c if (!microcode_ops->collect_cpu_info(cpu, &uci->cpu_sig)) cpu 296 arch/x86/kernel/microcode_core.c struct ucode_cpu_info *uci = ucode_cpu_info + cpu; cpu 299 arch/x86/kernel/microcode_core.c pr_debug("microcode: CPU%d resumed\n", cpu); cpu 308 arch/x86/kernel/microcode_core.c if (microcode_ops->collect_cpu_info(cpu, &nsig)) { cpu 309 arch/x86/kernel/microcode_core.c microcode_fini_cpu(cpu); cpu 314 arch/x86/kernel/microcode_core.c microcode_fini_cpu(cpu); cpu 324 arch/x86/kernel/microcode_core.c struct ucode_cpu_info *uci = ucode_cpu_info + cpu; cpu 332 arch/x86/kernel/microcode_core.c err = microcode_resume_cpu(cpu); cpu 334 arch/x86/kernel/microcode_core.c collect_cpu_info(cpu); cpu 336 arch/x86/kernel/microcode_core.c err = microcode_ops->request_microcode_fw(cpu, cpu 340 arch/x86/kernel/microcode_core.c microcode_ops->apply_microcode(cpu); cpu 347 arch/x86/kernel/microcode_core.c set_cpus_allowed_ptr(current, &cpumask_of_cpu(cpu)); cpu 349 arch/x86/kernel/microcode_core.c BUG_ON(raw_smp_processor_id() != cpu); cpu 352 arch/x86/kernel/microcode_core.c microcode_update_cpu(cpu); cpu 360 arch/x86/kernel/microcode_core.c int err, cpu = sys_dev->id; cpu 361 arch/x86/kernel/microcode_core.c struct ucode_cpu_info *uci = ucode_cpu_info + cpu; cpu 363 arch/x86/kernel/microcode_core.c if (!cpu_online(cpu)) cpu 366 arch/x86/kernel/microcode_core.c pr_debug("microcode: CPU%d added\n", cpu); cpu 373 arch/x86/kernel/microcode_core.c microcode_init_cpu(cpu); cpu 379 arch/x86/kernel/microcode_core.c int cpu = sys_dev->id; cpu 381 arch/x86/kernel/microcode_core.c if (!cpu_online(cpu)) cpu 384 arch/x86/kernel/microcode_core.c pr_debug("microcode: CPU%d removed\n", cpu); cpu 385 arch/x86/kernel/microcode_core.c microcode_fini_cpu(cpu); cpu 392 arch/x86/kernel/microcode_core.c int cpu = dev->id; cpu 394 arch/x86/kernel/microcode_core.c if (!cpu_online(cpu)) cpu 411 arch/x86/kernel/microcode_core.c unsigned int cpu = (unsigned long)hcpu; cpu 414 arch/x86/kernel/microcode_core.c sys_dev = get_cpu_sysdev(cpu); cpu 418 arch/x86/kernel/microcode_core.c microcode_init_cpu(cpu); cpu 421 arch/x86/kernel/microcode_core.c pr_debug("microcode: CPU%d added\n", cpu); cpu 424 arch/x86/kernel/microcode_core.c "group for CPU%d\n", cpu); cpu 430 arch/x86/kernel/microcode_core.c pr_debug("microcode: CPU%d removed\n", cpu); cpu 435 arch/x86/kernel/microcode_core.c microcode_fini_cpu(cpu); cpu 318 arch/x86/kernel/microcode_intel.c struct ucode_cpu_info *uci = ucode_cpu_info + cpu; cpu 322 arch/x86/kernel/microcode_intel.c BUG_ON(cpu_num != cpu); cpu 360 arch/x86/kernel/microcode_intel.c struct ucode_cpu_info *uci = ucode_cpu_info + cpu; cpu 408 arch/x86/kernel/microcode_intel.c cpu, new_rev, uci->cpu_sig.rev); cpu 425 arch/x86/kernel/microcode_intel.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu 430 arch/x86/kernel/microcode_intel.c BUG_ON(cpu != raw_smp_processor_id()); cpu 439 arch/x86/kernel/microcode_intel.c ret = generic_load_microcode(cpu, (void*)firmware->data, firmware->size, cpu 455 arch/x86/kernel/microcode_intel.c BUG_ON(cpu != raw_smp_processor_id()); cpu 457 arch/x86/kernel/microcode_intel.c return generic_load_microcode(cpu, (void*)buf, size, &get_ucode_user); cpu 462 arch/x86/kernel/microcode_intel.c struct ucode_cpu_info *uci = ucode_cpu_info + cpu; cpu 74 arch/x86/kernel/msr.c int cpu = iminor(file->f_path.dentry->d_inode); cpu 82 arch/x86/kernel/msr.c err = rdmsr_safe_on_cpu(cpu, reg, &data[0], &data[1]); cpu 105 arch/x86/kernel/msr.c int cpu = iminor(file->f_path.dentry->d_inode); cpu 117 arch/x86/kernel/msr.c err = wrmsr_safe_on_cpu(cpu, reg, data[0], data[1]); cpu 132 arch/x86/kernel/msr.c unsigned int cpu = iminor(file->f_path.dentry->d_inode); cpu 133 arch/x86/kernel/msr.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu 137 arch/x86/kernel/msr.c cpu = iminor(file->f_path.dentry->d_inode); cpu 139 arch/x86/kernel/msr.c if (cpu >= NR_CPUS || !cpu_online(cpu)) { cpu 143 arch/x86/kernel/msr.c c = &cpu_data(cpu); cpu 166 arch/x86/kernel/msr.c dev = device_create_drvdata(msr_class, NULL, MKDEV(MSR_MAJOR, cpu), cpu 167 arch/x86/kernel/msr.c NULL, "msr%d", cpu); cpu 173 arch/x86/kernel/msr.c device_destroy(msr_class, MKDEV(MSR_MAJOR, cpu)); cpu 179 arch/x86/kernel/msr.c unsigned int cpu = (unsigned long)hcpu; cpu 184 arch/x86/kernel/msr.c err = msr_device_create(cpu); cpu 189 arch/x86/kernel/msr.c msr_device_destroy(cpu); cpu 238 arch/x86/kernel/msr.c int cpu = 0; cpu 239 arch/x86/kernel/msr.c for_each_online_cpu(cpu) cpu 240 arch/x86/kernel/msr.c msr_device_destroy(cpu); cpu 67 arch/x86/kernel/nmi.c return cpu_pda(cpu)->__nmi_count; cpu 69 arch/x86/kernel/nmi.c return nmi_count(cpu); cpu 90 arch/x86/kernel/nmi.c return per_cpu(irq_stat, cpu).apic_timer_irqs + cpu 91 arch/x86/kernel/nmi.c per_cpu(irq_stat, cpu).irq0_irqs; cpu 123 arch/x86/kernel/nmi.c cpu, prev_nmi_count[cpu], get_nmi_count(cpu)); cpu 130 arch/x86/kernel/nmi.c per_cpu(wd_enabled, cpu) = 0; cpu 137 arch/x86/kernel/nmi.c int cpu; cpu 153 arch/x86/kernel/nmi.c for_each_possible_cpu(cpu) cpu 154 arch/x86/kernel/nmi.c prev_nmi_count[cpu] = get_nmi_count(cpu); cpu 158 arch/x86/kernel/nmi.c for_each_online_cpu(cpu) { cpu 159 arch/x86/kernel/nmi.c if (!per_cpu(wd_enabled, cpu)) cpu 161 arch/x86/kernel/nmi.c if (get_nmi_count(cpu) - prev_nmi_count[cpu] <= 5) cpu 162 arch/x86/kernel/nmi.c report_broken_nmi(cpu, prev_nmi_count); cpu 368 arch/x86/kernel/nmi.c unsigned cpu; cpu 375 arch/x86/kernel/nmi.c for_each_present_cpu(cpu) { cpu 376 arch/x86/kernel/nmi.c if (per_cpu(nmi_touch, cpu) != 1) cpu 377 arch/x86/kernel/nmi.c per_cpu(nmi_touch, cpu) = 1; cpu 398 arch/x86/kernel/nmi.c int cpu = smp_processor_id(); cpu 408 arch/x86/kernel/nmi.c sum = get_timer_irqs(cpu); cpu 415 arch/x86/kernel/nmi.c if (cpu_isset(cpu, backtrace_mask)) { cpu 419 arch/x86/kernel/nmi.c printk(KERN_WARNING "NMI backtrace for cpu %d\n", cpu); cpu 422 arch/x86/kernel/nmi.c cpu_clear(cpu, backtrace_mask); cpu 524 arch/x86/kernel/nmi.c return unknown_nmi_panic_callback(regs, cpu); cpu 254 arch/x86/kernel/process.c cpu_clear(cpu, c1e_mask); cpu 281 arch/x86/kernel/process.c int cpu = smp_processor_id(); cpu 283 arch/x86/kernel/process.c if (!cpu_isset(cpu, c1e_mask)) { cpu 284 arch/x86/kernel/process.c cpu_set(cpu, c1e_mask); cpu 292 arch/x86/kernel/process.c &cpu); cpu 294 arch/x86/kernel/process.c cpu); cpu 297 arch/x86/kernel/process.c clockevents_notify(CLOCK_EVT_NOTIFY_BROADCAST_ENTER, &cpu); cpu 306 arch/x86/kernel/process.c clockevents_notify(CLOCK_EVT_NOTIFY_BROADCAST_EXIT, &cpu); cpu 94 arch/x86/kernel/process_32.c int cpu = smp_processor_id(); cpu 106 arch/x86/kernel/process_32.c if (rcu_pending(cpu)) cpu 107 arch/x86/kernel/process_32.c rcu_check_callbacks(cpu, 0); cpu 109 arch/x86/kernel/process_32.c if (cpu_is_offline(cpu)) cpu 237 arch/x86/kernel/process_32.c int cpu = get_cpu(); cpu 238 arch/x86/kernel/process_32.c struct tss_struct *tss = &per_cpu(init_tss, cpu); cpu 555 arch/x86/kernel/process_32.c int cpu = smp_processor_id(); cpu 556 arch/x86/kernel/process_32.c struct tss_struct *tss = &per_cpu(init_tss, cpu); cpu 587 arch/x86/kernel/process_32.c load_TLS(next, cpu); cpu 553 arch/x86/kernel/process_64.c int cpu = smp_processor_id(); cpu 554 arch/x86/kernel/process_64.c struct tss_struct *tss = &per_cpu(init_tss, cpu); cpu 587 arch/x86/kernel/process_64.c load_TLS(next, cpu); cpu 761 arch/x86/kernel/process_64.c int cpu; cpu 767 arch/x86/kernel/process_64.c cpu = get_cpu(); cpu 773 arch/x86/kernel/process_64.c load_TLS(&task->thread, cpu); cpu 793 arch/x86/kernel/process_64.c cpu = get_cpu(); cpu 799 arch/x86/kernel/process_64.c load_TLS(&task->thread, cpu); cpu 62 arch/x86/kernel/setup_percpu.c int cpu; cpu 64 arch/x86/kernel/setup_percpu.c for_each_possible_cpu(cpu) { cpu 65 arch/x86/kernel/setup_percpu.c per_cpu(x86_cpu_to_apicid, cpu) = cpu 66 arch/x86/kernel/setup_percpu.c early_per_cpu_map(x86_cpu_to_apicid, cpu); cpu 67 arch/x86/kernel/setup_percpu.c per_cpu(x86_bios_cpu_apicid, cpu) = cpu 68 arch/x86/kernel/setup_percpu.c early_per_cpu_map(x86_bios_cpu_apicid, cpu); cpu 70 arch/x86/kernel/setup_percpu.c per_cpu(x86_cpu_to_node_map, cpu) = cpu 71 arch/x86/kernel/setup_percpu.c early_per_cpu_map(x86_cpu_to_node_map, cpu); cpu 105 arch/x86/kernel/setup_percpu.c int cpu; cpu 120 arch/x86/kernel/setup_percpu.c for_each_possible_cpu(cpu) { cpu 121 arch/x86/kernel/setup_percpu.c if (cpu == 0) { cpu 126 arch/x86/kernel/setup_percpu.c new_cpu_pda[cpu] = (struct x8664_pda *)pda; cpu 127 arch/x86/kernel/setup_percpu.c new_cpu_pda[cpu]->in_bootmem = 1; cpu 145 arch/x86/kernel/setup_percpu.c int cpu; cpu 155 arch/x86/kernel/setup_percpu.c for_each_possible_cpu(cpu) { cpu 159 arch/x86/kernel/setup_percpu.c int node = early_cpu_to_node(cpu); cpu 164 arch/x86/kernel/setup_percpu.c cpu, node); cpu 167 arch/x86/kernel/setup_percpu.c cpu, __pa(ptr)); cpu 173 arch/x86/kernel/setup_percpu.c cpu, node, __pa(ptr)); cpu 176 arch/x86/kernel/setup_percpu.c per_cpu_offset(cpu) = ptr - __per_cpu_start; cpu 227 arch/x86/kernel/setup_percpu.c if (cpu_pda(cpu) && node != NUMA_NO_NODE) cpu 228 arch/x86/kernel/setup_percpu.c cpu_pda(cpu)->nodenumber = node; cpu 231 arch/x86/kernel/setup_percpu.c cpu_to_node_map[cpu] = node; cpu 233 arch/x86/kernel/setup_percpu.c else if (per_cpu_offset(cpu)) cpu 234 arch/x86/kernel/setup_percpu.c per_cpu(x86_cpu_to_node_map, cpu) = node; cpu 237 arch/x86/kernel/setup_percpu.c pr_debug("Setting node for non-present cpu %d\n", cpu); cpu 242 arch/x86/kernel/setup_percpu.c numa_set_node(cpu, NUMA_NO_NODE); cpu 249 arch/x86/kernel/setup_percpu.c cpu_set(cpu, node_to_cpumask_map[early_cpu_to_node(cpu)]); cpu 254 arch/x86/kernel/setup_percpu.c cpu_clear(cpu, node_to_cpumask_map[cpu_to_node(cpu)]); cpu 264 arch/x86/kernel/setup_percpu.c int node = cpu_to_node(cpu); cpu 276 arch/x86/kernel/setup_percpu.c cpu_set(cpu, *mask); cpu 278 arch/x86/kernel/setup_percpu.c cpu_clear(cpu, *mask); cpu 282 arch/x86/kernel/setup_percpu.c enable? "numa_add_cpu":"numa_remove_cpu", cpu, node, buf); cpu 287 arch/x86/kernel/setup_percpu.c numa_set_cpumask(cpu, 1); cpu 292 arch/x86/kernel/setup_percpu.c numa_set_cpumask(cpu, 0); cpu 299 arch/x86/kernel/setup_percpu.c "cpu_to_node(%d): usage too early!\n", cpu); cpu 301 arch/x86/kernel/setup_percpu.c return early_per_cpu_ptr(x86_cpu_to_node_map)[cpu]; cpu 303 arch/x86/kernel/setup_percpu.c return per_cpu(x86_cpu_to_node_map, cpu); cpu 314 arch/x86/kernel/setup_percpu.c return early_per_cpu_ptr(x86_cpu_to_node_map)[cpu]; cpu 316 arch/x86/kernel/setup_percpu.c if (!per_cpu_offset(cpu)) { cpu 318 arch/x86/kernel/setup_percpu.c "early_cpu_to_node(%d): no per_cpu area!\n", cpu); cpu 322 arch/x86/kernel/setup_percpu.c return per_cpu(x86_cpu_to_node_map, cpu); cpu 117 arch/x86/kernel/smp.c if (unlikely(cpu_is_offline(cpu))) { cpu 121 arch/x86/kernel/smp.c send_IPI_mask(cpumask_of_cpu(cpu), RESCHEDULE_VECTOR); cpu 126 arch/x86/kernel/smp.c send_IPI_mask(cpumask_of_cpu(cpu), CALL_FUNCTION_SINGLE_VECTOR); cpu 147 arch/x86/kernel/smpboot.c printk(KERN_INFO "Mapping cpu %d to node %d\n", cpu, node); cpu 148 arch/x86/kernel/smpboot.c cpu_set(cpu, node_to_cpumask_map[node]); cpu 149 arch/x86/kernel/smpboot.c cpu_to_node_map[cpu] = node; cpu 157 arch/x86/kernel/smpboot.c printk(KERN_INFO "Unmapping cpu %d from all nodes\n", cpu); cpu 159 arch/x86/kernel/smpboot.c cpu_clear(cpu, node_to_cpumask_map[node]); cpu 160 arch/x86/kernel/smpboot.c cpu_to_node_map[cpu] = 0; cpu 175 arch/x86/kernel/smpboot.c int cpu = smp_processor_id(); cpu 182 arch/x86/kernel/smpboot.c cpu_2_logical_apicid[cpu] = apicid; cpu 183 arch/x86/kernel/smpboot.c map_cpu_to_node(cpu, node); cpu 188 arch/x86/kernel/smpboot.c cpu_2_logical_apicid[cpu] = BAD_APICID; cpu 189 arch/x86/kernel/smpboot.c unmap_cpu_to_node(cpu); cpu 446 arch/x86/kernel/smpboot.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu 448 arch/x86/kernel/smpboot.c cpu_set(cpu, cpu_sibling_setup_map); cpu 454 arch/x86/kernel/smpboot.c cpu_set(i, per_cpu(cpu_sibling_map, cpu)); cpu 455 arch/x86/kernel/smpboot.c cpu_set(cpu, per_cpu(cpu_sibling_map, i)); cpu 456 arch/x86/kernel/smpboot.c cpu_set(i, per_cpu(cpu_core_map, cpu)); cpu 457 arch/x86/kernel/smpboot.c cpu_set(cpu, per_cpu(cpu_core_map, i)); cpu 459 arch/x86/kernel/smpboot.c cpu_set(cpu, cpu_data(i).llc_shared_map); cpu 463 arch/x86/kernel/smpboot.c cpu_set(cpu, per_cpu(cpu_sibling_map, cpu)); cpu 466 arch/x86/kernel/smpboot.c cpu_set(cpu, c->llc_shared_map); cpu 469 arch/x86/kernel/smpboot.c per_cpu(cpu_core_map, cpu) = per_cpu(cpu_sibling_map, cpu); cpu 475 arch/x86/kernel/smpboot.c if (per_cpu(cpu_llc_id, cpu) != BAD_APICID && cpu 476 arch/x86/kernel/smpboot.c per_cpu(cpu_llc_id, cpu) == per_cpu(cpu_llc_id, i)) { cpu 478 arch/x86/kernel/smpboot.c cpu_set(cpu, cpu_data(i).llc_shared_map); cpu 481 arch/x86/kernel/smpboot.c cpu_set(i, per_cpu(cpu_core_map, cpu)); cpu 482 arch/x86/kernel/smpboot.c cpu_set(cpu, per_cpu(cpu_core_map, i)); cpu 486 arch/x86/kernel/smpboot.c if (cpus_weight(per_cpu(cpu_sibling_map, cpu)) == 1) { cpu 497 arch/x86/kernel/smpboot.c if (i != cpu) cpu 499 arch/x86/kernel/smpboot.c } else if (i != cpu && !c->booted_cores) cpu 508 arch/x86/kernel/smpboot.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu 514 arch/x86/kernel/smpboot.c return per_cpu(cpu_core_map, cpu); cpu 521 arch/x86/kernel/smpboot.c int cpu; cpu 527 arch/x86/kernel/smpboot.c for_each_possible_cpu(cpu) cpu 528 arch/x86/kernel/smpboot.c if (cpu_isset(cpu, cpu_callout_map)) cpu 529 arch/x86/kernel/smpboot.c bogosum += cpu_data(cpu).loops_per_jiffy; cpu 746 arch/x86/kernel/smpboot.c int cpu; cpu 754 arch/x86/kernel/smpboot.c c_idle->idle = fork_idle(c_idle->cpu); cpu 776 arch/x86/kernel/smpboot.c int node = cpu_to_node(cpu); cpu 778 arch/x86/kernel/smpboot.c if (cpu_pda(cpu) && !cpu_pda(cpu)->in_bootmem) cpu 781 arch/x86/kernel/smpboot.c oldpda = cpu_pda(cpu); cpu 785 arch/x86/kernel/smpboot.c "for CPU %d on node %d\n", cpu, node); cpu 799 arch/x86/kernel/smpboot.c cpu_pda(cpu) = newpda; cpu 816 arch/x86/kernel/smpboot.c .cpu = cpu, cpu 823 arch/x86/kernel/smpboot.c if (cpu > 0) { cpu 824 arch/x86/kernel/smpboot.c boot_error = get_local_pda(cpu); cpu 833 arch/x86/kernel/smpboot.c c_idle.idle = get_idle_for_cpu(cpu); cpu 842 arch/x86/kernel/smpboot.c init_idle(c_idle.idle, cpu); cpu 854 arch/x86/kernel/smpboot.c printk("failed fork for CPU %d\n", cpu); cpu 858 arch/x86/kernel/smpboot.c set_idle_for_cpu(cpu, c_idle.idle); cpu 861 arch/x86/kernel/smpboot.c per_cpu(current_task, cpu) = c_idle.idle; cpu 862 arch/x86/kernel/smpboot.c init_gdt(cpu); cpu 864 arch/x86/kernel/smpboot.c irq_ctx_init(cpu); cpu 866 arch/x86/kernel/smpboot.c cpu_pda(cpu)->pcurrent = c_idle.idle; cpu 869 arch/x86/kernel/smpboot.c early_gdt_descr.address = (unsigned long)get_cpu_gdt_table(cpu); cpu 878 arch/x86/kernel/smpboot.c cpu, apicid, start_ip); cpu 910 arch/x86/kernel/smpboot.c pr_debug("Before Callout %d.\n", cpu); cpu 911 arch/x86/kernel/smpboot.c cpu_set(cpu, cpu_callout_map); cpu 912 arch/x86/kernel/smpboot.c pr_debug("After Callout %d.\n", cpu); cpu 918 arch/x86/kernel/smpboot.c if (cpu_isset(cpu, cpu_callin_map)) cpu 923 arch/x86/kernel/smpboot.c if (cpu_isset(cpu, cpu_callin_map)) { cpu 926 arch/x86/kernel/smpboot.c printk(KERN_INFO "CPU%d: ", cpu); cpu 927 arch/x86/kernel/smpboot.c print_cpu_info(&cpu_data(cpu)); cpu 947 arch/x86/kernel/smpboot.c numa_remove_cpu(cpu); /* was set by numa_add_cpu */ cpu 948 arch/x86/kernel/smpboot.c cpu_clear(cpu, cpu_callout_map); /* was set by do_boot_cpu() */ cpu 949 arch/x86/kernel/smpboot.c cpu_clear(cpu, cpu_initialized); /* was set by cpu_init() */ cpu 950 arch/x86/kernel/smpboot.c cpu_clear(cpu, cpu_present_map); cpu 951 arch/x86/kernel/smpboot.c per_cpu(x86_cpu_to_apicid, cpu) = BAD_APICID; cpu 967 arch/x86/kernel/smpboot.c int apicid = cpu_present_to_apicid(cpu); cpu 973 arch/x86/kernel/smpboot.c pr_debug("++++++++++++++++++++=_---CPU UP %u\n", cpu); cpu 977 arch/x86/kernel/smpboot.c printk(KERN_ERR "%s: bad cpu %d\n", __func__, cpu); cpu 984 arch/x86/kernel/smpboot.c if (cpu_isset(cpu, cpu_callin_map)) { cpu 985 arch/x86/kernel/smpboot.c pr_debug("do_boot_cpu %d Already started\n", cpu); cpu 995 arch/x86/kernel/smpboot.c per_cpu(cpu_state, cpu) = CPU_UP_PREPARE; cpu 1004 arch/x86/kernel/smpboot.c err = do_boot_cpu(apicid, cpu); cpu 1009 arch/x86/kernel/smpboot.c err = do_boot_cpu(apicid, cpu); cpu 1021 arch/x86/kernel/smpboot.c check_tsc_sync_source(cpu); cpu 1024 arch/x86/kernel/smpboot.c while (!cpu_online(cpu)) { cpu 1061 arch/x86/kernel/smpboot.c unsigned int cpu; cpu 1069 arch/x86/kernel/smpboot.c for_each_present_cpu(cpu) { cpu 1071 arch/x86/kernel/smpboot.c cpu_clear(cpu, cpu_present_map); cpu 1076 arch/x86/kernel/smpboot.c for_each_possible_cpu(cpu) { cpu 1078 arch/x86/kernel/smpboot.c cpu_clear(cpu, cpu_possible_map); cpu 1181 arch/x86/kernel/smpboot.c current_thread_info()->cpu = 0; /* needed? */ cpu 1307 arch/x86/kernel/smpboot.c struct cpuinfo_x86 *c = &cpu_data(cpu); cpu 1309 arch/x86/kernel/smpboot.c for_each_cpu_mask_nr(sibling, per_cpu(cpu_core_map, cpu)) { cpu 1310 arch/x86/kernel/smpboot.c cpu_clear(cpu, per_cpu(cpu_core_map, sibling)); cpu 1314 arch/x86/kernel/smpboot.c if (cpus_weight(per_cpu(cpu_sibling_map, cpu)) == 1) cpu 1318 arch/x86/kernel/smpboot.c for_each_cpu_mask_nr(sibling, per_cpu(cpu_sibling_map, cpu)) cpu 1319 arch/x86/kernel/smpboot.c cpu_clear(cpu, per_cpu(cpu_sibling_map, sibling)); cpu 1320 arch/x86/kernel/smpboot.c cpus_clear(per_cpu(cpu_sibling_map, cpu)); cpu 1321 arch/x86/kernel/smpboot.c cpus_clear(per_cpu(cpu_core_map, cpu)); cpu 1324 arch/x86/kernel/smpboot.c cpu_clear(cpu, cpu_sibling_setup_map); cpu 1329 arch/x86/kernel/smpboot.c cpu_clear(cpu, cpu_online_map); cpu 1330 arch/x86/kernel/smpboot.c cpu_clear(cpu, cpu_callout_map); cpu 1331 arch/x86/kernel/smpboot.c cpu_clear(cpu, cpu_callin_map); cpu 1333 arch/x86/kernel/smpboot.c cpu_clear(cpu, cpu_initialized); cpu 1334 arch/x86/kernel/smpboot.c numa_remove_cpu(cpu); cpu 1339 arch/x86/kernel/smpboot.c int cpu = smp_processor_id(); cpu 1350 arch/x86/kernel/smpboot.c remove_siblinginfo(cpu); cpu 1354 arch/x86/kernel/smpboot.c remove_cpu_from_maps(cpu); cpu 1361 arch/x86/kernel/smpboot.c int cpu = smp_processor_id(); cpu 1371 arch/x86/kernel/smpboot.c if (cpu == 0) cpu 1389 arch/x86/kernel/smpboot.c if (per_cpu(cpu_state, cpu) == CPU_DEAD) { cpu 1390 arch/x86/kernel/smpboot.c printk(KERN_INFO "CPU %d is now offline\n", cpu); cpu 1397 arch/x86/kernel/smpboot.c printk(KERN_ERR "CPU %u didn't die...\n", cpu); cpu 20 arch/x86/kernel/smpcommon.c pack_descriptor(&gdt, __per_cpu_offset[cpu], 0xFFFFF, cpu 24 arch/x86/kernel/smpcommon.c write_gdt_entry(get_cpu_gdt_table(cpu), cpu 27 arch/x86/kernel/smpcommon.c per_cpu(this_cpu_off, cpu) = __per_cpu_offset[cpu]; cpu 28 arch/x86/kernel/smpcommon.c per_cpu(cpu_number, cpu) = cpu; cpu 37 arch/x86/kernel/tlb_32.c if (per_cpu(cpu_tlbstate, cpu).state == TLBSTATE_OK) cpu 39 arch/x86/kernel/tlb_32.c cpu_clear(cpu, per_cpu(cpu_tlbstate, cpu).active_mm->cpu_vm_mask); cpu 92 arch/x86/kernel/tlb_32.c unsigned long cpu; cpu 94 arch/x86/kernel/tlb_32.c cpu = get_cpu(); cpu 96 arch/x86/kernel/tlb_32.c if (!cpu_isset(cpu, flush_cpumask)) cpu 107 arch/x86/kernel/tlb_32.c if (flush_mm == per_cpu(cpu_tlbstate, cpu).active_mm) { cpu 108 arch/x86/kernel/tlb_32.c if (per_cpu(cpu_tlbstate, cpu).state == TLBSTATE_OK) { cpu 114 arch/x86/kernel/tlb_32.c leave_mm(cpu); cpu 118 arch/x86/kernel/tlb_32.c cpu_clear(cpu, flush_cpumask); cpu 232 arch/x86/kernel/tlb_32.c unsigned long cpu = smp_processor_id(); cpu 235 arch/x86/kernel/tlb_32.c if (per_cpu(cpu_tlbstate, cpu).state == TLBSTATE_LAZY) cpu 236 arch/x86/kernel/tlb_32.c leave_mm(cpu); cpu 246 arch/x86/kernel/tlb_32.c int cpu = raw_smp_processor_id(); cpu 248 arch/x86/kernel/tlb_32.c per_cpu(cpu_tlbstate, cpu).state = 0; cpu 249 arch/x86/kernel/tlb_32.c per_cpu(cpu_tlbstate, cpu).active_mm = &init_mm; cpu 67 arch/x86/kernel/tlb_64.c cpu_clear(cpu, read_pda(active_mm)->cpu_vm_mask); cpu 122 arch/x86/kernel/tlb_64.c int cpu; cpu 126 arch/x86/kernel/tlb_64.c cpu = smp_processor_id(); cpu 134 arch/x86/kernel/tlb_64.c if (!cpu_isset(cpu, f->flush_cpumask)) cpu 152 arch/x86/kernel/tlb_64.c leave_mm(cpu); cpu 156 arch/x86/kernel/tlb_64.c cpu_clear(cpu, f->flush_cpumask); cpu 269 arch/x86/kernel/tlb_64.c unsigned long cpu = smp_processor_id(); cpu 273 arch/x86/kernel/tlb_64.c leave_mm(cpu); cpu 66 arch/x86/kernel/tlb_uv.c int cpu; cpu 69 arch/x86/kernel/tlb_uv.c cpu = uv_blade_processor_id(); cpu 72 arch/x86/kernel/tlb_uv.c this_cpu_mask = 1UL << cpu; cpu 231 arch/x86/kernel/tlb_uv.c if (cpu < UV_CPUS_PER_ACT_STATUS) { cpu 233 arch/x86/kernel/tlb_uv.c right_shift = cpu * UV_ACT_STATUS_SIZE; cpu 237 arch/x86/kernel/tlb_uv.c ((cpu - UV_CPUS_PER_ACT_STATUS) * UV_ACT_STATUS_SIZE); cpu 243 arch/x86/kernel/tlb_uv.c cpu; cpu 305 arch/x86/kernel/tlb_uv.c int cpu; cpu 310 arch/x86/kernel/tlb_uv.c cpu = uv_blade_processor_id(); cpu 313 arch/x86/kernel/tlb_uv.c bau_desc += UV_ITEMS_PER_DESCRIPTOR * cpu; cpu 343 arch/x86/kernel/tlb_uv.c return uv_flush_send_and_wait(cpu, this_blade, bau_desc, cpumaskp); cpu 458 arch/x86/kernel/tlb_uv.c int cpu; cpu 460 arch/x86/kernel/tlb_uv.c cpu = *(loff_t *)data; cpu 462 arch/x86/kernel/tlb_uv.c if (!cpu) { cpu 468 arch/x86/kernel/tlb_uv.c if (cpu < num_possible_cpus() && cpu_online(cpu)) { cpu 469 arch/x86/kernel/tlb_uv.c stat = &per_cpu(ptcstats, cpu); cpu 471 arch/x86/kernel/tlb_uv.c cpu, stat->requestor, cpu 476 arch/x86/kernel/tlb_uv.c (uv_cpu_to_blade_id(cpu)), cpu 36 arch/x86/kernel/tls.c int cpu; cpu 41 arch/x86/kernel/tls.c cpu = get_cpu(); cpu 53 arch/x86/kernel/tls.c load_TLS(t, cpu); cpu 49 arch/x86/kernel/topology.c per_cpu(cpu_devices, num).cpu.hotpluggable = 1; cpu 50 arch/x86/kernel/topology.c return register_cpu(&per_cpu(cpu_devices, num).cpu, num); cpu 56 arch/x86/kernel/topology.c unregister_cpu(&per_cpu(cpu_devices, num).cpu); cpu 62 arch/x86/kernel/topology.c return register_cpu(&per_cpu(cpu_devices, num).cpu, num); cpu 132 arch/x86/kernel/traps.c int cpu; cpu 134 arch/x86/kernel/traps.c cpu = get_cpu(); cpu 135 arch/x86/kernel/traps.c tss = &per_cpu(init_tss, cpu); cpu 435 arch/x86/kernel/traps.c int cpu; cpu 437 arch/x86/kernel/traps.c cpu = smp_processor_id(); cpu 440 arch/x86/kernel/traps.c if (!cpu) cpu 454 arch/x86/kernel/traps.c if (!do_nmi_callback(regs, cpu)) cpu 485 arch/x86/kernel/traps.c { int cpu; cpu = smp_processor_id(); ++nmi_count(cpu); } cpu 566 arch/x86/kernel/tsc.c scale = &per_cpu(cyc2ns, cpu); cpu 601 arch/x86/kernel/tsc.c if (cpu_has(&cpu_data(freq->cpu), X86_FEATURE_CONSTANT_TSC)) cpu 607 arch/x86/kernel/tsc.c lpj = &cpu_data(freq->cpu).loops_per_jiffy; cpu 627 arch/x86/kernel/tsc.c set_cyc2ns_scale(tsc_khz, freq->cpu); cpu 797 arch/x86/kernel/tsc.c int cpu; cpu 830 arch/x86/kernel/tsc.c for_each_possible_cpu(cpu) cpu 831 arch/x86/kernel/tsc.c set_cyc2ns_scale(cpu_khz, cpu); cpu 112 arch/x86/kernel/tsc_sync.c smp_processor_id(), cpu); cpu 198 arch/x86/kernel/vmi_32.c struct desc_struct *gdt = get_cpu_gdt_table(cpu); cpu 206 arch/x86/kernel/vmi_32.c unsigned cpu = smp_processor_id(); cpu 212 arch/x86/kernel/vmi_32.c write_gdt_entry(get_cpu_gdt_table(cpu), GDT_ENTRY_LDT, &desc, DESC_LDT); cpu 214 arch/x86/kernel/vmiclock_32.c int cpu = smp_processor_id(); cpu 229 arch/x86/kernel/vmiclock_32.c evt->cpumask = cpumask_of_cpu(cpu); cpu 213 arch/x86/kernel/vsyscall_64.c if (cpu) cpu 214 arch/x86/kernel/vsyscall_64.c *cpu = p & 0xfff; cpu 256 arch/x86/kernel/vsyscall_64.c node = cpu_to_node(cpu); cpu 258 arch/x86/kernel/vsyscall_64.c if (cpu_has(&cpu_data(cpu), X86_FEATURE_RDTSCP)) cpu 259 arch/x86/kernel/vsyscall_64.c write_rdtscp_aux((node << 12) | cpu); cpu 265 arch/x86/kernel/vsyscall_64.c d |= cpu; cpu 268 arch/x86/kernel/vsyscall_64.c write_gdt_entry(get_cpu_gdt_table(cpu), GDT_ENTRY_PER_CPU, &d, DESCTYPE_S); cpu 280 arch/x86/kernel/vsyscall_64.c long cpu = (long)arg; cpu 282 arch/x86/kernel/vsyscall_64.c smp_call_function_single(cpu, cpu_vsyscall_init, NULL, 1); cpu 84 arch/x86/kvm/svm.c int cpu; cpu 98 arch/x86/kvm/svm.c int cpu; cpu 340 arch/x86/kvm/svm.c svm_data->cpu = cpu; cpu 346 arch/x86/kvm/svm.c per_cpu(svm_data, cpu) = svm_data; cpu 420 arch/x86/kvm/svm.c int cpu; cpu 438 arch/x86/kvm/svm.c for_each_online_cpu(cpu) { cpu 439 arch/x86/kvm/svm.c r = svm_cpu_init(cpu); cpu 470 arch/x86/kvm/svm.c int cpu; cpu 472 arch/x86/kvm/svm.c for_each_online_cpu(cpu) cpu 473 arch/x86/kvm/svm.c svm_cpu_uninit(cpu); cpu 694 arch/x86/kvm/svm.c if (unlikely(cpu != vcpu->cpu)) { cpu 704 arch/x86/kvm/svm.c vcpu->cpu = cpu; cpu 959 arch/x86/kvm/svm.c svm->vcpu.cpu = svm_data->cpu; cpu 1515 arch/x86/kvm/svm.c int cpu = raw_smp_processor_id(); cpu 1517 arch/x86/kvm/svm.c struct svm_cpu_data *svm_data = per_cpu(svm_data, cpu); cpu 1524 arch/x86/kvm/svm.c int cpu = raw_smp_processor_id(); cpu 1526 arch/x86/kvm/svm.c struct svm_cpu_data *svm_data = per_cpu(svm_data, cpu); cpu 1529 arch/x86/kvm/svm.c if (svm->vcpu.cpu != cpu || cpu 335 arch/x86/kvm/vmx.c int cpu = raw_smp_processor_id(); cpu 337 arch/x86/kvm/vmx.c if (vmx->vcpu.cpu == cpu) cpu 339 arch/x86/kvm/vmx.c if (per_cpu(current_vmcs, cpu) == vmx->vmcs) cpu 340 arch/x86/kvm/vmx.c per_cpu(current_vmcs, cpu) = NULL; cpu 343 arch/x86/kvm/vmx.c vmx->vcpu.cpu = -1; cpu 349 arch/x86/kvm/vmx.c if (vmx->vcpu.cpu == -1) cpu 351 arch/x86/kvm/vmx.c smp_call_function_single(vmx->vcpu.cpu, __vcpu_clear, vmx, 1); cpu 626 arch/x86/kvm/vmx.c if (vcpu->cpu != cpu) { cpu 632 arch/x86/kvm/vmx.c &per_cpu(vcpus_on_cpu, cpu)); cpu 636 arch/x86/kvm/vmx.c if (per_cpu(current_vmcs, cpu) != vmx->vmcs) { cpu 639 arch/x86/kvm/vmx.c per_cpu(current_vmcs, cpu) = vmx->vmcs; cpu 648 arch/x86/kvm/vmx.c if (vcpu->cpu != cpu) { cpu 652 arch/x86/kvm/vmx.c vcpu->cpu = cpu; cpu 1042 arch/x86/kvm/vmx.c int cpu = raw_smp_processor_id(); cpu 1043 arch/x86/kvm/vmx.c u64 phys_addr = __pa(per_cpu(vmxarea, cpu)); cpu 1046 arch/x86/kvm/vmx.c INIT_LIST_HEAD(&per_cpu(vcpus_on_cpu, cpu)); cpu 1064 arch/x86/kvm/vmx.c int cpu = raw_smp_processor_id(); cpu 1067 arch/x86/kvm/vmx.c list_for_each_entry_safe(vmx, n, &per_cpu(vcpus_on_cpu, cpu), cpu 1207 arch/x86/kvm/vmx.c int node = cpu_to_node(cpu); cpu 1232 arch/x86/kvm/vmx.c int cpu; cpu 1234 arch/x86/kvm/vmx.c for_each_online_cpu(cpu) cpu 1235 arch/x86/kvm/vmx.c free_vmcs(per_cpu(vmxarea, cpu)); cpu 1240 arch/x86/kvm/vmx.c int cpu; cpu 1242 arch/x86/kvm/vmx.c for_each_online_cpu(cpu) { cpu 1245 arch/x86/kvm/vmx.c vmcs = alloc_vmcs_cpu(cpu); cpu 1251 arch/x86/kvm/vmx.c per_cpu(vmxarea, cpu) = vmcs; cpu 3115 arch/x86/kvm/vmx.c int cpu; cpu 3142 arch/x86/kvm/vmx.c cpu = get_cpu(); cpu 3143 arch/x86/kvm/vmx.c vmx_vcpu_load(&vmx->vcpu, cpu); cpu 973 arch/x86/kvm/x86.c kvm_x86_ops->vcpu_load(vcpu, cpu); cpu 4043 arch/x86/kvm/x86.c int ipi_pcpu = vcpu->cpu; cpu 4044 arch/x86/kvm/x86.c int cpu = get_cpu(); cpu 4054 arch/x86/kvm/x86.c if (vcpu->guest_mode && vcpu->cpu != cpu) cpu 282 arch/x86/lguest/boot.c lazy_hcall(LHCALL_LOAD_TLS, __pa(&t->tls_array), cpu, 0); cpu 54 arch/x86/lib/delay.c int cpu; cpu 57 arch/x86/lib/delay.c cpu = smp_processor_id(); cpu 78 arch/x86/lib/delay.c if (unlikely(cpu != smp_processor_id())) { cpu 80 arch/x86/lib/delay.c cpu = smp_processor_id(); cpu 32 arch/x86/lib/msr-on-cpu.c err = smp_call_function_single(cpu, __rdmsr_on_cpu, &rv, 1); cpu 47 arch/x86/lib/msr-on-cpu.c err = smp_call_function_single(cpu, __wrmsr_on_cpu, &rv, 1); cpu 74 arch/x86/lib/msr-on-cpu.c err = smp_call_function_single(cpu, __rdmsr_safe_on_cpu, &rv, 1); cpu 89 arch/x86/lib/msr-on-cpu.c err = smp_call_function_single(cpu, __wrmsr_safe_on_cpu, &rv, 1); cpu 731 arch/x86/mach-voyager/voyager_cat.c __u8 cpu = i & 0x07; cpu 735 arch/x86/mach-voyager/voyager_cat.c voyager_extended_vic_processors |= (1 << cpu); cpu 736 arch/x86/mach-voyager/voyager_cat.c cpu += 4; cpu 737 arch/x86/mach-voyager/voyager_cat.c voyager_extended_vic_processors |= (1 << cpu); cpu 883 arch/x86/mach-voyager/voyager_cat.c __u8 cpu; cpu 889 arch/x86/mach-voyager/voyager_cat.c cpu = (i & 0x07) + j * 8; cpu 891 arch/x86/mach-voyager/voyager_cat.c cpu = (i & 0x03) + j * 4; cpu 894 arch/x86/mach-voyager/voyager_cat.c voyager_extended_vic_processors |= (1 << cpu); cpu 901 arch/x86/mach-voyager/voyager_cat.c voyager_extended_vic_processors |= (1 << cpu); cpu 903 arch/x86/mach-voyager/voyager_cat.c (~(1 << cpu)); cpu 906 arch/x86/mach-voyager/voyager_cat.c voyager_quad_processors |= (1 << cpu); cpu 907 arch/x86/mach-voyager/voyager_cat.c voyager_quad_cpi_addr[cpu] = (struct voyager_qic_cpi *) cpu 909 arch/x86/mach-voyager/voyager_cat.c CDEBUG(("CPU%d: CPI address 0x%lx\n", cpu, cpu 910 arch/x86/mach-voyager/voyager_cat.c (unsigned long)voyager_quad_cpi_addr[cpu])); cpu 100 arch/x86/mach-voyager/voyager_smp.c voyager_quad_cpi_addr[cpu]->qic_cpi[cpi].cpi = cpu 106 arch/x86/mach-voyager/voyager_smp.c int cpu; cpu 108 arch/x86/mach-voyager/voyager_smp.c for_each_online_cpu(cpu) { cpu 109 arch/x86/mach-voyager/voyager_smp.c if (cpuset & (1 << cpu)) { cpu 111 arch/x86/mach-voyager/voyager_smp.c if (!cpu_online(cpu)) cpu 114 arch/x86/mach-voyager/voyager_smp.c hard_smp_processor_id(), cpi, cpu)); cpu 116 arch/x86/mach-voyager/voyager_smp.c send_one_QIC_CPI(cpu, cpi - QIC_CPI_OFFSET); cpu 130 arch/x86/mach-voyager/voyager_smp.c if (voyager_quad_processors & (1 << cpu)) cpu 131 arch/x86/mach-voyager/voyager_smp.c send_one_QIC_CPI(cpu, cpi - QIC_CPI_OFFSET); cpu 133 arch/x86/mach-voyager/voyager_smp.c send_CPI(1 << cpu, cpi); cpu 138 arch/x86/mach-voyager/voyager_smp.c __u8 cpu = smp_processor_id(); cpu 139 arch/x86/mach-voyager/voyager_smp.c __u32 mask = cpus_addr(cpu_online_map)[0] & ~(1 << cpu); cpu 151 arch/x86/mach-voyager/voyager_smp.c __u8 cpu = hard_smp_processor_id(); cpu 153 arch/x86/mach-voyager/voyager_smp.c return (voyager_extended_vic_processors & (1 << cpu)); cpu 158 arch/x86/mach-voyager/voyager_smp.c __u8 cpu = hard_smp_processor_id(); cpu 161 arch/x86/mach-voyager/voyager_smp.c & voyager_allowed_boot_processors & (1 << cpu)); cpu 404 arch/x86/mach-voyager/voyager_smp.c current_thread_info()->cpu = boot_cpu_id; cpu 498 arch/x86/mach-voyager/voyager_smp.c int quad_boot = (1 << cpu) & voyager_quad_processors cpu 525 arch/x86/mach-voyager/voyager_smp.c idle = fork_idle(cpu); cpu 527 arch/x86/mach-voyager/voyager_smp.c panic("failed fork for CPU%d", cpu); cpu 532 arch/x86/mach-voyager/voyager_smp.c init_gdt(cpu); cpu 533 arch/x86/mach-voyager/voyager_smp.c per_cpu(current_task, cpu) = idle; cpu 534 arch/x86/mach-voyager/voyager_smp.c early_gdt_descr.address = (unsigned long)get_cpu_gdt_table(cpu); cpu 535 arch/x86/mach-voyager/voyager_smp.c irq_ctx_init(cpu); cpu 538 arch/x86/mach-voyager/voyager_smp.c VDEBUG(("VOYAGER SMP: Booting CPU%d at 0x%lx[%x:%x], stack %p\n", cpu, cpu 548 arch/x86/mach-voyager/voyager_smp.c printk("CPU %d: non extended Quad boot\n", cpu); cpu 554 arch/x86/mach-voyager/voyager_smp.c printk("CPU%d: extended VIC boot\n", cpu); cpu 572 arch/x86/mach-voyager/voyager_smp.c send_one_QIC_CPI(cpu, VIC_CPU_BOOT_CPI); cpu 574 arch/x86/mach-voyager/voyager_smp.c outb(VIC_CPU_MASQUERADE_ENABLE | cpu, VIC_PROCESSOR_ID); cpu 583 arch/x86/mach-voyager/voyager_smp.c send_CPI((1 << cpu), VIC_CPU_BOOT_CPI); cpu 599 arch/x86/mach-voyager/voyager_smp.c cpu, smp_processor_id())); cpu 601 arch/x86/mach-voyager/voyager_smp.c printk("CPU%d: ", cpu); cpu 602 arch/x86/mach-voyager/voyager_smp.c print_cpu_info(&cpu_data(cpu)); cpu 604 arch/x86/mach-voyager/voyager_smp.c cpu_set(cpu, cpu_callout_map); cpu 605 arch/x86/mach-voyager/voyager_smp.c cpu_set(cpu, cpu_present_map); cpu 607 arch/x86/mach-voyager/voyager_smp.c printk("CPU%d FAILED TO BOOT: ", cpu); cpu 783 arch/x86/mach-voyager/voyager_smp.c if (per_cpu(cpu_tlbstate, cpu).state == TLBSTATE_OK) cpu 785 arch/x86/mach-voyager/voyager_smp.c cpu_clear(cpu, per_cpu(cpu_tlbstate, cpu).active_mm->cpu_vm_mask); cpu 794 arch/x86/mach-voyager/voyager_smp.c __u8 cpu = smp_processor_id(); cpu 796 arch/x86/mach-voyager/voyager_smp.c if (!test_bit(cpu, &smp_invalidate_needed)) cpu 804 arch/x86/mach-voyager/voyager_smp.c if (flush_mm == per_cpu(cpu_tlbstate, cpu).active_mm) { cpu 805 arch/x86/mach-voyager/voyager_smp.c if (per_cpu(cpu_tlbstate, cpu).state == TLBSTATE_OK) { cpu 811 arch/x86/mach-voyager/voyager_smp.c voyager_leave_mm(cpu); cpu 814 arch/x86/mach-voyager/voyager_smp.c clear_bit(cpu, &smp_invalidate_needed); cpu 927 arch/x86/mach-voyager/voyager_smp.c __u8 cpu = get_cpu(); cpu 929 arch/x86/mach-voyager/voyager_smp.c VDEBUG(("VOYAGER SMP: CPU%d enabling irq mask 0x%x\n", cpu, cpu 930 arch/x86/mach-voyager/voyager_smp.c vic_irq_enable_mask[cpu])); cpu 934 arch/x86/mach-voyager/voyager_smp.c if (vic_irq_enable_mask[cpu] & (1 << irq)) cpu 937 arch/x86/mach-voyager/voyager_smp.c vic_irq_enable_mask[cpu] = 0; cpu 1038 arch/x86/mach-voyager/voyager_smp.c __u8 cpu = smp_processor_id(); cpu 1045 arch/x86/mach-voyager/voyager_smp.c if (test_and_clear_bit(VIC_TIMER_CPI, &vic_cpi_mailbox[cpu])) cpu 1047 arch/x86/mach-voyager/voyager_smp.c if (test_and_clear_bit(VIC_INVALIDATE_CPI, &vic_cpi_mailbox[cpu])) cpu 1049 arch/x86/mach-voyager/voyager_smp.c if (test_and_clear_bit(VIC_RESCHEDULE_CPI, &vic_cpi_mailbox[cpu])) cpu 1051 arch/x86/mach-voyager/voyager_smp.c if (test_and_clear_bit(VIC_ENABLE_IRQ_CPI, &vic_cpi_mailbox[cpu])) cpu 1053 arch/x86/mach-voyager/voyager_smp.c if (test_and_clear_bit(VIC_CALL_FUNCTION_CPI, &vic_cpi_mailbox[cpu])) cpu 1055 arch/x86/mach-voyager/voyager_smp.c if (test_and_clear_bit(VIC_CALL_FUNCTION_SINGLE_CPI, &vic_cpi_mailbox[cpu])) cpu 1062 arch/x86/mach-voyager/voyager_smp.c unsigned long cpu = smp_processor_id(); cpu 1065 arch/x86/mach-voyager/voyager_smp.c if (per_cpu(cpu_tlbstate, cpu).state == TLBSTATE_LAZY) cpu 1066 arch/x86/mach-voyager/voyager_smp.c voyager_leave_mm(cpu); cpu 1078 arch/x86/mach-voyager/voyager_smp.c send_one_CPI(cpu, VIC_RESCHEDULE_CPI); cpu 1125 arch/x86/mach-voyager/voyager_smp.c int cpu = smp_processor_id(); cpu 1129 arch/x86/mach-voyager/voyager_smp.c if (--per_cpu(prof_counter, cpu) <= 0) { cpu 1138 arch/x86/mach-voyager/voyager_smp.c per_cpu(prof_counter, cpu) = per_cpu(prof_multiplier, cpu); cpu 1139 arch/x86/mach-voyager/voyager_smp.c if (per_cpu(prof_counter, cpu) != cpu 1140 arch/x86/mach-voyager/voyager_smp.c per_cpu(prof_old_multiplier, cpu)) { cpu 1142 arch/x86/mach-voyager/voyager_smp.c per_cpu(prof_old_multiplier, cpu) = cpu 1143 arch/x86/mach-voyager/voyager_smp.c per_cpu(prof_counter, cpu); cpu 1149 arch/x86/mach-voyager/voyager_smp.c if (((1 << cpu) & voyager_extended_vic_processors) == 0) cpu 1165 arch/x86/mach-voyager/voyager_smp.c if ((++vic_tick[cpu] & 0x7) != 0) cpu 1194 arch/x86/mach-voyager/voyager_smp.c weight = (vic_intr_count[cpu] * voyager_extended_cpus cpu 1205 arch/x86/mach-voyager/voyager_smp.c if ((vic_tick[cpu] & 0xFFF) == 0) { cpu 1208 arch/x86/mach-voyager/voyager_smp.c cpu, vic_tick[cpu], weight); cpu 1282 arch/x86/mach-voyager/voyager_smp.c int cpu; cpu 1297 arch/x86/mach-voyager/voyager_smp.c for_each_online_cpu(cpu) { cpu 1298 arch/x86/mach-voyager/voyager_smp.c if (cpuset & (1 << cpu)) cpu 1299 arch/x86/mach-voyager/voyager_smp.c set_bit(cpi, &vic_cpi_mailbox[cpu]); cpu 1313 arch/x86/mach-voyager/voyager_smp.c __u8 cpu = hard_smp_processor_id(); cpu 1318 arch/x86/mach-voyager/voyager_smp.c return voyager_quad_cpi_addr[cpu]->qic_cpi[cpi].cpi; cpu 1341 arch/x86/mach-voyager/voyager_smp.c __u8 cpu = smp_processor_id(); cpu 1346 arch/x86/mach-voyager/voyager_smp.c printk("VOYAGER SMP: CPU%d lost CPI%d\n", cpu, cpi); cpu 1356 arch/x86/mach-voyager/voyager_smp.c printk("VOYAGER SMP: CPU%d still asserting CPI%d\n", cpu, cpi); cpu 1364 arch/x86/mach-voyager/voyager_smp.c #define cached_21(cpu) (__byte(0,vic_irq_mask[cpu])) cpu 1365 arch/x86/mach-voyager/voyager_smp.c #define cached_A1(cpu) (__byte(1,vic_irq_mask[cpu])) cpu 1399 arch/x86/mach-voyager/voyager_smp.c int cpu = smp_processor_id(), real_cpu; cpu 1405 arch/x86/mach-voyager/voyager_smp.c irq, cpu, cpu_irq_affinity[cpu])); cpu 1414 arch/x86/mach-voyager/voyager_smp.c if (real_cpu == cpu) { cpu 1433 arch/x86/mach-voyager/voyager_smp.c __u8 cpu = smp_processor_id(); cpu 1435 arch/x86/mach-voyager/voyager_smp.c __u16 old_mask = vic_irq_mask[cpu]; cpu 1437 arch/x86/mach-voyager/voyager_smp.c vic_irq_mask[cpu] &= mask; cpu 1438 arch/x86/mach-voyager/voyager_smp.c if (vic_irq_mask[cpu] == old_mask) cpu 1442 arch/x86/mach-voyager/voyager_smp.c irq, cpu)); cpu 1445 arch/x86/mach-voyager/voyager_smp.c outb_p(cached_A1(cpu), 0xA1); cpu 1448 arch/x86/mach-voyager/voyager_smp.c outb_p(cached_21(cpu), 0x21); cpu 1455 arch/x86/mach-voyager/voyager_smp.c __u8 cpu = smp_processor_id(); cpu 1457 arch/x86/mach-voyager/voyager_smp.c __u16 old_mask = vic_irq_mask[cpu]; cpu 1462 arch/x86/mach-voyager/voyager_smp.c vic_irq_mask[cpu] |= mask; cpu 1463 arch/x86/mach-voyager/voyager_smp.c if (old_mask == vic_irq_mask[cpu]) cpu 1467 arch/x86/mach-voyager/voyager_smp.c irq, cpu)); cpu 1470 arch/x86/mach-voyager/voyager_smp.c outb_p(cached_A1(cpu), 0xA1); cpu 1473 arch/x86/mach-voyager/voyager_smp.c outb_p(cached_21(cpu), 0x21); cpu 1487 arch/x86/mach-voyager/voyager_smp.c __u8 cpu = smp_processor_id(); cpu 1491 arch/x86/mach-voyager/voyager_smp.c vic_intr_count[cpu]++; cpu 1493 arch/x86/mach-voyager/voyager_smp.c if (!(cpu_irq_affinity[cpu] & (1 << irq))) { cpu 1497 arch/x86/mach-voyager/voyager_smp.c "on cpu %d\n", irq, cpu)); cpu 1508 arch/x86/mach-voyager/voyager_smp.c irq, cpu)); cpu 1538 arch/x86/mach-voyager/voyager_smp.c __u8 cpu = smp_processor_id(); cpu 1543 arch/x86/mach-voyager/voyager_smp.c cpu, irq); cpu 1555 arch/x86/mach-voyager/voyager_smp.c outb(cpu, VIC_PROCESSOR_ID); cpu 1604 arch/x86/mach-voyager/voyager_smp.c int cpu; cpu 1628 arch/x86/mach-voyager/voyager_smp.c for_each_online_cpu(cpu) { cpu 1629 arch/x86/mach-voyager/voyager_smp.c unsigned long cpu_mask = 1 << cpu; cpu 1633 arch/x86/mach-voyager/voyager_smp.c cpu_irq_affinity[cpu] |= irq_mask; cpu 1636 arch/x86/mach-voyager/voyager_smp.c cpu_irq_affinity[cpu] &= ~irq_mask; cpu 1665 arch/x86/mach-voyager/voyager_smp.c __u8 cpu = smp_processor_id(); cpu 1668 arch/x86/mach-voyager/voyager_smp.c vic_irq_mask[cpu] = vic_irq_mask[boot_cpu_id]; cpu 1679 arch/x86/mach-voyager/voyager_smp.c cpu, QIC_CPI_ENABLE)); cpu 1683 arch/x86/mach-voyager/voyager_smp.c cpu, vic_irq_mask[cpu])); cpu 1688 arch/x86/mach-voyager/voyager_smp.c int old_cpu = smp_processor_id(), cpu; cpu 1691 arch/x86/mach-voyager/voyager_smp.c for_each_online_cpu(cpu) { cpu 1696 arch/x86/mach-voyager/voyager_smp.c outb(VIC_CPU_MASQUERADE_ENABLE | cpu, VIC_PROCESSOR_ID); cpu 1709 arch/x86/mach-voyager/voyager_smp.c cpu, vic_irq_mask[cpu], imr, irr, isr); cpu 1717 arch/x86/mach-voyager/voyager_smp.c cpu, irq); cpu 1719 arch/x86/mach-voyager/voyager_smp.c outb(VIC_CPU_MASQUERADE_ENABLE | cpu, cpu 1759 arch/x86/mach-voyager/voyager_smp.c if (cpu_isset(cpu, smp_commenced_mask)) cpu 1763 arch/x86/mach-voyager/voyager_smp.c if (!cpu_isset(cpu, cpu_callin_map)) cpu 1766 arch/x86/mach-voyager/voyager_smp.c cpu_set(cpu, smp_commenced_mask); cpu 1767 arch/x86/mach-voyager/voyager_smp.c while (!cpu_online(cpu)) cpu 1779 arch/x86/mach-voyager/voyager_smp.c current_thread_info()->cpu = hard_smp_processor_id(); cpu 411 arch/x86/mm/mmio-mod.c int cpu; cpu 421 arch/x86/mm/mmio-mod.c for_each_cpu_mask(cpu, downed_cpus) { cpu 422 arch/x86/mm/mmio-mod.c err = cpu_down(cpu); cpu 424 arch/x86/mm/mmio-mod.c pr_info(NAME "CPU%d is down.\n", cpu); cpu 426 arch/x86/mm/mmio-mod.c pr_err(NAME "Error taking CPU%d down: %d\n", cpu, err); cpu 437 arch/x86/mm/mmio-mod.c int cpu; cpu 443 arch/x86/mm/mmio-mod.c for_each_cpu_mask(cpu, downed_cpus) { cpu 444 arch/x86/mm/mmio-mod.c err = cpu_up(cpu); cpu 446 arch/x86/mm/mmio-mod.c pr_info(NAME "enabled CPU%d.\n", cpu); cpu 448 arch/x86/mm/mmio-mod.c pr_err(NAME "cannot re-enable CPU%d: %d\n", cpu, err); cpu 621 arch/x86/mm/numa_64.c int cpu; cpu 626 arch/x86/mm/numa_64.c for_each_possible_cpu(cpu) { cpu 628 arch/x86/mm/numa_64.c u16 apicid = cpu_to_apicid[cpu]; cpu 637 arch/x86/mm/numa_64.c numa_set_node(cpu, node); cpu 43 arch/x86/oprofile/nmi_int.c int cpu = (unsigned long)data; cpu 47 arch/x86/oprofile/nmi_int.c smp_call_function_single(cpu, nmi_cpu_start, NULL, 0); cpu 50 arch/x86/oprofile/nmi_int.c smp_call_function_single(cpu, nmi_cpu_stop, NULL, 1); cpu 115 arch/x86/oprofile/nmi_int.c int cpu = smp_processor_id(); cpu 119 arch/x86/oprofile/nmi_int.c if (model->check_ctrs(args->regs, &per_cpu(cpu_msrs, cpu))) cpu 155 arch/x86/oprofile/nmi_int.c int cpu = smp_processor_id(); cpu 156 arch/x86/oprofile/nmi_int.c struct op_msrs *msrs = &per_cpu(cpu_msrs, cpu); cpu 201 arch/x86/oprofile/nmi_int.c int cpu = smp_processor_id(); cpu 202 arch/x86/oprofile/nmi_int.c struct op_msrs *msrs = &per_cpu(cpu_msrs, cpu); cpu 206 arch/x86/oprofile/nmi_int.c per_cpu(saved_lvtpc, cpu) = apic_read(APIC_LVTPC); cpu 219 arch/x86/oprofile/nmi_int.c int cpu; cpu 236 arch/x86/oprofile/nmi_int.c for_each_possible_cpu(cpu) { cpu 237 arch/x86/oprofile/nmi_int.c if (cpu != 0) { cpu 238 arch/x86/oprofile/nmi_int.c memcpy(per_cpu(cpu_msrs, cpu).counters, cpu 242 arch/x86/oprofile/nmi_int.c memcpy(per_cpu(cpu_msrs, cpu).controls, cpu 282 arch/x86/oprofile/nmi_int.c int cpu = smp_processor_id(); cpu 292 arch/x86/oprofile/nmi_int.c apic_write(APIC_LVTPC, per_cpu(saved_lvtpc, cpu)); cpu 382 arch/x86/oprofile/op_model_p4.c int cpu = smp_processor_id(); cpu 383 arch/x86/oprofile/op_model_p4.c return (cpu != first_cpu(per_cpu(cpu_sibling_map, cpu))); cpu 582 arch/x86/pci/amd_bus.c int cpu = (long)hcpu; cpu 586 arch/x86/pci/amd_bus.c smp_call_function_single(cpu, enable_pci_io_ecs, NULL, 0); cpu 600 arch/x86/pci/amd_bus.c int cpu; cpu 607 arch/x86/pci/amd_bus.c for_each_online_cpu(cpu) cpu 609 arch/x86/pci/amd_bus.c (void *)(long)cpu); cpu 51 arch/x86/pci/mmconfig_32.c int cpu = smp_processor_id(); cpu 53 arch/x86/pci/mmconfig_32.c cpu != mmcfg_last_accessed_cpu) { cpu 55 arch/x86/pci/mmconfig_32.c mmcfg_last_accessed_cpu = cpu; cpu 69 arch/x86/power/cpu_32.c int cpu = smp_processor_id(); cpu 70 arch/x86/power/cpu_32.c struct tss_struct *t = &per_cpu(init_tss, cpu); cpu 72 arch/x86/power/cpu_32.c set_tss_desc(cpu, t); /* cpu 145 arch/x86/power/cpu_64.c int cpu = smp_processor_id(); cpu 146 arch/x86/power/cpu_64.c struct tss_struct *t = &per_cpu(init_tss, cpu); cpu 153 arch/x86/power/cpu_64.c set_tss_desc(cpu, t); cpu 155 arch/x86/power/cpu_64.c get_cpu_gdt_table(cpu)[GDT_ENTRY_TSS].type = 9; cpu 228 arch/x86/vdso/vdso32-setup.c int cpu = get_cpu(); cpu 229 arch/x86/vdso/vdso32-setup.c struct tss_struct *tss = &per_cpu(init_tss, cpu); cpu 28 arch/x86/vdso/vgetcpu.c if (cpu) cpu 29 arch/x86/vdso/vgetcpu.c *cpu = p & 0xfff; cpu 132 arch/x86/xen/enlighten.c per_cpu(xen_vcpu, cpu) = &HYPERVISOR_shared_info->vcpu_info[cpu]; cpu 137 arch/x86/xen/enlighten.c vcpup = &per_cpu(xen_vcpu_info, cpu); cpu 143 arch/x86/xen/enlighten.c cpu, vcpup, info.mfn, info.offset); cpu 148 arch/x86/xen/enlighten.c err = HYPERVISOR_vcpu_op(VCPUOP_register_vcpu_info, cpu, &info); cpu 156 arch/x86/xen/enlighten.c per_cpu(xen_vcpu, cpu) = vcpup; cpu 159 arch/x86/xen/enlighten.c cpu, vcpup); cpu 171 arch/x86/xen/enlighten.c int cpu; cpu 173 arch/x86/xen/enlighten.c for_each_online_cpu(cpu) { cpu 174 arch/x86/xen/enlighten.c bool other_cpu = (cpu != smp_processor_id()); cpu 177 arch/x86/xen/enlighten.c HYPERVISOR_vcpu_op(VCPUOP_down, cpu, NULL)) cpu 180 arch/x86/xen/enlighten.c xen_vcpu_setup(cpu); cpu 183 arch/x86/xen/enlighten.c HYPERVISOR_vcpu_op(VCPUOP_up, cpu, NULL)) cpu 349 arch/x86/xen/enlighten.c struct desc_struct *gdt = get_cpu_gdt_table(cpu); cpu 385 arch/x86/xen/enlighten.c load_TLS_descriptor(t, cpu, 0); cpu 386 arch/x86/xen/enlighten.c load_TLS_descriptor(t, cpu, 1); cpu 387 arch/x86/xen/enlighten.c load_TLS_descriptor(t, cpu, 2); cpu 1062 arch/x86/xen/enlighten.c int cpu; cpu 1064 arch/x86/xen/enlighten.c for_each_possible_cpu(cpu) cpu 1065 arch/x86/xen/enlighten.c xen_vcpu_setup(cpu); cpu 29 arch/x86/xen/irq.c int cpu; cpu 31 arch/x86/xen/irq.c for_each_possible_cpu(cpu) cpu 32 arch/x86/xen/irq.c per_cpu(vector_irq, cpu)[i] = i; cpu 1062 arch/x86/xen/mmu.c unsigned cpu; cpu 1080 arch/x86/xen/mmu.c for_each_online_cpu(cpu) { cpu 1081 arch/x86/xen/mmu.c if (per_cpu(xen_current_cr3, cpu) == __pa(mm->pgd)) cpu 1082 arch/x86/xen/mmu.c cpu_set(cpu, mask); cpu 64 arch/x86/xen/smp.c int cpu = smp_processor_id(); cpu 73 arch/x86/xen/smp.c cpu = smp_processor_id(); cpu 74 arch/x86/xen/smp.c smp_store_cpu_info(cpu); cpu 75 arch/x86/xen/smp.c cpu_data(cpu).x86_max_cores = 1; cpu 76 arch/x86/xen/smp.c set_cpu_sibling_map(cpu); cpu 80 arch/x86/xen/smp.c cpu_set(cpu, cpu_online_map); cpu 101 arch/x86/xen/smp.c resched_name = kasprintf(GFP_KERNEL, "resched%d", cpu); cpu 103 arch/x86/xen/smp.c cpu, cpu 110 arch/x86/xen/smp.c per_cpu(resched_irq, cpu) = rc; cpu 112 arch/x86/xen/smp.c callfunc_name = kasprintf(GFP_KERNEL, "callfunc%d", cpu); cpu 114 arch/x86/xen/smp.c cpu, cpu 121 arch/x86/xen/smp.c per_cpu(callfunc_irq, cpu) = rc; cpu 123 arch/x86/xen/smp.c debug_name = kasprintf(GFP_KERNEL, "debug%d", cpu); cpu 124 arch/x86/xen/smp.c rc = bind_virq_to_irqhandler(VIRQ_DEBUG, cpu, xen_debug_interrupt, cpu 129 arch/x86/xen/smp.c per_cpu(debug_irq, cpu) = rc; cpu 131 arch/x86/xen/smp.c callfunc_name = kasprintf(GFP_KERNEL, "callfuncsingle%d", cpu); cpu 133 arch/x86/xen/smp.c cpu, cpu 140 arch/x86/xen/smp.c per_cpu(callfuncsingle_irq, cpu) = rc; cpu 145 arch/x86/xen/smp.c if (per_cpu(resched_irq, cpu) >= 0) cpu 146 arch/x86/xen/smp.c unbind_from_irqhandler(per_cpu(resched_irq, cpu), NULL); cpu 147 arch/x86/xen/smp.c if (per_cpu(callfunc_irq, cpu) >= 0) cpu 148 arch/x86/xen/smp.c unbind_from_irqhandler(per_cpu(callfunc_irq, cpu), NULL); cpu 149 arch/x86/xen/smp.c if (per_cpu(debug_irq, cpu) >= 0) cpu 150 arch/x86/xen/smp.c unbind_from_irqhandler(per_cpu(debug_irq, cpu), NULL); cpu 151 arch/x86/xen/smp.c if (per_cpu(callfuncsingle_irq, cpu) >= 0) cpu 152 arch/x86/xen/smp.c unbind_from_irqhandler(per_cpu(callfuncsingle_irq, cpu), NULL); cpu 184 arch/x86/xen/smp.c unsigned cpu; cpu 199 arch/x86/xen/smp.c for (cpu = NR_CPUS - 1; !cpu_possible(cpu); cpu--) cpu 201 arch/x86/xen/smp.c cpu_clear(cpu, cpu_possible_map); cpu 204 arch/x86/xen/smp.c for_each_possible_cpu (cpu) { cpu 207 arch/x86/xen/smp.c if (cpu == 0) cpu 210 arch/x86/xen/smp.c idle = fork_idle(cpu); cpu 212 arch/x86/xen/smp.c panic("failed fork for CPU %d", cpu); cpu 214 arch/x86/xen/smp.c cpu_set(cpu, cpu_present_map); cpu 224 arch/x86/xen/smp.c if (cpu_test_and_set(cpu, xen_cpu_initialized_map)) cpu 231 arch/x86/xen/smp.c gdt = get_cpu_gdt_table(cpu); cpu 268 arch/x86/xen/smp.c per_cpu(xen_cr3, cpu) = __pa(swapper_pg_dir); cpu 271 arch/x86/xen/smp.c if (HYPERVISOR_vcpu_op(VCPUOP_initialise, cpu, ctxt)) cpu 280 arch/x86/xen/smp.c struct task_struct *idle = idle_task(cpu); cpu 285 arch/x86/xen/smp.c WARN_ON(cpu == 0); cpu 286 arch/x86/xen/smp.c if (cpu > 0) { cpu 287 arch/x86/xen/smp.c rc = get_local_pda(cpu); cpu 294 arch/x86/xen/smp.c init_gdt(cpu); cpu 295 arch/x86/xen/smp.c per_cpu(current_task, cpu) = idle; cpu 296 arch/x86/xen/smp.c irq_ctx_init(cpu); cpu 298 arch/x86/xen/smp.c cpu_pda(cpu)->pcurrent = idle; cpu 301 arch/x86/xen/smp.c xen_setup_timer(cpu); cpu 302 arch/x86/xen/smp.c xen_init_lock_cpu(cpu); cpu 304 arch/x86/xen/smp.c per_cpu(cpu_state, cpu) = CPU_UP_PREPARE; cpu 307 arch/x86/xen/smp.c per_cpu(xen_vcpu, cpu)->evtchn_upcall_mask = 1; cpu 309 arch/x86/xen/smp.c rc = cpu_initialize_context(cpu, idle); cpu 316 arch/x86/xen/smp.c rc = xen_smp_intr_init(cpu); cpu 320 arch/x86/xen/smp.c rc = HYPERVISOR_vcpu_op(VCPUOP_up, cpu, NULL); cpu 323 arch/x86/xen/smp.c while(per_cpu(cpu_state, cpu) != CPU_ONLINE) { cpu 338 arch/x86/xen/smp.c unsigned int cpu = smp_processor_id(); cpu 339 arch/x86/xen/smp.c if (cpu == 0) cpu 350 arch/x86/xen/smp.c while (HYPERVISOR_vcpu_op(VCPUOP_is_up, cpu, NULL)) { cpu 354 arch/x86/xen/smp.c unbind_from_irqhandler(per_cpu(resched_irq, cpu), NULL); cpu 355 arch/x86/xen/smp.c unbind_from_irqhandler(per_cpu(callfunc_irq, cpu), NULL); cpu 356 arch/x86/xen/smp.c unbind_from_irqhandler(per_cpu(debug_irq, cpu), NULL); cpu 357 arch/x86/xen/smp.c unbind_from_irqhandler(per_cpu(callfuncsingle_irq, cpu), NULL); cpu 358 arch/x86/xen/smp.c xen_uninit_lock_cpu(cpu); cpu 359 arch/x86/xen/smp.c xen_teardown_timer(cpu); cpu 391 arch/x86/xen/smp.c int cpu = smp_processor_id(); cpu 397 arch/x86/xen/smp.c HYPERVISOR_vcpu_op(VCPUOP_down, cpu, NULL); cpu 408 arch/x86/xen/smp.c xen_send_IPI_one(cpu, XEN_RESCHEDULE_VECTOR); cpu 413 arch/x86/xen/smp.c unsigned cpu; cpu 417 arch/x86/xen/smp.c for_each_cpu_mask_nr(cpu, mask) cpu 418 arch/x86/xen/smp.c xen_send_IPI_one(cpu, vector); cpu 423 arch/x86/xen/smp.c int cpu; cpu 428 arch/x86/xen/smp.c for_each_cpu_mask_nr(cpu, mask) { cpu 429 arch/x86/xen/smp.c if (xen_vcpu_stolen(cpu)) { cpu 438 arch/x86/xen/smp.c xen_send_IPI_mask(cpumask_of_cpu(cpu), XEN_CALL_FUNCTION_SINGLE_VECTOR); cpu 303 arch/x86/xen/spinlock.c int cpu; cpu 307 arch/x86/xen/spinlock.c for_each_online_cpu(cpu) { cpu 309 arch/x86/xen/spinlock.c if (per_cpu(lock_spinners, cpu) == xl) { cpu 311 arch/x86/xen/spinlock.c xen_send_IPI_one(cpu, XEN_SPIN_UNLOCK_VECTOR); cpu 344 arch/x86/xen/spinlock.c name = kasprintf(GFP_KERNEL, "spinlock%d", cpu); cpu 346 arch/x86/xen/spinlock.c cpu, cpu 354 arch/x86/xen/spinlock.c per_cpu(lock_kicker_irq, cpu) = irq; cpu 357 arch/x86/xen/spinlock.c printk("cpu %d spinlock event irq %d\n", cpu, irq); cpu 362 arch/x86/xen/spinlock.c unbind_from_irqhandler(per_cpu(lock_kicker_irq, cpu), NULL); cpu 107 arch/x86/xen/time.c area.addr.v = &per_cpu(runstate, cpu); cpu 110 arch/x86/xen/time.c cpu, &area)) cpu 351 arch/x86/xen/time.c int cpu = smp_processor_id(); cpu 359 arch/x86/xen/time.c if (HYPERVISOR_vcpu_op(VCPUOP_stop_periodic_timer, cpu, NULL)) cpu 365 arch/x86/xen/time.c if (HYPERVISOR_vcpu_op(VCPUOP_stop_singleshot_timer, cpu, NULL) || cpu 366 arch/x86/xen/time.c HYPERVISOR_vcpu_op(VCPUOP_stop_periodic_timer, cpu, NULL)) cpu 377 arch/x86/xen/time.c int cpu = smp_processor_id(); cpu 386 arch/x86/xen/time.c ret = HYPERVISOR_vcpu_op(VCPUOP_set_singleshot_timer, cpu, &single); cpu 434 arch/x86/xen/time.c printk(KERN_INFO "installing Xen timer for CPU %d\n", cpu); cpu 436 arch/x86/xen/time.c name = kasprintf(GFP_KERNEL, "timer%d", cpu); cpu 440 arch/x86/xen/time.c irq = bind_virq_to_irqhandler(VIRQ_TIMER, cpu, xen_timer_interrupt, cpu 444 arch/x86/xen/time.c evt = &per_cpu(xen_clock_events, cpu); cpu 447 arch/x86/xen/time.c evt->cpumask = cpumask_of_cpu(cpu); cpu 450 arch/x86/xen/time.c setup_runstate_info(cpu); cpu 456 arch/x86/xen/time.c BUG_ON(cpu == 0); cpu 457 arch/x86/xen/time.c evt = &per_cpu(xen_clock_events, cpu); cpu 470 arch/x86/xen/time.c int cpu; cpu 475 arch/x86/xen/time.c for_each_online_cpu(cpu) { cpu 476 arch/x86/xen/time.c if (HYPERVISOR_vcpu_op(VCPUOP_stop_periodic_timer, cpu, NULL)) cpu 483 arch/x86/xen/time.c int cpu = smp_processor_id(); cpu 487 arch/x86/xen/time.c if (HYPERVISOR_vcpu_op(VCPUOP_stop_periodic_timer, cpu, NULL) == 0) { cpu 501 arch/x86/xen/time.c xen_setup_timer(cpu); cpu 55 block/blk-core.c int cpu; cpu 60 block/blk-core.c cpu = part_stat_lock(); cpu 64 block/blk-core.c part_stat_inc(cpu, part, merges[rw]); cpu 66 block/blk-core.c part_round_stats(cpu, part); cpu 114 block/blk-core.c rq->cpu = -1; cpu 998 block/blk-core.c __part_stat_add(cpu, part, time_in_queue, cpu 1000 block/blk-core.c __part_stat_add(cpu, part, io_ticks, (now - part->stamp)); cpu 1025 block/blk-core.c part_round_stats_single(cpu, &part_to_disk(part)->part0, now); cpu 1026 block/blk-core.c part_round_stats_single(cpu, part, now); cpu 1072 block/blk-core.c req->cpu = bio->bi_comp_cpu; cpu 1154 block/blk-core.c req->cpu = bio->bi_comp_cpu; cpu 1183 block/blk-core.c req->cpu = bio->bi_comp_cpu; cpu 1221 block/blk-core.c req->cpu = blk_cpu_to_group(smp_processor_id()); cpu 1654 block/blk-core.c int cpu; cpu 1656 block/blk-core.c cpu = part_stat_lock(); cpu 1658 block/blk-core.c part_stat_add(cpu, part, sectors[rw], nr_bytes >> 9); cpu 1770 block/blk-core.c int cpu; cpu 1772 block/blk-core.c cpu = part_stat_lock(); cpu 1775 block/blk-core.c part_stat_inc(cpu, part, ios[rw]); cpu 1776 block/blk-core.c part_stat_add(cpu, part, ticks[rw], duration); cpu 1777 block/blk-core.c part_round_stats(cpu, part); cpu 391 block/blk-merge.c int cpu; cpu 393 block/blk-merge.c cpu = part_stat_lock(); cpu 396 block/blk-merge.c part_round_stats(cpu, part); cpu 404 block/blk-merge.c req->cpu = next->cpu; cpu 60 block/blk-softirq.c if (cpu_online(cpu)) { cpu 67 block/blk-softirq.c __smp_call_function_single(cpu, data); cpu 88 block/blk-softirq.c int cpu = (unsigned long) hcpu; cpu 91 block/blk-softirq.c list_splice_init(&per_cpu(blk_cpu_done, cpu), cpu 108 block/blk-softirq.c int ccpu, cpu, group_cpu; cpu 113 block/blk-softirq.c cpu = smp_processor_id(); cpu 114 block/blk-softirq.c group_cpu = blk_cpu_to_group(cpu); cpu 119 block/blk-softirq.c if (test_bit(QUEUE_FLAG_SAME_COMP, &q->queue_flags) && req->cpu != -1) cpu 120 block/blk-softirq.c ccpu = req->cpu; cpu 122 block/blk-softirq.c ccpu = cpu; cpu 124 block/blk-softirq.c if (ccpu == cpu || ccpu == group_cpu) { cpu 101 block/blk.h cpumask_t mask = cpu_coregroup_map(cpu); cpu 104 block/blk.h return first_cpu(per_cpu(cpu_sibling_map, cpu)); cpu 106 block/blk.h return cpu; cpu 40 block/blktrace.c const int cpu = smp_processor_id(); cpu 47 block/blktrace.c t->cpu = cpu; cpu 130 block/blktrace.c int cpu; cpu 158 block/blktrace.c cpu = smp_processor_id(); cpu 159 block/blktrace.c sequence = per_cpu_ptr(bt->sequence, cpu); cpu 169 block/blktrace.c t->cpu = cpu; cpu 959 block/genhd.c int cpu; cpu 971 block/genhd.c cpu = part_stat_lock(); cpu 972 block/genhd.c part_round_stats(cpu, hd); cpu 214 crypto/async_tx/async_tx.c int cpu, cap, cpu_idx = 0; cpu 224 crypto/async_tx/async_tx.c for_each_possible_cpu(cpu) { cpu 226 crypto/async_tx/async_tx.c per_cpu_ptr(channel_table[cap], cpu)->ref; cpu 229 crypto/async_tx/async_tx.c per_cpu_ptr(channel_table[cap], cpu)->ref = cpu 235 crypto/async_tx/async_tx.c for_each_online_cpu(cpu) { cpu 242 crypto/async_tx/async_tx.c per_cpu_ptr(channel_table[cap], cpu)->ref = new; cpu 396 crypto/async_tx/async_tx.c int cpu = get_cpu(); cpu 397 crypto/async_tx/async_tx.c ref = per_cpu_ptr(channel_table[tx_type], cpu)->ref; cpu 3259 fs/buffer.c struct bh_lru *b = &per_cpu(bh_lrus, cpu); cpu 3265 fs/buffer.c get_cpu_var(bh_accounting).nr += per_cpu(bh_accounting, cpu).nr; cpu 3266 fs/buffer.c per_cpu(bh_accounting, cpu).nr = 0; cpu 409 fs/file.c struct fdtable_defer *fddef = &per_cpu(fdtable_defer_list, cpu); cpu 187 fs/namespace.c int cpu; cpu 188 fs/namespace.c for_each_possible_cpu(cpu) { cpu 189 fs/namespace.c struct mnt_writer *writer = &per_cpu(mnt_writers, cpu); cpu 200 fs/namespace.c int cpu; cpu 203 fs/namespace.c for_each_possible_cpu(cpu) { cpu 204 fs/namespace.c cpu_writer = &per_cpu(mnt_writers, cpu); cpu 274 fs/namespace.c int cpu; cpu 277 fs/namespace.c for_each_possible_cpu(cpu) { cpu 278 fs/namespace.c cpu_writer = &per_cpu(mnt_writers, cpu); cpu 607 fs/namespace.c int cpu; cpu 615 fs/namespace.c for_each_possible_cpu(cpu) { cpu 616 fs/namespace.c struct mnt_writer *cpu_writer = &per_cpu(mnt_writers, cpu); cpu 26 fs/nfs/iostat.h int cpu; cpu 28 fs/nfs/iostat.h cpu = get_cpu(); cpu 29 fs/nfs/iostat.h iostats = per_cpu_ptr(server->io_stats, cpu); cpu 45 fs/nfs/iostat.h int cpu; cpu 47 fs/nfs/iostat.h cpu = get_cpu(); cpu 48 fs/nfs/iostat.h iostats = per_cpu_ptr(server->io_stats, cpu); cpu 576 fs/nfs/super.c int i, cpu; cpu 621 fs/nfs/super.c for_each_possible_cpu(cpu) { cpu 625 fs/nfs/super.c stats = per_cpu_ptr(nfss->io_stats, cpu); cpu 217 fs/partitions/check.c int cpu; cpu 219 fs/partitions/check.c cpu = part_stat_lock(); cpu 220 fs/partitions/check.c part_round_stats(cpu, p); cpu 2319 fs/xfs/xfs_mount.c int cpu, ret = 0; cpu 2323 fs/xfs/xfs_mount.c cpu = get_cpu(); cpu 2324 fs/xfs/xfs_mount.c icsbp = (xfs_icsb_cnts_t *)per_cpu_ptr(mp->m_sb_cnts, cpu); cpu 101 include/asm-cris/arch-v32/hwregs/config_defs.h unsigned int cpu : 1; cpu 96 include/asm-cris/arch-v32/mach-a3/hwregs/clkgen_defs.h unsigned int cpu : 1; cpu 101 include/asm-cris/arch-v32/mach-fs/hwregs/config_defs.h unsigned int cpu : 1; cpu 9 include/asm-cris/smp.h #define raw_smp_processor_id() (current_thread_info()->cpu) cpu 35 include/asm-cris/thread_info.h __u32 cpu; /* current CPU */ cpu 62 include/asm-cris/thread_info.h .cpu = 0, \ cpu 37 include/asm-frv/thread_info.h __u32 cpu; /* current CPU */ cpu 69 include/asm-frv/thread_info.h .cpu = 0, \ cpu 57 include/asm-generic/percpu.h (*SHIFT_PERCPU_PTR(&per_cpu_var(var), per_cpu_offset(cpu))) cpu 70 include/asm-generic/percpu.h #define per_cpu(var, cpu) (*((void)(cpu), &per_cpu_var(var))) cpu 35 include/asm-generic/topology.h #define cpu_to_node(cpu) ((void)(cpu),0) cpu 25 include/asm-generic/vmlinux.lds.h #define CPU_KEEP(sec) *(.cpu##sec) cpu 29 include/asm-generic/vmlinux.lds.h #define CPU_DISCARD(sec) *(.cpu##sec) cpu 125 include/asm-m32r/mmu_context.h int cpu = smp_processor_id(); cpu 130 include/asm-m32r/mmu_context.h cpu_set(cpu, next->cpu_vm_mask); cpu 138 include/asm-m32r/mmu_context.h if (!cpu_test_and_set(cpu, next->cpu_vm_mask)) cpu 63 include/asm-m32r/smp.h #define raw_smp_processor_id() (current_thread_info()->cpu) cpu 76 include/asm-m32r/smp.h return cpu; cpu 81 include/asm-m32r/smp.h return cpu; cpu 30 include/asm-m32r/thread_info.h __u32 cpu; /* current CPU */ cpu 70 include/asm-m32r/thread_info.h .cpu = 0, \ cpu 12 include/asm-m68k/thread_info.h __u32 cpu; /* should always be 0 on m68k */ cpu 42 include/asm-mn10300/mmu_context.h cpu_set((cpu), (task)->cpu_vm_mask) cpu 44 include/asm-mn10300/mmu_context.h cpu_test_and_set((cpu), (task)->cpu_vm_mask) cpu 124 include/asm-mn10300/mmu_context.h int cpu = smp_processor_id(); cpu 127 include/asm-mn10300/mmu_context.h cpu_ran_vm(cpu, next); cpu 128 include/asm-mn10300/mmu_context.h activate_context(next, cpu); cpu 130 include/asm-mn10300/mmu_context.h } else if (!cpu_maybe_ran_vm(cpu, next)) { cpu 131 include/asm-mn10300/mmu_context.h activate_context(next, cpu); cpu 46 include/asm-mn10300/thread_info.h __u32 cpu; /* current CPU */ cpu 78 include/asm-mn10300/thread_info.h .cpu = 0, \ cpu 27 include/asm-parisc/smp.h #define cpu_number_map(cpu) (cpu) cpu 28 include/asm-parisc/smp.h #define cpu_logical_map(cpu) (cpu) cpu 49 include/asm-parisc/smp.h #define raw_smp_processor_id() (current_thread_info()->cpu) cpu 14 include/asm-parisc/thread_info.h __u32 cpu; /* current CPU */ cpu 24 include/asm-parisc/thread_info.h .cpu = 0, \ cpu 35 include/asm-um/mmu_context.h unsigned cpu = smp_processor_id(); cpu 38 include/asm-um/mmu_context.h cpu_clear(cpu, prev->cpu_vm_mask); cpu 39 include/asm-um/mmu_context.h cpu_set(cpu, next->cpu_vm_mask); cpu 10 include/asm-um/smp.h #define raw_smp_processor_id() (current_thread->cpu) cpu 19 include/asm-um/thread_info.h __u32 cpu; /* current CPU */ cpu 34 include/asm-um/thread_info.h .cpu = 0, \ cpu 4 include/asm-x86/bigsmp/apic.h #define xapic_phys_to_log_apicid(cpu) (per_cpu(x86_bios_cpu_apicid, cpu)) cpu 15 include/asm-x86/bigsmp/apic.h static unsigned long cpu = NR_CPUS; cpu 17 include/asm-x86/bigsmp/apic.h if (cpu >= NR_CPUS) cpu 18 include/asm-x86/bigsmp/apic.h cpu = first_cpu(cpu_online_map); cpu 20 include/asm-x86/bigsmp/apic.h cpu = next_cpu(cpu, cpu_online_map); cpu 21 include/asm-x86/bigsmp/apic.h } while (cpu >= NR_CPUS); cpu 22 include/asm-x86/bigsmp/apic.h return cpumask_of_cpu(cpu); cpu 49 include/asm-x86/bigsmp/apic.h id = xapic_phys_to_log_apicid(cpu); cpu 64 include/asm-x86/bigsmp/apic.h int cpu = smp_processor_id(); cpu 67 include/asm-x86/bigsmp/apic.h val = calculate_ldr(cpu); cpu 104 include/asm-x86/bigsmp/apic.h if (cpu >= NR_CPUS) cpu 106 include/asm-x86/bigsmp/apic.h return cpu_physical_id(cpu); cpu 131 include/asm-x86/bigsmp/apic.h int cpu; cpu 134 include/asm-x86/bigsmp/apic.h cpu = first_cpu(cpumask); cpu 135 include/asm-x86/bigsmp/apic.h apicid = cpu_to_logical_apicid(cpu); cpu 11 include/asm-x86/cpu.h struct cpu cpu; cpu 44 include/asm-x86/desc.h return per_cpu(gdt_page, cpu).gdt; cpu 96 include/asm-x86/desc.h #define load_TLS(t, cpu) native_load_tls(t, cpu) cpu 178 include/asm-x86/desc.h struct desc_struct *d = get_cpu_gdt_table(cpu); cpu 194 include/asm-x86/desc.h #define set_tss_desc(cpu, addr) __set_tss_desc(cpu, GDT_ENTRY_TSS, addr) cpu 201 include/asm-x86/desc.h unsigned cpu = smp_processor_id(); cpu 206 include/asm-x86/desc.h write_gdt_entry(get_cpu_gdt_table(cpu), GDT_ENTRY_LDT, cpu 247 include/asm-x86/desc.h struct desc_struct *gdt = get_cpu_gdt_table(cpu); cpu 4 include/asm-x86/es7000/apic.h #define xapic_phys_to_log_apicid(cpu) per_cpu(x86_bios_cpu_apicid, cpu) cpu 53 include/asm-x86/es7000/apic.h id = xapic_phys_to_log_apicid(cpu); cpu 67 include/asm-x86/es7000/apic.h int cpu = smp_processor_id(); cpu 70 include/asm-x86/es7000/apic.h val = calculate_ldr(cpu); cpu 122 include/asm-x86/es7000/apic.h if (cpu >= NR_CPUS) cpu 124 include/asm-x86/es7000/apic.h return (int)cpu_2_logical_apicid[cpu]; cpu 152 include/asm-x86/es7000/apic.h int cpu; cpu 167 include/asm-x86/es7000/apic.h cpu = first_cpu(cpumask); cpu 168 include/asm-x86/es7000/apic.h apicid = cpu_to_logical_apicid(cpu); cpu 170 include/asm-x86/es7000/apic.h if (cpu_isset(cpu, cpumask)) { cpu 171 include/asm-x86/es7000/apic.h int new_apicid = cpu_to_logical_apicid(cpu); cpu 184 include/asm-x86/es7000/apic.h cpu++; cpu 41 include/asm-x86/genapic_32.h int (*cpu_to_logical_apicid)(int cpu); cpu 22 include/asm-x86/genapic_64.h cpumask_t (*vector_allocation_domain)(int cpu); cpu 23 include/asm-x86/hardirq_32.h #define __IRQ_STAT(cpu, member) (per_cpu(irq_stat, cpu).member) cpu 390 include/asm-x86/kvm_host.h void (*vcpu_load)(struct kvm_vcpu *vcpu, int cpu); cpu 113 include/asm-x86/mach-default/mach_apic.h return 1 << cpu; cpu 39 include/asm-x86/mce.h __u8 cpu; /* cpu that raised the error */ cpu 95 include/asm-x86/mce.h extern void (*threshold_cpu_callback)(unsigned long action, unsigned int cpu); cpu 13 include/asm-x86/microcode.h int (*request_microcode_user) (int cpu, const void __user *buf, size_t size); cpu 14 include/asm-x86/microcode.h int (*request_microcode_fw) (int cpu, struct device *device); cpu 16 include/asm-x86/microcode.h void (*apply_microcode) (int cpu); cpu 18 include/asm-x86/microcode.h int (*collect_cpu_info) (int cpu, struct cpu_signature *csig); cpu 19 include/asm-x86/microcode.h void (*microcode_fini_cpu) (int cpu); cpu 7 include/asm-x86/mmu_context_32.h unsigned cpu = smp_processor_id(); cpu 8 include/asm-x86/mmu_context_32.h if (per_cpu(cpu_tlbstate, cpu).state == TLBSTATE_OK) cpu 9 include/asm-x86/mmu_context_32.h per_cpu(cpu_tlbstate, cpu).state = TLBSTATE_LAZY; cpu 17 include/asm-x86/mmu_context_32.h int cpu = smp_processor_id(); cpu 21 include/asm-x86/mmu_context_32.h cpu_clear(cpu, prev->cpu_vm_mask); cpu 23 include/asm-x86/mmu_context_32.h per_cpu(cpu_tlbstate, cpu).state = TLBSTATE_OK; cpu 24 include/asm-x86/mmu_context_32.h per_cpu(cpu_tlbstate, cpu).active_mm = next; cpu 26 include/asm-x86/mmu_context_32.h cpu_set(cpu, next->cpu_vm_mask); cpu 39 include/asm-x86/mmu_context_32.h per_cpu(cpu_tlbstate, cpu).state = TLBSTATE_OK; cpu 40 include/asm-x86/mmu_context_32.h BUG_ON(per_cpu(cpu_tlbstate, cpu).active_mm != next); cpu 42 include/asm-x86/mmu_context_32.h if (!cpu_test_and_set(cpu, next->cpu_vm_mask)) { cpu 17 include/asm-x86/mmu_context_64.h unsigned cpu = smp_processor_id(); cpu 20 include/asm-x86/mmu_context_64.h cpu_clear(cpu, prev->cpu_vm_mask); cpu 25 include/asm-x86/mmu_context_64.h cpu_set(cpu, next->cpu_vm_mask); cpu 36 include/asm-x86/mmu_context_64.h if (!cpu_test_and_set(cpu, next->cpu_vm_mask)) { cpu 68 include/asm-x86/numaq/apic.h if (cpu >= NR_CPUS) cpu 70 include/asm-x86/numaq/apic.h return (int)cpu_2_logical_apicid[cpu]; cpu 94 include/asm-x86/numaq/apic.h int cpu = __ffs(logical_apicid & 0xf); cpu 96 include/asm-x86/numaq/apic.h return physid_mask_of_physid(cpu + 4*node); cpu 117 include/asm-x86/paravirt.h void (*load_tls)(struct thread_struct *t, unsigned int cpu); cpu 878 include/asm-x86/paravirt.h PVOP_VCALL2(pv_cpu_ops.load_tls, t, cpu); cpu 15 include/asm-x86/percpu.h #define __per_cpu_offset(cpu) (cpu_pda(cpu)->data_offset) cpu 137 include/asm-x86/processor.h #define cpu_data(cpu) per_cpu(cpu_info, cpu) cpu 149 include/asm-x86/processor.h return cpu_data(cpu).hlt_works_ok; cpu 56 include/asm-x86/smp.h void (*smp_send_reschedule)(int cpu); cpu 58 include/asm-x86/smp.h int (*cpu_up)(unsigned cpu); cpu 60 include/asm-x86/smp.h void (*cpu_die)(unsigned int cpu); cpu 64 include/asm-x86/smp.h void (*send_call_func_single_ipi)(int cpu); cpu 98 include/asm-x86/smp.h return smp_ops.cpu_up(cpu); cpu 108 include/asm-x86/smp.h smp_ops.cpu_die(cpu); cpu 118 include/asm-x86/smp.h smp_ops.smp_send_reschedule(cpu); cpu 123 include/asm-x86/smp.h smp_ops.send_call_func_single_ipi(cpu); cpu 147 include/asm-x86/smp.h #define cpu_physical_id(cpu) per_cpu(x86_cpu_to_apicid, cpu) cpu 178 include/asm-x86/smp.h ti->cpu; \ cpu 101 include/asm-x86/summit/apic.h if (cpu >= NR_CPUS) cpu 103 include/asm-x86/summit/apic.h return (int)cpu_2_logical_apicid[cpu]; cpu 145 include/asm-x86/summit/apic.h int cpu; cpu 156 include/asm-x86/summit/apic.h cpu = first_cpu(cpumask); cpu 157 include/asm-x86/summit/apic.h apicid = cpu_to_logical_apicid(cpu); cpu 159 include/asm-x86/summit/apic.h if (cpu_isset(cpu, cpumask)) { cpu 160 include/asm-x86/summit/apic.h int new_apicid = cpu_to_logical_apicid(cpu); cpu 169 include/asm-x86/summit/apic.h cpu++; cpu 29 include/asm-x86/thread_info.h __u32 cpu; /* current CPU */ cpu 48 include/asm-x86/thread_info.h .cpu = 0, \ cpu 56 include/asm-x86/topology.h return cpu_to_node_map[cpu]; cpu 58 include/asm-x86/topology.h #define early_cpu_to_node(cpu) cpu_to_node(cpu) cpu 93 include/asm-x86/topology.h return per_cpu(x86_cpu_to_node_map, cpu); cpu 100 include/asm-x86/topology.h return early_per_cpu_ptr(x86_cpu_to_node_map)[cpu]; cpu 102 include/asm-x86/topology.h return per_cpu(x86_cpu_to_node_map, cpu); cpu 224 include/asm-x86/topology.h #define topology_physical_package_id(cpu) (cpu_data(cpu).phys_proc_id) cpu 225 include/asm-x86/topology.h #define topology_core_id(cpu) (cpu_data(cpu).cpu_core_id) cpu 226 include/asm-x86/topology.h #define topology_core_siblings(cpu) (per_cpu(cpu_core_map, cpu)) cpu 227 include/asm-x86/topology.h #define topology_thread_siblings(cpu) (per_cpu(cpu_sibling_map, cpu)) cpu 326 include/asm-x86/uv/uv_bau.h test_bit((cpu), (bau_local_cpumask).bits) cpu 136 include/asm-x86/uv/uv_hub.h #define uv_cpu_hub_info(cpu) (&per_cpu(__uv_hub_info, cpu)) cpu 308 include/asm-x86/uv/uv_hub.h return uv_cpu_to_blade[cpu]; cpu 338 include/asm-x86/uv/uv_hub.h return uv_blade_info[uv_cpu_to_blade_id(cpu)].pnode; cpu 25 include/asm-xtensa/smp.h #define cpu_logical_map(cpu) (cpu) cpu 50 include/asm-xtensa/thread_info.h __u32 cpu; /* current CPU */ cpu 94 include/asm-xtensa/thread_info.h .cpu = 0, \ cpu 368 include/linux/bio.h bio->bi_comp_cpu = cpu; cpu 145 include/linux/blkdev.h int cpu; cpu 573 include/linux/blkdev.h #define blk_rq_cpu_valid(rq) ((rq)->cpu != -1) cpu 105 include/linux/blktrace_api.h __u32 cpu; /* on what cpu did it happen */ cpu 59 include/linux/cn_proc.h __u32 cpu; cpu 87 include/linux/cpufreq.h unsigned int cpu; /* cpu nr of registered CPU */ cpu 124 include/linux/cpufreq.h unsigned int cpu; /* cpu nr */ cpu 227 include/linux/cpufreq.h unsigned int (*get) (unsigned int cpu); cpu 231 include/linux/cpufreq.h unsigned int cpu); cpu 87 include/linux/cpuidle.h unsigned int cpu; cpu 144 include/linux/cpumask.h #define cpu_set(cpu, dst) __cpu_set((cpu), &(dst)) cpu 147 include/linux/cpumask.h set_bit(cpu, dstp->bits); cpu 150 include/linux/cpumask.h #define cpu_clear(cpu, dst) __cpu_clear((cpu), &(dst)) cpu 153 include/linux/cpumask.h clear_bit(cpu, dstp->bits); cpu 169 include/linux/cpumask.h #define cpu_isset(cpu, cpumask) test_bit((cpu), (cpumask).bits) cpu 171 include/linux/cpumask.h #define cpu_test_and_set(cpu, cpumask) __cpu_test_and_set((cpu), &(cpumask)) cpu 174 include/linux/cpumask.h return test_and_set_bit(cpu, addr->bits); cpu 280 include/linux/cpumask.h const unsigned long *p = cpu_bit_bitmap[1 + cpu % BITS_PER_LONG]; cpu 281 include/linux/cpumask.h p -= cpu / BITS_PER_LONG; cpu 290 include/linux/cpumask.h #define cpumask_of_cpu(cpu) (*get_cpu_mask(cpu)) cpu 408 include/linux/cpumask.h for ((cpu) = 0; (cpu) < 1; (cpu)++, (void)mask) cpu 421 include/linux/cpumask.h for ((cpu) = -1; \ cpu 422 include/linux/cpumask.h (cpu) = next_cpu((cpu), (mask)), \ cpu 423 include/linux/cpumask.h (cpu) < NR_CPUS; ) cpu 430 include/linux/cpumask.h #define for_each_cpu_mask_nr(cpu, mask) for_each_cpu_mask(cpu, mask) cpu 438 include/linux/cpumask.h for ((cpu) = -1; \ cpu 439 include/linux/cpumask.h (cpu) = next_cpu_nr((cpu), (mask)), \ cpu 440 include/linux/cpumask.h (cpu) < nr_cpu_ids; ) cpu 510 include/linux/cpumask.h #define cpu_online(cpu) cpu_isset((cpu), cpu_online_map) cpu 511 include/linux/cpumask.h #define cpu_possible(cpu) cpu_isset((cpu), cpu_possible_map) cpu 512 include/linux/cpumask.h #define cpu_present(cpu) cpu_isset((cpu), cpu_present_map) cpu 513 include/linux/cpumask.h #define cpu_active(cpu) cpu_isset((cpu), cpu_active_map) cpu 518 include/linux/cpumask.h #define cpu_online(cpu) ((cpu) == 0) cpu 519 include/linux/cpumask.h #define cpu_possible(cpu) ((cpu) == 0) cpu 520 include/linux/cpumask.h #define cpu_present(cpu) ((cpu) == 0) cpu 521 include/linux/cpumask.h #define cpu_active(cpu) ((cpu) == 0) cpu 524 include/linux/cpumask.h #define cpu_is_offline(cpu) unlikely(!cpu_online(cpu)) cpu 526 include/linux/cpumask.h #define for_each_possible_cpu(cpu) for_each_cpu_mask_nr((cpu), cpu_possible_map) cpu 527 include/linux/cpumask.h #define for_each_online_cpu(cpu) for_each_cpu_mask_nr((cpu), cpu_online_map) cpu 528 include/linux/cpumask.h #define for_each_present_cpu(cpu) for_each_cpu_mask_nr((cpu), cpu_present_map) cpu 23 include/linux/dca.h int cpu); cpu 252 include/linux/genhd.h (per_cpu_ptr((part)->dkstats, (cpu))->field += (addnd)) cpu 311 include/linux/genhd.h __part_stat_add((cpu), (part), field, addnd); \ cpu 313 include/linux/genhd.h __part_stat_add((cpu), &part_to_disk((part))->part0, \ cpu 318 include/linux/genhd.h part_stat_add(cpu, gendiskp, field, -1) cpu 320 include/linux/genhd.h part_stat_add(cpu, gendiskp, field, 1) cpu 322 include/linux/genhd.h part_stat_add(cpu, gendiskp, field, -subnd) cpu 171 include/linux/irq.h unsigned int cpu; cpu 21 include/linux/irq_cpustat.h #define __IRQ_STAT(cpu, member) (irq_stat[cpu].member) cpu 29 include/linux/irq_cpustat.h #define nmi_count(cpu) __IRQ_STAT((cpu), __nmi_count) /* i386 */ cpu 36 include/linux/kernel_stat.h #define kstat_cpu(cpu) per_cpu(kstat, cpu) cpu 47 include/linux/kernel_stat.h int cpu, sum = 0; cpu 49 include/linux/kernel_stat.h for_each_possible_cpu(cpu) cpu 50 include/linux/kernel_stat.h sum += kstat_cpu(cpu).irqs[irq]; cpu 67 include/linux/kvm_host.h int cpu; cpu 167 include/linux/lockdep.h int cpu; cpu 401 include/linux/module.h unsigned int cpu = get_cpu(); cpu 403 include/linux/module.h local_inc(&module->ref[cpu].count); cpu 1500 include/linux/netdevice.h txq->xmit_lock_owner = cpu; cpu 1539 include/linux/netdevice.h int cpu; cpu 1542 include/linux/netdevice.h cpu = smp_processor_id(); cpu 1552 include/linux/netdevice.h __netif_tx_lock(txq, cpu); cpu 1590 include/linux/netdevice.h __netif_tx_lock(txq, cpu); \ cpu 1603 include/linux/netdevice.h int cpu; cpu 1606 include/linux/netdevice.h cpu = smp_processor_id(); cpu 1610 include/linux/netdevice.h __netif_tx_lock(txq, cpu); cpu 81 include/linux/percpu.h (__typeof__(ptr))__p->ptrs[(cpu)]; \ cpu 89 include/linux/percpu.h #define percpu_ptr(ptr, cpu) ({ (void)(cpu); (ptr); }) cpu 114 include/linux/percpu.h #define per_cpu_ptr(ptr, cpu) percpu_ptr((ptr), (cpu)) cpu 9 include/linux/posix-timers.h cputime_t cpu; cpu 57 include/linux/posix-timers.h struct cpu_timer_list cpu; cpu 110 include/linux/preempt.h void (*sched_in)(struct preempt_notifier *notifier, int cpu); cpu 107 include/linux/rcuclassic.h int cpu; cpu 122 include/linux/rcuclassic.h struct rcu_data *rdp = &per_cpu(rcu_data, cpu); cpu 127 include/linux/rcuclassic.h struct rcu_data *rdp = &per_cpu(rcu_bh_data, cpu); cpu 55 include/linux/rcupreempt.h struct rcu_dyntick_sched *rdssp = &per_cpu(rcu_dyntick_sched, cpu); cpu 52 include/linux/relay.h unsigned int cpu; /* this buf's cpu */ cpu 2105 include/linux/sched.h return task_thread_info(p)->cpu; cpu 183 include/linux/topology.h #define topology_physical_package_id(cpu) ((void)(cpu), -1) cpu 186 include/linux/topology.h #define topology_core_id(cpu) ((void)(cpu), 0) cpu 189 include/linux/topology.h #define topology_thread_siblings(cpu) cpumask_of_cpu(cpu) cpu 192 include/linux/topology.h #define topology_core_siblings(cpu) cpumask_of_cpu(cpu) cpu 56 include/media/saa7146.h __le32 *cpu; cpu 1160 include/net/tcp.h int cpu = get_cpu(); cpu 1161 include/net/tcp.h struct tcp_md5sig_pool *ret = __tcp_get_md5sig_pool(cpu); cpu 379 init/main.c int cpu, highest_cpu = 0; cpu 381 init/main.c for_each_possible_cpu(cpu) cpu 382 init/main.c highest_cpu = cpu; cpu 413 init/main.c unsigned int cpu; cpu 419 init/main.c cpu = smp_processor_id(); cpu 420 init/main.c cpu_set(cpu, cpu_active_map); cpu 423 init/main.c for_each_present_cpu(cpu) { cpu 426 init/main.c if (!cpu_online(cpu)) cpu 427 init/main.c cpu_up(cpu); cpu 524 init/main.c int cpu = smp_processor_id(); cpu 526 init/main.c cpu_set(cpu, cpu_online_map); cpu 527 init/main.c cpu_set(cpu, cpu_present_map); cpu 528 init/main.c cpu_set(cpu, cpu_possible_map); cpu 180 kernel/cpu.c if (task_cpu(p) == cpu && cpu 185 kernel/cpu.c p->comm, task_pid_nr(p), cpu, cpu 221 kernel/cpu.c void *hcpu = (void *)(long)cpu; cpu 231 kernel/cpu.c if (!cpu_online(cpu)) cpu 242 kernel/cpu.c __func__, cpu); cpu 250 kernel/cpu.c cpu_clear(cpu, tmp); cpu 252 kernel/cpu.c tmp = cpumask_of_cpu(cpu); cpu 263 kernel/cpu.c BUG_ON(cpu_online(cpu)); cpu 266 kernel/cpu.c while (!idle_cpu(cpu)) cpu 270 kernel/cpu.c __cpu_die(cpu); cpu 277 kernel/cpu.c check_for_tasks(cpu); cpu 302 kernel/cpu.c cpu_clear(cpu, cpu_active_map); cpu 314 kernel/cpu.c err = _cpu_down(cpu, 0); cpu 316 kernel/cpu.c if (cpu_online(cpu)) cpu 317 kernel/cpu.c cpu_set(cpu, cpu_active_map); cpu 330 kernel/cpu.c void *hcpu = (void *)(long)cpu; cpu 333 kernel/cpu.c if (cpu_online(cpu) || !cpu_present(cpu)) cpu 342 kernel/cpu.c __func__, cpu); cpu 348 kernel/cpu.c ret = __cpu_up(cpu); cpu 351 kernel/cpu.c BUG_ON(!cpu_online(cpu)); cpu 353 kernel/cpu.c cpu_set(cpu, cpu_active_map); cpu 370 kernel/cpu.c if (!cpu_isset(cpu, cpu_possible_map)) { cpu 372 kernel/cpu.c "configured as may-hotadd at boot time\n", cpu); cpu 387 kernel/cpu.c err = _cpu_up(cpu, 0); cpu 399 kernel/cpu.c int cpu, first_cpu, error = 0; cpu 408 kernel/cpu.c for_each_online_cpu(cpu) { cpu 409 kernel/cpu.c if (cpu == first_cpu) cpu 411 kernel/cpu.c error = _cpu_down(cpu, 1); cpu 413 kernel/cpu.c cpu_set(cpu, frozen_cpus); cpu 414 kernel/cpu.c printk("CPU%d is down\n", cpu); cpu 417 kernel/cpu.c cpu, error); cpu 434 kernel/cpu.c int cpu, error; cpu 443 kernel/cpu.c for_each_cpu_mask_nr(cpu, frozen_cpus) { cpu 444 kernel/cpu.c error = _cpu_up(cpu, 1); cpu 446 kernel/cpu.c printk("CPU%d is up\n", cpu); cpu 449 kernel/cpu.c printk(KERN_WARNING "Error taking CPU%d up: %d\n", cpu, error); cpu 470 kernel/cpu.c if (cpu_isset(cpu, frozen_cpus)) cpu 473 kernel/cpu.c raw_notifier_call_chain(&cpu_chain, val, (void *)(long)cpu); cpu 83 kernel/fork.c int cpu; cpu 86 kernel/fork.c for_each_online_cpu(cpu) cpu 87 kernel/fork.c total += per_cpu(process_counts, cpu); cpu 1313 kernel/fork.c init_idle(task, cpu); cpu 707 kernel/hrtimer.c int cpu = smp_processor_id(); cpu 708 kernel/hrtimer.c struct hrtimer_cpu_base *base = &per_cpu(hrtimer_bases, cpu); cpu 719 kernel/hrtimer.c "mode on CPU %d\n", cpu); cpu 1582 kernel/hrtimer.c struct hrtimer_cpu_base *cpu_base = &per_cpu(hrtimer_bases, cpu); cpu 1686 kernel/hrtimer.c BUG_ON(cpu_online(cpu)); cpu 1687 kernel/hrtimer.c old_base = &per_cpu(hrtimer_bases, cpu); cpu 1690 kernel/hrtimer.c tick_cancel_sched_timer(cpu); cpu 1698 kernel/hrtimer.c &new_base->clock_base[i], cpu)) cpu 1718 kernel/hrtimer.c unsigned int cpu = (long)hcpu; cpu 1724 kernel/hrtimer.c init_hrtimers_cpu(cpu); cpu 1730 kernel/hrtimer.c clockevents_notify(CLOCK_EVT_NOTIFY_CPU_DEAD, &cpu); cpu 1731 kernel/hrtimer.c migrate_hrtimers(cpu); cpu 308 kernel/irq/chip.c const unsigned int cpu = smp_processor_id(); cpu 315 kernel/irq/chip.c kstat_cpu(cpu).irqs[irq]++; cpu 347 kernel/irq/chip.c unsigned int cpu = smp_processor_id(); cpu 357 kernel/irq/chip.c kstat_cpu(cpu).irqs[irq]++; cpu 395 kernel/irq/chip.c unsigned int cpu = smp_processor_id(); cpu 405 kernel/irq/chip.c kstat_cpu(cpu).irqs[irq]++; cpu 454 kernel/irq/chip.c const unsigned int cpu = smp_processor_id(); cpu 472 kernel/irq/chip.c kstat_cpu(cpu).irqs[irq]++; cpu 1118 kernel/kexec.c if ((cpu < 0) || (cpu >= NR_CPUS)) cpu 1128 kernel/kexec.c buf = (u32*)per_cpu_ptr(crash_notes, cpu); cpu 64 kernel/kgdb.c int cpu; cpu 567 kernel/kgdb.c int cpu; cpu 570 kernel/kgdb.c cpu = raw_smp_processor_id(); cpu 571 kernel/kgdb.c kgdb_info[cpu].debuggerinfo = regs; cpu 572 kernel/kgdb.c kgdb_info[cpu].task = current; cpu 578 kernel/kgdb.c atomic_set(&cpu_in_kgdb[cpu], 1); cpu 581 kernel/kgdb.c while (atomic_read(&passive_cpu_wait[cpu])) cpu 584 kernel/kgdb.c kgdb_info[cpu].debuggerinfo = NULL; cpu 585 kernel/kgdb.c kgdb_info[cpu].task = NULL; cpu 592 kernel/kgdb.c atomic_set(&cpu_in_kgdb[cpu], 0); cpu 855 kernel/kgdb.c thread = kgdb_info[ks->cpu].task; cpu 856 kernel/kgdb.c local_debuggerinfo = kgdb_info[ks->cpu].debuggerinfo; cpu 998 kernel/kgdb.c int cpu; cpu 1014 kernel/kgdb.c for_each_online_cpu(cpu) { cpu 1016 kernel/kgdb.c int_to_threadref(thref, -cpu - 2); cpu 1241 kernel/kgdb.c kgdb_usethread = kgdb_info[ks->cpu].task; cpu 1242 kernel/kgdb.c ks->kgdb_usethreadid = shadow_pid(kgdb_info[ks->cpu].task->pid); cpu 1399 kernel/kgdb.c int i, cpu; cpu 1401 kernel/kgdb.c ks->cpu = raw_smp_processor_id(); cpu 1419 kernel/kgdb.c cpu = raw_smp_processor_id(); cpu 1424 kernel/kgdb.c while (atomic_cmpxchg(&kgdb_active, -1, cpu) != -1) cpu 1433 kernel/kgdb.c atomic_read(&kgdb_cpu_doing_single_step) != cpu) { cpu 1458 kernel/kgdb.c kgdb_info[ks->cpu].debuggerinfo = ks->linux_regs; cpu 1459 kernel/kgdb.c kgdb_info[ks->cpu].task = current; cpu 1476 kernel/kgdb.c atomic_set(&cpu_in_kgdb[ks->cpu], 1); cpu 1509 kernel/kgdb.c kgdb_info[ks->cpu].debuggerinfo = NULL; cpu 1510 kernel/kgdb.c kgdb_info[ks->cpu].task = NULL; cpu 1511 kernel/kgdb.c atomic_set(&cpu_in_kgdb[ks->cpu], 0); cpu 1539 kernel/kgdb.c if (!atomic_read(&cpu_in_kgdb[cpu]) && cpu 1540 kernel/kgdb.c atomic_read(&kgdb_active) != cpu && cpu 180 kernel/kthread.c set_task_cpu(k, cpu); cpu 181 kernel/kthread.c k->cpus_allowed = cpumask_of_cpu(cpu); cpu 178 kernel/lockdep.c int cpu, i; cpu 181 kernel/lockdep.c for_each_possible_cpu(cpu) { cpu 183 kernel/lockdep.c &per_cpu(lock_stats, cpu)[class - lock_classes]; cpu 203 kernel/lockdep.c int cpu; cpu 205 kernel/lockdep.c for_each_possible_cpu(cpu) { cpu 207 kernel/lockdep.c &per_cpu(lock_stats, cpu)[class - lock_classes]; cpu 2508 kernel/lockdep.c lock->cpu = raw_smp_processor_id(); cpu 3033 kernel/lockdep.c if (lock->cpu != smp_processor_id()) cpu 3047 kernel/lockdep.c int i, cpu; cpu 3069 kernel/lockdep.c cpu = smp_processor_id(); cpu 3083 kernel/lockdep.c if (lock->cpu != cpu) cpu 3087 kernel/lockdep.c lock->cpu = cpu; cpu 464 kernel/module.c int cpu; cpu 466 kernel/module.c for_each_possible_cpu(cpu) cpu 467 kernel/module.c memcpy(pcpudest + per_cpu_offset(cpu), from, size); cpu 856 kernel/module.c unsigned int cpu = get_cpu(); cpu 857 kernel/module.c local_dec(&module->ref[cpu].count); cpu 42 kernel/posix-cpu-timers.c ret.cpu = timespec_to_cputime(tp); cpu 52 kernel/posix-cpu-timers.c *tp = ns_to_timespec(cpu.sched); cpu 54 kernel/posix-cpu-timers.c cputime_to_timespec(cpu.cpu, tp); cpu 64 kernel/posix-cpu-timers.c return cputime_lt(now.cpu, then.cpu); cpu 74 kernel/posix-cpu-timers.c acc->cpu = cputime_add(acc->cpu, val.cpu); cpu 84 kernel/posix-cpu-timers.c a.cpu = cputime_sub(a.cpu, b.cpu); cpu 111 kernel/posix-cpu-timers.c if (timer->it.cpu.incr.sched == 0) cpu 117 kernel/posix-cpu-timers.c if (now.sched < timer->it.cpu.expires.sched) cpu 119 kernel/posix-cpu-timers.c incr = timer->it.cpu.incr.sched; cpu 120 kernel/posix-cpu-timers.c delta = now.sched + incr - timer->it.cpu.expires.sched; cpu 127 kernel/posix-cpu-timers.c timer->it.cpu.expires.sched += incr; cpu 134 kernel/posix-cpu-timers.c if (cputime_lt(now.cpu, timer->it.cpu.expires.cpu)) cpu 136 kernel/posix-cpu-timers.c incr = timer->it.cpu.incr.cpu; cpu 137 kernel/posix-cpu-timers.c delta = cputime_sub(cputime_add(now.cpu, incr), cpu 138 kernel/posix-cpu-timers.c timer->it.cpu.expires.cpu); cpu 145 kernel/posix-cpu-timers.c timer->it.cpu.expires.cpu = cpu 146 kernel/posix-cpu-timers.c cputime_add(timer->it.cpu.expires.cpu, incr); cpu 208 kernel/posix-cpu-timers.c cpu->cpu = prof_ticks(p); cpu 211 kernel/posix-cpu-timers.c cpu->cpu = virt_ticks(p); cpu 214 kernel/posix-cpu-timers.c cpu->sched = sched_ns(p); cpu 234 kernel/posix-cpu-timers.c cpu->cpu = cputime_add(p->signal->utime, p->signal->stime); cpu 236 kernel/posix-cpu-timers.c cpu->cpu = cputime_add(cpu->cpu, prof_ticks(t)); cpu 241 kernel/posix-cpu-timers.c cpu->cpu = p->signal->utime; cpu 243 kernel/posix-cpu-timers.c cpu->cpu = cputime_add(cpu->cpu, virt_ticks(t)); cpu 248 kernel/posix-cpu-timers.c cpu->sched = p->signal->sum_sched_runtime; cpu 251 kernel/posix-cpu-timers.c cpu->sched += t->se.sum_exec_runtime; cpu 253 kernel/posix-cpu-timers.c cpu->sched += sched_ns(p); cpu 271 kernel/posix-cpu-timers.c cpu); cpu 347 kernel/posix-cpu-timers.c INIT_LIST_HEAD(&new_timer->it.cpu.entry); cpu 348 kernel/posix-cpu-timers.c new_timer->it.cpu.incr.sched = 0; cpu 349 kernel/posix-cpu-timers.c new_timer->it.cpu.expires.sched = 0; cpu 369 kernel/posix-cpu-timers.c new_timer->it.cpu.task = p; cpu 388 kernel/posix-cpu-timers.c struct task_struct *p = timer->it.cpu.task; cpu 398 kernel/posix-cpu-timers.c BUG_ON(!list_empty(&timer->it.cpu.entry)); cpu 401 kernel/posix-cpu-timers.c if (timer->it.cpu.firing) cpu 404 kernel/posix-cpu-timers.c list_del(&timer->it.cpu.entry); cpu 431 kernel/posix-cpu-timers.c if (cputime_lt(timer->expires.cpu, ptime)) { cpu 432 kernel/posix-cpu-timers.c timer->expires.cpu = cputime_zero; cpu 434 kernel/posix-cpu-timers.c timer->expires.cpu = cputime_sub(timer->expires.cpu, cpu 442 kernel/posix-cpu-timers.c if (cputime_lt(timer->expires.cpu, utime)) { cpu 443 kernel/posix-cpu-timers.c timer->expires.cpu = cputime_zero; cpu 445 kernel/posix-cpu-timers.c timer->expires.cpu = cputime_sub(timer->expires.cpu, cpu 503 kernel/posix-cpu-timers.c left = cputime_div_non_zero(cputime_sub(expires.cpu, val.cpu), cpu 518 kernel/posix-cpu-timers.c left = cputime_div_non_zero(cputime_sub(expires.cpu, val.cpu), cpu 556 kernel/posix-cpu-timers.c put_task_struct(timer->it.cpu.task); cpu 557 kernel/posix-cpu-timers.c timer->it.cpu.task = NULL; cpu 558 kernel/posix-cpu-timers.c timer->it.cpu.expires = cpu_time_sub(timer->it_clock, cpu 559 kernel/posix-cpu-timers.c timer->it.cpu.expires, cpu 570 kernel/posix-cpu-timers.c struct task_struct *p = timer->it.cpu.task; cpu 572 kernel/posix-cpu-timers.c struct cpu_timer_list *const nt = &timer->it.cpu; cpu 592 kernel/posix-cpu-timers.c if (cputime_gt(next->expires.cpu, nt->expires.cpu)) cpu 614 kernel/posix-cpu-timers.c nt->expires.cpu)) cpu 615 kernel/posix-cpu-timers.c p->it_prof_expires = nt->expires.cpu; cpu 621 kernel/posix-cpu-timers.c nt->expires.cpu)) cpu 622 kernel/posix-cpu-timers.c p->it_virt_expires = nt->expires.cpu; cpu 642 kernel/posix-cpu-timers.c timer->it.cpu.expires.cpu)) cpu 649 kernel/posix-cpu-timers.c timer->it.cpu.expires.cpu)) cpu 653 kernel/posix-cpu-timers.c i <= cputime_to_secs(timer->it.cpu.expires.cpu)) cpu 659 kernel/posix-cpu-timers.c timer->it.cpu.task, cpu 661 kernel/posix-cpu-timers.c timer->it.cpu.expires, now); cpu 681 kernel/posix-cpu-timers.c timer->it.cpu.expires.sched = 0; cpu 682 kernel/posix-cpu-timers.c } else if (timer->it.cpu.incr.sched == 0) { cpu 687 kernel/posix-cpu-timers.c timer->it.cpu.expires.sched = 0; cpu 708 kernel/posix-cpu-timers.c struct task_struct *p = timer->it.cpu.task; cpu 730 kernel/posix-cpu-timers.c timer->it.cpu.task = NULL; cpu 741 kernel/posix-cpu-timers.c old_expires = timer->it.cpu.expires; cpu 742 kernel/posix-cpu-timers.c if (unlikely(timer->it.cpu.firing)) { cpu 743 kernel/posix-cpu-timers.c timer->it.cpu.firing = -1; cpu 746 kernel/posix-cpu-timers.c list_del_init(&timer->it.cpu.entry); cpu 780 kernel/posix-cpu-timers.c timer->it.cpu.expires)) { cpu 783 kernel/posix-cpu-timers.c timer->it.cpu.expires, val); cpu 814 kernel/posix-cpu-timers.c timer->it.cpu.expires = new_expires; cpu 827 kernel/posix-cpu-timers.c timer->it.cpu.incr = timespec_to_sample(timer->it_clock, cpu 855 kernel/posix-cpu-timers.c timer->it.cpu.incr, &old->it_interval); cpu 863 kernel/posix-cpu-timers.c struct task_struct *p = timer->it.cpu.task; cpu 870 kernel/posix-cpu-timers.c timer->it.cpu.incr, &itp->it_interval); cpu 872 kernel/posix-cpu-timers.c if (timer->it.cpu.expires.sched == 0) { /* Timer not armed at all. */ cpu 883 kernel/posix-cpu-timers.c sample_to_timespec(timer->it_clock, timer->it.cpu.expires, cpu 903 kernel/posix-cpu-timers.c timer->it.cpu.task = NULL; cpu 904 kernel/posix-cpu-timers.c timer->it.cpu.expires.sched = 0; cpu 916 kernel/posix-cpu-timers.c if (timer->it.cpu.incr.sched == 0 && cpu 918 kernel/posix-cpu-timers.c timer->it.cpu.expires, now)) { cpu 923 kernel/posix-cpu-timers.c timer->it.cpu.expires.sched = 0; cpu 944 kernel/posix-cpu-timers.c if (cpu_time_before(timer->it_clock, now, timer->it.cpu.expires)) { cpu 947 kernel/posix-cpu-timers.c timer->it.cpu.expires, now), cpu 977 kernel/posix-cpu-timers.c if (!--maxfire || cputime_lt(prof_ticks(tsk), t->expires.cpu)) { cpu 978 kernel/posix-cpu-timers.c tsk->it_prof_expires = t->expires.cpu; cpu 992 kernel/posix-cpu-timers.c if (!--maxfire || cputime_lt(virt_ticks(tsk), t->expires.cpu)) { cpu 993 kernel/posix-cpu-timers.c tsk->it_virt_expires = t->expires.cpu; cpu 1095 kernel/posix-cpu-timers.c if (!--maxfire || cputime_lt(ptime, tl->expires.cpu)) { cpu 1096 kernel/posix-cpu-timers.c prof_expires = tl->expires.cpu; cpu 1110 kernel/posix-cpu-timers.c if (!--maxfire || cputime_lt(utime, tl->expires.cpu)) { cpu 1111 kernel/posix-cpu-timers.c virt_expires = tl->expires.cpu; cpu 1258 kernel/posix-cpu-timers.c struct task_struct *p = timer->it.cpu.task; cpu 1286 kernel/posix-cpu-timers.c timer->it.cpu.task = p = NULL; cpu 1287 kernel/posix-cpu-timers.c timer->it.cpu.expires.sched = 0; cpu 1372 kernel/posix-cpu-timers.c list_for_each_entry_safe(timer, next, &firing, it.cpu.entry) { cpu 1375 kernel/posix-cpu-timers.c list_del_init(&timer->it.cpu.entry); cpu 1376 kernel/posix-cpu-timers.c firing = timer->it.cpu.firing; cpu 1377 kernel/posix-cpu-timers.c timer->it.cpu.firing = 0; cpu 1408 kernel/posix-cpu-timers.c if (cputime_le(*oldval, now.cpu)) { cpu 1412 kernel/posix-cpu-timers.c *oldval = cputime_sub(*oldval, now.cpu); cpu 1418 kernel/posix-cpu-timers.c *newval = cputime_add(*newval, now.cpu); cpu 1436 kernel/posix-cpu-timers.c struct cpu_timer_list, entry)->expires.cpu, cpu 1443 kernel/posix-cpu-timers.c expires.cpu = *newval; cpu 1477 kernel/posix-cpu-timers.c if (timer.it.cpu.expires.sched == 0) { cpu 1497 kernel/posix-cpu-timers.c sample_to_timespec(which_clock, timer.it.cpu.expires, rqtp); cpu 623 kernel/printk.c return cpu_online(cpu) || have_callable_console(); cpu 649 kernel/printk.c if (!can_use_console(cpu)) { cpu 239 kernel/profile.c int cpu = smp_processor_id(); cpu 241 kernel/profile.c per_cpu(cpu_profile_flip, cpu) = !per_cpu(cpu_profile_flip, cpu); cpu 246 kernel/profile.c int i, j, cpu; cpu 252 kernel/profile.c for_each_online_cpu(cpu) { cpu 253 kernel/profile.c struct profile_hit *hits = per_cpu(cpu_profile_hits, cpu)[j]; cpu 269 kernel/profile.c int i, cpu; cpu 275 kernel/profile.c for_each_online_cpu(cpu) { cpu 276 kernel/profile.c struct profile_hit *hits = per_cpu(cpu_profile_hits, cpu)[i]; cpu 285 kernel/profile.c int i, j, cpu; cpu 293 kernel/profile.c cpu = get_cpu(); cpu 294 kernel/profile.c hits = per_cpu(cpu_profile_hits, cpu)[per_cpu(cpu_profile_flip, cpu)]; cpu 336 kernel/profile.c int node, cpu = (unsigned long)__cpu; cpu 342 kernel/profile.c node = cpu_to_node(cpu); cpu 343 kernel/profile.c per_cpu(cpu_profile_flip, cpu) = 0; cpu 344 kernel/profile.c if (!per_cpu(cpu_profile_hits, cpu)[1]) { cpu 350 kernel/profile.c per_cpu(cpu_profile_hits, cpu)[1] = page_address(page); cpu 352 kernel/profile.c if (!per_cpu(cpu_profile_hits, cpu)[0]) { cpu 358 kernel/profile.c per_cpu(cpu_profile_hits, cpu)[0] = page_address(page); cpu 362 kernel/profile.c page = virt_to_page(per_cpu(cpu_profile_hits, cpu)[1]); cpu 363 kernel/profile.c per_cpu(cpu_profile_hits, cpu)[1] = NULL; cpu 368 kernel/profile.c cpu_set(cpu, prof_cpu_mask); cpu 374 kernel/profile.c cpu_clear(cpu, prof_cpu_mask); cpu 375 kernel/profile.c if (per_cpu(cpu_profile_hits, cpu)[0]) { cpu 376 kernel/profile.c page = virt_to_page(per_cpu(cpu_profile_hits, cpu)[0]); cpu 377 kernel/profile.c per_cpu(cpu_profile_hits, cpu)[0] = NULL; cpu 380 kernel/profile.c if (per_cpu(cpu_profile_hits, cpu)[1]) { cpu 381 kernel/profile.c page = virt_to_page(per_cpu(cpu_profile_hits, cpu)[1]); cpu 382 kernel/profile.c per_cpu(cpu_profile_hits, cpu)[1] = NULL; cpu 532 kernel/profile.c int cpu; cpu 534 kernel/profile.c for_each_online_cpu(cpu) { cpu 535 kernel/profile.c int node = cpu_to_node(cpu); cpu 543 kernel/profile.c per_cpu(cpu_profile_hits, cpu)[1] cpu 550 kernel/profile.c per_cpu(cpu_profile_hits, cpu)[0] cpu 558 kernel/profile.c for_each_online_cpu(cpu) { cpu 561 kernel/profile.c if (per_cpu(cpu_profile_hits, cpu)[0]) { cpu 562 kernel/profile.c page = virt_to_page(per_cpu(cpu_profile_hits, cpu)[0]); cpu 563 kernel/profile.c per_cpu(cpu_profile_hits, cpu)[0] = NULL; cpu 566 kernel/profile.c if (per_cpu(cpu_profile_hits, cpu)[1]) { cpu 567 kernel/profile.c page = virt_to_page(per_cpu(cpu_profile_hits, cpu)[1]); cpu 568 kernel/profile.c per_cpu(cpu_profile_hits, cpu)[1] = NULL; cpu 87 kernel/rcuclassic.c int cpu; cpu 114 kernel/rcuclassic.c cpu_clear(rdp->cpu, cpumask); cpu 115 kernel/rcuclassic.c for_each_cpu_mask_nr(cpu, cpumask) cpu 116 kernel/rcuclassic.c smp_send_reschedule(cpu); cpu 177 kernel/rcuclassic.c int cpu; cpu 195 kernel/rcuclassic.c for_each_possible_cpu(cpu) { cpu 196 kernel/rcuclassic.c if (cpu_isset(cpu, rcp->cpumask)) cpu 197 kernel/rcuclassic.c printk(" %d", cpu); cpu 409 kernel/rcuclassic.c cpu_clear(cpu, rcp->cpumask); cpu 456 kernel/rcuclassic.c cpu_quiet(rdp->cpu, rcp); cpu 494 kernel/rcuclassic.c cpu_quiet(rdp->cpu, rcp); cpu 509 kernel/rcuclassic.c &per_cpu(rcu_data, cpu)); cpu 511 kernel/rcuclassic.c &per_cpu(rcu_bh_data, cpu)); cpu 653 kernel/rcuclassic.c return __rcu_pending(&rcu_ctrlblk, &per_cpu(rcu_data, cpu)) || cpu 654 kernel/rcuclassic.c __rcu_pending(&rcu_bh_ctrlblk, &per_cpu(rcu_bh_data, cpu)); cpu 665 kernel/rcuclassic.c struct rcu_data *rdp = &per_cpu(rcu_data, cpu); cpu 666 kernel/rcuclassic.c struct rcu_data *rdp_bh = &per_cpu(rcu_bh_data, cpu); cpu 668 kernel/rcuclassic.c return !!rdp->nxtlist || !!rdp_bh->nxtlist || rcu_pending(cpu); cpu 680 kernel/rcuclassic.c (idle_cpu(cpu) && !in_softirq() && cpu 698 kernel/rcuclassic.c rcu_qsctr_inc(cpu); cpu 699 kernel/rcuclassic.c rcu_bh_qsctr_inc(cpu); cpu 712 kernel/rcuclassic.c rcu_bh_qsctr_inc(cpu); cpu 728 kernel/rcuclassic.c rdp->cpu = cpu; cpu 735 kernel/rcuclassic.c struct rcu_data *rdp = &per_cpu(rcu_data, cpu); cpu 736 kernel/rcuclassic.c struct rcu_data *bh_rdp = &per_cpu(rcu_bh_data, cpu); cpu 738 kernel/rcuclassic.c rcu_init_percpu_data(cpu, &rcu_ctrlblk, rdp); cpu 739 kernel/rcuclassic.c rcu_init_percpu_data(cpu, &rcu_bh_ctrlblk, bh_rdp); cpu 746 kernel/rcuclassic.c long cpu = (long)hcpu; cpu 751 kernel/rcuclassic.c rcu_online_cpu(cpu); cpu 755 kernel/rcuclassic.c rcu_offline_cpu(cpu); cpu 95 kernel/rcupdate.c int cpu = smp_processor_id(); cpu 96 kernel/rcupdate.c struct rcu_head *head = &per_cpu(rcu_barrier_head, cpu); cpu 203 kernel/rcupreempt.c #define RCU_DATA_CPU(cpu) (&per_cpu(rcu_data, cpu)) cpu 209 kernel/rcupreempt.c #define RCU_TRACE_CPU(f, cpu) RCU_TRACE(f, &(RCU_DATA_CPU(cpu)->trace)); cpu 378 kernel/rcupreempt.c int cpu; cpu 420 kernel/rcupreempt.c cpu = raw_smp_processor_id(); cpu 421 kernel/rcupreempt.c if (per_cpu(rcu_flip_flag, cpu) == rcu_flipped) { cpu 423 kernel/rcupreempt.c per_cpu(rcu_flip_flag, cpu) = rcu_flip_seen; cpu 445 kernel/rcupreempt.c int cpu = smp_processor_id(); cpu 446 kernel/rcupreempt.c struct rcu_dyntick_sched *rdssp = &per_cpu(rcu_dyntick_sched, cpu); cpu 448 kernel/rcupreempt.c if (per_cpu(rcu_update_flag, cpu)) cpu 449 kernel/rcupreempt.c per_cpu(rcu_update_flag, cpu)++; cpu 496 kernel/rcupreempt.c per_cpu(rcu_update_flag, cpu)++; cpu 515 kernel/rcupreempt.c int cpu = smp_processor_id(); cpu 516 kernel/rcupreempt.c struct rcu_dyntick_sched *rdssp = &per_cpu(rcu_dyntick_sched, cpu); cpu 527 kernel/rcupreempt.c if (per_cpu(rcu_update_flag, cpu)) { cpu 528 kernel/rcupreempt.c if (--per_cpu(rcu_update_flag, cpu)) cpu 556 kernel/rcupreempt.c struct rcu_dyntick_sched *rdssp = &per_cpu(rcu_dyntick_sched, cpu); cpu 566 kernel/rcupreempt.c struct rcu_dyntick_sched *rdssp = &per_cpu(rcu_dyntick_sched, cpu); cpu 603 kernel/rcupreempt.c struct rcu_dyntick_sched *rdssp = &per_cpu(rcu_dyntick_sched, cpu); cpu 635 kernel/rcupreempt.c struct rcu_dyntick_sched *rdssp = &per_cpu(rcu_dyntick_sched, cpu); cpu 644 kernel/rcupreempt.c struct rcu_dyntick_sched *rdssp = &per_cpu(rcu_dyntick_sched, cpu); cpu 689 kernel/rcupreempt.c struct rcu_dyntick_sched *rdssp = &per_cpu(rcu_dyntick_sched, cpu); cpu 696 kernel/rcupreempt.c struct rcu_dyntick_sched *rdssp = &per_cpu(rcu_dyntick_sched, cpu); cpu 726 kernel/rcupreempt.c int cpu; cpu 751 kernel/rcupreempt.c for_each_cpu_mask_nr(cpu, rcu_cpu_online_map) { cpu 752 kernel/rcupreempt.c per_cpu(rcu_flip_flag, cpu) = rcu_flipped; cpu 753 kernel/rcupreempt.c dyntick_save_progress_counter(cpu); cpu 766 kernel/rcupreempt.c int cpu; cpu 769 kernel/rcupreempt.c for_each_cpu_mask_nr(cpu, rcu_cpu_online_map) cpu 770 kernel/rcupreempt.c if (rcu_try_flip_waitack_needed(cpu) && cpu 771 kernel/rcupreempt.c per_cpu(rcu_flip_flag, cpu) != rcu_flip_seen) { cpu 794 kernel/rcupreempt.c int cpu; cpu 801 kernel/rcupreempt.c for_each_cpu_mask_nr(cpu, rcu_cpu_online_map) cpu 802 kernel/rcupreempt.c sum += RCU_DATA_CPU(cpu)->rcu_flipctr[lastidx]; cpu 816 kernel/rcupreempt.c for_each_cpu_mask_nr(cpu, rcu_cpu_online_map) { cpu 817 kernel/rcupreempt.c per_cpu(rcu_mb_flag, cpu) = rcu_mb_needed; cpu 818 kernel/rcupreempt.c dyntick_save_progress_counter(cpu); cpu 833 kernel/rcupreempt.c int cpu; cpu 836 kernel/rcupreempt.c for_each_cpu_mask_nr(cpu, rcu_cpu_online_map) cpu 837 kernel/rcupreempt.c if (rcu_try_flip_waitmb_needed(cpu) && cpu 838 kernel/rcupreempt.c per_cpu(rcu_mb_flag, cpu) != rcu_mb_done) { cpu 904 kernel/rcupreempt.c if (per_cpu(rcu_mb_flag, cpu) == rcu_mb_needed) { cpu 906 kernel/rcupreempt.c per_cpu(rcu_mb_flag, cpu) = rcu_mb_done; cpu 913 kernel/rcupreempt.c struct rcu_data *rdp = RCU_DATA_CPU(cpu); cpu 929 kernel/rcupreempt.c (idle_cpu(cpu) && !in_softirq() && cpu 932 kernel/rcupreempt.c rcu_qsctr_inc(cpu); cpu 935 kernel/rcupreempt.c rcu_check_mb(cpu); cpu 956 kernel/rcupreempt.c struct rcu_data *rdp = RCU_DATA_CPU(cpu); cpu 984 kernel/rcupreempt.c struct rcu_data *rdp = RCU_DATA_CPU(cpu); cpu 1011 kernel/rcupreempt.c rcu_check_mb(cpu); cpu 1012 kernel/rcupreempt.c if (per_cpu(rcu_flip_flag, cpu) == rcu_flipped) { cpu 1014 kernel/rcupreempt.c per_cpu(rcu_flip_flag, cpu) = rcu_flip_seen; cpu 1019 kernel/rcupreempt.c RCU_DATA_ME()->rcu_flipctr[0] += RCU_DATA_CPU(cpu)->rcu_flipctr[0]; cpu 1020 kernel/rcupreempt.c RCU_DATA_ME()->rcu_flipctr[1] += RCU_DATA_CPU(cpu)->rcu_flipctr[1]; cpu 1022 kernel/rcupreempt.c RCU_DATA_CPU(cpu)->rcu_flipctr[0] = 0; cpu 1023 kernel/rcupreempt.c RCU_DATA_CPU(cpu)->rcu_flipctr[1] = 0; cpu 1025 kernel/rcupreempt.c cpu_clear(cpu, rcu_cpu_online_map); cpu 1065 kernel/rcupreempt.c cpu_set(cpu, rcu_cpu_online_map); cpu 1077 kernel/rcupreempt.c rdp = RCU_DATA_CPU(cpu); cpu 1179 kernel/rcupreempt.c int cpu; cpu 1191 kernel/rcupreempt.c for_each_online_cpu(cpu) { cpu 1192 kernel/rcupreempt.c dyntick_save_progress_counter_sched(cpu); cpu 1193 kernel/rcupreempt.c save_qsctr_sched(cpu); cpu 1218 kernel/rcupreempt.c for_each_online_cpu(cpu) { cpu 1219 kernel/rcupreempt.c while (rcu_qsctr_inc_needed(cpu) && cpu 1220 kernel/rcupreempt.c rcu_qsctr_inc_needed_dyntick(cpu)) { cpu 1228 kernel/rcupreempt.c for_each_online_cpu(cpu) { cpu 1230 kernel/rcupreempt.c rdp = RCU_DATA_CPU(cpu); cpu 1325 kernel/rcupreempt.c struct rcu_data *rdp = RCU_DATA_CPU(cpu); cpu 1336 kernel/rcupreempt.c struct rcu_data *rdp = RCU_DATA_CPU(cpu); cpu 1349 kernel/rcupreempt.c if ((per_cpu(rcu_flip_flag, cpu) == rcu_flipped) || cpu 1350 kernel/rcupreempt.c (per_cpu(rcu_mb_flag, cpu) == rcu_mb_needed)) cpu 1366 kernel/rcupreempt.c long cpu = (long)hcpu; cpu 1371 kernel/rcupreempt.c rcu_online_cpu(cpu); cpu 1377 kernel/rcupreempt.c rcu_offline_cpu(cpu); cpu 1391 kernel/rcupreempt.c int cpu; cpu 1396 kernel/rcupreempt.c for_each_possible_cpu(cpu) { cpu 1397 kernel/rcupreempt.c rdp = RCU_DATA_CPU(cpu); cpu 1431 kernel/rcupreempt.c for_each_online_cpu(cpu) cpu 1432 kernel/rcupreempt.c rcu_cpu_notify(&rcu_nb, CPU_UP_PREPARE, (void *)(long) cpu); cpu 1452 kernel/rcupreempt.c return &RCU_DATA_CPU(cpu)->rcu_flipctr[0]; cpu 1458 kernel/rcupreempt.c return per_cpu(rcu_flip_flag, cpu); cpu 1464 kernel/rcupreempt.c return per_cpu(rcu_mb_flag, cpu); cpu 1476 kernel/rcupreempt.c struct rcu_data *rdp = RCU_DATA_CPU(cpu); cpu 140 kernel/rcupreempt_trace.c int cpu; cpu 143 kernel/rcupreempt_trace.c for_each_possible_cpu(cpu) { cpu 144 kernel/rcupreempt_trace.c cp = rcupreempt_trace_cpu(cpu); cpu 231 kernel/rcupreempt_trace.c int cpu; cpu 239 kernel/rcupreempt_trace.c for_each_online_cpu(cpu) { cpu 240 kernel/rcupreempt_trace.c long *flipctr = rcupreempt_flipctr(cpu); cpu 244 kernel/rcupreempt_trace.c cpu, cpu 247 kernel/rcupreempt_trace.c rcupreempt_flip_flag(cpu), cpu 248 kernel/rcupreempt_trace.c rcupreempt_mb_flag(cpu)); cpu 475 kernel/rcutorture.c int cpu; cpu 480 kernel/rcutorture.c for_each_possible_cpu(cpu) { cpu 481 kernel/rcutorture.c cnt += sprintf(&page[cnt], " %d(%d,%d)", cpu, cpu 482 kernel/rcutorture.c per_cpu_ptr(srcu_ctl.per_cpu_ref, cpu)->c[!idx], cpu 483 kernel/rcutorture.c per_cpu_ptr(srcu_ctl.per_cpu_ref, cpu)->c[idx]); cpu 749 kernel/rcutorture.c int cpu; cpu 754 kernel/rcutorture.c for_each_possible_cpu(cpu) { cpu 756 kernel/rcutorture.c pipesummary[i] += per_cpu(rcu_torture_count, cpu)[i]; cpu 757 kernel/rcutorture.c batchsummary[i] += per_cpu(rcu_torture_batch, cpu)[i]; cpu 1012 kernel/rcutorture.c int cpu; cpu 1059 kernel/rcutorture.c for_each_possible_cpu(cpu) { cpu 1061 kernel/rcutorture.c per_cpu(rcu_torture_count, cpu)[i] = 0; cpu 1062 kernel/rcutorture.c per_cpu(rcu_torture_batch, cpu)[i] = 0; cpu 223 kernel/relay.c chan->buf[buf->cpu] = NULL; cpu 427 kernel/relay.c snprintf(tmpname, NAME_MAX, "%s%d", chan->base_filename, cpu); cpu 457 kernel/relay.c dentry = relay_create_buf_file(chan, buf, cpu); cpu 463 kernel/relay.c buf->cpu = cpu; cpu 468 kernel/relay.c buf->cpu = 0; cpu 799 kernel/relay.c if (cpu >= NR_CPUS || !chan->buf[cpu]) cpu 802 kernel/relay.c buf = chan->buf[cpu]; cpu 952 kernel/relay.c relay_subbufs_consumed(buf->chan, buf->cpu, 1); cpu 965 kernel/relay.c relay_subbufs_consumed(buf->chan, buf->cpu, 1); cpu 1186 kernel/relay.c relay_subbufs_consumed(rbuf->chan, rbuf->cpu, 1); cpu 361 kernel/sched.c p->se.cfs_rq = task_group(p)->cfs_rq[cpu]; cpu 362 kernel/sched.c p->se.parent = task_group(p)->se[cpu]; cpu 366 kernel/sched.c p->rt.rt_rq = task_group(p)->rt_rq[cpu]; cpu 367 kernel/sched.c p->rt.parent = task_group(p)->rt_se[cpu]; cpu 569 kernel/sched.c int cpu; cpu 620 kernel/sched.c return rq->cpu; cpu 634 kernel/sched.c for (__sd = rcu_dereference(cpu_rq(cpu)->sd); __sd; __sd = __sd->parent) cpu 636 kernel/sched.c #define cpu_rq(cpu) (&per_cpu(runqueues, (cpu))) cpu 639 kernel/sched.c #define cpu_curr(cpu) (cpu_rq(cpu)->curr) cpu 664 kernel/sched.c int cpu = get_cpu(); cpu 665 kernel/sched.c struct rq *rq = cpu_rq(cpu); cpu 1079 kernel/sched.c int cpu = (int)(long)hcpu; cpu 1088 kernel/sched.c hrtick_clear(cpu_rq(cpu)); cpu 1158 kernel/sched.c int cpu; cpu 1167 kernel/sched.c cpu = task_cpu(p); cpu 1168 kernel/sched.c if (cpu == smp_processor_id()) cpu 1174 kernel/sched.c smp_send_reschedule(cpu); cpu 1179 kernel/sched.c struct rq *rq = cpu_rq(cpu); cpu 1184 kernel/sched.c resched_task(cpu_curr(cpu)); cpu 1201 kernel/sched.c struct rq *rq = cpu_rq(cpu); cpu 1203 kernel/sched.c if (cpu == smp_processor_id()) cpu 1226 kernel/sched.c smp_send_reschedule(cpu); cpu 1440 kernel/sched.c struct rq *rq = cpu_rq(cpu); cpu 1463 kernel/sched.c if (!tg->se[cpu]) cpu 1466 kernel/sched.c rq_weight = tg->cfs_rq[cpu]->load.weight; cpu 1492 kernel/sched.c tg->cfs_rq[cpu]->shares = boost ? 0 : shares; cpu 1493 kernel/sched.c tg->cfs_rq[cpu]->rq_weight = rq_weight; cpu 1500 kernel/sched.c __set_se_shares(tg->se[cpu], shares); cpu 1549 kernel/sched.c long cpu = (long)data; cpu 1552 kernel/sched.c load = cpu_rq(cpu)->load.weight; cpu 1554 kernel/sched.c load = tg->parent->cfs_rq[cpu]->h_load; cpu 1555 kernel/sched.c load *= tg->cfs_rq[cpu]->shares; cpu 1556 kernel/sched.c load /= tg->parent->cfs_rq[cpu]->load.weight + 1; cpu 1559 kernel/sched.c tg->cfs_rq[cpu]->h_load = load; cpu 1584 kernel/sched.c walk_tg_tree(tg_load_down, tg_nop, (void *)cpu); cpu 1760 kernel/sched.c set_task_rq(p, cpu); cpu 1768 kernel/sched.c task_thread_info(p)->cpu = cpu; cpu 1789 kernel/sched.c return cpu_rq(cpu)->load.weight; cpu 2003 kernel/sched.c int cpu; cpu 2006 kernel/sched.c cpu = task_cpu(p); cpu 2007 kernel/sched.c if ((cpu != smp_processor_id()) && task_curr(p)) cpu 2008 kernel/sched.c smp_send_reschedule(cpu); cpu 2021 kernel/sched.c struct rq *rq = cpu_rq(cpu); cpu 2022 kernel/sched.c unsigned long total = weighted_cpuload(cpu); cpu 2036 kernel/sched.c struct rq *rq = cpu_rq(cpu); cpu 2037 kernel/sched.c unsigned long total = weighted_cpuload(cpu); cpu 2141 kernel/sched.c for_each_domain(cpu, tmp) { cpu 2165 kernel/sched.c group = find_idlest_group(sd, t, cpu); cpu 2171 kernel/sched.c new_cpu = find_idlest_cpu(group, t, cpu, &tmpmask); cpu 2172 kernel/sched.c if (new_cpu == -1 || new_cpu == cpu) { cpu 2179 kernel/sched.c cpu = new_cpu; cpu 2182 kernel/sched.c for_each_domain(cpu, tmp) { cpu 2191 kernel/sched.c return cpu; cpu 2212 kernel/sched.c int cpu, orig_cpu, this_cpu, success = 0; cpu 2225 kernel/sched.c cpu = task_cpu(p); cpu 2228 kernel/sched.c if (cpu_isset(cpu, sd->span)) { cpu 2245 kernel/sched.c cpu = task_cpu(p); cpu 2246 kernel/sched.c orig_cpu = cpu; cpu 2253 kernel/sched.c cpu = p->sched_class->select_task_rq(p, sync); cpu 2254 kernel/sched.c if (cpu != orig_cpu) { cpu 2255 kernel/sched.c set_task_cpu(p, cpu); cpu 2266 kernel/sched.c cpu = task_cpu(p); cpu 2271 kernel/sched.c if (cpu == this_cpu) cpu 2276 kernel/sched.c if (cpu_isset(cpu, sd->span)) { cpu 2289 kernel/sched.c if (orig_cpu != cpu) cpu 2291 kernel/sched.c if (cpu == this_cpu) cpu 2377 kernel/sched.c int cpu = get_cpu(); cpu 2382 kernel/sched.c cpu = sched_balance_self(cpu, SD_BALANCE_FORK); cpu 2384 kernel/sched.c set_task_cpu(p, cpu); cpu 3810 kernel/sched.c int cpu = smp_processor_id(); cpu 3813 kernel/sched.c cpu_set(cpu, nohz.cpu_mask); cpu 3814 kernel/sched.c cpu_rq(cpu)->in_nohz_recently = 1; cpu 3819 kernel/sched.c if (!cpu_active(cpu) && cpu 3820 kernel/sched.c atomic_read(&nohz.load_balancer) == cpu) { cpu 3821 kernel/sched.c if (atomic_cmpxchg(&nohz.load_balancer, cpu, -1) != cpu) cpu 3828 kernel/sched.c if (atomic_read(&nohz.load_balancer) == cpu) cpu 3835 kernel/sched.c if (atomic_cmpxchg(&nohz.load_balancer, -1, cpu) == -1) cpu 3837 kernel/sched.c } else if (atomic_read(&nohz.load_balancer) == cpu) cpu 3840 kernel/sched.c if (!cpu_isset(cpu, nohz.cpu_mask)) cpu 3843 kernel/sched.c cpu_clear(cpu, nohz.cpu_mask); cpu 3845 kernel/sched.c if (atomic_read(&nohz.load_balancer) == cpu) cpu 3846 kernel/sched.c if (atomic_cmpxchg(&nohz.load_balancer, cpu, -1) != cpu) cpu 3864 kernel/sched.c struct rq *rq = cpu_rq(cpu); cpu 3873 kernel/sched.c for_each_domain(cpu, sd) { cpu 3896 kernel/sched.c if (load_balance(cpu, rq, sd, idle, &balance, &tmp)) { cpu 3996 kernel/sched.c if (atomic_read(&nohz.load_balancer) == cpu) { cpu 3997 kernel/sched.c cpu_clear(cpu, nohz.cpu_mask); cpu 4021 kernel/sched.c if (rq->idle_at_tick && atomic_read(&nohz.load_balancer) == cpu && cpu 4023 kernel/sched.c resched_cpu(cpu); cpu 4031 kernel/sched.c if (rq->idle_at_tick && atomic_read(&nohz.load_balancer) != cpu && cpu 4032 kernel/sched.c cpu_isset(cpu, nohz.cpu_mask)) cpu 4264 kernel/sched.c int cpu = smp_processor_id(); cpu 4265 kernel/sched.c struct rq *rq = cpu_rq(cpu); cpu 4277 kernel/sched.c rq->idle_at_tick = idle_cpu(cpu); cpu 4278 kernel/sched.c trigger_load_balance(rq, cpu); cpu 4426 kernel/sched.c int cpu; cpu 4430 kernel/sched.c cpu = smp_processor_id(); cpu 4431 kernel/sched.c rq = cpu_rq(cpu); cpu 4432 kernel/sched.c rcu_qsctr_inc(cpu); cpu 4466 kernel/sched.c idle_balance(cpu, rq); cpu 4483 kernel/sched.c cpu = smp_processor_id(); cpu 4484 kernel/sched.c rq = cpu_rq(cpu); cpu 5075 kernel/sched.c return cpu_curr(cpu) == cpu_rq(cpu)->idle; cpu 5084 kernel/sched.c return cpu_rq(cpu)->idle; cpu 5856 kernel/sched.c struct rq *rq = cpu_rq(cpu); cpu 5863 kernel/sched.c idle->cpus_allowed = cpumask_of_cpu(cpu); cpu 5864 kernel/sched.c __set_task_cpu(idle, cpu); cpu 6045 kernel/sched.c int cpu = (long)data; cpu 6048 kernel/sched.c rq = cpu_rq(cpu); cpu 6058 kernel/sched.c if (cpu_is_offline(cpu)) { cpu 6064 kernel/sched.c active_load_balance(rq, cpu); cpu 6080 kernel/sched.c __migrate_task(req->task, cpu, req->dest_cpu); cpu 6396 kernel/sched.c for_each_domain(cpu, sd) cpu 6403 kernel/sched.c for_each_domain(cpu, sd) { cpu 6462 kernel/sched.c cpu_set(rq->cpu, rq->rd->online); cpu 6482 kernel/sched.c cpu_clear(rq->cpu, rq->rd->online); cpu 6495 kernel/sched.c int cpu = (long)hcpu; cpu 6503 kernel/sched.c p = kthread_create(migration_thread, hcpu, "migration/%d", cpu); cpu 6506 kernel/sched.c kthread_bind(p, cpu); cpu 6511 kernel/sched.c cpu_rq(cpu)->migration_thread = p; cpu 6517 kernel/sched.c wake_up_process(cpu_rq(cpu)->migration_thread); cpu 6520 kernel/sched.c rq = cpu_rq(cpu); cpu 6523 kernel/sched.c BUG_ON(!cpu_isset(cpu, rq->rd->span)); cpu 6533 kernel/sched.c if (!cpu_rq(cpu)->migration_thread) cpu 6536 kernel/sched.c kthread_bind(cpu_rq(cpu)->migration_thread, cpu 6538 kernel/sched.c kthread_stop(cpu_rq(cpu)->migration_thread); cpu 6539 kernel/sched.c cpu_rq(cpu)->migration_thread = NULL; cpu 6545 kernel/sched.c migrate_live_tasks(cpu); cpu 6546 kernel/sched.c rq = cpu_rq(cpu); cpu 6556 kernel/sched.c migrate_dead_tasks(cpu); cpu 6582 kernel/sched.c rq = cpu_rq(cpu); cpu 6585 kernel/sched.c BUG_ON(!cpu_isset(cpu, rq->rd->span)); cpu 6605 kernel/sched.c void *cpu = (void *)(long)smp_processor_id(); cpu 6609 kernel/sched.c err = migration_call(&migration_notifier, CPU_UP_PREPARE, cpu); cpu 6611 kernel/sched.c migration_call(&migration_notifier, CPU_ONLINE, cpu); cpu 6667 kernel/sched.c if (!cpu_isset(cpu, sd->span)) { cpu 6669 kernel/sched.c "CPU%d\n", cpu); cpu 6671 kernel/sched.c if (!cpu_isset(cpu, group->cpumask)) { cpu 6673 kernel/sched.c " CPU%d\n", cpu); cpu 6727 kernel/sched.c printk(KERN_DEBUG "CPU%d attaching NULL sched-domain.\n", cpu); cpu 6731 kernel/sched.c printk(KERN_DEBUG "CPU%d attaching sched-domain:\n", cpu); cpu 6740 kernel/sched.c if (sched_domain_debug_one(sd, cpu, level, groupmask)) cpu 6817 kernel/sched.c if (cpu_isset(rq->cpu, old_rd->online)) cpu 6820 kernel/sched.c cpu_clear(rq->cpu, old_rd->span); cpu 6829 kernel/sched.c cpu_set(rq->cpu, rd->span); cpu 6830 kernel/sched.c if (cpu_isset(rq->cpu, cpu_online_map)) cpu 6872 kernel/sched.c struct rq *rq = cpu_rq(cpu); cpu 6893 kernel/sched.c sched_domain_debug(sd, cpu); cpu 7055 kernel/sched.c *sg = &per_cpu(sched_group_cpus, cpu); cpu 7056 kernel/sched.c return cpu; cpu 7075 kernel/sched.c *mask = per_cpu(cpu_sibling_map, cpu); cpu 7088 kernel/sched.c *sg = &per_cpu(sched_group_core, cpu); cpu 7089 kernel/sched.c return cpu; cpu 7102 kernel/sched.c *mask = cpu_coregroup_map(cpu); cpu 7106 kernel/sched.c *mask = per_cpu(cpu_sibling_map, cpu); cpu 7110 kernel/sched.c group = cpu; cpu 7134 kernel/sched.c *nodemask = node_to_cpumask(cpu_to_node(cpu)); cpu 7174 kernel/sched.c int cpu, i; cpu 7176 kernel/sched.c for_each_cpu_mask_nr(cpu, *cpu_map) { cpu 7178 kernel/sched.c = sched_group_nodes_bycpu[cpu]; cpu 7202 kernel/sched.c sched_group_nodes_bycpu[cpu] = NULL; cpu 7232 kernel/sched.c if (cpu != first_cpu(sd->groups->cpumask)) cpu 7949 kernel/sched.c int cpu = (int)(long)hcpu; cpu 7954 kernel/sched.c disable_runtime(cpu_rq(cpu)); cpu 7961 kernel/sched.c enable_runtime(cpu_rq(cpu)); cpu 8063 kernel/sched.c struct rq *rq = cpu_rq(cpu); cpu 8064 kernel/sched.c tg->cfs_rq[cpu] = cfs_rq; cpu 8070 kernel/sched.c tg->se[cpu] = se; cpu 8092 kernel/sched.c struct rq *rq = cpu_rq(cpu); cpu 8094 kernel/sched.c tg->rt_rq[cpu] = rt_rq; cpu 8102 kernel/sched.c tg->rt_se[cpu] = rt_se; cpu 8273 kernel/sched.c rq->cpu = i; cpu 8428 kernel/sched.c return cpu_curr(cpu); cpu 8448 kernel/sched.c cpu_curr(cpu) = p; cpu 8511 kernel/sched.c list_add_rcu(&tg->cfs_rq[cpu]->leaf_cfs_rq_list, cpu 8512 kernel/sched.c &cpu_rq(cpu)->leaf_cfs_rq_list); cpu 8517 kernel/sched.c list_del_rcu(&tg->cfs_rq[cpu]->leaf_cfs_rq_list); cpu 8600 kernel/sched.c list_add_rcu(&tg->rt_rq[cpu]->leaf_rt_rq_list, cpu 8601 kernel/sched.c &cpu_rq(cpu)->leaf_rt_rq_list); cpu 8606 kernel/sched.c list_del_rcu(&tg->rt_rq[cpu]->leaf_rt_rq_list); cpu 69 kernel/sched_clock.c return &per_cpu(sched_clock_data, cpu); cpu 75 kernel/sched_clock.c int cpu; cpu 77 kernel/sched_clock.c for_each_possible_cpu(cpu) { cpu 78 kernel/sched_clock.c struct sched_clock_data *scd = cpu_sdc(cpu); cpu 151 kernel/sched_clock.c struct sched_clock_data *scd = cpu_sdc(cpu); cpu 160 kernel/sched_clock.c if (cpu != raw_smp_processor_id()) { cpu 258 kernel/sched_clock.c clock = sched_clock_cpu(cpu); cpu 106 kernel/sched_cpupri.c int *currpri = &cp->cpu_to_pri[cpu]; cpu 129 kernel/sched_cpupri.c cpu_clear(cpu, vec->mask); cpu 139 kernel/sched_cpupri.c cpu_set(cpu, vec->mask); cpu 118 kernel/sched_debug.c struct rq *rq = &per_cpu(runqueues, cpu); cpu 133 kernel/sched_debug.c SEQ_printf(m, "\ncfs_rq[%d]:%s\n", cpu, path); cpu 135 kernel/sched_debug.c SEQ_printf(m, "\ncfs_rq[%d]:\n", cpu); cpu 205 kernel/sched_debug.c SEQ_printf(m, "\nrt_rq[%d]:%s\n", cpu, path); cpu 207 kernel/sched_debug.c SEQ_printf(m, "\nrt_rq[%d]:\n", cpu); cpu 227 kernel/sched_debug.c struct rq *rq = &per_cpu(runqueues, cpu); cpu 234 kernel/sched_debug.c cpu, freq / 1000, (freq % 1000)); cpu 237 kernel/sched_debug.c SEQ_printf(m, "\ncpu#%d\n", cpu); cpu 263 kernel/sched_debug.c print_cfs_stats(m, cpu); cpu 264 kernel/sched_debug.c print_rt_stats(m, cpu); cpu 266 kernel/sched_debug.c print_rq(m, rq, cpu); cpu 272 kernel/sched_debug.c int cpu; cpu 293 kernel/sched_debug.c for_each_online_cpu(cpu) cpu 294 kernel/sched_debug.c print_cpu(m, cpu); cpu 971 kernel/sched_fair.c if (idle_cpu(cpu) || cpu_rq(cpu)->cfs.nr_running > 1) cpu 972 kernel/sched_fair.c return cpu; cpu 974 kernel/sched_fair.c for_each_domain(cpu, sd) { cpu 993 kernel/sched_fair.c return cpu; cpu 998 kernel/sched_fair.c return cpu; cpu 1031 kernel/sched_fair.c struct sched_entity *se = tg->se[cpu]; cpu 1610 kernel/sched_fair.c for_each_leaf_cfs_rq(cpu_rq(cpu), cfs_rq) cpu 1611 kernel/sched_fair.c print_cfs_rq(m, cpu, cfs_rq); cpu 18 kernel/sched_rt.c cpu_set(rq->cpu, rq->rd->rto_mask); cpu 37 kernel/sched_rt.c cpu_clear(rq->cpu, rq->rd->rto_mask); cpu 156 kernel/sched_rt.c return container_of(rt_b, struct task_group, rt_bandwidth)->rt_rq[cpu]; cpu 223 kernel/sched_rt.c return &cpu_rq(cpu)->rt; cpu 562 kernel/sched_rt.c cpupri_set(&rq->rd->cpupri, rq->cpu, cpu 621 kernel/sched_rt.c cpupri_set(&rq->rd->cpupri, rq->cpu, cpu 792 kernel/sched_rt.c int cpu = find_lowest_rq(p); cpu 794 kernel/sched_rt.c return (cpu == -1) ? task_cpu(p) : cpu; cpu 918 kernel/sched_rt.c (cpu < 0 || cpu_isset(cpu, p->cpus_allowed)) && cpu 943 kernel/sched_rt.c if (pick_rt_task(rq, p, cpu)) { cpu 979 kernel/sched_rt.c int cpu = task_cpu(task); cpu 1002 kernel/sched_rt.c if (cpu_isset(cpu, *lowest_mask)) cpu 1003 kernel/sched_rt.c return cpu; cpu 1009 kernel/sched_rt.c if (this_cpu == cpu) cpu 1012 kernel/sched_rt.c for_each_domain(cpu, sd) { cpu 1039 kernel/sched_rt.c int cpu; cpu 1042 kernel/sched_rt.c cpu = find_lowest_rq(task); cpu 1044 kernel/sched_rt.c if ((cpu == -1) || (cpu == rq->cpu)) cpu 1047 kernel/sched_rt.c lowest_rq = cpu_rq(cpu); cpu 1058 kernel/sched_rt.c !cpu_isset(lowest_rq->cpu, cpu 1138 kernel/sched_rt.c set_task_cpu(next_task, lowest_rq->cpu); cpu 1171 kernel/sched_rt.c int this_cpu = this_rq->cpu, ret = 0, cpu; cpu 1180 kernel/sched_rt.c for_each_cpu_mask_nr(cpu, this_rq->rd->rto_mask) { cpu 1181 kernel/sched_rt.c if (this_cpu == cpu) cpu 1184 kernel/sched_rt.c src_rq = cpu_rq(cpu); cpu 1344 kernel/sched_rt.c cpupri_set(&rq->rd->cpupri, rq->cpu, rq->rt.highest_prio); cpu 1355 kernel/sched_rt.c cpupri_set(&rq->rd->cpupri, rq->cpu, CPUPRI_INVALID); cpu 1541 kernel/sched_rt.c for_each_leaf_rt_rq(rt_rq, cpu_rq(cpu)) cpu 1542 kernel/sched_rt.c print_rt_rq(m, cpu, rt_rq); cpu 11 kernel/sched_stats.h int cpu; cpu 20 kernel/sched_stats.h for_each_online_cpu(cpu) { cpu 21 kernel/sched_stats.h struct rq *rq = cpu_rq(cpu); cpu 30 kernel/sched_stats.h cpu, rq->yld_both_empty, cpu 42 kernel/sched_stats.h for_each_domain(cpu, sd) { cpu 70 kernel/smp.c struct call_single_queue *dst = &per_cpu(call_single_queue, cpu); cpu 80 kernel/smp.c arch_send_call_function_single_ipi(cpu); cpu 102 kernel/smp.c int cpu = get_cpu(); cpu 112 kernel/smp.c if (!cpu_isset(cpu, data->cpumask)) cpu 118 kernel/smp.c cpu_clear(cpu, data->cpumask); cpu 221 kernel/smp.c if (cpu == me) { cpu 225 kernel/smp.c } else if ((unsigned)cpu < NR_CPUS && cpu_online(cpu)) { cpu 240 kernel/smp.c generic_exec_single(cpu, data); cpu 265 kernel/smp.c generic_exec_single(cpu, data); cpu 293 kernel/smp.c int cpu; cpu 298 kernel/smp.c for_each_cpu_mask(cpu, mask) { cpu 300 kernel/smp.c generic_exec_single(cpu, &data); cpu 328 kernel/smp.c int cpu, num_cpus; cpu 334 kernel/smp.c cpu = smp_processor_id(); cpu 336 kernel/smp.c cpu_clear(cpu, allbutself); cpu 347 kernel/smp.c cpu = first_cpu(mask); cpu 348 kernel/smp.c return smp_call_function_single(cpu, func, info, wait); cpu 189 kernel/softirq.c int cpu; cpu 197 kernel/softirq.c cpu = smp_processor_id(); cpu 209 kernel/softirq.c rcu_bh_qsctr_inc(cpu); cpu 258 kernel/softirq.c int cpu = smp_processor_id(); cpu 259 kernel/softirq.c if (idle_cpu(cpu) && !in_interrupt()) cpu 260 kernel/softirq.c tick_nohz_stop_idle(cpu); cpu 264 kernel/softirq.c if (idle_cpu(cpu)) cpu 468 kernel/softirq.c int cpu; cpu 470 kernel/softirq.c for_each_possible_cpu(cpu) { cpu 471 kernel/softirq.c per_cpu(tasklet_vec, cpu).tail = cpu 472 kernel/softirq.c &per_cpu(tasklet_vec, cpu).head; cpu 473 kernel/softirq.c per_cpu(tasklet_hi_vec, cpu).tail = cpu 474 kernel/softirq.c &per_cpu(tasklet_hi_vec, cpu).head; cpu 538 kernel/softirq.c BUG_ON(cpu_online(cpu)); cpu 545 kernel/softirq.c for (i = &per_cpu(tasklet_vec, cpu).head; *i; i = &(*i)->next) { cpu 550 kernel/softirq.c per_cpu(tasklet_vec, cpu).tail = i; cpu 563 kernel/softirq.c if (&per_cpu(tasklet_vec, cpu).head != per_cpu(tasklet_vec, cpu).tail) { cpu 564 kernel/softirq.c *(__get_cpu_var(tasklet_vec).tail) = per_cpu(tasklet_vec, cpu).head; cpu 565 kernel/softirq.c __get_cpu_var(tasklet_vec).tail = per_cpu(tasklet_vec, cpu).tail; cpu 566 kernel/softirq.c per_cpu(tasklet_vec, cpu).head = NULL; cpu 567 kernel/softirq.c per_cpu(tasklet_vec, cpu).tail = &per_cpu(tasklet_vec, cpu).head; cpu 571 kernel/softirq.c if (&per_cpu(tasklet_hi_vec, cpu).head != per_cpu(tasklet_hi_vec, cpu).tail) { cpu 572 kernel/softirq.c *__get_cpu_var(tasklet_hi_vec).tail = per_cpu(tasklet_hi_vec, cpu).head; cpu 573 kernel/softirq.c __get_cpu_var(tasklet_hi_vec).tail = per_cpu(tasklet_hi_vec, cpu).tail; cpu 574 kernel/softirq.c per_cpu(tasklet_hi_vec, cpu).head = NULL; cpu 575 kernel/softirq.c per_cpu(tasklet_hi_vec, cpu).tail = &per_cpu(tasklet_hi_vec, cpu).head; cpu 635 kernel/softirq.c void *cpu = (void *)(long)smp_processor_id(); cpu 636 kernel/softirq.c int err = cpu_callback(&cpu_nfb, CPU_UP_PREPARE, cpu); cpu 639 kernel/softirq.c cpu_callback(&cpu_nfb, CPU_ONLINE, cpu); cpu 83 kernel/softlockup.c int cpu; cpu 86 kernel/softlockup.c for_each_online_cpu(cpu) cpu 87 kernel/softlockup.c per_cpu(touch_timestamp, cpu) = 0; cpu 353 kernel/softlockup.c void *cpu = (void *)(long)smp_processor_id(); cpu 359 kernel/softlockup.c err = cpu_callback(&cpu_nfb, CPU_UP_PREPARE, cpu); cpu 364 kernel/softlockup.c cpu_callback(&cpu_nfb, CPU_ONLINE, cpu); cpu 60 kernel/srcu.c int cpu; cpu 64 kernel/srcu.c for_each_possible_cpu(cpu) cpu 65 kernel/srcu.c sum += per_cpu_ptr(sp->per_cpu_ref, cpu)->c[idx]; cpu 1739 kernel/sys.c int cpu = raw_smp_processor_id(); cpu 1741 kernel/sys.c err |= put_user(cpu, cpup); cpu 1743 kernel/sys.c err |= put_user(cpu_to_node(cpu), nodep); cpu 297 kernel/taskstats.c unsigned int cpu; cpu 304 kernel/taskstats.c for_each_cpu_mask_nr(cpu, mask) { cpu 306 kernel/taskstats.c cpu_to_node(cpu)); cpu 313 kernel/taskstats.c listeners = &per_cpu(listener_array, cpu); cpu 323 kernel/taskstats.c for_each_cpu_mask_nr(cpu, mask) { cpu 324 kernel/taskstats.c listeners = &per_cpu(listener_array, cpu); cpu 107 kernel/time/tick-broadcast.c cpu_set(cpu, tick_broadcast_mask); cpu 117 kernel/time/tick-broadcast.c int cpu = smp_processor_id(); cpu 119 kernel/time/tick-broadcast.c cpu_clear(cpu, tick_broadcast_mask); cpu 120 kernel/time/tick-broadcast.c tick_broadcast_clear_oneshot(cpu); cpu 132 kernel/time/tick-broadcast.c int cpu = smp_processor_id(); cpu 138 kernel/time/tick-broadcast.c if (cpu_isset(cpu, mask)) { cpu 139 kernel/time/tick-broadcast.c cpu_clear(cpu, mask); cpu 140 kernel/time/tick-broadcast.c td = &per_cpu(tick_cpu_device, cpu); cpu 151 kernel/time/tick-broadcast.c cpu = first_cpu(mask); cpu 152 kernel/time/tick-broadcast.c td = &per_cpu(tick_cpu_device, cpu); cpu 213 kernel/time/tick-broadcast.c int cpu, bc_stopped; cpu 217 kernel/time/tick-broadcast.c cpu = smp_processor_id(); cpu 218 kernel/time/tick-broadcast.c td = &per_cpu(tick_cpu_device, cpu); cpu 236 kernel/time/tick-broadcast.c if (!cpu_isset(cpu, tick_broadcast_mask)) { cpu 237 kernel/time/tick-broadcast.c cpu_set(cpu, tick_broadcast_mask); cpu 247 kernel/time/tick-broadcast.c cpu_isset(cpu, tick_broadcast_mask)) { cpu 248 kernel/time/tick-broadcast.c cpu_clear(cpu, tick_broadcast_mask); cpu 301 kernel/time/tick-broadcast.c unsigned int cpu = *cpup; cpu 306 kernel/time/tick-broadcast.c cpu_clear(cpu, tick_broadcast_mask); cpu 394 kernel/time/tick-broadcast.c int cpu; cpu 403 kernel/time/tick-broadcast.c for_each_cpu_mask_nr(cpu, tick_broadcast_oneshot_mask) { cpu 404 kernel/time/tick-broadcast.c td = &per_cpu(tick_cpu_device, cpu); cpu 406 kernel/time/tick-broadcast.c cpu_set(cpu, mask); cpu 446 kernel/time/tick-broadcast.c int cpu; cpu 458 kernel/time/tick-broadcast.c cpu = smp_processor_id(); cpu 459 kernel/time/tick-broadcast.c td = &per_cpu(tick_cpu_device, cpu); cpu 466 kernel/time/tick-broadcast.c if (!cpu_isset(cpu, tick_broadcast_oneshot_mask)) { cpu 467 kernel/time/tick-broadcast.c cpu_set(cpu, tick_broadcast_oneshot_mask); cpu 473 kernel/time/tick-broadcast.c if (cpu_isset(cpu, tick_broadcast_oneshot_mask)) { cpu 474 kernel/time/tick-broadcast.c cpu_clear(cpu, tick_broadcast_oneshot_mask); cpu 492 kernel/time/tick-broadcast.c cpu_clear(cpu, tick_broadcast_oneshot_mask); cpu 498 kernel/time/tick-broadcast.c int cpu; cpu 500 kernel/time/tick-broadcast.c for_each_cpu_mask_nr(cpu, *mask) { cpu 501 kernel/time/tick-broadcast.c td = &per_cpu(tick_cpu_device, cpu); cpu 515 kernel/time/tick-broadcast.c int cpu = smp_processor_id(); cpu 522 kernel/time/tick-broadcast.c tick_do_timer_cpu = cpu; cpu 531 kernel/time/tick-broadcast.c cpu_clear(cpu, mask); cpu 567 kernel/time/tick-broadcast.c unsigned int cpu = *cpup; cpu 575 kernel/time/tick-broadcast.c cpu_clear(cpu, tick_broadcast_oneshot_mask); cpu 44 kernel/time/tick-common.c return &per_cpu(tick_cpu_device, cpu); cpu 62 kernel/time/tick-common.c if (tick_do_timer_cpu == cpu) { cpu 81 kernel/time/tick-common.c int cpu = smp_processor_id(); cpu 84 kernel/time/tick-common.c tick_periodic(cpu); cpu 96 kernel/time/tick-common.c tick_periodic(cpu); cpu 153 kernel/time/tick-common.c tick_do_timer_cpu = cpu; cpu 183 kernel/time/tick-common.c if (tick_device_uses_broadcast(newdev, cpu)) cpu 199 kernel/time/tick-common.c int cpu, ret = NOTIFY_OK; cpu 204 kernel/time/tick-common.c cpu = smp_processor_id(); cpu 205 kernel/time/tick-common.c if (!cpu_isset(cpu, newdev->cpumask)) cpu 208 kernel/time/tick-common.c td = &per_cpu(tick_cpu_device, cpu); cpu 212 kernel/time/tick-common.c if (!cpus_equal(newdev->cpumask, cpumask_of_cpu(cpu))) { cpu 225 kernel/time/tick-common.c if (curdev && cpus_equal(curdev->cpumask, cpumask_of_cpu(cpu))) cpu 257 kernel/time/tick-common.c tick_setup_device(td, newdev, cpu, &cpumask_of_cpu(cpu)); cpu 302 kernel/time/tick-common.c int cpu = first_cpu(cpu_online_map); cpu 304 kernel/time/tick-common.c tick_do_timer_cpu = (cpu != NR_CPUS) ? cpu : cpu 41 kernel/time/tick-sched.c return &per_cpu(tick_cpu_sched, cpu); cpu 139 kernel/time/tick-sched.c int cpu = smp_processor_id(); cpu 140 kernel/time/tick-sched.c struct tick_sched *ts = &per_cpu(tick_cpu_sched, cpu); cpu 147 kernel/time/tick-sched.c cpu_clear(cpu, nohz_cpu_mask); cpu 160 kernel/time/tick-sched.c struct tick_sched *ts = &per_cpu(tick_cpu_sched, cpu); cpu 192 kernel/time/tick-sched.c struct tick_sched *ts = &per_cpu(tick_cpu_sched, cpu); cpu 219 kernel/time/tick-sched.c int cpu; cpu 223 kernel/time/tick-sched.c cpu = smp_processor_id(); cpu 224 kernel/time/tick-sched.c ts = &per_cpu(tick_cpu_sched, cpu); cpu 234 kernel/time/tick-sched.c if (unlikely(!cpu_online(cpu))) { cpu 235 kernel/time/tick-sched.c if (cpu == tick_do_timer_cpu) cpu 273 kernel/time/tick-sched.c if (rcu_needs_cpu(cpu)) cpu 286 kernel/time/tick-sched.c cpu_set(cpu, nohz_cpu_mask); cpu 299 kernel/time/tick-sched.c cpu_clear(cpu, nohz_cpu_mask); cpu 317 kernel/time/tick-sched.c if (cpu == tick_do_timer_cpu) cpu 357 kernel/time/tick-sched.c cpu_clear(cpu, nohz_cpu_mask); cpu 387 kernel/time/tick-sched.c int cpu = smp_processor_id(); cpu 388 kernel/time/tick-sched.c struct tick_sched *ts = &per_cpu(tick_cpu_sched, cpu); cpu 393 kernel/time/tick-sched.c tick_nohz_stop_idle(cpu); cpu 409 kernel/time/tick-sched.c cpu_clear(cpu, nohz_cpu_mask); cpu 471 kernel/time/tick-sched.c int cpu = smp_processor_id(); cpu 484 kernel/time/tick-sched.c tick_do_timer_cpu = cpu; cpu 487 kernel/time/tick-sched.c if (tick_do_timer_cpu == cpu) cpu 575 kernel/time/tick-sched.c int cpu = smp_processor_id(); cpu 586 kernel/time/tick-sched.c tick_do_timer_cpu = cpu; cpu 590 kernel/time/tick-sched.c if (tick_do_timer_cpu == cpu) cpu 666 kernel/time/tick-sched.c struct tick_sched *ts = &per_cpu(tick_cpu_sched, cpu); cpu 682 kernel/time/tick-sched.c int cpu; cpu 684 kernel/time/tick-sched.c for_each_possible_cpu(cpu) cpu 685 kernel/time/tick-sched.c set_bit(0, &per_cpu(tick_cpu_sched, cpu).check_clocks); cpu 129 kernel/time/timer_list.c struct hrtimer_cpu_base *cpu_base = &per_cpu(hrtimer_bases, cpu); cpu 133 kernel/time/timer_list.c SEQ_printf(m, "cpu: %d\n", cpu); cpu 161 kernel/time/timer_list.c struct tick_sched *ts = tick_get_tick_sched(cpu); cpu 222 kernel/time/timer_list.c int cpu; cpu 234 kernel/time/timer_list.c for_each_online_cpu(cpu) cpu 235 kernel/time/timer_list.c print_tickdevice(m, tick_get_device(cpu)); cpu 245 kernel/time/timer_list.c int cpu; cpu 251 kernel/time/timer_list.c for_each_online_cpu(cpu) cpu 252 kernel/time/timer_list.c print_cpu(m, cpu, now); cpu 348 kernel/time/timer_stats.c int cpu; cpu 350 kernel/time/timer_stats.c for_each_online_cpu(cpu) { cpu 351 kernel/time/timer_stats.c spin_lock_irqsave(&per_cpu(lookup_lock, cpu), flags); cpu 353 kernel/time/timer_stats.c spin_unlock_irqrestore(&per_cpu(lookup_lock, cpu), flags); cpu 408 kernel/time/timer_stats.c int cpu; cpu 410 kernel/time/timer_stats.c for_each_possible_cpu(cpu) cpu 411 kernel/time/timer_stats.c spin_lock_init(&per_cpu(lookup_lock, cpu)); cpu 148 kernel/timer.c j += cpu * 3; cpu 164 kernel/timer.c j -= cpu * 3; cpu 200 kernel/timer.c return __round_jiffies(j + jiffies, cpu) - jiffies; cpu 579 kernel/timer.c struct tvec_base *base = per_cpu(tvec_bases, cpu); cpu 596 kernel/timer.c wake_up_idle_cpu(cpu); cpu 974 kernel/timer.c int cpu = smp_processor_id(); cpu 979 kernel/timer.c if (rcu_pending(cpu)) cpu 980 kernel/timer.c rcu_check_callbacks(cpu, user_tick); cpu 1366 kernel/timer.c if (!tvec_base_done[cpu]) { cpu 1375 kernel/timer.c cpu_to_node(cpu)); cpu 1385 kernel/timer.c per_cpu(tvec_bases, cpu) = base; cpu 1396 kernel/timer.c tvec_base_done[cpu] = 1; cpu 1398 kernel/timer.c base = per_cpu(tvec_bases, cpu); cpu 1435 kernel/timer.c BUG_ON(cpu_online(cpu)); cpu 1436 kernel/timer.c old_base = per_cpu(tvec_bases, cpu); cpu 1464 kernel/timer.c long cpu = (long)hcpu; cpu 1468 kernel/timer.c if (init_timers_cpu(cpu) < 0) cpu 1474 kernel/timer.c migrate_timers(cpu); cpu 342 kernel/trace/ftrace.c int cpu; cpu 356 kernel/trace/ftrace.c cpu = raw_smp_processor_id(); cpu 357 kernel/trace/ftrace.c per_cpu(ftrace_shutdown_disable_cpu, cpu)++; cpu 358 kernel/trace/ftrace.c if (per_cpu(ftrace_shutdown_disable_cpu, cpu) != 1) cpu 392 kernel/trace/ftrace.c per_cpu(ftrace_shutdown_disable_cpu, cpu)--; cpu 44 kernel/trace/trace.c for_each_cpu_mask(cpu, tracing_buffer_mask) cpu 63 kernel/trace/trace.c return cpu_clock(cpu); cpu 138 kernel/trace/trace.c int cpu; cpu 142 kernel/trace/trace.c for_each_online_cpu(cpu) cpu 143 kernel/trace/trace.c tracing_reset(tr->data[cpu]); cpu 250 kernel/trace/trace.c struct trace_array_cpu *data = tr->data[cpu]; cpu 252 kernel/trace/trace.c max_tr.cpu = cpu; cpu 255 kernel/trace/trace.c data = max_tr.data[cpu]; cpu 515 kernel/trace/trace.c __update_max_tr(tr, tsk, cpu); cpu 530 kernel/trace/trace.c struct trace_array_cpu *data = tr->data[cpu]; cpu 538 kernel/trace/trace.c flip_trace(max_tr.data[cpu], data); cpu 541 kernel/trace/trace.c __update_max_tr(tr, tsk, cpu); cpu 1007 kernel/trace/trace.c int cpu; cpu 1013 kernel/trace/trace.c cpu = raw_smp_processor_id(); cpu 1014 kernel/trace/trace.c data = tr->data[cpu]; cpu 1032 kernel/trace/trace.c int cpu; cpu 1041 kernel/trace/trace.c cpu = raw_smp_processor_id(); cpu 1042 kernel/trace/trace.c data = tr->data[cpu]; cpu 1083 kernel/trace/trace.c if (iter->next_idx[cpu] >= tr->entries || cpu 1084 kernel/trace/trace.c iter->next_idx[cpu] >= data->trace_idx || cpu 1089 kernel/trace/trace.c if (!iter->next_page[cpu]) { cpu 1093 kernel/trace/trace.c iter->next_page[cpu] = &page->lru; cpu 1094 kernel/trace/trace.c iter->next_page_idx[cpu] = data->trace_tail_idx; cpu 1097 kernel/trace/trace.c page = list_entry(iter->next_page[cpu], struct page, lru); cpu 1102 kernel/trace/trace.c WARN_ON(iter->next_page_idx[cpu] >= ENTRIES_PER_PAGE); cpu 1103 kernel/trace/trace.c return &array[iter->next_page_idx[cpu]]; cpu 1112 kernel/trace/trace.c int cpu; cpu 1114 kernel/trace/trace.c for_each_tracing_cpu(cpu) { cpu 1115 kernel/trace/trace.c if (!head_page(tr->data[cpu])) cpu 1117 kernel/trace/trace.c ent = trace_entry_idx(tr, tr->data[cpu], iter, cpu); cpu 1123 kernel/trace/trace.c next_cpu = cpu; cpu 1136 kernel/trace/trace.c iter->next_idx[iter->cpu]++; cpu 1137 kernel/trace/trace.c iter->next_page_idx[iter->cpu]++; cpu 1139 kernel/trace/trace.c if (iter->next_page_idx[iter->cpu] >= ENTRIES_PER_PAGE) { cpu 1140 kernel/trace/trace.c struct trace_array_cpu *data = iter->tr->data[iter->cpu]; cpu 1142 kernel/trace/trace.c iter->next_page_idx[iter->cpu] = 0; cpu 1143 kernel/trace/trace.c iter->next_page[iter->cpu] = cpu 1144 kernel/trace/trace.c trace_next_list(data, iter->next_page[iter->cpu]); cpu 1150 kernel/trace/trace.c struct trace_array_cpu *data = iter->tr->data[iter->cpu]; cpu 1172 kernel/trace/trace.c iter->prev_cpu = iter->cpu; cpu 1175 kernel/trace/trace.c iter->cpu = next_cpu; cpu 1230 kernel/trace/trace.c iter->cpu = 0; cpu 1356 kernel/trace/trace.c struct trace_array_cpu *data = tr->data[tr->cpu]; cpu 1360 kernel/trace/trace.c int cpu; cpu 1366 kernel/trace/trace.c for_each_tracing_cpu(cpu) { cpu 1367 kernel/trace/trace.c if (head_page(tr->data[cpu])) { cpu 1368 kernel/trace/trace.c total += tr->data[cpu]->trace_idx; cpu 1369 kernel/trace/trace.c if (tr->data[cpu]->trace_idx > tr->entries) cpu 1372 kernel/trace/trace.c entries += tr->data[cpu]->trace_idx; cpu 1385 kernel/trace/trace.c tr->cpu, cpu 1431 kernel/trace/trace.c trace_seq_printf(s, "%d", cpu); cpu 1499 kernel/trace/trace.c entry->pid, cpu, entry->flags, cpu 1506 kernel/trace/trace.c lat_print_generic(s, entry, cpu); cpu 1579 kernel/trace/trace.c ret = trace_seq_printf(s, "[%02d] ", iter->cpu); cpu 1663 kernel/trace/trace.c entry->pid, iter->cpu, entry->t); cpu 1727 kernel/trace/trace.c SEQ_PUT_HEX_FIELD_RET(s, iter->cpu); cpu 1771 kernel/trace/trace.c SEQ_PUT_FIELD_RET(s, entry->cpu); cpu 1800 kernel/trace/trace.c int cpu; cpu 1802 kernel/trace/trace.c for_each_tracing_cpu(cpu) { cpu 1803 kernel/trace/trace.c data = iter->tr->data[cpu]; cpu 1828 kernel/trace/trace.c return print_lat_fmt(iter, iter->idx, iter->cpu); cpu 2113 kernel/trace/trace.c int err, cpu; cpu 2122 kernel/trace/trace.c for_each_tracing_cpu(cpu) { cpu 2127 kernel/trace/trace.c if (cpu_isset(cpu, tracing_cpumask) && cpu 2128 kernel/trace/trace.c !cpu_isset(cpu, tracing_cpumask_new)) { cpu 2129 kernel/trace/trace.c atomic_inc(&global_trace.data[cpu]->disabled); cpu 2131 kernel/trace/trace.c if (!cpu_isset(cpu, tracing_cpumask) && cpu 2132 kernel/trace/trace.c cpu_isset(cpu, tracing_cpumask_new)) { cpu 2133 kernel/trace/trace.c atomic_dec(&global_trace.data[cpu]->disabled); cpu 2509 kernel/trace/trace.c int cpu; cpu 2605 kernel/trace/trace.c for_each_tracing_cpu(cpu) { cpu 2606 kernel/trace/trace.c data = iter->tr->data[cpu]; cpu 2612 kernel/trace/trace.c cpu_set(cpu, mask); cpu 2615 kernel/trace/trace.c for_each_cpu_mask(cpu, mask) { cpu 2616 kernel/trace/trace.c data = iter->tr->data[cpu]; cpu 2619 kernel/trace/trace.c if (data->overrun > iter->last_overrun[cpu]) cpu 2620 kernel/trace/trace.c iter->overrun[cpu] += cpu 2621 kernel/trace/trace.c data->overrun - iter->last_overrun[cpu]; cpu 2622 kernel/trace/trace.c iter->last_overrun[cpu] = data->overrun; cpu 2642 kernel/trace/trace.c for_each_cpu_mask(cpu, mask) { cpu 2643 kernel/trace/trace.c data = iter->tr->data[cpu]; cpu 2647 kernel/trace/trace.c for_each_cpu_mask(cpu, mask) { cpu 2648 kernel/trace/trace.c data = iter->tr->data[cpu]; cpu 71 kernel/trace/trace.h char cpu; cpu 129 kernel/trace/trace.h int cpu; cpu 180 kernel/trace/trace.h int cpu; cpu 21 kernel/trace/trace_functions.c int cpu; cpu 23 kernel/trace/trace_functions.c tr->time_start = ftrace_now(tr->cpu); cpu 25 kernel/trace/trace_functions.c for_each_online_cpu(cpu) cpu 26 kernel/trace/trace_functions.c tracing_reset(tr->data[cpu]); cpu 31 kernel/trace/trace_functions.c tr->cpu = get_cpu(); cpu 77 kernel/trace/trace_irqsoff.c int cpu; cpu 85 kernel/trace/trace_irqsoff.c cpu = raw_smp_processor_id(); cpu 86 kernel/trace/trace_irqsoff.c if (likely(!per_cpu(tracing_cpu, cpu))) cpu 94 kernel/trace/trace_irqsoff.c data = tr->data[cpu]; cpu 139 kernel/trace/trace_irqsoff.c T1 = ftrace_now(cpu); cpu 166 kernel/trace/trace_irqsoff.c update_max_tr_single(tr, current, cpu); cpu 175 kernel/trace/trace_irqsoff.c data->preempt_timestamp = ftrace_now(cpu); cpu 183 kernel/trace/trace_irqsoff.c int cpu; cpu 191 kernel/trace/trace_irqsoff.c cpu = raw_smp_processor_id(); cpu 193 kernel/trace/trace_irqsoff.c if (per_cpu(tracing_cpu, cpu)) cpu 196 kernel/trace/trace_irqsoff.c data = tr->data[cpu]; cpu 204 kernel/trace/trace_irqsoff.c data->preempt_timestamp = ftrace_now(cpu); cpu 212 kernel/trace/trace_irqsoff.c per_cpu(tracing_cpu, cpu) = 1; cpu 220 kernel/trace/trace_irqsoff.c int cpu; cpu 225 kernel/trace/trace_irqsoff.c cpu = raw_smp_processor_id(); cpu 227 kernel/trace/trace_irqsoff.c if (unlikely(per_cpu(tracing_cpu, cpu))) cpu 228 kernel/trace/trace_irqsoff.c per_cpu(tracing_cpu, cpu) = 0; cpu 235 kernel/trace/trace_irqsoff.c data = tr->data[cpu]; cpu 245 kernel/trace/trace_irqsoff.c check_critical_timing(tr, data, parent_ip ? : ip, cpu); cpu 24 kernel/trace/trace_mmiotrace.c int cpu; cpu 27 kernel/trace/trace_mmiotrace.c tr->time_start = ftrace_now(tr->cpu); cpu 29 kernel/trace/trace_mmiotrace.c for_each_online_cpu(cpu) cpu 30 kernel/trace/trace_mmiotrace.c tracing_reset(tr->data[cpu]); cpu 131 kernel/trace/trace_mmiotrace.c int cpu; cpu 133 kernel/trace/trace_mmiotrace.c for_each_online_cpu(cpu) { cpu 134 kernel/trace/trace_mmiotrace.c cnt += iter->overrun[cpu]; cpu 135 kernel/trace/trace_mmiotrace.c iter->overrun[cpu] = 0; cpu 30 kernel/trace/trace_sched_switch.c int cpu; cpu 39 kernel/trace/trace_sched_switch.c cpu = raw_smp_processor_id(); cpu 40 kernel/trace/trace_sched_switch.c data = tr->data[cpu]; cpu 85 kernel/trace/trace_sched_switch.c int cpu; cpu 93 kernel/trace/trace_sched_switch.c cpu = raw_smp_processor_id(); cpu 94 kernel/trace/trace_sched_switch.c data = tr->data[cpu]; cpu 131 kernel/trace/trace_sched_switch.c int cpu; cpu 133 kernel/trace/trace_sched_switch.c tr->time_start = ftrace_now(tr->cpu); cpu 135 kernel/trace/trace_sched_switch.c for_each_online_cpu(cpu) cpu 136 kernel/trace/trace_sched_switch.c tracing_reset(tr->data[cpu]); cpu 46 kernel/trace/trace_sched_wakeup.c int cpu; cpu 54 kernel/trace/trace_sched_wakeup.c cpu = raw_smp_processor_id(); cpu 55 kernel/trace/trace_sched_wakeup.c data = tr->data[cpu]; cpu 70 kernel/trace/trace_sched_wakeup.c if (task_cpu(wakeup_task) != cpu) cpu 125 kernel/trace/trace_sched_wakeup.c int cpu; cpu 146 kernel/trace/trace_sched_wakeup.c cpu = raw_smp_processor_id(); cpu 147 kernel/trace/trace_sched_wakeup.c disabled = atomic_inc_return(&tr->data[cpu]->disabled); cpu 165 kernel/trace/trace_sched_wakeup.c T1 = ftrace_now(cpu); cpu 184 kernel/trace/trace_sched_wakeup.c atomic_dec(&tr->data[cpu]->disabled); cpu 215 kernel/trace/trace_sched_wakeup.c int cpu; cpu 217 kernel/trace/trace_sched_wakeup.c for_each_possible_cpu(cpu) { cpu 218 kernel/trace/trace_sched_wakeup.c data = tr->data[cpu]; cpu 246 kernel/trace/trace_sched_wakeup.c int cpu = smp_processor_id(); cpu 255 kernel/trace/trace_sched_wakeup.c disabled = atomic_inc_return(&tr->data[cpu]->disabled); cpu 277 kernel/trace/trace_sched_wakeup.c tr->data[wakeup_cpu]->preempt_timestamp = ftrace_now(cpu); cpu 284 kernel/trace/trace_sched_wakeup.c atomic_dec(&tr->data[cpu]->disabled); cpu 87 kernel/trace/trace_selftest.c int cpu, ret = 0; cpu 92 kernel/trace/trace_selftest.c for_each_possible_cpu(cpu) { cpu 93 kernel/trace/trace_selftest.c if (!head_page(tr->data[cpu])) cpu 96 kernel/trace/trace_selftest.c cnt += tr->data[cpu]->trace_idx; cpu 98 kernel/trace/trace_selftest.c ret = trace_test_buffer_cpu(tr, tr->data[cpu]); cpu 137 kernel/trace/trace_sysprof.c data = tr->data[cpu]; cpu 201 kernel/trace/trace_sysprof.c struct hrtimer *hrtimer = &per_cpu(stack_trace_hrtimer, cpu); cpu 213 kernel/trace/trace_sysprof.c int cpu; cpu 215 kernel/trace/trace_sysprof.c for_each_online_cpu(cpu) { cpu 216 kernel/trace/trace_sysprof.c set_cpus_allowed_ptr(current, &cpumask_of_cpu(cpu)); cpu 217 kernel/trace/trace_sysprof.c start_stack_timer(cpu); cpu 224 kernel/trace/trace_sysprof.c struct hrtimer *hrtimer = &per_cpu(stack_trace_hrtimer, cpu); cpu 231 kernel/trace/trace_sysprof.c int cpu; cpu 233 kernel/trace/trace_sysprof.c for_each_online_cpu(cpu) cpu 234 kernel/trace/trace_sysprof.c stop_stack_timer(cpu); cpu 239 kernel/trace/trace_sysprof.c int cpu; cpu 241 kernel/trace/trace_sysprof.c tr->time_start = ftrace_now(tr->cpu); cpu 243 kernel/trace/trace_sysprof.c for_each_online_cpu(cpu) cpu 244 kernel/trace/trace_sysprof.c tracing_reset(tr->data[cpu]); cpu 101 kernel/workqueue.c cpu = singlethread_cpu; cpu 102 kernel/workqueue.c return per_cpu_ptr(wq->cpu_wq, cpu); cpu 189 kernel/workqueue.c __queue_work(wq_per_cpu(wq, cpu), work); cpu 251 kernel/workqueue.c if (unlikely(cpu >= 0)) cpu 252 kernel/workqueue.c add_timer_on(timer, cpu); cpu 413 kernel/workqueue.c int cpu; cpu 418 kernel/workqueue.c for_each_cpu_mask_nr(cpu, *cpu_map) cpu 419 kernel/workqueue.c flush_cpu_workqueue(per_cpu_ptr(wq->cpu_wq, cpu)); cpu 535 kernel/workqueue.c int cpu; cpu 549 kernel/workqueue.c for_each_cpu_mask_nr(cpu, *cpu_map) cpu 550 kernel/workqueue.c wait_on_cpu_work(per_cpu_ptr(wq->cpu_wq, cpu), work); cpu 634 kernel/workqueue.c return queue_work_on(cpu, keventd_wq, work); cpu 665 kernel/workqueue.c return queue_delayed_work_on(cpu, keventd_wq, dwork, delay); cpu 680 kernel/workqueue.c int cpu; cpu 688 kernel/workqueue.c for_each_online_cpu(cpu) { cpu 689 kernel/workqueue.c struct work_struct *work = per_cpu_ptr(works, cpu); cpu 692 kernel/workqueue.c schedule_work_on(cpu, work); cpu 694 kernel/workqueue.c for_each_online_cpu(cpu) cpu 695 kernel/workqueue.c flush_work(per_cpu_ptr(works, cpu)); cpu 741 kernel/workqueue.c int cpu = raw_smp_processor_id(); /* preempt-safe: keventd is per-cpu */ cpu 746 kernel/workqueue.c cwq = per_cpu_ptr(keventd_wq->cpu_wq, cpu); cpu 757 kernel/workqueue.c struct cpu_workqueue_struct *cwq = per_cpu_ptr(wq->cpu_wq, cpu); cpu 773 kernel/workqueue.c p = kthread_create(worker_thread, cwq, fmt, wq->name, cpu); cpu 795 kernel/workqueue.c if (cpu >= 0) cpu 796 kernel/workqueue.c kthread_bind(p, cpu); cpu 809 kernel/workqueue.c int err = 0, cpu; cpu 848 kernel/workqueue.c for_each_possible_cpu(cpu) { cpu 849 kernel/workqueue.c cwq = init_cpu_workqueue(wq, cpu); cpu 850 kernel/workqueue.c if (err || !cpu_online(cpu)) cpu 852 kernel/workqueue.c err = create_workqueue_thread(cwq, cpu); cpu 853 kernel/workqueue.c start_workqueue_thread(cwq, cpu); cpu 902 kernel/workqueue.c int cpu; cpu 909 kernel/workqueue.c for_each_cpu_mask_nr(cpu, *cpu_map) cpu 910 kernel/workqueue.c cleanup_workqueue_thread(per_cpu_ptr(wq->cpu_wq, cpu)); cpu 922 kernel/workqueue.c unsigned int cpu = (unsigned long)hcpu; cpu 931 kernel/workqueue.c cpu_set(cpu, cpu_populated_map); cpu 935 kernel/workqueue.c cwq = per_cpu_ptr(wq->cpu_wq, cpu); cpu 939 kernel/workqueue.c if (!create_workqueue_thread(cwq, cpu)) cpu 942 kernel/workqueue.c wq->name, cpu); cpu 948 kernel/workqueue.c start_workqueue_thread(cwq, cpu); cpu 962 kernel/workqueue.c cpu_clear(cpu, cpu_populated_map); cpu 29 lib/cpumask.c int cpu; cpu 31 lib/cpumask.c for_each_cpu_mask(cpu, *mask) { cpu 32 lib/cpumask.c if (cpu_online(cpu)) cpu 35 lib/cpumask.c return cpu; cpu 19 lib/percpu_counter.c int cpu; cpu 22 lib/percpu_counter.c for_each_possible_cpu(cpu) { cpu 23 lib/percpu_counter.c s32 *pcount = per_cpu_ptr(fbc->counters, cpu); cpu 35 lib/percpu_counter.c int cpu = get_cpu(); cpu 37 lib/percpu_counter.c pcount = per_cpu_ptr(fbc->counters, cpu); cpu 58 lib/percpu_counter.c int cpu; cpu 62 lib/percpu_counter.c for_each_online_cpu(cpu) { cpu 63 lib/percpu_counter.c s32 *pcount = per_cpu_ptr(fbc->counters, cpu); cpu 121 lib/percpu_counter.c unsigned int cpu; cpu 127 lib/percpu_counter.c cpu = (unsigned long)hcpu; cpu 134 lib/percpu_counter.c pcount = per_cpu_ptr(fbc->counters, cpu); cpu 1215 lib/radix-tree.c int cpu = (long)hcpu; cpu 1220 lib/radix-tree.c rtp = &per_cpu(radix_tree_preloads, cpu); cpu 25 mm/allocpercpu.c kfree(pdata->ptrs[cpu]); cpu 26 mm/allocpercpu.c pdata->ptrs[cpu] = NULL; cpu 36 mm/allocpercpu.c int cpu; cpu 37 mm/allocpercpu.c for_each_cpu_mask_nr(cpu, *mask) cpu 38 mm/allocpercpu.c percpu_depopulate(__pdata, cpu); cpu 58 mm/allocpercpu.c int node = cpu_to_node(cpu); cpu 65 mm/allocpercpu.c BUG_ON(pdata->ptrs[cpu]); cpu 67 mm/allocpercpu.c pdata->ptrs[cpu] = kmalloc_node(size, gfp|__GFP_ZERO, node); cpu 69 mm/allocpercpu.c pdata->ptrs[cpu] = kzalloc(size, gfp); cpu 70 mm/allocpercpu.c return pdata->ptrs[cpu]; cpu 86 mm/allocpercpu.c int cpu; cpu 89 mm/allocpercpu.c for_each_cpu_mask_nr(cpu, *mask) cpu 90 mm/allocpercpu.c if (unlikely(!percpu_populate(__pdata, size, gfp, cpu))) { cpu 94 mm/allocpercpu.c cpu_set(cpu, populated); cpu 71 mm/memcontrol.c int cpu = smp_processor_id(); cpu 72 mm/memcontrol.c stat->cpustat[cpu].count[idx] += val; cpu 78 mm/memcontrol.c int cpu; cpu 80 mm/memcontrol.c for_each_possible_cpu(cpu) cpu 81 mm/memcontrol.c ret += stat->cpustat[cpu].count[idx]; cpu 906 mm/page_alloc.c pset = zone_pcp(zone, cpu); cpu 1045 mm/page_alloc.c int cpu; cpu 1049 mm/page_alloc.c cpu = get_cpu(); cpu 1053 mm/page_alloc.c pcp = &zone_pcp(zone, cpu)->pcp; cpu 1844 mm/page_alloc.c int cpu; cpu 1854 mm/page_alloc.c for_each_online_cpu(cpu) { cpu 1857 mm/page_alloc.c pageset = zone_pcp(zone, cpu); cpu 1860 mm/page_alloc.c cpu, pageset->pcp.high, cpu 2724 mm/page_alloc.c int node = cpu_to_node(cpu); cpu 2733 mm/page_alloc.c zone_pcp(zone, cpu) = kmalloc_node(sizeof(struct per_cpu_pageset), cpu 2735 mm/page_alloc.c if (!zone_pcp(zone, cpu)) cpu 2738 mm/page_alloc.c setup_pageset(zone_pcp(zone, cpu), zone_batchsize(zone)); cpu 2741 mm/page_alloc.c setup_pagelist_highmark(zone_pcp(zone, cpu), cpu 2752 mm/page_alloc.c kfree(zone_pcp(dzone, cpu)); cpu 2753 mm/page_alloc.c zone_pcp(dzone, cpu) = NULL; cpu 2763 mm/page_alloc.c struct per_cpu_pageset *pset = zone_pcp(zone, cpu); cpu 2766 mm/page_alloc.c if (pset != &boot_pageset[cpu]) cpu 2768 mm/page_alloc.c zone_pcp(zone, cpu) = NULL; cpu 2776 mm/page_alloc.c int cpu = (long)hcpu; cpu 2782 mm/page_alloc.c if (process_zones(cpu)) cpu 2789 mm/page_alloc.c free_zone_pagesets(cpu); cpu 2860 mm/page_alloc.c int cpu; cpu 2863 mm/page_alloc.c for (cpu = 0; cpu < NR_CPUS; cpu++) { cpu 2866 mm/page_alloc.c zone_pcp(zone, cpu) = &boot_pageset[cpu]; cpu 2867 mm/page_alloc.c setup_pageset(&boot_pageset[cpu],0); cpu 2869 mm/page_alloc.c setup_pageset(zone_pcp(zone,cpu), batch); cpu 4090 mm/page_alloc.c int cpu = (unsigned long)hcpu; cpu 4093 mm/page_alloc.c drain_pages(cpu); cpu 4101 mm/page_alloc.c vm_events_fold_cpu(cpu); cpu 4110 mm/page_alloc.c refresh_cpu_vm_stats(cpu); cpu 4368 mm/page_alloc.c unsigned int cpu; cpu 4375 mm/page_alloc.c for_each_online_cpu(cpu) { cpu 4378 mm/page_alloc.c setup_pagelist_highmark(zone_pcp(zone, cpu), high); cpu 93 mm/quicklist.c int cpu; cpu 96 mm/quicklist.c for_each_online_cpu(cpu) { cpu 97 mm/quicklist.c ql = per_cpu(quicklist, cpu); cpu 904 mm/slab.c node = next_node(cpu_to_node(cpu), node_online_map); cpu 908 mm/slab.c per_cpu(reap_node, cpu) = node; cpu 935 mm/slab.c struct delayed_work *reap_work = &per_cpu(reap_work, cpu); cpu 943 mm/slab.c init_reap_node(cpu); cpu 945 mm/slab.c schedule_delayed_work_on(cpu, reap_work, cpu 946 mm/slab.c __round_jiffies_relative(HZ, cpu)); cpu 1161 mm/slab.c int node = cpu_to_node(cpu); cpu 1170 mm/slab.c nc = cachep->array[cpu]; cpu 1171 mm/slab.c cachep->array[cpu] = NULL; cpu 1226 mm/slab.c int node = cpu_to_node(cpu); cpu 1295 mm/slab.c cachep->array[cpu] = nc; cpu 1320 mm/slab.c cpuup_canceled(cpu); cpu 1327 mm/slab.c long cpu = (long)hcpu; cpu 1334 mm/slab.c err = cpuup_prepare(cpu); cpu 1339 mm/slab.c start_cpu_timer(cpu); cpu 1350 mm/slab.c cancel_rearming_delayed_work(&per_cpu(reap_work, cpu)); cpu 1352 mm/slab.c per_cpu(reap_work, cpu).work.func = NULL; cpu 1356 mm/slab.c start_cpu_timer(cpu); cpu 1373 mm/slab.c cpuup_canceled(cpu); cpu 1646 mm/slab.c int cpu; cpu 1651 mm/slab.c for_each_online_cpu(cpu) cpu 1652 mm/slab.c start_cpu_timer(cpu); cpu 181 mm/slub.c int cpu; /* Was running on cpu */ cpu 232 mm/slub.c return s->cpu_slab[cpu]; cpu 381 mm/slub.c p->cpu = smp_processor_id(); cpu 403 mm/slub.c s, t->addr, jiffies - t->when, t->cpu, t->pid); cpu 1452 mm/slub.c struct kmem_cache_cpu *c = get_cpu_slab(s, cpu); cpu 1967 mm/slub.c struct kmem_cache_cpu *c = per_cpu(kmem_cache_cpu_free, cpu); cpu 1970 mm/slub.c per_cpu(kmem_cache_cpu_free, cpu) = cpu 1976 mm/slub.c flags, cpu_to_node(cpu)); cpu 1987 mm/slub.c if (c < per_cpu(kmem_cache_cpu, cpu) || cpu 1988 mm/slub.c c > per_cpu(kmem_cache_cpu, cpu) + NR_KMEM_CACHE_CPU) { cpu 1992 mm/slub.c c->freelist = (void *)per_cpu(kmem_cache_cpu_free, cpu); cpu 1993 mm/slub.c per_cpu(kmem_cache_cpu_free, cpu) = c; cpu 1998 mm/slub.c int cpu; cpu 2000 mm/slub.c for_each_online_cpu(cpu) { cpu 2001 mm/slub.c struct kmem_cache_cpu *c = get_cpu_slab(s, cpu); cpu 2004 mm/slub.c s->cpu_slab[cpu] = NULL; cpu 2005 mm/slub.c free_kmem_cache_cpu(c, cpu); cpu 2012 mm/slub.c int cpu; cpu 2014 mm/slub.c for_each_online_cpu(cpu) { cpu 2015 mm/slub.c struct kmem_cache_cpu *c = get_cpu_slab(s, cpu); cpu 2020 mm/slub.c c = alloc_kmem_cache_cpu(s, cpu, flags); cpu 2025 mm/slub.c s->cpu_slab[cpu] = c; cpu 2037 mm/slub.c if (cpu_isset(cpu, kmem_cach_cpu_free_init_once)) cpu 2041 mm/slub.c free_kmem_cache_cpu(&per_cpu(kmem_cache_cpu, cpu)[i], cpu); cpu 2043 mm/slub.c cpu_set(cpu, kmem_cach_cpu_free_init_once); cpu 2048 mm/slub.c int cpu; cpu 2050 mm/slub.c for_each_online_cpu(cpu) cpu 2051 mm/slub.c init_alloc_cpu_cpu(cpu); cpu 3104 mm/slub.c int cpu; cpu 3117 mm/slub.c for_each_online_cpu(cpu) cpu 3118 mm/slub.c get_cpu_slab(s, cpu)->objsize = s->objsize; cpu 3159 mm/slub.c long cpu = (long)hcpu; cpu 3166 mm/slub.c init_alloc_cpu_cpu(cpu); cpu 3169 mm/slub.c s->cpu_slab[cpu] = alloc_kmem_cache_cpu(s, cpu, cpu 3180 mm/slub.c struct kmem_cache_cpu *c = get_cpu_slab(s, cpu); cpu 3183 mm/slub.c __flush_cpu_slab(s, cpu); cpu 3185 mm/slub.c free_kmem_cache_cpu(c, cpu); cpu 3186 mm/slub.c s->cpu_slab[cpu] = NULL; cpu 3510 mm/slub.c cpu_set(track->cpu, l->cpus); cpu 3541 mm/slub.c cpu_set(track->cpu, l->cpus); cpu 3674 mm/slub.c int cpu; cpu 3676 mm/slub.c for_each_possible_cpu(cpu) { cpu 3677 mm/slub.c struct kmem_cache_cpu *c = get_cpu_slab(s, cpu); cpu 4074 mm/slub.c int cpu; cpu 4081 mm/slub.c for_each_online_cpu(cpu) { cpu 4082 mm/slub.c unsigned x = get_cpu_slab(s, cpu)->stat[si]; cpu 4084 mm/slub.c data[cpu] = x; cpu 4091 mm/slub.c for_each_online_cpu(cpu) { cpu 4092 mm/slub.c if (data[cpu] && len < PAGE_SIZE - 20) cpu 4093 mm/slub.c len += sprintf(buf + len, " C%d=%u", cpu, data[cpu]); cpu 222 mm/swap.c pvec = &per_cpu(lru_add_pvecs, cpu); cpu 226 mm/swap.c pvec = &per_cpu(lru_add_active_pvecs, cpu); cpu 230 mm/swap.c pvec = &per_cpu(lru_rotate_pvecs, cpu); cpu 25 mm/vmstat.c int cpu; cpu 30 mm/vmstat.c for_each_cpu_mask_nr(cpu, *cpumask) { cpu 31 mm/vmstat.c struct vm_event_state *this = &per_cpu(vm_event_states, cpu); cpu 60 mm/vmstat.c struct vm_event_state *fold_state = &per_cpu(vm_event_states, cpu); cpu 135 mm/vmstat.c int cpu; cpu 145 mm/vmstat.c for_each_online_cpu(cpu) cpu 146 mm/vmstat.c zone_pcp(zone, cpu)->stat_threshold = threshold; cpu 310 mm/vmstat.c p = zone_pcp(zone, cpu); cpu 838 mm/vmstat.c struct delayed_work *vmstat_work = &per_cpu(vmstat_work, cpu); cpu 841 mm/vmstat.c schedule_delayed_work_on(cpu, vmstat_work, HZ + cpu); cpu 852 mm/vmstat.c long cpu = (long)hcpu; cpu 857 mm/vmstat.c start_cpu_timer(cpu); cpu 861 mm/vmstat.c cancel_rearming_delayed_work(&per_cpu(vmstat_work, cpu)); cpu 862 mm/vmstat.c per_cpu(vmstat_work, cpu).work.func = NULL; cpu 866 mm/vmstat.c start_cpu_timer(cpu); cpu 883 mm/vmstat.c int cpu; cpu 888 mm/vmstat.c for_each_online_cpu(cpu) cpu 889 mm/vmstat.c start_cpu_timer(cpu); cpu 53 net/bridge/netfilter/ebtables.c COUNTER_OFFSET(n) * cpu)) cpu 936 net/bridge/netfilter/ebtables.c int i, cpu; cpu 944 net/bridge/netfilter/ebtables.c for_each_possible_cpu(cpu) { cpu 945 net/bridge/netfilter/ebtables.c if (cpu == 0) cpu 947 net/bridge/netfilter/ebtables.c counter_base = COUNTER_BASE(oldcounters, nentries, cpu); cpu 1865 net/core/dev.c int cpu = smp_processor_id(); /* ok because BHs are off */ cpu 1867 net/core/dev.c if (txq->xmit_lock_owner != cpu) { cpu 1869 net/core/dev.c HARD_TX_LOCK(dev, txq, cpu); cpu 4548 net/core/dev.c unsigned int cpu, oldcpu = (unsigned long)ocpu; cpu 4555 net/core/dev.c cpu = smp_processor_id(); cpu 4556 net/core/dev.c sd = &per_cpu(softnet_data, cpu); cpu 4596 net/core/dev.c unsigned int cpu, i, n, chan_idx; cpu 4600 net/core/dev.c for_each_online_cpu(cpu) cpu 4601 net/core/dev.c rcu_assign_pointer(per_cpu(softnet_data, cpu).net_dma, NULL); cpu 4606 net/core/dev.c cpu = first_cpu(cpu_online_map); cpu 4616 net/core/dev.c per_cpu(softnet_data, cpu).net_dma = chan; cpu 4617 net/core/dev.c cpu = next_cpu(cpu, cpu_online_map); cpu 44 net/core/flow.c #define flow_table(cpu) (per_cpu(flow_tables, cpu)) cpu 58 net/core/flow.c (per_cpu(flow_hash_info, cpu).hash_rnd_recalc) cpu 60 net/core/flow.c (per_cpu(flow_hash_info, cpu).hash_rnd) cpu 62 net/core/flow.c (per_cpu(flow_hash_info, cpu).count) cpu 74 net/core/flow.c #define flow_flush_tasklet(cpu) (&per_cpu(flow_flush_tasklets, cpu)) cpu 92 net/core/flow.c flow_count(cpu)--; cpu 103 net/core/flow.c flp = &flow_table(cpu)[i]; cpu 110 net/core/flow.c flow_entry_kill(cpu, fle); cpu 119 net/core/flow.c __flow_cache_shrink(cpu, shrink_to); cpu 124 net/core/flow.c get_random_bytes(&flow_hash_rnd(cpu), sizeof(u32)); cpu 125 net/core/flow.c flow_hash_rnd_recalc(cpu) = 0; cpu 127 net/core/flow.c __flow_cache_shrink(cpu, 0); cpu 134 net/core/flow.c return (jhash2(k, (sizeof(*key) / sizeof(u32)), flow_hash_rnd(cpu)) & cpu 173 net/core/flow.c int cpu; cpu 176 net/core/flow.c cpu = smp_processor_id(); cpu 181 net/core/flow.c if (!flow_table(cpu)) cpu 184 net/core/flow.c if (flow_hash_rnd_recalc(cpu)) cpu 185 net/core/flow.c flow_new_hash_rnd(cpu); cpu 186 net/core/flow.c hash = flow_hash_code(key, cpu); cpu 188 net/core/flow.c head = &flow_table(cpu)[hash]; cpu 207 net/core/flow.c if (flow_count(cpu) > flow_hwm) cpu 208 net/core/flow.c flow_cache_shrink(cpu); cpu 218 net/core/flow.c flow_count(cpu)++; cpu 253 net/core/flow.c int cpu; cpu 255 net/core/flow.c cpu = smp_processor_id(); cpu 259 net/core/flow.c fle = flow_table(cpu)[i]; cpu 279 net/core/flow.c int cpu; cpu 282 net/core/flow.c cpu = smp_processor_id(); cpu 284 net/core/flow.c tasklet = flow_flush_tasklet(cpu); cpu 321 net/core/flow.c flow_table(cpu) = (struct flow_cache_entry **) cpu 323 net/core/flow.c if (!flow_table(cpu)) cpu 326 net/core/flow.c flow_hash_rnd_recalc(cpu) = 1; cpu 327 net/core/flow.c flow_count(cpu) = 0; cpu 329 net/core/flow.c tasklet = flow_flush_tasklet(cpu); cpu 1761 net/core/neighbour.c int cpu; cpu 1766 net/core/neighbour.c for_each_possible_cpu(cpu) { cpu 1769 net/core/neighbour.c st = per_cpu_ptr(tbl->stats, cpu); cpu 2421 net/core/neighbour.c int cpu; cpu 2426 net/core/neighbour.c for (cpu = *pos-1; cpu < NR_CPUS; ++cpu) { cpu 2427 net/core/neighbour.c if (!cpu_possible(cpu)) cpu 2429 net/core/neighbour.c *pos = cpu+1; cpu 2430 net/core/neighbour.c return per_cpu_ptr(tbl->stats, cpu); cpu 2439 net/core/neighbour.c int cpu; cpu 2441 net/core/neighbour.c for (cpu = *pos; cpu < NR_CPUS; ++cpu) { cpu 2442 net/core/neighbour.c if (!cpu_possible(cpu)) cpu 2444 net/core/neighbour.c *pos = cpu+1; cpu 2445 net/core/neighbour.c return per_cpu_ptr(tbl->stats, cpu); cpu 391 net/core/pktgen.c int cpu; cpu 3501 net/core/pktgen.c int cpu = t->cpu; cpu 3503 net/core/pktgen.c BUG_ON(smp_processor_id() != cpu); cpu 3508 net/core/pktgen.c pr_debug("pktgen: starting pktgen/%d: pid=%d\n", cpu, task_pid_nr(current)); cpu 3703 net/core/pktgen.c t->cpu = cpu; cpu 3710 net/core/pktgen.c p = kthread_create(pktgen_thread_worker, t, "kpktgend_%d", cpu); cpu 3713 net/core/pktgen.c "for cpu %d\n", t->cpu); cpu 3718 net/core/pktgen.c kthread_bind(p, cpu); cpu 3791 net/core/pktgen.c int cpu; cpu 3812 net/core/pktgen.c for_each_online_cpu(cpu) { cpu 3815 net/core/pktgen.c err = pktgen_create_thread(cpu); cpu 3818 net/core/pktgen.c "thread for cpu %d (%d)\n", cpu, err); cpu 1947 net/core/sock.c int cpu = smp_processor_id(); cpu 1948 net/core/sock.c per_cpu_ptr(net->core.inuse, cpu)->val[prot->inuse_idx] += val; cpu 1954 net/core/sock.c int cpu, idx = prot->inuse_idx; cpu 1957 net/core/sock.c for_each_possible_cpu(cpu) cpu 1958 net/core/sock.c res += per_cpu_ptr(net->core.inuse, cpu)->val[idx]; cpu 2000 net/core/sock.c int cpu, idx = prot->inuse_idx; cpu 2003 net/core/sock.c for_each_possible_cpu(cpu) cpu 2004 net/core/sock.c res += per_cpu(prot_inuse, cpu).val[idx]; cpu 687 net/ipv4/netfilter/arp_tables.c unsigned int cpu; cpu 705 net/ipv4/netfilter/arp_tables.c for_each_possible_cpu(cpu) { cpu 706 net/ipv4/netfilter/arp_tables.c if (cpu == curcpu) cpu 709 net/ipv4/netfilter/arp_tables.c ARPT_ENTRY_ITERATE(t->entries[cpu], cpu 903 net/ipv4/netfilter/ip_tables.c unsigned int cpu; cpu 921 net/ipv4/netfilter/ip_tables.c for_each_possible_cpu(cpu) { cpu 922 net/ipv4/netfilter/ip_tables.c if (cpu == curcpu) cpu 925 net/ipv4/netfilter/ip_tables.c IPT_ENTRY_ITERATE(t->entries[cpu], cpu 289 net/ipv4/netfilter/nf_conntrack_l3proto_ipv4_compat.c int cpu; cpu 294 net/ipv4/netfilter/nf_conntrack_l3proto_ipv4_compat.c for (cpu = *pos-1; cpu < NR_CPUS; ++cpu) { cpu 295 net/ipv4/netfilter/nf_conntrack_l3proto_ipv4_compat.c if (!cpu_possible(cpu)) cpu 297 net/ipv4/netfilter/nf_conntrack_l3proto_ipv4_compat.c *pos = cpu+1; cpu 298 net/ipv4/netfilter/nf_conntrack_l3proto_ipv4_compat.c return per_cpu_ptr(net->ct.stat, cpu); cpu 307 net/ipv4/netfilter/nf_conntrack_l3proto_ipv4_compat.c int cpu; cpu 309 net/ipv4/netfilter/nf_conntrack_l3proto_ipv4_compat.c for (cpu = *pos; cpu < NR_CPUS; ++cpu) { cpu 310 net/ipv4/netfilter/nf_conntrack_l3proto_ipv4_compat.c if (!cpu_possible(cpu)) cpu 312 net/ipv4/netfilter/nf_conntrack_l3proto_ipv4_compat.c *pos = cpu+1; cpu 313 net/ipv4/netfilter/nf_conntrack_l3proto_ipv4_compat.c return per_cpu_ptr(net->ct.stat, cpu); cpu 425 net/ipv4/route.c int cpu; cpu 430 net/ipv4/route.c for (cpu = *pos-1; cpu < NR_CPUS; ++cpu) { cpu 431 net/ipv4/route.c if (!cpu_possible(cpu)) cpu 433 net/ipv4/route.c *pos = cpu+1; cpu 434 net/ipv4/route.c return &per_cpu(rt_cache_stat, cpu); cpu 441 net/ipv4/route.c int cpu; cpu 443 net/ipv4/route.c for (cpu = *pos; cpu < NR_CPUS; ++cpu) { cpu 444 net/ipv4/route.c if (!cpu_possible(cpu)) cpu 446 net/ipv4/route.c *pos = cpu+1; cpu 447 net/ipv4/route.c return &per_cpu(rt_cache_stat, cpu); cpu 2472 net/ipv4/tcp.c int cpu; cpu 2473 net/ipv4/tcp.c for_each_possible_cpu(cpu) { cpu 2474 net/ipv4/tcp.c struct tcp_md5sig_pool *p = *per_cpu_ptr(pool, cpu); cpu 2503 net/ipv4/tcp.c int cpu; cpu 2510 net/ipv4/tcp.c for_each_possible_cpu(cpu) { cpu 2517 net/ipv4/tcp.c *per_cpu_ptr(pool, cpu) = p; cpu 2582 net/ipv4/tcp.c return (p ? *per_cpu_ptr(p, cpu) : NULL); cpu 928 net/ipv6/netfilter/ip6_tables.c unsigned int cpu; cpu 946 net/ipv6/netfilter/ip6_tables.c for_each_possible_cpu(cpu) { cpu 947 net/ipv6/netfilter/ip6_tables.c if (cpu == curcpu) cpu 950 net/ipv6/netfilter/ip6_tables.c IP6T_ENTRY_ITERATE(t->entries[cpu], cpu 347 net/iucv/iucv.c int cpu = smp_processor_id(); cpu 359 net/iucv/iucv.c parm = iucv_param[cpu]; cpu 365 net/iucv/iucv.c cpu_set(cpu, iucv_irq_cpumask); cpu 376 net/iucv/iucv.c int cpu = smp_processor_id(); cpu 380 net/iucv/iucv.c parm = iucv_param[cpu]; cpu 385 net/iucv/iucv.c cpu_clear(cpu, iucv_irq_cpumask); cpu 396 net/iucv/iucv.c int cpu = smp_processor_id(); cpu 400 net/iucv/iucv.c if (cpu_isset(cpu, iucv_buffer_cpumask)) cpu 404 net/iucv/iucv.c parm = iucv_param[cpu]; cpu 406 net/iucv/iucv.c parm->db.ipbfadr1 = virt_to_phys(iucv_irq_data[cpu]); cpu 428 net/iucv/iucv.c "on cpu %i returned error 0x%02x (%s)\n", cpu, rc, err); cpu 433 net/iucv/iucv.c cpu_set(cpu, iucv_buffer_cpumask); cpu 451 net/iucv/iucv.c int cpu = smp_processor_id(); cpu 454 net/iucv/iucv.c if (!cpu_isset(cpu, iucv_buffer_cpumask)) cpu 461 net/iucv/iucv.c parm = iucv_param[cpu]; cpu 465 net/iucv/iucv.c cpu_clear(cpu, iucv_buffer_cpumask); cpu 475 net/iucv/iucv.c int cpu; cpu 478 net/iucv/iucv.c for_each_online_cpu(cpu) cpu 480 net/iucv/iucv.c if (cpu_isset(cpu, iucv_buffer_cpumask) && cpu 481 net/iucv/iucv.c !cpu_isset(cpu, iucv_irq_cpumask)) cpu 482 net/iucv/iucv.c smp_call_function_single(cpu, iucv_allow_cpu, cpu 495 net/iucv/iucv.c int cpu; cpu 500 net/iucv/iucv.c for_each_cpu_mask_nr(cpu, cpumask) cpu 501 net/iucv/iucv.c smp_call_function_single(cpu, iucv_block_cpu, NULL, 1); cpu 515 net/iucv/iucv.c int cpu, rc; cpu 525 net/iucv/iucv.c for_each_online_cpu(cpu) cpu 526 net/iucv/iucv.c smp_call_function_single(cpu, iucv_declare_cpu, NULL, 1); cpu 559 net/iucv/iucv.c long cpu = (long) hcpu; cpu 564 net/iucv/iucv.c iucv_irq_data[cpu] = kmalloc_node(sizeof(struct iucv_irq_data), cpu 565 net/iucv/iucv.c GFP_KERNEL|GFP_DMA, cpu_to_node(cpu)); cpu 566 net/iucv/iucv.c if (!iucv_irq_data[cpu]) cpu 568 net/iucv/iucv.c iucv_param[cpu] = kmalloc_node(sizeof(union iucv_param), cpu 569 net/iucv/iucv.c GFP_KERNEL|GFP_DMA, cpu_to_node(cpu)); cpu 570 net/iucv/iucv.c if (!iucv_param[cpu]) { cpu 571 net/iucv/iucv.c kfree(iucv_irq_data[cpu]); cpu 572 net/iucv/iucv.c iucv_irq_data[cpu] = NULL; cpu 580 net/iucv/iucv.c kfree(iucv_param[cpu]); cpu 581 net/iucv/iucv.c iucv_param[cpu] = NULL; cpu 582 net/iucv/iucv.c kfree(iucv_irq_data[cpu]); cpu 583 net/iucv/iucv.c iucv_irq_data[cpu] = NULL; cpu 589 net/iucv/iucv.c smp_call_function_single(cpu, iucv_declare_cpu, NULL, 1); cpu 594 net/iucv/iucv.c cpu_clear(cpu, cpumask); cpu 598 net/iucv/iucv.c smp_call_function_single(cpu, iucv_retrieve_cpu, NULL, 1); cpu 1600 net/iucv/iucv.c int cpu; cpu 1618 net/iucv/iucv.c for_each_online_cpu(cpu) { cpu 1620 net/iucv/iucv.c iucv_irq_data[cpu] = kmalloc_node(sizeof(struct iucv_irq_data), cpu 1621 net/iucv/iucv.c GFP_KERNEL|GFP_DMA, cpu_to_node(cpu)); cpu 1622 net/iucv/iucv.c if (!iucv_irq_data[cpu]) { cpu 1628 net/iucv/iucv.c iucv_param[cpu] = kmalloc_node(sizeof(union iucv_param), cpu 1629 net/iucv/iucv.c GFP_KERNEL|GFP_DMA, cpu_to_node(cpu)); cpu 1630 net/iucv/iucv.c if (!iucv_param[cpu]) { cpu 1650 net/iucv/iucv.c for_each_possible_cpu(cpu) { cpu 1651 net/iucv/iucv.c kfree(iucv_param[cpu]); cpu 1652 net/iucv/iucv.c iucv_param[cpu] = NULL; cpu 1653 net/iucv/iucv.c kfree(iucv_irq_data[cpu]); cpu 1654 net/iucv/iucv.c iucv_irq_data[cpu] = NULL; cpu 1671 net/iucv/iucv.c int cpu; cpu 1680 net/iucv/iucv.c for_each_possible_cpu(cpu) { cpu 1681 net/iucv/iucv.c kfree(iucv_param[cpu]); cpu 1682 net/iucv/iucv.c iucv_param[cpu] = NULL; cpu 1683 net/iucv/iucv.c kfree(iucv_irq_data[cpu]); cpu 1684 net/iucv/iucv.c iucv_irq_data[cpu] = NULL; cpu 84 net/netfilter/nf_conntrack_ecache.c int cpu; cpu 86 net/netfilter/nf_conntrack_ecache.c for_each_possible_cpu(cpu) { cpu 87 net/netfilter/nf_conntrack_ecache.c ecache = per_cpu_ptr(net->ct.ecache, cpu); cpu 198 net/netfilter/nf_conntrack_standalone.c int cpu; cpu 203 net/netfilter/nf_conntrack_standalone.c for (cpu = *pos-1; cpu < NR_CPUS; ++cpu) { cpu 204 net/netfilter/nf_conntrack_standalone.c if (!cpu_possible(cpu)) cpu 206 net/netfilter/nf_conntrack_standalone.c *pos = cpu + 1; cpu 207 net/netfilter/nf_conntrack_standalone.c return per_cpu_ptr(net->ct.stat, cpu); cpu 216 net/netfilter/nf_conntrack_standalone.c int cpu; cpu 218 net/netfilter/nf_conntrack_standalone.c for (cpu = *pos; cpu < NR_CPUS; ++cpu) { cpu 219 net/netfilter/nf_conntrack_standalone.c if (!cpu_possible(cpu)) cpu 221 net/netfilter/nf_conntrack_standalone.c *pos = cpu + 1; cpu 222 net/netfilter/nf_conntrack_standalone.c return per_cpu_ptr(net->ct.stat, cpu); cpu 575 net/netfilter/x_tables.c int cpu; cpu 587 net/netfilter/x_tables.c for_each_possible_cpu(cpu) { cpu 589 net/netfilter/x_tables.c newinfo->entries[cpu] = kmalloc_node(size, cpu 591 net/netfilter/x_tables.c cpu_to_node(cpu)); cpu 593 net/netfilter/x_tables.c newinfo->entries[cpu] = vmalloc_node(size, cpu 594 net/netfilter/x_tables.c cpu_to_node(cpu)); cpu 596 net/netfilter/x_tables.c if (newinfo->entries[cpu] == NULL) { cpu 608 net/netfilter/x_tables.c int cpu; cpu 610 net/netfilter/x_tables.c for_each_possible_cpu(cpu) { cpu 612 net/netfilter/x_tables.c kfree(info->entries[cpu]); cpu 614 net/netfilter/x_tables.c vfree(info->entries[cpu]); cpu 2306 net/socket.c int cpu; cpu 2309 net/socket.c for_each_possible_cpu(cpu) cpu 2310 net/socket.c counter += per_cpu(sockets_in_use, cpu); cpu 181 net/sunrpc/svc.c unsigned int cpu; cpu 188 net/sunrpc/svc.c for_each_online_cpu(cpu) { cpu 190 net/sunrpc/svc.c m->to_pool[cpu] = pidx; cpu 191 net/sunrpc/svc.c m->pool_to[pidx] = cpu; cpu 346 net/sunrpc/svc.c pidx = m->to_pool[cpu]; cpu 349 net/sunrpc/svc.c pidx = m->to_pool[cpu_to_node(cpu)]; cpu 300 net/sunrpc/svc_xprt.c int cpu; cpu 306 net/sunrpc/svc_xprt.c cpu = get_cpu(); cpu 307 net/sunrpc/svc_xprt.c pool = svc_pool_for_cpu(xprt->xpt_server, cpu); cpu 48 net/xfrm/xfrm_ipcomp.c const int cpu = get_cpu(); cpu 49 net/xfrm/xfrm_ipcomp.c u8 *scratch = *per_cpu_ptr(ipcomp_scratches, cpu); cpu 50 net/xfrm/xfrm_ipcomp.c struct crypto_comp *tfm = *per_cpu_ptr(ipcd->tfms, cpu); cpu 142 net/xfrm/xfrm_ipcomp.c const int cpu = get_cpu(); cpu 143 net/xfrm/xfrm_ipcomp.c u8 *scratch = *per_cpu_ptr(ipcomp_scratches, cpu); cpu 144 net/xfrm/xfrm_ipcomp.c struct crypto_comp *tfm = *per_cpu_ptr(ipcd->tfms, cpu); cpu 246 net/xfrm/xfrm_ipcomp.c int cpu; cpu 264 net/xfrm/xfrm_ipcomp.c for_each_possible_cpu(cpu) { cpu 265 net/xfrm/xfrm_ipcomp.c struct crypto_comp *tfm = *per_cpu_ptr(tfms, cpu); cpu 275 net/xfrm/xfrm_ipcomp.c int cpu; cpu 278 net/xfrm/xfrm_ipcomp.c cpu = raw_smp_processor_id(); cpu 284 net/xfrm/xfrm_ipcomp.c tfm = *per_cpu_ptr(tfms, cpu); cpu 304 net/xfrm/xfrm_ipcomp.c for_each_possible_cpu(cpu) { cpu 309 net/xfrm/xfrm_ipcomp.c *per_cpu_ptr(tfms, cpu) = tfm; cpu 1207 security/selinux/selinuxfs.c int cpu; cpu 1209 security/selinux/selinuxfs.c for (cpu = *idx; cpu < NR_CPUS; ++cpu) { cpu 1210 security/selinux/selinuxfs.c if (!cpu_possible(cpu)) cpu 1212 security/selinux/selinuxfs.c *idx = cpu + 1; cpu 1213 security/selinux/selinuxfs.c return &per_cpu(avc_cache_stats, cpu); cpu 84 virt/kvm/kvm_main.c int cpu; cpu 87 virt/kvm/kvm_main.c cpu = get_cpu(); cpu 89 virt/kvm/kvm_main.c kvm_arch_vcpu_load(vcpu, cpu); cpu 108 virt/kvm/kvm_main.c int i, cpu, me; cpu 120 virt/kvm/kvm_main.c cpu = vcpu->cpu; cpu 121 virt/kvm/kvm_main.c if (cpu != -1 && cpu != me) cpu 122 virt/kvm/kvm_main.c cpu_set(cpu, cpus); cpu 134 virt/kvm/kvm_main.c int i, cpu, me; cpu 146 virt/kvm/kvm_main.c cpu = vcpu->cpu; cpu 147 virt/kvm/kvm_main.c if (cpu != -1 && cpu != me) cpu 148 virt/kvm/kvm_main.c cpu_set(cpu, cpus); cpu 164 virt/kvm/kvm_main.c vcpu->cpu = -1; cpu 1476 virt/kvm/kvm_main.c int cpu = raw_smp_processor_id(); cpu 1478 virt/kvm/kvm_main.c if (cpu_isset(cpu, cpus_hardware_enabled)) cpu 1480 virt/kvm/kvm_main.c cpu_set(cpu, cpus_hardware_enabled); cpu 1486 virt/kvm/kvm_main.c int cpu = raw_smp_processor_id(); cpu 1488 virt/kvm/kvm_main.c if (!cpu_isset(cpu, cpus_hardware_enabled)) cpu 1490 virt/kvm/kvm_main.c cpu_clear(cpu, cpus_hardware_enabled); cpu 1497 virt/kvm/kvm_main.c int cpu = (long)v; cpu 1503 virt/kvm/kvm_main.c cpu); cpu 1508 virt/kvm/kvm_main.c cpu); cpu 1509 virt/kvm/kvm_main.c smp_call_function_single(cpu, hardware_disable, NULL, 1); cpu 1513 virt/kvm/kvm_main.c cpu); cpu 1514 virt/kvm/kvm_main.c smp_call_function_single(cpu, hardware_enable, NULL, 1); cpu 1692 virt/kvm/kvm_main.c kvm_arch_vcpu_load(vcpu, cpu); cpu 1707 virt/kvm/kvm_main.c int cpu; cpu 1728 virt/kvm/kvm_main.c for_each_online_cpu(cpu) { cpu 1729 virt/kvm/kvm_main.c smp_call_function_single(cpu,