cpumask_t 540 arch/x86/kernel/acpi/boot.c cpumask_t tmp_map, new_map; cpumask_t 75 arch/x86/kernel/acpi/cstate.c cpumask_t saved_mask; cpumask_t 94 arch/x86/kernel/apic_32.c static void lapic_timer_broadcast(cpumask_t mask); cpumask_t 375 arch/x86/kernel/apic_32.c static void lapic_timer_broadcast(cpumask_t mask) cpumask_t 1493 arch/x86/kernel/apic_32.c cpumask_t tmp_map; cpumask_t 81 arch/x86/kernel/apic_64.c static void lapic_timer_broadcast(cpumask_t mask); cpumask_t 387 arch/x86/kernel/apic_64.c static void lapic_timer_broadcast(cpumask_t mask) cpumask_t 1431 arch/x86/kernel/apic_64.c cpumask_t tmp_map; cpumask_t 496 arch/x86/kernel/apm_32.c static cpumask_t apm_save_cpus(void) cpumask_t 498 arch/x86/kernel/apm_32.c cpumask_t x = current->cpus_allowed; cpumask_t 505 arch/x86/kernel/apm_32.c static inline void apm_restore_cpus(cpumask_t mask) cpumask_t 600 arch/x86/kernel/apm_32.c cpumask_t cpus; cpumask_t 643 arch/x86/kernel/apm_32.c cpumask_t cpus; cpumask_t 852 arch/x86/kernel/cpu/common.c cpumask_t cpu_initialized __cpuinitdata = CPU_MASK_NONE; cpumask_t 147 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c cpumask_t mask; cpumask_t 192 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c cpumask_t saved_mask = current->cpus_allowed; cpumask_t 202 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c cpumask_t saved_mask = current->cpus_allowed; cpumask_t 214 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c static u32 get_cur_val(const cpumask_t *mask) cpumask_t 270 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c cpumask_t saved_mask; cpumask_t 367 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c static unsigned int check_freqs(const cpumask_t *mask, unsigned int freq, cpumask_t 388 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c cpumask_t online_policy_cpus; cpumask_t 60 arch/x86/kernel/cpu/cpufreq/powernow-k8.c DEFINE_PER_CPU(cpumask_t, cpu_core_map); cpumask_t 478 arch/x86/kernel/cpu/cpufreq/powernow-k8.c cpumask_t oldmask; cpumask_t 1015 arch/x86/kernel/cpu/cpufreq/powernow-k8.c cpumask_t oldmask; cpumask_t 1104 arch/x86/kernel/cpu/cpufreq/powernow-k8.c cpumask_t oldmask; cpumask_t 1243 arch/x86/kernel/cpu/cpufreq/powernow-k8.c cpumask_t oldmask = current->cpus_allowed; cpumask_t 41 arch/x86/kernel/cpu/cpufreq/powernow-k8.h cpumask_t *available_cores; cpumask_t 216 arch/x86/kernel/cpu/cpufreq/powernow-k8.h static inline void define_siblings(int cpu, cpumask_t cpu_sharedcore_mask[]) cpumask_t 220 arch/x86/kernel/cpu/cpufreq/powernow-k8.h static inline void define_siblings(int cpu, cpumask_t cpu_sharedcore_mask[]) cpumask_t 326 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c cpumask_t saved_mask; cpumask_t 462 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c cpumask_t online_policy_cpus; cpumask_t 463 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c cpumask_t saved_mask; cpumask_t 464 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c cpumask_t set_mask; cpumask_t 465 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c cpumask_t covered_cpus; cpumask_t 232 arch/x86/kernel/cpu/cpufreq/speedstep-ich.c static unsigned int _speedstep_get(const cpumask_t *cpus) cpumask_t 235 arch/x86/kernel/cpu/cpufreq/speedstep-ich.c cpumask_t cpus_allowed; cpumask_t 264 arch/x86/kernel/cpu/cpufreq/speedstep-ich.c cpumask_t cpus_allowed; cpumask_t 321 arch/x86/kernel/cpu/cpufreq/speedstep-ich.c cpumask_t cpus_allowed; cpumask_t 135 arch/x86/kernel/cpu/intel_cacheinfo.c cpumask_t shared_cpu_map; /* future?: only cpus/node is needed */ cpumask_t 542 arch/x86/kernel/cpu/intel_cacheinfo.c cpumask_t oldmask; cpumask_t 626 arch/x86/kernel/cpu/intel_cacheinfo.c cpumask_t *mask = &this_leaf->shared_cpu_map; cpumask_t 872 arch/x86/kernel/cpu/intel_cacheinfo.c static cpumask_t cache_dev_map = CPU_MASK_NONE; cpumask_t 513 arch/x86/kernel/cpu/mcheck/mce_64.c static cpumask_t mce_cpus = CPU_MASK_NONE; cpumask_t 825 arch/x86/kernel/cpu/mcheck/mce_64.c static cpumask_t mce_device_initialized = CPU_MASK_NONE; cpumask_t 70 arch/x86/kernel/cpu/mcheck/mce_amd_64.c cpumask_t cpus; cpumask_t 254 arch/x86/kernel/cpu/mcheck/mce_amd_64.c static void affinity_set(unsigned int cpu, cpumask_t *oldmask, cpumask_t 255 arch/x86/kernel/cpu/mcheck/mce_amd_64.c cpumask_t *newmask) cpumask_t 263 arch/x86/kernel/cpu/mcheck/mce_amd_64.c static void affinity_restore(const cpumask_t *oldmask) cpumask_t 280 arch/x86/kernel/cpu/mcheck/mce_amd_64.c cpumask_t oldmask, newmask; cpumask_t 297 arch/x86/kernel/cpu/mcheck/mce_amd_64.c cpumask_t oldmask, newmask; cpumask_t 319 arch/x86/kernel/cpu/mcheck/mce_amd_64.c cpumask_t oldmask, newmask; cpumask_t 330 arch/x86/kernel/cpu/mcheck/mce_amd_64.c cpumask_t oldmask, newmask; cpumask_t 471 arch/x86/kernel/cpu/mcheck/mce_amd_64.c cpumask_t oldmask, newmask; cpumask_t 80 arch/x86/kernel/crash.c cpumask_t mask = cpu_online_map; cpumask_t 33 arch/x86/kernel/genapic_flat_64.c static cpumask_t flat_target_cpus(void) cpumask_t 38 arch/x86/kernel/genapic_flat_64.c static cpumask_t flat_vector_allocation_domain(int cpu) cpumask_t 48 arch/x86/kernel/genapic_flat_64.c cpumask_t domain = { { [0] = APIC_ALL_CPUS, } }; cpumask_t 72 arch/x86/kernel/genapic_flat_64.c static void flat_send_IPI_mask(cpumask_t cpumask, int vector) cpumask_t 90 arch/x86/kernel/genapic_flat_64.c cpumask_t allbutme = cpu_online_map; cpumask_t 138 arch/x86/kernel/genapic_flat_64.c static unsigned int flat_cpu_mask_to_apicid(cpumask_t cpumask) cpumask_t 189 arch/x86/kernel/genapic_flat_64.c static cpumask_t physflat_target_cpus(void) cpumask_t 194 arch/x86/kernel/genapic_flat_64.c static cpumask_t physflat_vector_allocation_domain(int cpu) cpumask_t 199 arch/x86/kernel/genapic_flat_64.c static void physflat_send_IPI_mask(cpumask_t cpumask, int vector) cpumask_t 206 arch/x86/kernel/genapic_flat_64.c cpumask_t allbutme = cpu_online_map; cpumask_t 217 arch/x86/kernel/genapic_flat_64.c static unsigned int physflat_cpu_mask_to_apicid(cpumask_t cpumask) cpumask_t 25 arch/x86/kernel/genx2apic_cluster.c static cpumask_t x2apic_target_cpus(void) cpumask_t 33 arch/x86/kernel/genx2apic_cluster.c static cpumask_t x2apic_vector_allocation_domain(int cpu) cpumask_t 35 arch/x86/kernel/genx2apic_cluster.c cpumask_t domain = CPU_MASK_NONE; cpumask_t 59 arch/x86/kernel/genx2apic_cluster.c static void x2apic_send_IPI_mask(cpumask_t mask, int vector) cpumask_t 74 arch/x86/kernel/genx2apic_cluster.c cpumask_t mask = cpu_online_map; cpumask_t 92 arch/x86/kernel/genx2apic_cluster.c static unsigned int x2apic_cpu_mask_to_apicid(cpumask_t cpumask) cpumask_t 32 arch/x86/kernel/genx2apic_phys.c static cpumask_t x2apic_target_cpus(void) cpumask_t 37 arch/x86/kernel/genx2apic_phys.c static cpumask_t x2apic_vector_allocation_domain(int cpu) cpumask_t 39 arch/x86/kernel/genx2apic_phys.c cpumask_t domain = CPU_MASK_NONE; cpumask_t 57 arch/x86/kernel/genx2apic_phys.c static void x2apic_send_IPI_mask(cpumask_t mask, int vector) cpumask_t 72 arch/x86/kernel/genx2apic_phys.c cpumask_t mask = cpu_online_map; cpumask_t 90 arch/x86/kernel/genx2apic_phys.c static unsigned int x2apic_cpu_mask_to_apicid(cpumask_t cpumask) cpumask_t 79 arch/x86/kernel/genx2apic_uv_x.c static cpumask_t uv_target_cpus(void) cpumask_t 84 arch/x86/kernel/genx2apic_uv_x.c static cpumask_t uv_vector_allocation_domain(int cpu) cpumask_t 86 arch/x86/kernel/genx2apic_uv_x.c cpumask_t domain = CPU_MASK_NONE; cpumask_t 127 arch/x86/kernel/genx2apic_uv_x.c static void uv_send_IPI_mask(cpumask_t mask, int vector) cpumask_t 138 arch/x86/kernel/genx2apic_uv_x.c cpumask_t mask = cpu_online_map; cpumask_t 160 arch/x86/kernel/genx2apic_uv_x.c static unsigned int uv_cpu_mask_to_apicid(cpumask_t cpumask) cpumask_t 340 arch/x86/kernel/io_apic_32.c static void set_ioapic_affinity_irq(unsigned int irq, cpumask_t cpumask) cpumask_t 346 arch/x86/kernel/io_apic_32.c cpumask_t tmp; cpumask_t 404 arch/x86/kernel/io_apic_32.c static cpumask_t balance_irq_affinity[NR_IRQS] = { cpumask_t 408 arch/x86/kernel/io_apic_32.c void set_balance_irq_affinity(unsigned int irq, cpumask_t mask) cpumask_t 413 arch/x86/kernel/io_apic_32.c static unsigned long move(int curr_cpu, cpumask_t allowed_mask, cpumask_t 443 arch/x86/kernel/io_apic_32.c cpumask_t allowed_mask; cpumask_t 487 arch/x86/kernel/io_apic_32.c cpumask_t allowed_mask, target_cpu_mask, tmp; cpumask_t 692 arch/x86/kernel/io_apic_32.c cpumask_t tmp; cpumask_t 2519 arch/x86/kernel/io_apic_32.c static void set_msi_irq_affinity(unsigned int irq, cpumask_t mask) cpumask_t 2523 arch/x86/kernel/io_apic_32.c cpumask_t tmp; cpumask_t 2614 arch/x86/kernel/io_apic_32.c static void set_ht_irq_affinity(unsigned int irq, cpumask_t mask) cpumask_t 2617 arch/x86/kernel/io_apic_32.c cpumask_t tmp; cpumask_t 2651 arch/x86/kernel/io_apic_32.c cpumask_t tmp; cpumask_t 61 arch/x86/kernel/io_apic_64.c cpumask_t domain; cpumask_t 62 arch/x86/kernel/io_apic_64.c cpumask_t old_domain; cpumask_t 88 arch/x86/kernel/io_apic_64.c static int assign_irq_vector(int irq, cpumask_t mask); cpumask_t 327 arch/x86/kernel/io_apic_64.c static void set_ioapic_affinity_irq(unsigned int irq, cpumask_t mask) cpumask_t 332 arch/x86/kernel/io_apic_64.c cpumask_t tmp; cpumask_t 786 arch/x86/kernel/io_apic_64.c static int __assign_irq_vector(int irq, cpumask_t mask) cpumask_t 815 arch/x86/kernel/io_apic_64.c cpumask_t tmp; cpumask_t 822 arch/x86/kernel/io_apic_64.c cpumask_t domain, new_mask; cpumask_t 861 arch/x86/kernel/io_apic_64.c static int assign_irq_vector(int irq, cpumask_t mask) cpumask_t 875 arch/x86/kernel/io_apic_64.c cpumask_t mask; cpumask_t 1014 arch/x86/kernel/io_apic_64.c cpumask_t mask; cpumask_t 1581 arch/x86/kernel/io_apic_64.c static void migrate_ioapic_irq(int irq, cpumask_t mask) cpumask_t 1585 arch/x86/kernel/io_apic_64.c cpumask_t tmp, cleanup_mask; cpumask_t 1684 arch/x86/kernel/io_apic_64.c static void set_ir_ioapic_affinity_irq(unsigned int irq, cpumask_t mask) cpumask_t 1742 arch/x86/kernel/io_apic_64.c cpumask_t cleanup_mask; cpumask_t 2322 arch/x86/kernel/io_apic_64.c cpumask_t tmp; cpumask_t 2384 arch/x86/kernel/io_apic_64.c static void set_msi_irq_affinity(unsigned int irq, cpumask_t mask) cpumask_t 2389 arch/x86/kernel/io_apic_64.c cpumask_t tmp; cpumask_t 2417 arch/x86/kernel/io_apic_64.c static void ir_set_msi_irq_affinity(unsigned int irq, cpumask_t mask) cpumask_t 2421 arch/x86/kernel/io_apic_64.c cpumask_t tmp, cleanup_mask; cpumask_t 2637 arch/x86/kernel/io_apic_64.c static void dmar_msi_set_affinity(unsigned int irq, cpumask_t mask) cpumask_t 2642 arch/x86/kernel/io_apic_64.c cpumask_t tmp; cpumask_t 2714 arch/x86/kernel/io_apic_64.c static void set_ht_irq_affinity(unsigned int irq, cpumask_t mask) cpumask_t 2718 arch/x86/kernel/io_apic_64.c cpumask_t tmp; cpumask_t 2750 arch/x86/kernel/io_apic_64.c cpumask_t tmp; cpumask_t 119 arch/x86/kernel/ipi.c void send_IPI_mask_bitmask(cpumask_t cpumask, int vector) cpumask_t 130 arch/x86/kernel/ipi.c void send_IPI_mask_sequence(cpumask_t mask, int vector) cpumask_t 394 arch/x86/kernel/irq_32.c void fixup_irqs(cpumask_t map) cpumask_t 400 arch/x86/kernel/irq_32.c cpumask_t mask; cpumask_t 222 arch/x86/kernel/irq_64.c void fixup_irqs(cpumask_t map) cpumask_t 228 arch/x86/kernel/irq_64.c cpumask_t mask; cpumask_t 113 arch/x86/kernel/microcode_core.c cpumask_t old; cpumask_t 221 arch/x86/kernel/microcode_core.c cpumask_t old = current->cpus_allowed; cpumask_t 345 arch/x86/kernel/microcode_core.c cpumask_t old = current->cpus_allowed; cpumask_t 44 arch/x86/kernel/nmi.c static cpumask_t backtrace_mask = CPU_MASK_NONE; cpumask_t 249 arch/x86/kernel/process.c static cpumask_t c1e_mask = CPU_MASK_NONE; cpumask_t 44 arch/x86/kernel/setup_percpu.c cpumask_t *node_to_cpumask_map; cpumask_t 204 arch/x86/kernel/setup_percpu.c cpumask_t *map; cpumask_t 214 arch/x86/kernel/setup_percpu.c map = alloc_bootmem_low(nr_node_ids * sizeof(cpumask_t)); cpumask_t 265 arch/x86/kernel/setup_percpu.c cpumask_t *mask; cpumask_t 327 arch/x86/kernel/setup_percpu.c static const cpumask_t cpu_mask_none; cpumask_t 332 arch/x86/kernel/setup_percpu.c const cpumask_t *_node_to_cpumask_ptr(int node) cpumask_t 339 arch/x86/kernel/setup_percpu.c return (const cpumask_t *)&cpu_online_map; cpumask_t 359 arch/x86/kernel/setup_percpu.c cpumask_t node_to_cpumask(int node) cpumask_t 129 arch/x86/kernel/smp.c void native_send_call_func_ipi(cpumask_t mask) cpumask_t 131 arch/x86/kernel/smp.c cpumask_t allbutself; cpumask_t 105 arch/x86/kernel/smpboot.c cpumask_t cpu_online_map __read_mostly; cpumask_t 108 arch/x86/kernel/smpboot.c cpumask_t cpu_callin_map; cpumask_t 109 arch/x86/kernel/smpboot.c cpumask_t cpu_callout_map; cpumask_t 110 arch/x86/kernel/smpboot.c cpumask_t cpu_possible_map; cpumask_t 114 arch/x86/kernel/smpboot.c DEFINE_PER_CPU(cpumask_t, cpu_sibling_map); cpumask_t 118 arch/x86/kernel/smpboot.c DEFINE_PER_CPU(cpumask_t, cpu_core_map); cpumask_t 129 arch/x86/kernel/smpboot.c static cpumask_t cpu_sibling_setup_map; cpumask_t 137 arch/x86/kernel/smpboot.c cpumask_t node_to_cpumask_map[MAX_NUMNODES] __read_mostly = cpumask_t 506 arch/x86/kernel/smpboot.c cpumask_t cpu_coregroup_map(int cpu) cpumask_t 23 arch/x86/kernel/tlb_32.c static cpumask_t flush_cpumask; cpumask_t 125 arch/x86/kernel/tlb_32.c void native_flush_tlb_others(const cpumask_t *cpumaskp, struct mm_struct *mm, cpumask_t 128 arch/x86/kernel/tlb_32.c cpumask_t cpumask = *cpumaskp; cpumask_t 175 arch/x86/kernel/tlb_32.c cpumask_t cpu_mask; cpumask_t 189 arch/x86/kernel/tlb_32.c cpumask_t cpu_mask; cpumask_t 210 arch/x86/kernel/tlb_32.c cpumask_t cpu_mask; cpumask_t 46 arch/x86/kernel/tlb_64.c cpumask_t flush_cpumask; cpumask_t 160 arch/x86/kernel/tlb_64.c void native_flush_tlb_others(const cpumask_t *cpumaskp, struct mm_struct *mm, cpumask_t 165 arch/x86/kernel/tlb_64.c cpumask_t cpumask = *cpumaskp; cpumask_t 213 arch/x86/kernel/tlb_64.c cpumask_t cpu_mask; cpumask_t 227 arch/x86/kernel/tlb_64.c cpumask_t cpu_mask; cpumask_t 248 arch/x86/kernel/tlb_64.c cpumask_t cpu_mask; cpumask_t 219 arch/x86/kernel/tlb_uv.c cpumask_t *cpumaskp) cpumask_t 299 arch/x86/kernel/tlb_uv.c int uv_flush_tlb_others(cpumask_t *cpumaskp, struct mm_struct *mm, cpumask_t 67 arch/x86/mach-voyager/voyager_smp.c cpumask_t cpu_online_map = CPU_MASK_NONE; cpumask_t 72 arch/x86/mach-voyager/voyager_smp.c cpumask_t phys_cpu_present_map = CPU_MASK_NONE; cpumask_t 88 arch/x86/mach-voyager/voyager_smp.c static void set_vic_irq_affinity(unsigned int irq, cpumask_t mask); cpumask_t 214 arch/x86/mach-voyager/voyager_smp.c static cpumask_t smp_commenced_mask = CPU_MASK_NONE; cpumask_t 217 arch/x86/mach-voyager/voyager_smp.c cpumask_t cpu_callin_map = CPU_MASK_NONE; cpumask_t 218 arch/x86/mach-voyager/voyager_smp.c cpumask_t cpu_callout_map = CPU_MASK_NONE; cpumask_t 219 arch/x86/mach-voyager/voyager_smp.c cpumask_t cpu_possible_map = CPU_MASK_NONE; cpumask_t 1599 arch/x86/mach-voyager/voyager_smp.c void set_vic_irq_affinity(unsigned int irq, cpumask_t mask) cpumask_t 407 arch/x86/mm/mmio-mod.c static cpumask_t downed_cpus; cpumask_t 636 arch/x86/xen/enlighten.c static void xen_flush_tlb_others(const cpumask_t *cpus, struct mm_struct *mm, cpumask_t 641 arch/x86/xen/enlighten.c cpumask_t mask; cpumask_t 643 arch/x86/xen/enlighten.c cpumask_t cpumask = *cpus; cpumask_t 1061 arch/x86/xen/mmu.c cpumask_t mask; cpumask_t 36 arch/x86/xen/smp.c cpumask_t xen_cpu_initialized_map; cpumask_t 411 arch/x86/xen/smp.c static void xen_send_IPI_mask(cpumask_t mask, enum ipi_vector vector) cpumask_t 421 arch/x86/xen/smp.c static void xen_smp_send_call_function_ipi(cpumask_t mask) cpumask_t 61 arch/x86/xen/xen-ops.h extern cpumask_t xen_cpu_initialized_map; cpumask_t 101 block/blk.h cpumask_t mask = cpu_coregroup_map(cpu); cpumask_t 130 include/acpi/processor.h cpumask_t shared_cpu_map; cpumask_t 175 include/acpi/processor.h cpumask_t shared_cpu_map; cpumask_t 6 include/asm-cris/smp.h extern cpumask_t phys_cpu_present_map; cpumask_t 7 include/asm-cris/smp.h extern cpumask_t cpu_possible_map; cpumask_t 63 include/asm-generic/topology.h cpumask_t _##v = node_to_cpumask(node); \ cpumask_t 64 include/asm-generic/topology.h const cpumask_t *v = &_##v cpumask_t 65 include/asm-m32r/smp.h extern cpumask_t cpu_callout_map; cpumask_t 66 include/asm-m32r/smp.h extern cpumask_t cpu_possible_map; cpumask_t 67 include/asm-m32r/smp.h extern cpumask_t cpu_present_map; cpumask_t 90 include/asm-m32r/smp.h extern unsigned long send_IPI_mask_phys(cpumask_t, int, int); cpumask_t 93 include/asm-m32r/smp.h extern void arch_send_call_function_ipi(cpumask_t mask); cpumask_t 52 include/asm-parisc/irq.h extern int cpu_check_affinity(unsigned int irq, cpumask_t *dest); cpumask_t 19 include/asm-parisc/smp.h extern cpumask_t cpu_online_map; cpumask_t 34 include/asm-parisc/smp.h extern void arch_send_call_function_ipi(cpumask_t mask); cpumask_t 13 include/asm-x86/bigsmp/apic.h static inline cpumask_t target_cpus(void) cpumask_t 129 include/asm-x86/bigsmp/apic.h static inline unsigned int cpu_mask_to_apicid(cpumask_t cpumask) cpumask_t 4 include/asm-x86/bigsmp/ipi.h void send_IPI_mask_sequence(cpumask_t mask, int vector); cpumask_t 6 include/asm-x86/bigsmp/ipi.h static inline void send_IPI_mask(cpumask_t mask, int vector) cpumask_t 13 include/asm-x86/bigsmp/ipi.h cpumask_t mask = cpu_online_map; cpumask_t 12 include/asm-x86/es7000/apic.h static inline cpumask_t target_cpus(void) cpumask_t 148 include/asm-x86/es7000/apic.h static inline unsigned int cpu_mask_to_apicid(cpumask_t cpumask) cpumask_t 4 include/asm-x86/es7000/ipi.h void send_IPI_mask_sequence(cpumask_t mask, int vector); cpumask_t 6 include/asm-x86/es7000/ipi.h static inline void send_IPI_mask(cpumask_t mask, int vector) cpumask_t 13 include/asm-x86/es7000/ipi.h cpumask_t mask = cpu_online_map; cpumask_t 26 include/asm-x86/genapic_32.h cpumask_t (*target_cpus)(void); cpumask_t 59 include/asm-x86/genapic_32.h unsigned int (*cpu_mask_to_apicid)(cpumask_t cpumask); cpumask_t 63 include/asm-x86/genapic_32.h void (*send_IPI_mask)(cpumask_t mask, int vector); cpumask_t 21 include/asm-x86/genapic_64.h cpumask_t (*target_cpus)(void); cpumask_t 22 include/asm-x86/genapic_64.h cpumask_t (*vector_allocation_domain)(int cpu); cpumask_t 25 include/asm-x86/genapic_64.h void (*send_IPI_mask)(cpumask_t mask, int vector); cpumask_t 30 include/asm-x86/genapic_64.h unsigned int (*cpu_mask_to_apicid)(cpumask_t cpumask); cpumask_t 120 include/asm-x86/ipi.h static inline void send_IPI_mask_sequence(cpumask_t mask, int vector) cpumask_t 40 include/asm-x86/irq.h extern void fixup_irqs(cpumask_t map); cpumask_t 11 include/asm-x86/mach-default/mach_apic.h static inline cpumask_t target_cpus(void) cpumask_t 62 include/asm-x86/mach-default/mach_apic.h static inline unsigned int cpu_mask_to_apicid(cpumask_t cpumask) cpumask_t 7 include/asm-x86/mach-default/mach_ipi.h void send_IPI_mask_bitmask(cpumask_t mask, int vector); cpumask_t 16 include/asm-x86/mach-default/mach_ipi.h static inline void send_IPI_mask(cpumask_t mask, int vector) cpumask_t 25 include/asm-x86/mach-default/mach_ipi.h cpumask_t mask = cpu_online_map; cpumask_t 10 include/asm-x86/numaq/apic.h static inline cpumask_t target_cpus(void) cpumask_t 127 include/asm-x86/numaq/apic.h static inline unsigned int cpu_mask_to_apicid(cpumask_t cpumask) cpumask_t 4 include/asm-x86/numaq/ipi.h void send_IPI_mask_sequence(cpumask_t, int vector); cpumask_t 6 include/asm-x86/numaq/ipi.h static inline void send_IPI_mask(cpumask_t mask, int vector) cpumask_t 13 include/asm-x86/numaq/ipi.h cpumask_t mask = cpu_online_map; cpumask_t 247 include/asm-x86/paravirt.h void (*flush_tlb_others)(const cpumask_t *cpus, struct mm_struct *mm, cpumask_t 987 include/asm-x86/paravirt.h static inline void flush_tlb_others(cpumask_t cpumask, struct mm_struct *mm, cpumask_t 108 include/asm-x86/pci.h static inline cpumask_t __pcibus_to_cpumask(struct pci_bus *bus) cpumask_t 96 include/asm-x86/processor.h cpumask_t llc_shared_map; cpumask_t 21 include/asm-x86/smp.h extern cpumask_t cpu_callout_map; cpumask_t 22 include/asm-x86/smp.h extern cpumask_t cpu_initialized; cpumask_t 23 include/asm-x86/smp.h extern cpumask_t cpu_callin_map; cpumask_t 32 include/asm-x86/smp.h extern cpumask_t cpu_initialized; cpumask_t 34 include/asm-x86/smp.h DECLARE_PER_CPU(cpumask_t, cpu_sibling_map); cpumask_t 35 include/asm-x86/smp.h DECLARE_PER_CPU(cpumask_t, cpu_core_map); cpumask_t 63 include/asm-x86/smp.h void (*send_call_func_ipi)(cpumask_t mask); cpumask_t 126 include/asm-x86/smp.h static inline void arch_send_call_function_ipi(cpumask_t mask) cpumask_t 141 include/asm-x86/smp.h void native_send_call_func_ipi(cpumask_t mask); cpumask_t 17 include/asm-x86/summit/apic.h static inline cpumask_t target_cpus(void) cpumask_t 141 include/asm-x86/summit/apic.h static inline unsigned int cpu_mask_to_apicid(cpumask_t cpumask) cpumask_t 4 include/asm-x86/summit/ipi.h void send_IPI_mask_sequence(cpumask_t mask, int vector); cpumask_t 6 include/asm-x86/summit/ipi.h static inline void send_IPI_mask(cpumask_t mask, int vector) cpumask_t 13 include/asm-x86/summit/ipi.h cpumask_t mask = cpu_online_map; cpumask_t 116 include/asm-x86/tlbflush.h static inline void native_flush_tlb_others(const cpumask_t *cpumask, cpumask_t 145 include/asm-x86/tlbflush.h void native_flush_tlb_others(const cpumask_t *cpumask, struct mm_struct *mm, cpumask_t 48 include/asm-x86/topology.h extern cpumask_t node_to_cpumask_map[]; cpumask_t 66 include/asm-x86/topology.h static inline cpumask_t node_to_cpumask(int node) cpumask_t 74 include/asm-x86/topology.h extern cpumask_t *node_to_cpumask_map; cpumask_t 85 include/asm-x86/topology.h extern const cpumask_t *_node_to_cpumask_ptr(int node); cpumask_t 86 include/asm-x86/topology.h extern cpumask_t node_to_cpumask(int node); cpumask_t 106 include/asm-x86/topology.h static inline const cpumask_t *_node_to_cpumask_ptr(int node) cpumask_t 112 include/asm-x86/topology.h static inline cpumask_t node_to_cpumask(int node) cpumask_t 121 include/asm-x86/topology.h const cpumask_t *v = _node_to_cpumask_ptr(node) cpumask_t 189 include/asm-x86/topology.h static inline const cpumask_t *_node_to_cpumask_ptr(int node) cpumask_t 193 include/asm-x86/topology.h static inline cpumask_t node_to_cpumask(int node) cpumask_t 204 include/asm-x86/topology.h const cpumask_t *v = _node_to_cpumask_ptr(node) cpumask_t 221 include/asm-x86/topology.h extern cpumask_t cpu_coregroup_map(int cpu); cpumask_t 328 include/asm-x86/uv/uv_bau.h extern int uv_flush_tlb_others(cpumask_t *, struct mm_struct *, unsigned long); cpumask_t 85 include/linux/clockchips.h cpumask_t cpumask; cpumask_t 91 include/linux/clockchips.h void (*broadcast)(cpumask_t mask); cpumask_t 83 include/linux/cpufreq.h cpumask_t cpus; /* CPUs requiring sw coordination */ cpumask_t 84 include/linux/cpufreq.h cpumask_t related_cpus; /* CPUs with any coordination */ cpumask_t 142 include/linux/cpumask.h extern cpumask_t _unused_cpumask_arg_; cpumask_t 145 include/linux/cpumask.h static inline void __cpu_set(int cpu, volatile cpumask_t *dstp) cpumask_t 151 include/linux/cpumask.h static inline void __cpu_clear(int cpu, volatile cpumask_t *dstp) cpumask_t 157 include/linux/cpumask.h static inline void __cpus_setall(cpumask_t *dstp, int nbits) cpumask_t 163 include/linux/cpumask.h static inline void __cpus_clear(cpumask_t *dstp, int nbits) cpumask_t 172 include/linux/cpumask.h static inline int __cpu_test_and_set(int cpu, cpumask_t *addr) cpumask_t 178 include/linux/cpumask.h static inline void __cpus_and(cpumask_t *dstp, const cpumask_t *src1p, cpumask_t 179 include/linux/cpumask.h const cpumask_t *src2p, int nbits) cpumask_t 185 include/linux/cpumask.h static inline void __cpus_or(cpumask_t *dstp, const cpumask_t *src1p, cpumask_t 186 include/linux/cpumask.h const cpumask_t *src2p, int nbits) cpumask_t 192 include/linux/cpumask.h static inline void __cpus_xor(cpumask_t *dstp, const cpumask_t *src1p, cpumask_t 193 include/linux/cpumask.h const cpumask_t *src2p, int nbits) cpumask_t 200 include/linux/cpumask.h static inline void __cpus_andnot(cpumask_t *dstp, const cpumask_t *src1p, cpumask_t 201 include/linux/cpumask.h const cpumask_t *src2p, int nbits) cpumask_t 207 include/linux/cpumask.h static inline void __cpus_complement(cpumask_t *dstp, cpumask_t 208 include/linux/cpumask.h const cpumask_t *srcp, int nbits) cpumask_t 214 include/linux/cpumask.h static inline int __cpus_equal(const cpumask_t *src1p, cpumask_t 215 include/linux/cpumask.h const cpumask_t *src2p, int nbits) cpumask_t 221 include/linux/cpumask.h static inline int __cpus_intersects(const cpumask_t *src1p, cpumask_t 222 include/linux/cpumask.h const cpumask_t *src2p, int nbits) cpumask_t 228 include/linux/cpumask.h static inline int __cpus_subset(const cpumask_t *src1p, cpumask_t 229 include/linux/cpumask.h const cpumask_t *src2p, int nbits) cpumask_t 235 include/linux/cpumask.h static inline int __cpus_empty(const cpumask_t *srcp, int nbits) cpumask_t 241 include/linux/cpumask.h static inline int __cpus_full(const cpumask_t *srcp, int nbits) cpumask_t 247 include/linux/cpumask.h static inline int __cpus_weight(const cpumask_t *srcp, int nbits) cpumask_t 254 include/linux/cpumask.h static inline void __cpus_shift_right(cpumask_t *dstp, cpumask_t 255 include/linux/cpumask.h const cpumask_t *srcp, int n, int nbits) cpumask_t 262 include/linux/cpumask.h static inline void __cpus_shift_left(cpumask_t *dstp, cpumask_t 263 include/linux/cpumask.h const cpumask_t *srcp, int n, int nbits) cpumask_t 278 include/linux/cpumask.h static inline const cpumask_t *get_cpu_mask(unsigned int cpu) cpumask_t 282 include/linux/cpumask.h return (const cpumask_t *)p; cpumask_t 298 include/linux/cpumask.h (cpumask_t) { { \ cpumask_t 307 include/linux/cpumask.h (cpumask_t) { { \ cpumask_t 313 include/linux/cpumask.h extern cpumask_t cpu_mask_all; cpumask_t 319 include/linux/cpumask.h (cpumask_t) { { \ cpumask_t 324 include/linux/cpumask.h (cpumask_t) { { \ cpumask_t 337 include/linux/cpumask.h #define CPUMASK_PTR(v, m) cpumask_t *v = &(m->v) cpumask_t 342 include/linux/cpumask.h const cpumask_t *srcp, int nbits) cpumask_t 350 include/linux/cpumask.h cpumask_t *dstp, int nbits) cpumask_t 358 include/linux/cpumask.h const cpumask_t *srcp, int nbits) cpumask_t 364 include/linux/cpumask.h static inline int __cpulist_parse(const char *buf, cpumask_t *dstp, int nbits) cpumask_t 372 include/linux/cpumask.h const cpumask_t *oldp, const cpumask_t *newp, int nbits) cpumask_t 379 include/linux/cpumask.h static inline void __cpus_remap(cpumask_t *dstp, const cpumask_t *srcp, cpumask_t 380 include/linux/cpumask.h const cpumask_t *oldp, const cpumask_t *newp, int nbits) cpumask_t 387 include/linux/cpumask.h static inline void __cpus_onto(cpumask_t *dstp, const cpumask_t *origp, cpumask_t 388 include/linux/cpumask.h const cpumask_t *relmapp, int nbits) cpumask_t 395 include/linux/cpumask.h static inline void __cpus_fold(cpumask_t *dstp, const cpumask_t *origp, cpumask_t 413 include/linux/cpumask.h int __first_cpu(const cpumask_t *srcp); cpumask_t 414 include/linux/cpumask.h int __next_cpu(int n, const cpumask_t *srcp); cpumask_t 415 include/linux/cpumask.h int __any_online_cpu(const cpumask_t *mask); cpumask_t 434 include/linux/cpumask.h int __next_cpu_nr(int n, const cpumask_t *srcp); cpumask_t 501 include/linux/cpumask.h extern cpumask_t cpu_possible_map; cpumask_t 502 include/linux/cpumask.h extern cpumask_t cpu_online_map; cpumask_t 503 include/linux/cpumask.h extern cpumask_t cpu_present_map; cpumask_t 504 include/linux/cpumask.h extern cpumask_t cpu_active_map; cpumask_t 23 include/linux/cpuset.h extern void cpuset_cpus_allowed(struct task_struct *p, cpumask_t *mask); cpumask_t 24 include/linux/cpuset.h extern void cpuset_cpus_allowed_locked(struct task_struct *p, cpumask_t *mask); cpumask_t 89 include/linux/cpuset.h static inline void cpuset_cpus_allowed(struct task_struct *p, cpumask_t *mask) cpumask_t 94 include/linux/cpuset.h cpumask_t *mask) cpumask_t 63 include/linux/interrupt.h cpumask_t mask; cpumask_t 107 include/linux/interrupt.h extern cpumask_t irq_default_affinity; cpumask_t 109 include/linux/interrupt.h extern int irq_set_affinity(unsigned int irq, cpumask_t cpumask); cpumask_t 115 include/linux/interrupt.h static inline int irq_set_affinity(unsigned int irq, cpumask_t cpumask) cpumask_t 114 include/linux/irq.h void (*set_affinity)(unsigned int irq, cpumask_t dest); cpumask_t 170 include/linux/irq.h cpumask_t affinity; cpumask_t 174 include/linux/irq.h cpumask_t pending_mask; cpumask_t 209 include/linux/irq.h void set_pending_irq(unsigned int irq, cpumask_t mask); cpumask_t 227 include/linux/irq.h static inline void set_pending_irq(unsigned int irq, cpumask_t mask) cpumask_t 241 include/linux/irq.h extern void set_balance_irq_affinity(unsigned int irq, cpumask_t mask); cpumask_t 243 include/linux/irq.h static inline void set_balance_irq_affinity(unsigned int irq, cpumask_t mask) cpumask_t 217 include/linux/mm_types.h cpumask_t cpu_vm_mask; cpumask_t 84 include/linux/percpu.h extern void *__percpu_alloc_mask(size_t size, gfp_t gfp, cpumask_t *mask); cpumask_t 91 include/linux/percpu.h static __always_inline void *__percpu_alloc_mask(size_t size, gfp_t gfp, cpumask_t *mask) cpumask_t 62 include/linux/rcuclassic.h cpumask_t cpumask; /* CPUs that need to switch in order */ cpumask_t 251 include/linux/sched.h extern cpumask_t nohz_cpu_mask; cpumask_t 737 include/linux/sched.h cpumask_t cpumask; cpumask_t 775 include/linux/sched.h cpumask_t span; /* span of all CPUs in this domain */ cpumask_t 832 include/linux/sched.h extern void partition_sched_domains(int ndoms_new, cpumask_t *doms_new, cpumask_t 841 include/linux/sched.h partition_sched_domains(int ndoms_new, cpumask_t *doms_new, cpumask_t 926 include/linux/sched.h const cpumask_t *newmask); cpumask_t 1071 include/linux/sched.h cpumask_t cpus_allowed; cpumask_t 1545 include/linux/sched.h const cpumask_t *new_mask); cpumask_t 1548 include/linux/sched.h const cpumask_t *new_mask) cpumask_t 1555 include/linux/sched.h static inline int set_cpus_allowed(struct task_struct *p, cpumask_t new_mask) cpumask_t 2137 include/linux/sched.h extern long sched_setaffinity(pid_t pid, const cpumask_t *new_mask); cpumask_t 2138 include/linux/sched.h extern long sched_getaffinity(pid_t pid, cpumask_t *mask); cpumask_t 53 include/linux/seq_file.h static inline int seq_cpumask(struct seq_file *m, cpumask_t *mask) cpumask_t 65 include/linux/smp.h int smp_call_function_mask(cpumask_t mask, void(*func)(void *info), void *info, cpumask_t 26 include/linux/stop_machine.h int stop_machine(int (*fn)(void *), void *data, const cpumask_t *cpus); cpumask_t 37 include/linux/stop_machine.h int __stop_machine(int (*fn)(void *), void *data, const cpumask_t *cpus); cpumask_t 41 include/linux/stop_machine.h const cpumask_t *cpus) cpumask_t 87 include/linux/tick.h extern cpumask_t *tick_get_broadcast_mask(void); cpumask_t 90 include/linux/tick.h extern cpumask_t *tick_get_broadcast_oneshot_mask(void); cpumask_t 368 init/main.c cpumask_t cpu_mask_all __read_mostly = CPU_MASK_ALL; cpumask_t 424 kernel/compat.c unsigned len, cpumask_t *new_mask) cpumask_t 428 kernel/compat.c if (len < sizeof(cpumask_t)) cpumask_t 429 kernel/compat.c memset(new_mask, 0, sizeof(cpumask_t)); cpumask_t 430 kernel/compat.c else if (len > sizeof(cpumask_t)) cpumask_t 431 kernel/compat.c len = sizeof(cpumask_t); cpumask_t 441 kernel/compat.c cpumask_t new_mask; cpumask_t 455 kernel/compat.c cpumask_t mask; cpumask_t 457 kernel/compat.c unsigned int min_length = sizeof(cpumask_t); cpumask_t 24 kernel/cpu.c cpumask_t cpu_present_map __read_mostly; cpumask_t 32 kernel/cpu.c cpumask_t cpu_online_map __read_mostly = CPU_MASK_ALL; cpumask_t 35 kernel/cpu.c cpumask_t cpu_possible_map __read_mostly = CPU_MASK_ALL; cpumask_t 67 kernel/cpu.c cpumask_t cpu_active_map; cpumask_t 220 kernel/cpu.c cpumask_t old_allowed, tmp; cpumask_t 395 kernel/cpu.c static cpumask_t frozen_cpus; cpumask_t 86 kernel/cpuset.c cpumask_t cpus_allowed; /* CPUs allowed to tasks in cpuset */ cpumask_t 282 kernel/cpuset.c static void guarantee_online_cpus(const struct cpuset *cs, cpumask_t *pmask) cpumask_t 577 kernel/cpuset.c static int generate_sched_domains(cpumask_t **domains, cpumask_t 585 kernel/cpuset.c cpumask_t *doms; /* resulting partition; i.e. sched domains */ cpumask_t 597 kernel/cpuset.c doms = kmalloc(sizeof(cpumask_t), GFP_KERNEL); cpumask_t 676 kernel/cpuset.c doms = kmalloc(ndoms * sizeof(cpumask_t), GFP_KERNEL); cpumask_t 690 kernel/cpuset.c cpumask_t *dp; cpumask_t 753 kernel/cpuset.c cpumask_t *doms; cpumask_t 1308 kernel/cpuset.c cpumask_t mask; cpumask_t 1324 kernel/cpuset.c cpumask_t cpus; cpumask_t 1475 kernel/cpuset.c cpumask_t mask; cpumask_t 1983 kernel/cpuset.c cpumask_t *doms; cpumask_t 2049 kernel/cpuset.c void cpuset_cpus_allowed(struct task_struct *tsk, cpumask_t *pmask) cpumask_t 2060 kernel/cpuset.c void cpuset_cpus_allowed_locked(struct task_struct *tsk, cpumask_t *pmask) cpumask_t 20 kernel/irq/manage.c cpumask_t irq_default_affinity = CPU_MASK_ALL; cpumask_t 82 kernel/irq/manage.c int irq_set_affinity(unsigned int irq, cpumask_t cpumask) cpumask_t 113 kernel/irq/manage.c cpumask_t mask; cpumask_t 4 kernel/irq/migration.c void set_pending_irq(unsigned int irq, cpumask_t mask) cpumask_t 18 kernel/irq/migration.c cpumask_t tmp; cpumask_t 23 kernel/irq/proc.c cpumask_t *mask = &desc->affinity; cpumask_t 43 kernel/irq/proc.c cpumask_t new_value; cpumask_t 95 kernel/irq/proc.c cpumask_t new_value; cpumask_t 46 kernel/profile.c static cpumask_t prof_cpu_mask = CPU_MASK_ALL; cpumask_t 424 kernel/profile.c int len = cpumask_scnprintf(page, count, *(cpumask_t *)data); cpumask_t 434 kernel/profile.c cpumask_t *mask = (cpumask_t *)data; cpumask_t 436 kernel/profile.c cpumask_t new_value; cpumask_t 88 kernel/rcuclassic.c cpumask_t cpumask; cpumask_t 167 kernel/rcupreempt.c static cpumask_t rcu_cpu_online_map __read_mostly = CPU_MASK_NONE; cpumask_t 846 kernel/rcutorture.c cpumask_t tmp_mask; cpumask_t 484 kernel/sched.c cpumask_t span; cpumask_t 485 kernel/sched.c cpumask_t online; cpumask_t 491 kernel/sched.c cpumask_t rto_mask; cpumask_t 2104 kernel/sched.c cpumask_t *tmp) cpumask_t 2155 kernel/sched.c cpumask_t span, tmpmask; cpumask_t 3090 kernel/sched.c int *sd_idle, const cpumask_t *cpus, int *balance) cpumask_t 3411 kernel/sched.c unsigned long imbalance, const cpumask_t *cpus) cpumask_t 3450 kernel/sched.c int *balance, cpumask_t *cpus) cpumask_t 3611 kernel/sched.c cpumask_t *cpus) cpumask_t 3700 kernel/sched.c cpumask_t tmpmask; cpumask_t 3782 kernel/sched.c cpumask_t cpu_mask; cpumask_t 3871 kernel/sched.c cpumask_t tmp; cpumask_t 3954 kernel/sched.c cpumask_t cpus = nohz.cpu_mask; cpumask_t 5388 kernel/sched.c long sched_setaffinity(pid_t pid, const cpumask_t *in_mask) cpumask_t 5390 kernel/sched.c cpumask_t cpus_allowed; cpumask_t 5391 kernel/sched.c cpumask_t new_mask = *in_mask; cpumask_t 5446 kernel/sched.c cpumask_t *new_mask) cpumask_t 5448 kernel/sched.c if (len < sizeof(cpumask_t)) { cpumask_t 5449 kernel/sched.c memset(new_mask, 0, sizeof(cpumask_t)); cpumask_t 5450 kernel/sched.c } else if (len > sizeof(cpumask_t)) { cpumask_t 5451 kernel/sched.c len = sizeof(cpumask_t); cpumask_t 5465 kernel/sched.c cpumask_t new_mask; cpumask_t 5475 kernel/sched.c long sched_getaffinity(pid_t pid, cpumask_t *mask) cpumask_t 5511 kernel/sched.c cpumask_t mask; cpumask_t 5513 kernel/sched.c if (len < sizeof(cpumask_t)) cpumask_t 5520 kernel/sched.c if (copy_to_user(user_mask_ptr, &mask, sizeof(cpumask_t))) cpumask_t 5523 kernel/sched.c return sizeof(cpumask_t); cpumask_t 5892 kernel/sched.c cpumask_t nohz_cpu_mask = CPU_MASK_NONE; cpumask_t 5947 kernel/sched.c int set_cpus_allowed_ptr(struct task_struct *p, const cpumask_t *new_mask) cpumask_t 6118 kernel/sched.c cpumask_t mask; cpumask_t 6134 kernel/sched.c cpumask_t cpus_allowed; cpumask_t 6646 kernel/sched.c cpumask_t *groupmask) cpumask_t 6723 kernel/sched.c cpumask_t *groupmask; cpumask_t 6733 kernel/sched.c groupmask = kmalloc(sizeof(cpumask_t), GFP_KERNEL); cpumask_t 6900 kernel/sched.c static cpumask_t cpu_isolated_map = CPU_MASK_NONE; cpumask_t 6929 kernel/sched.c init_sched_build_groups(const cpumask_t *span, const cpumask_t *cpu_map, cpumask_t 6930 kernel/sched.c int (*group_fn)(int cpu, const cpumask_t *cpu_map, cpumask_t 6932 kernel/sched.c cpumask_t *tmpmask), cpumask_t 6933 kernel/sched.c cpumask_t *covered, cpumask_t *tmpmask) cpumask_t 7020 kernel/sched.c static void sched_domain_node_span(int node, cpumask_t *span) cpumask_t 7051 kernel/sched.c cpu_to_cpu_group(int cpu, const cpumask_t *cpu_map, struct sched_group **sg, cpumask_t 7052 kernel/sched.c cpumask_t *unused) cpumask_t 7070 kernel/sched.c cpu_to_core_group(int cpu, const cpumask_t *cpu_map, struct sched_group **sg, cpumask_t 7071 kernel/sched.c cpumask_t *mask) cpumask_t 7084 kernel/sched.c cpu_to_core_group(int cpu, const cpumask_t *cpu_map, struct sched_group **sg, cpumask_t 7085 kernel/sched.c cpumask_t *unused) cpumask_t 7097 kernel/sched.c cpu_to_phys_group(int cpu, const cpumask_t *cpu_map, struct sched_group **sg, cpumask_t 7098 kernel/sched.c cpumask_t *mask) cpumask_t 7129 kernel/sched.c static int cpu_to_allnodes_group(int cpu, const cpumask_t *cpu_map, cpumask_t 7130 kernel/sched.c struct sched_group **sg, cpumask_t *nodemask) cpumask_t 7172 kernel/sched.c static void free_sched_groups(const cpumask_t *cpu_map, cpumask_t *nodemask) cpumask_t 7206 kernel/sched.c static void free_sched_groups(const cpumask_t *cpu_map, cpumask_t *nodemask) cpumask_t 7303 kernel/sched.c cpumask_t tmpmask; /* make this one first */ cpumask_t 7305 kernel/sched.c cpumask_t nodemask; cpumask_t 7306 kernel/sched.c cpumask_t this_sibling_map; cpumask_t 7307 kernel/sched.c cpumask_t this_core_map; cpumask_t 7309 kernel/sched.c cpumask_t send_covered; cpumask_t 7312 kernel/sched.c cpumask_t domainspan; cpumask_t 7313 kernel/sched.c cpumask_t covered; cpumask_t 7314 kernel/sched.c cpumask_t notcovered; cpumask_t 7328 kernel/sched.c #define SCHED_CPUMASK_VAR(v, a) cpumask_t *v = (cpumask_t *) \ cpumask_t 7370 kernel/sched.c static int __build_sched_domains(const cpumask_t *cpu_map, cpumask_t 7376 kernel/sched.c cpumask_t *tmpmask; cpumask_t 7413 kernel/sched.c tmpmask = (cpumask_t *)allmasks; cpumask_t 7677 kernel/sched.c static int build_sched_domains(const cpumask_t *cpu_map) cpumask_t 7682 kernel/sched.c static cpumask_t *doms_cur; /* current sched domains */ cpumask_t 7692 kernel/sched.c static cpumask_t fallback_doms; cpumask_t 7703 kernel/sched.c static int arch_init_sched_domains(const cpumask_t *cpu_map) cpumask_t 7709 kernel/sched.c doms_cur = kmalloc(sizeof(cpumask_t), GFP_KERNEL); cpumask_t 7720 kernel/sched.c static void arch_destroy_sched_domains(const cpumask_t *cpu_map, cpumask_t 7721 kernel/sched.c cpumask_t *tmpmask) cpumask_t 7730 kernel/sched.c static void detach_destroy_domains(const cpumask_t *cpu_map) cpumask_t 7732 kernel/sched.c cpumask_t tmpmask; cpumask_t 7784 kernel/sched.c void partition_sched_domains(int ndoms_new, cpumask_t *doms_new, cpumask_t 7971 kernel/sched.c cpumask_t non_isolated_cpus; cpumask_t 70 kernel/sched_cpupri.c cpumask_t *lowest_mask) cpumask_t 77 kernel/sched_cpupri.c cpumask_t mask; cpumask_t 17 kernel/sched_cpupri.h cpumask_t mask; cpumask_t 28 kernel/sched_cpupri.h struct task_struct *p, cpumask_t *lowest_mask); cpumask_t 958 kernel/sched_fair.c cpumask_t tmp; cpumask_t 142 kernel/sched_rt.c static inline cpumask_t sched_rt_period_mask(void) cpumask_t 147 kernel/sched_rt.c static inline cpumask_t sched_rt_period_mask(void) cpumask_t 215 kernel/sched_rt.c static inline cpumask_t sched_rt_period_mask(void) cpumask_t 432 kernel/sched_rt.c cpumask_t span; cpumask_t 806 kernel/sched_rt.c cpumask_t mask; cpumask_t 957 kernel/sched_rt.c static DEFINE_PER_CPU(cpumask_t, local_cpu_mask); cpumask_t 959 kernel/sched_rt.c static inline int pick_optimal_cpu(int this_cpu, cpumask_t *mask) cpumask_t 977 kernel/sched_rt.c cpumask_t *lowest_mask = &__get_cpu_var(local_cpu_mask); cpumask_t 1014 kernel/sched_rt.c cpumask_t domain_mask; cpumask_t 1309 kernel/sched_rt.c const cpumask_t *new_mask) cpumask_t 27 kernel/smp.c cpumask_t cpumask; cpumask_t 290 kernel/smp.c static void smp_call_function_mask_quiesce_stack(cpumask_t mask) cpumask_t 321 kernel/smp.c int smp_call_function_mask(cpumask_t mask, void (*func)(void *), void *info, cpumask_t 326 kernel/smp.c cpumask_t allbutself; cpumask_t 313 kernel/softlockup.c cpumask_t temp_cpu_online_map = cpu_online_map; cpumask_t 102 kernel/stop_machine.c int __stop_machine(int (*fn)(void *), void *data, const cpumask_t *cpus) cpumask_t 178 kernel/stop_machine.c int stop_machine(int (*fn)(void *), void *data, const cpumask_t *cpus) cpumask_t 293 kernel/taskstats.c static int add_del_listener(pid_t pid, cpumask_t *maskp, int isadd) cpumask_t 298 kernel/taskstats.c cpumask_t mask = *maskp; cpumask_t 338 kernel/taskstats.c static int parse(struct nlattr *na, cpumask_t *mask) cpumask_t 435 kernel/taskstats.c cpumask_t mask; cpumask_t 31 kernel/time/tick-broadcast.c static cpumask_t tick_broadcast_mask; cpumask_t 49 kernel/time/tick-broadcast.c cpumask_t *tick_get_broadcast_mask(void) cpumask_t 130 kernel/time/tick-broadcast.c static void tick_do_broadcast(cpumask_t mask) cpumask_t 163 kernel/time/tick-broadcast.c cpumask_t mask; cpumask_t 363 kernel/time/tick-broadcast.c static cpumask_t tick_broadcast_oneshot_mask; cpumask_t 368 kernel/time/tick-broadcast.c cpumask_t *tick_get_broadcast_oneshot_mask(void) cpumask_t 392 kernel/time/tick-broadcast.c cpumask_t mask; cpumask_t 495 kernel/time/tick-broadcast.c static void tick_broadcast_init_next_event(cpumask_t *mask, ktime_t expires) cpumask_t 516 kernel/time/tick-broadcast.c cpumask_t mask; cpumask_t 139 kernel/time/tick-common.c const cpumask_t *cpumask) cpumask_t 41 kernel/trace/trace.c static cpumask_t __read_mostly tracing_buffer_mask; cpumask_t 2067 kernel/trace/trace.c static cpumask_t tracing_cpumask = CPU_MASK_ALL; cpumask_t 2073 kernel/trace/trace.c static cpumask_t tracing_cpumask_new; cpumask_t 2504 kernel/trace/trace.c static cpumask_t mask; cpumask_t 212 kernel/trace/trace_sysprof.c cpumask_t saved_mask = current->cpus_allowed; cpumask_t 75 kernel/workqueue.c static cpumask_t cpu_singlethread_map __read_mostly; cpumask_t 83 kernel/workqueue.c static cpumask_t cpu_populated_map __read_mostly; cpumask_t 91 kernel/workqueue.c static const cpumask_t *wq_cpu_map(struct workqueue_struct *wq) cpumask_t 412 kernel/workqueue.c const cpumask_t *cpu_map = wq_cpu_map(wq); cpumask_t 534 kernel/workqueue.c const cpumask_t *cpu_map; cpumask_t 901 kernel/workqueue.c const cpumask_t *cpu_map = wq_cpu_map(wq); cpumask_t 6 lib/cpumask.c int __first_cpu(const cpumask_t *srcp) cpumask_t 12 lib/cpumask.c int __next_cpu(int n, const cpumask_t *srcp) cpumask_t 19 lib/cpumask.c int __next_cpu_nr(int n, const cpumask_t *srcp) cpumask_t 27 lib/cpumask.c int __any_online_cpu(const cpumask_t *mask) cpumask_t 34 mm/allocpercpu.c static void __percpu_depopulate_mask(void *__pdata, cpumask_t *mask) cpumask_t 83 mm/allocpercpu.c cpumask_t *mask) cpumask_t 85 mm/allocpercpu.c cpumask_t populated; cpumask_t 111 mm/allocpercpu.c void *__percpu_alloc_mask(size_t size, gfp_t gfp, cpumask_t *mask) cpumask_t 175 mm/pdflush.c cpumask_t cpus_allowed; cpumask_t 1962 mm/slub.c static cpumask_t kmem_cach_cpu_free_init_once = CPU_MASK_NONE; cpumask_t 3435 mm/slub.c cpumask_t cpus; cpumask_t 23 mm/vmstat.c static void sum_vm_events(unsigned long *ret, cpumask_t *cpumask) cpumask_t 172 net/core/dev.c cpumask_t channel_mask; cpumask_t 101 net/iucv/iucv.c static cpumask_t iucv_buffer_cpumask = CPU_MASK_NONE; cpumask_t 102 net/iucv/iucv.c static cpumask_t iucv_irq_cpumask = CPU_MASK_NONE; cpumask_t 494 net/iucv/iucv.c cpumask_t cpumask; cpumask_t 558 net/iucv/iucv.c cpumask_t cpumask; cpumask_t 60 virt/kvm/kvm_main.c static cpumask_t cpus_hardware_enabled; cpumask_t 109 virt/kvm/kvm_main.c cpumask_t cpus; cpumask_t 135 virt/kvm/kvm_main.c cpumask_t cpus;