cpumask_t         540 arch/x86/kernel/acpi/boot.c 	cpumask_t tmp_map, new_map;
cpumask_t          75 arch/x86/kernel/acpi/cstate.c 	cpumask_t saved_mask;
cpumask_t          94 arch/x86/kernel/apic_32.c static void lapic_timer_broadcast(cpumask_t mask);
cpumask_t         375 arch/x86/kernel/apic_32.c static void lapic_timer_broadcast(cpumask_t mask)
cpumask_t        1493 arch/x86/kernel/apic_32.c 	cpumask_t tmp_map;
cpumask_t          81 arch/x86/kernel/apic_64.c static void lapic_timer_broadcast(cpumask_t mask);
cpumask_t         387 arch/x86/kernel/apic_64.c static void lapic_timer_broadcast(cpumask_t mask)
cpumask_t        1431 arch/x86/kernel/apic_64.c 	cpumask_t tmp_map;
cpumask_t         496 arch/x86/kernel/apm_32.c static cpumask_t apm_save_cpus(void)
cpumask_t         498 arch/x86/kernel/apm_32.c 	cpumask_t x = current->cpus_allowed;
cpumask_t         505 arch/x86/kernel/apm_32.c static inline void apm_restore_cpus(cpumask_t mask)
cpumask_t         600 arch/x86/kernel/apm_32.c 	cpumask_t		cpus;
cpumask_t         643 arch/x86/kernel/apm_32.c 	cpumask_t		cpus;
cpumask_t         852 arch/x86/kernel/cpu/common.c cpumask_t cpu_initialized __cpuinitdata = CPU_MASK_NONE;
cpumask_t         147 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c 	cpumask_t mask;
cpumask_t         192 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c 	cpumask_t saved_mask = current->cpus_allowed;
cpumask_t         202 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c 	cpumask_t saved_mask = current->cpus_allowed;
cpumask_t         214 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c static u32 get_cur_val(const cpumask_t *mask)
cpumask_t         270 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c 	cpumask_t saved_mask;
cpumask_t         367 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c static unsigned int check_freqs(const cpumask_t *mask, unsigned int freq,
cpumask_t         388 arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c 	cpumask_t online_policy_cpus;
cpumask_t          60 arch/x86/kernel/cpu/cpufreq/powernow-k8.c DEFINE_PER_CPU(cpumask_t, cpu_core_map);
cpumask_t         478 arch/x86/kernel/cpu/cpufreq/powernow-k8.c 	cpumask_t oldmask;
cpumask_t        1015 arch/x86/kernel/cpu/cpufreq/powernow-k8.c 	cpumask_t oldmask;
cpumask_t        1104 arch/x86/kernel/cpu/cpufreq/powernow-k8.c 	cpumask_t oldmask;
cpumask_t        1243 arch/x86/kernel/cpu/cpufreq/powernow-k8.c 	cpumask_t oldmask = current->cpus_allowed;
cpumask_t          41 arch/x86/kernel/cpu/cpufreq/powernow-k8.h 	cpumask_t *available_cores;
cpumask_t         216 arch/x86/kernel/cpu/cpufreq/powernow-k8.h static inline void define_siblings(int cpu, cpumask_t cpu_sharedcore_mask[])
cpumask_t         220 arch/x86/kernel/cpu/cpufreq/powernow-k8.h static inline void define_siblings(int cpu, cpumask_t cpu_sharedcore_mask[])
cpumask_t         326 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c 	cpumask_t saved_mask;
cpumask_t         462 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c 	cpumask_t		online_policy_cpus;
cpumask_t         463 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c 	cpumask_t		saved_mask;
cpumask_t         464 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c 	cpumask_t		set_mask;
cpumask_t         465 arch/x86/kernel/cpu/cpufreq/speedstep-centrino.c 	cpumask_t		covered_cpus;
cpumask_t         232 arch/x86/kernel/cpu/cpufreq/speedstep-ich.c static unsigned int _speedstep_get(const cpumask_t *cpus)
cpumask_t         235 arch/x86/kernel/cpu/cpufreq/speedstep-ich.c 	cpumask_t cpus_allowed;
cpumask_t         264 arch/x86/kernel/cpu/cpufreq/speedstep-ich.c 	cpumask_t cpus_allowed;
cpumask_t         321 arch/x86/kernel/cpu/cpufreq/speedstep-ich.c 	cpumask_t cpus_allowed;
cpumask_t         135 arch/x86/kernel/cpu/intel_cacheinfo.c 	cpumask_t shared_cpu_map;	/* future?: only cpus/node is needed */
cpumask_t         542 arch/x86/kernel/cpu/intel_cacheinfo.c 	cpumask_t		oldmask;
cpumask_t         626 arch/x86/kernel/cpu/intel_cacheinfo.c 		cpumask_t *mask = &this_leaf->shared_cpu_map;
cpumask_t         872 arch/x86/kernel/cpu/intel_cacheinfo.c static cpumask_t cache_dev_map = CPU_MASK_NONE;
cpumask_t         513 arch/x86/kernel/cpu/mcheck/mce_64.c 	static cpumask_t mce_cpus = CPU_MASK_NONE;
cpumask_t         825 arch/x86/kernel/cpu/mcheck/mce_64.c static cpumask_t mce_device_initialized = CPU_MASK_NONE;
cpumask_t          70 arch/x86/kernel/cpu/mcheck/mce_amd_64.c 	cpumask_t cpus;
cpumask_t         254 arch/x86/kernel/cpu/mcheck/mce_amd_64.c static void affinity_set(unsigned int cpu, cpumask_t *oldmask,
cpumask_t         255 arch/x86/kernel/cpu/mcheck/mce_amd_64.c 					   cpumask_t *newmask)
cpumask_t         263 arch/x86/kernel/cpu/mcheck/mce_amd_64.c static void affinity_restore(const cpumask_t *oldmask)
cpumask_t         280 arch/x86/kernel/cpu/mcheck/mce_amd_64.c 	cpumask_t oldmask, newmask;
cpumask_t         297 arch/x86/kernel/cpu/mcheck/mce_amd_64.c 	cpumask_t oldmask, newmask;
cpumask_t         319 arch/x86/kernel/cpu/mcheck/mce_amd_64.c 	cpumask_t oldmask, newmask;
cpumask_t         330 arch/x86/kernel/cpu/mcheck/mce_amd_64.c 	cpumask_t oldmask, newmask;
cpumask_t         471 arch/x86/kernel/cpu/mcheck/mce_amd_64.c 	cpumask_t oldmask, newmask;
cpumask_t          80 arch/x86/kernel/crash.c 	cpumask_t mask = cpu_online_map;
cpumask_t          33 arch/x86/kernel/genapic_flat_64.c static cpumask_t flat_target_cpus(void)
cpumask_t          38 arch/x86/kernel/genapic_flat_64.c static cpumask_t flat_vector_allocation_domain(int cpu)
cpumask_t          48 arch/x86/kernel/genapic_flat_64.c 	cpumask_t domain = { { [0] = APIC_ALL_CPUS, } };
cpumask_t          72 arch/x86/kernel/genapic_flat_64.c static void flat_send_IPI_mask(cpumask_t cpumask, int vector)
cpumask_t          90 arch/x86/kernel/genapic_flat_64.c 		cpumask_t allbutme = cpu_online_map;
cpumask_t         138 arch/x86/kernel/genapic_flat_64.c static unsigned int flat_cpu_mask_to_apicid(cpumask_t cpumask)
cpumask_t         189 arch/x86/kernel/genapic_flat_64.c static cpumask_t physflat_target_cpus(void)
cpumask_t         194 arch/x86/kernel/genapic_flat_64.c static cpumask_t physflat_vector_allocation_domain(int cpu)
cpumask_t         199 arch/x86/kernel/genapic_flat_64.c static void physflat_send_IPI_mask(cpumask_t cpumask, int vector)
cpumask_t         206 arch/x86/kernel/genapic_flat_64.c 	cpumask_t allbutme = cpu_online_map;
cpumask_t         217 arch/x86/kernel/genapic_flat_64.c static unsigned int physflat_cpu_mask_to_apicid(cpumask_t cpumask)
cpumask_t          25 arch/x86/kernel/genx2apic_cluster.c static cpumask_t x2apic_target_cpus(void)
cpumask_t          33 arch/x86/kernel/genx2apic_cluster.c static cpumask_t x2apic_vector_allocation_domain(int cpu)
cpumask_t          35 arch/x86/kernel/genx2apic_cluster.c 	cpumask_t domain = CPU_MASK_NONE;
cpumask_t          59 arch/x86/kernel/genx2apic_cluster.c static void x2apic_send_IPI_mask(cpumask_t mask, int vector)
cpumask_t          74 arch/x86/kernel/genx2apic_cluster.c 	cpumask_t mask = cpu_online_map;
cpumask_t          92 arch/x86/kernel/genx2apic_cluster.c static unsigned int x2apic_cpu_mask_to_apicid(cpumask_t cpumask)
cpumask_t          32 arch/x86/kernel/genx2apic_phys.c static cpumask_t x2apic_target_cpus(void)
cpumask_t          37 arch/x86/kernel/genx2apic_phys.c static cpumask_t x2apic_vector_allocation_domain(int cpu)
cpumask_t          39 arch/x86/kernel/genx2apic_phys.c 	cpumask_t domain = CPU_MASK_NONE;
cpumask_t          57 arch/x86/kernel/genx2apic_phys.c static void x2apic_send_IPI_mask(cpumask_t mask, int vector)
cpumask_t          72 arch/x86/kernel/genx2apic_phys.c 	cpumask_t mask = cpu_online_map;
cpumask_t          90 arch/x86/kernel/genx2apic_phys.c static unsigned int x2apic_cpu_mask_to_apicid(cpumask_t cpumask)
cpumask_t          79 arch/x86/kernel/genx2apic_uv_x.c static cpumask_t uv_target_cpus(void)
cpumask_t          84 arch/x86/kernel/genx2apic_uv_x.c static cpumask_t uv_vector_allocation_domain(int cpu)
cpumask_t          86 arch/x86/kernel/genx2apic_uv_x.c 	cpumask_t domain = CPU_MASK_NONE;
cpumask_t         127 arch/x86/kernel/genx2apic_uv_x.c static void uv_send_IPI_mask(cpumask_t mask, int vector)
cpumask_t         138 arch/x86/kernel/genx2apic_uv_x.c 	cpumask_t mask = cpu_online_map;
cpumask_t         160 arch/x86/kernel/genx2apic_uv_x.c static unsigned int uv_cpu_mask_to_apicid(cpumask_t cpumask)
cpumask_t         340 arch/x86/kernel/io_apic_32.c static void set_ioapic_affinity_irq(unsigned int irq, cpumask_t cpumask)
cpumask_t         346 arch/x86/kernel/io_apic_32.c 	cpumask_t tmp;
cpumask_t         404 arch/x86/kernel/io_apic_32.c static cpumask_t balance_irq_affinity[NR_IRQS] = {
cpumask_t         408 arch/x86/kernel/io_apic_32.c void set_balance_irq_affinity(unsigned int irq, cpumask_t mask)
cpumask_t         413 arch/x86/kernel/io_apic_32.c static unsigned long move(int curr_cpu, cpumask_t allowed_mask,
cpumask_t         443 arch/x86/kernel/io_apic_32.c 	cpumask_t allowed_mask;
cpumask_t         487 arch/x86/kernel/io_apic_32.c 	cpumask_t allowed_mask, target_cpu_mask, tmp;
cpumask_t         692 arch/x86/kernel/io_apic_32.c 	cpumask_t tmp;
cpumask_t        2519 arch/x86/kernel/io_apic_32.c static void set_msi_irq_affinity(unsigned int irq, cpumask_t mask)
cpumask_t        2523 arch/x86/kernel/io_apic_32.c 	cpumask_t tmp;
cpumask_t        2614 arch/x86/kernel/io_apic_32.c static void set_ht_irq_affinity(unsigned int irq, cpumask_t mask)
cpumask_t        2617 arch/x86/kernel/io_apic_32.c 	cpumask_t tmp;
cpumask_t        2651 arch/x86/kernel/io_apic_32.c 		cpumask_t tmp;
cpumask_t          61 arch/x86/kernel/io_apic_64.c 	cpumask_t domain;
cpumask_t          62 arch/x86/kernel/io_apic_64.c 	cpumask_t old_domain;
cpumask_t          88 arch/x86/kernel/io_apic_64.c static int assign_irq_vector(int irq, cpumask_t mask);
cpumask_t         327 arch/x86/kernel/io_apic_64.c static void set_ioapic_affinity_irq(unsigned int irq, cpumask_t mask)
cpumask_t         332 arch/x86/kernel/io_apic_64.c 	cpumask_t tmp;
cpumask_t         786 arch/x86/kernel/io_apic_64.c static int __assign_irq_vector(int irq, cpumask_t mask)
cpumask_t         815 arch/x86/kernel/io_apic_64.c 		cpumask_t tmp;
cpumask_t         822 arch/x86/kernel/io_apic_64.c 		cpumask_t domain, new_mask;
cpumask_t         861 arch/x86/kernel/io_apic_64.c static int assign_irq_vector(int irq, cpumask_t mask)
cpumask_t         875 arch/x86/kernel/io_apic_64.c 	cpumask_t mask;
cpumask_t        1014 arch/x86/kernel/io_apic_64.c 	cpumask_t mask;
cpumask_t        1581 arch/x86/kernel/io_apic_64.c static void migrate_ioapic_irq(int irq, cpumask_t mask)
cpumask_t        1585 arch/x86/kernel/io_apic_64.c 	cpumask_t tmp, cleanup_mask;
cpumask_t        1684 arch/x86/kernel/io_apic_64.c static void set_ir_ioapic_affinity_irq(unsigned int irq, cpumask_t mask)
cpumask_t        1742 arch/x86/kernel/io_apic_64.c 		cpumask_t cleanup_mask;
cpumask_t        2322 arch/x86/kernel/io_apic_64.c 	cpumask_t tmp;
cpumask_t        2384 arch/x86/kernel/io_apic_64.c static void set_msi_irq_affinity(unsigned int irq, cpumask_t mask)
cpumask_t        2389 arch/x86/kernel/io_apic_64.c 	cpumask_t tmp;
cpumask_t        2417 arch/x86/kernel/io_apic_64.c static void ir_set_msi_irq_affinity(unsigned int irq, cpumask_t mask)
cpumask_t        2421 arch/x86/kernel/io_apic_64.c 	cpumask_t tmp, cleanup_mask;
cpumask_t        2637 arch/x86/kernel/io_apic_64.c static void dmar_msi_set_affinity(unsigned int irq, cpumask_t mask)
cpumask_t        2642 arch/x86/kernel/io_apic_64.c 	cpumask_t tmp;
cpumask_t        2714 arch/x86/kernel/io_apic_64.c static void set_ht_irq_affinity(unsigned int irq, cpumask_t mask)
cpumask_t        2718 arch/x86/kernel/io_apic_64.c 	cpumask_t tmp;
cpumask_t        2750 arch/x86/kernel/io_apic_64.c 	cpumask_t tmp;
cpumask_t         119 arch/x86/kernel/ipi.c void send_IPI_mask_bitmask(cpumask_t cpumask, int vector)
cpumask_t         130 arch/x86/kernel/ipi.c void send_IPI_mask_sequence(cpumask_t mask, int vector)
cpumask_t         394 arch/x86/kernel/irq_32.c void fixup_irqs(cpumask_t map)
cpumask_t         400 arch/x86/kernel/irq_32.c 		cpumask_t mask;
cpumask_t         222 arch/x86/kernel/irq_64.c void fixup_irqs(cpumask_t map)
cpumask_t         228 arch/x86/kernel/irq_64.c 		cpumask_t mask;
cpumask_t         113 arch/x86/kernel/microcode_core.c 	cpumask_t old;
cpumask_t         221 arch/x86/kernel/microcode_core.c 		cpumask_t old = current->cpus_allowed;
cpumask_t         345 arch/x86/kernel/microcode_core.c 	cpumask_t old = current->cpus_allowed;
cpumask_t          44 arch/x86/kernel/nmi.c static cpumask_t backtrace_mask = CPU_MASK_NONE;
cpumask_t         249 arch/x86/kernel/process.c static cpumask_t c1e_mask = CPU_MASK_NONE;
cpumask_t          44 arch/x86/kernel/setup_percpu.c cpumask_t *node_to_cpumask_map;
cpumask_t         204 arch/x86/kernel/setup_percpu.c 	cpumask_t *map;
cpumask_t         214 arch/x86/kernel/setup_percpu.c 	map = alloc_bootmem_low(nr_node_ids * sizeof(cpumask_t));
cpumask_t         265 arch/x86/kernel/setup_percpu.c 	cpumask_t *mask;
cpumask_t         327 arch/x86/kernel/setup_percpu.c static const cpumask_t cpu_mask_none;
cpumask_t         332 arch/x86/kernel/setup_percpu.c const cpumask_t *_node_to_cpumask_ptr(int node)
cpumask_t         339 arch/x86/kernel/setup_percpu.c 		return (const cpumask_t *)&cpu_online_map;
cpumask_t         359 arch/x86/kernel/setup_percpu.c cpumask_t node_to_cpumask(int node)
cpumask_t         129 arch/x86/kernel/smp.c void native_send_call_func_ipi(cpumask_t mask)
cpumask_t         131 arch/x86/kernel/smp.c 	cpumask_t allbutself;
cpumask_t         105 arch/x86/kernel/smpboot.c cpumask_t cpu_online_map __read_mostly;
cpumask_t         108 arch/x86/kernel/smpboot.c cpumask_t cpu_callin_map;
cpumask_t         109 arch/x86/kernel/smpboot.c cpumask_t cpu_callout_map;
cpumask_t         110 arch/x86/kernel/smpboot.c cpumask_t cpu_possible_map;
cpumask_t         114 arch/x86/kernel/smpboot.c DEFINE_PER_CPU(cpumask_t, cpu_sibling_map);
cpumask_t         118 arch/x86/kernel/smpboot.c DEFINE_PER_CPU(cpumask_t, cpu_core_map);
cpumask_t         129 arch/x86/kernel/smpboot.c static cpumask_t cpu_sibling_setup_map;
cpumask_t         137 arch/x86/kernel/smpboot.c cpumask_t node_to_cpumask_map[MAX_NUMNODES] __read_mostly =
cpumask_t         506 arch/x86/kernel/smpboot.c cpumask_t cpu_coregroup_map(int cpu)
cpumask_t          23 arch/x86/kernel/tlb_32.c static cpumask_t flush_cpumask;
cpumask_t         125 arch/x86/kernel/tlb_32.c void native_flush_tlb_others(const cpumask_t *cpumaskp, struct mm_struct *mm,
cpumask_t         128 arch/x86/kernel/tlb_32.c 	cpumask_t cpumask = *cpumaskp;
cpumask_t         175 arch/x86/kernel/tlb_32.c 	cpumask_t cpu_mask;
cpumask_t         189 arch/x86/kernel/tlb_32.c 	cpumask_t cpu_mask;
cpumask_t         210 arch/x86/kernel/tlb_32.c 	cpumask_t cpu_mask;
cpumask_t          46 arch/x86/kernel/tlb_64.c 		cpumask_t flush_cpumask;
cpumask_t         160 arch/x86/kernel/tlb_64.c void native_flush_tlb_others(const cpumask_t *cpumaskp, struct mm_struct *mm,
cpumask_t         165 arch/x86/kernel/tlb_64.c 	cpumask_t cpumask = *cpumaskp;
cpumask_t         213 arch/x86/kernel/tlb_64.c 	cpumask_t cpu_mask;
cpumask_t         227 arch/x86/kernel/tlb_64.c 	cpumask_t cpu_mask;
cpumask_t         248 arch/x86/kernel/tlb_64.c 	cpumask_t cpu_mask;
cpumask_t         219 arch/x86/kernel/tlb_uv.c 			   cpumask_t *cpumaskp)
cpumask_t         299 arch/x86/kernel/tlb_uv.c int uv_flush_tlb_others(cpumask_t *cpumaskp, struct mm_struct *mm,
cpumask_t          67 arch/x86/mach-voyager/voyager_smp.c cpumask_t cpu_online_map = CPU_MASK_NONE;
cpumask_t          72 arch/x86/mach-voyager/voyager_smp.c cpumask_t phys_cpu_present_map = CPU_MASK_NONE;
cpumask_t          88 arch/x86/mach-voyager/voyager_smp.c static void set_vic_irq_affinity(unsigned int irq, cpumask_t mask);
cpumask_t         214 arch/x86/mach-voyager/voyager_smp.c static cpumask_t smp_commenced_mask = CPU_MASK_NONE;
cpumask_t         217 arch/x86/mach-voyager/voyager_smp.c cpumask_t cpu_callin_map = CPU_MASK_NONE;
cpumask_t         218 arch/x86/mach-voyager/voyager_smp.c cpumask_t cpu_callout_map = CPU_MASK_NONE;
cpumask_t         219 arch/x86/mach-voyager/voyager_smp.c cpumask_t cpu_possible_map = CPU_MASK_NONE;
cpumask_t        1599 arch/x86/mach-voyager/voyager_smp.c void set_vic_irq_affinity(unsigned int irq, cpumask_t mask)
cpumask_t         407 arch/x86/mm/mmio-mod.c static cpumask_t downed_cpus;
cpumask_t         636 arch/x86/xen/enlighten.c static void xen_flush_tlb_others(const cpumask_t *cpus, struct mm_struct *mm,
cpumask_t         641 arch/x86/xen/enlighten.c 		cpumask_t mask;
cpumask_t         643 arch/x86/xen/enlighten.c 	cpumask_t cpumask = *cpus;
cpumask_t        1061 arch/x86/xen/mmu.c 	cpumask_t mask;
cpumask_t          36 arch/x86/xen/smp.c cpumask_t xen_cpu_initialized_map;
cpumask_t         411 arch/x86/xen/smp.c static void xen_send_IPI_mask(cpumask_t mask, enum ipi_vector vector)
cpumask_t         421 arch/x86/xen/smp.c static void xen_smp_send_call_function_ipi(cpumask_t mask)
cpumask_t          61 arch/x86/xen/xen-ops.h extern cpumask_t xen_cpu_initialized_map;
cpumask_t         101 block/blk.h    	cpumask_t mask = cpu_coregroup_map(cpu);
cpumask_t         130 include/acpi/processor.h 	cpumask_t shared_cpu_map;
cpumask_t         175 include/acpi/processor.h 	cpumask_t shared_cpu_map;
cpumask_t           6 include/asm-cris/smp.h extern cpumask_t phys_cpu_present_map;
cpumask_t           7 include/asm-cris/smp.h extern cpumask_t cpu_possible_map;
cpumask_t          63 include/asm-generic/topology.h 		cpumask_t _##v = node_to_cpumask(node);			\
cpumask_t          64 include/asm-generic/topology.h 		const cpumask_t *v = &_##v
cpumask_t          65 include/asm-m32r/smp.h extern cpumask_t cpu_callout_map;
cpumask_t          66 include/asm-m32r/smp.h extern cpumask_t cpu_possible_map;
cpumask_t          67 include/asm-m32r/smp.h extern cpumask_t cpu_present_map;
cpumask_t          90 include/asm-m32r/smp.h extern unsigned long send_IPI_mask_phys(cpumask_t, int, int);
cpumask_t          93 include/asm-m32r/smp.h extern void arch_send_call_function_ipi(cpumask_t mask);
cpumask_t          52 include/asm-parisc/irq.h extern int cpu_check_affinity(unsigned int irq, cpumask_t *dest);
cpumask_t          19 include/asm-parisc/smp.h extern cpumask_t cpu_online_map;
cpumask_t          34 include/asm-parisc/smp.h extern void arch_send_call_function_ipi(cpumask_t mask);
cpumask_t          13 include/asm-x86/bigsmp/apic.h static inline cpumask_t target_cpus(void)
cpumask_t         129 include/asm-x86/bigsmp/apic.h static inline unsigned int cpu_mask_to_apicid(cpumask_t cpumask)
cpumask_t           4 include/asm-x86/bigsmp/ipi.h void send_IPI_mask_sequence(cpumask_t mask, int vector);
cpumask_t           6 include/asm-x86/bigsmp/ipi.h static inline void send_IPI_mask(cpumask_t mask, int vector)
cpumask_t          13 include/asm-x86/bigsmp/ipi.h 	cpumask_t mask = cpu_online_map;
cpumask_t          12 include/asm-x86/es7000/apic.h static inline cpumask_t target_cpus(void)
cpumask_t         148 include/asm-x86/es7000/apic.h static inline unsigned int cpu_mask_to_apicid(cpumask_t cpumask)
cpumask_t           4 include/asm-x86/es7000/ipi.h void send_IPI_mask_sequence(cpumask_t mask, int vector);
cpumask_t           6 include/asm-x86/es7000/ipi.h static inline void send_IPI_mask(cpumask_t mask, int vector)
cpumask_t          13 include/asm-x86/es7000/ipi.h 	cpumask_t mask = cpu_online_map;
cpumask_t          26 include/asm-x86/genapic_32.h 	cpumask_t (*target_cpus)(void);
cpumask_t          59 include/asm-x86/genapic_32.h 	unsigned int (*cpu_mask_to_apicid)(cpumask_t cpumask);
cpumask_t          63 include/asm-x86/genapic_32.h 	void (*send_IPI_mask)(cpumask_t mask, int vector);
cpumask_t          21 include/asm-x86/genapic_64.h 	cpumask_t (*target_cpus)(void);
cpumask_t          22 include/asm-x86/genapic_64.h 	cpumask_t (*vector_allocation_domain)(int cpu);
cpumask_t          25 include/asm-x86/genapic_64.h 	void (*send_IPI_mask)(cpumask_t mask, int vector);
cpumask_t          30 include/asm-x86/genapic_64.h 	unsigned int (*cpu_mask_to_apicid)(cpumask_t cpumask);
cpumask_t         120 include/asm-x86/ipi.h static inline void send_IPI_mask_sequence(cpumask_t mask, int vector)
cpumask_t          40 include/asm-x86/irq.h extern void fixup_irqs(cpumask_t map);
cpumask_t          11 include/asm-x86/mach-default/mach_apic.h static inline cpumask_t target_cpus(void)
cpumask_t          62 include/asm-x86/mach-default/mach_apic.h static inline unsigned int cpu_mask_to_apicid(cpumask_t cpumask)
cpumask_t           7 include/asm-x86/mach-default/mach_ipi.h void send_IPI_mask_bitmask(cpumask_t mask, int vector);
cpumask_t          16 include/asm-x86/mach-default/mach_ipi.h static inline void send_IPI_mask(cpumask_t mask, int vector)
cpumask_t          25 include/asm-x86/mach-default/mach_ipi.h 		cpumask_t mask = cpu_online_map;
cpumask_t          10 include/asm-x86/numaq/apic.h static inline cpumask_t target_cpus(void)
cpumask_t         127 include/asm-x86/numaq/apic.h static inline unsigned int cpu_mask_to_apicid(cpumask_t cpumask)
cpumask_t           4 include/asm-x86/numaq/ipi.h void send_IPI_mask_sequence(cpumask_t, int vector);
cpumask_t           6 include/asm-x86/numaq/ipi.h static inline void send_IPI_mask(cpumask_t mask, int vector)
cpumask_t          13 include/asm-x86/numaq/ipi.h 	cpumask_t mask = cpu_online_map;
cpumask_t         247 include/asm-x86/paravirt.h 	void (*flush_tlb_others)(const cpumask_t *cpus, struct mm_struct *mm,
cpumask_t         987 include/asm-x86/paravirt.h static inline void flush_tlb_others(cpumask_t cpumask, struct mm_struct *mm,
cpumask_t         108 include/asm-x86/pci.h static inline cpumask_t __pcibus_to_cpumask(struct pci_bus *bus)
cpumask_t          96 include/asm-x86/processor.h 	cpumask_t		llc_shared_map;
cpumask_t          21 include/asm-x86/smp.h extern cpumask_t cpu_callout_map;
cpumask_t          22 include/asm-x86/smp.h extern cpumask_t cpu_initialized;
cpumask_t          23 include/asm-x86/smp.h extern cpumask_t cpu_callin_map;
cpumask_t          32 include/asm-x86/smp.h extern cpumask_t cpu_initialized;
cpumask_t          34 include/asm-x86/smp.h DECLARE_PER_CPU(cpumask_t, cpu_sibling_map);
cpumask_t          35 include/asm-x86/smp.h DECLARE_PER_CPU(cpumask_t, cpu_core_map);
cpumask_t          63 include/asm-x86/smp.h 	void (*send_call_func_ipi)(cpumask_t mask);
cpumask_t         126 include/asm-x86/smp.h static inline void arch_send_call_function_ipi(cpumask_t mask)
cpumask_t         141 include/asm-x86/smp.h void native_send_call_func_ipi(cpumask_t mask);
cpumask_t          17 include/asm-x86/summit/apic.h static inline cpumask_t target_cpus(void)
cpumask_t         141 include/asm-x86/summit/apic.h static inline unsigned int cpu_mask_to_apicid(cpumask_t cpumask)
cpumask_t           4 include/asm-x86/summit/ipi.h void send_IPI_mask_sequence(cpumask_t mask, int vector);
cpumask_t           6 include/asm-x86/summit/ipi.h static inline void send_IPI_mask(cpumask_t mask, int vector)
cpumask_t          13 include/asm-x86/summit/ipi.h 	cpumask_t mask = cpu_online_map;
cpumask_t         116 include/asm-x86/tlbflush.h static inline void native_flush_tlb_others(const cpumask_t *cpumask,
cpumask_t         145 include/asm-x86/tlbflush.h void native_flush_tlb_others(const cpumask_t *cpumask, struct mm_struct *mm,
cpumask_t          48 include/asm-x86/topology.h extern cpumask_t node_to_cpumask_map[];
cpumask_t          66 include/asm-x86/topology.h static inline cpumask_t node_to_cpumask(int node)
cpumask_t          74 include/asm-x86/topology.h extern cpumask_t *node_to_cpumask_map;
cpumask_t          85 include/asm-x86/topology.h extern const cpumask_t *_node_to_cpumask_ptr(int node);
cpumask_t          86 include/asm-x86/topology.h extern cpumask_t node_to_cpumask(int node);
cpumask_t         106 include/asm-x86/topology.h static inline const cpumask_t *_node_to_cpumask_ptr(int node)
cpumask_t         112 include/asm-x86/topology.h static inline cpumask_t node_to_cpumask(int node)
cpumask_t         121 include/asm-x86/topology.h 		const cpumask_t *v = _node_to_cpumask_ptr(node)
cpumask_t         189 include/asm-x86/topology.h static inline const cpumask_t *_node_to_cpumask_ptr(int node)
cpumask_t         193 include/asm-x86/topology.h static inline cpumask_t node_to_cpumask(int node)
cpumask_t         204 include/asm-x86/topology.h 		const cpumask_t *v = _node_to_cpumask_ptr(node)
cpumask_t         221 include/asm-x86/topology.h extern cpumask_t cpu_coregroup_map(int cpu);
cpumask_t         328 include/asm-x86/uv/uv_bau.h extern int uv_flush_tlb_others(cpumask_t *, struct mm_struct *, unsigned long);
cpumask_t          85 include/linux/clockchips.h 	cpumask_t		cpumask;
cpumask_t          91 include/linux/clockchips.h 	void			(*broadcast)(cpumask_t mask);
cpumask_t          83 include/linux/cpufreq.h 	cpumask_t		cpus;	/* CPUs requiring sw coordination */
cpumask_t          84 include/linux/cpufreq.h 	cpumask_t		related_cpus; /* CPUs with any coordination */
cpumask_t         142 include/linux/cpumask.h extern cpumask_t _unused_cpumask_arg_;
cpumask_t         145 include/linux/cpumask.h static inline void __cpu_set(int cpu, volatile cpumask_t *dstp)
cpumask_t         151 include/linux/cpumask.h static inline void __cpu_clear(int cpu, volatile cpumask_t *dstp)
cpumask_t         157 include/linux/cpumask.h static inline void __cpus_setall(cpumask_t *dstp, int nbits)
cpumask_t         163 include/linux/cpumask.h static inline void __cpus_clear(cpumask_t *dstp, int nbits)
cpumask_t         172 include/linux/cpumask.h static inline int __cpu_test_and_set(int cpu, cpumask_t *addr)
cpumask_t         178 include/linux/cpumask.h static inline void __cpus_and(cpumask_t *dstp, const cpumask_t *src1p,
cpumask_t         179 include/linux/cpumask.h 					const cpumask_t *src2p, int nbits)
cpumask_t         185 include/linux/cpumask.h static inline void __cpus_or(cpumask_t *dstp, const cpumask_t *src1p,
cpumask_t         186 include/linux/cpumask.h 					const cpumask_t *src2p, int nbits)
cpumask_t         192 include/linux/cpumask.h static inline void __cpus_xor(cpumask_t *dstp, const cpumask_t *src1p,
cpumask_t         193 include/linux/cpumask.h 					const cpumask_t *src2p, int nbits)
cpumask_t         200 include/linux/cpumask.h static inline void __cpus_andnot(cpumask_t *dstp, const cpumask_t *src1p,
cpumask_t         201 include/linux/cpumask.h 					const cpumask_t *src2p, int nbits)
cpumask_t         207 include/linux/cpumask.h static inline void __cpus_complement(cpumask_t *dstp,
cpumask_t         208 include/linux/cpumask.h 					const cpumask_t *srcp, int nbits)
cpumask_t         214 include/linux/cpumask.h static inline int __cpus_equal(const cpumask_t *src1p,
cpumask_t         215 include/linux/cpumask.h 					const cpumask_t *src2p, int nbits)
cpumask_t         221 include/linux/cpumask.h static inline int __cpus_intersects(const cpumask_t *src1p,
cpumask_t         222 include/linux/cpumask.h 					const cpumask_t *src2p, int nbits)
cpumask_t         228 include/linux/cpumask.h static inline int __cpus_subset(const cpumask_t *src1p,
cpumask_t         229 include/linux/cpumask.h 					const cpumask_t *src2p, int nbits)
cpumask_t         235 include/linux/cpumask.h static inline int __cpus_empty(const cpumask_t *srcp, int nbits)
cpumask_t         241 include/linux/cpumask.h static inline int __cpus_full(const cpumask_t *srcp, int nbits)
cpumask_t         247 include/linux/cpumask.h static inline int __cpus_weight(const cpumask_t *srcp, int nbits)
cpumask_t         254 include/linux/cpumask.h static inline void __cpus_shift_right(cpumask_t *dstp,
cpumask_t         255 include/linux/cpumask.h 					const cpumask_t *srcp, int n, int nbits)
cpumask_t         262 include/linux/cpumask.h static inline void __cpus_shift_left(cpumask_t *dstp,
cpumask_t         263 include/linux/cpumask.h 					const cpumask_t *srcp, int n, int nbits)
cpumask_t         278 include/linux/cpumask.h static inline const cpumask_t *get_cpu_mask(unsigned int cpu)
cpumask_t         282 include/linux/cpumask.h 	return (const cpumask_t *)p;
cpumask_t         298 include/linux/cpumask.h (cpumask_t) { {								\
cpumask_t         307 include/linux/cpumask.h (cpumask_t) { {								\
cpumask_t         313 include/linux/cpumask.h extern cpumask_t cpu_mask_all;
cpumask_t         319 include/linux/cpumask.h (cpumask_t) { {								\
cpumask_t         324 include/linux/cpumask.h (cpumask_t) { {								\
cpumask_t         337 include/linux/cpumask.h #define	CPUMASK_PTR(v, m) 	cpumask_t *v = &(m->v)
cpumask_t         342 include/linux/cpumask.h 					const cpumask_t *srcp, int nbits)
cpumask_t         350 include/linux/cpumask.h 					cpumask_t *dstp, int nbits)
cpumask_t         358 include/linux/cpumask.h 					const cpumask_t *srcp, int nbits)
cpumask_t         364 include/linux/cpumask.h static inline int __cpulist_parse(const char *buf, cpumask_t *dstp, int nbits)
cpumask_t         372 include/linux/cpumask.h 		const cpumask_t *oldp, const cpumask_t *newp, int nbits)
cpumask_t         379 include/linux/cpumask.h static inline void __cpus_remap(cpumask_t *dstp, const cpumask_t *srcp,
cpumask_t         380 include/linux/cpumask.h 		const cpumask_t *oldp, const cpumask_t *newp, int nbits)
cpumask_t         387 include/linux/cpumask.h static inline void __cpus_onto(cpumask_t *dstp, const cpumask_t *origp,
cpumask_t         388 include/linux/cpumask.h 		const cpumask_t *relmapp, int nbits)
cpumask_t         395 include/linux/cpumask.h static inline void __cpus_fold(cpumask_t *dstp, const cpumask_t *origp,
cpumask_t         413 include/linux/cpumask.h int __first_cpu(const cpumask_t *srcp);
cpumask_t         414 include/linux/cpumask.h int __next_cpu(int n, const cpumask_t *srcp);
cpumask_t         415 include/linux/cpumask.h int __any_online_cpu(const cpumask_t *mask);
cpumask_t         434 include/linux/cpumask.h int __next_cpu_nr(int n, const cpumask_t *srcp);
cpumask_t         501 include/linux/cpumask.h extern cpumask_t cpu_possible_map;
cpumask_t         502 include/linux/cpumask.h extern cpumask_t cpu_online_map;
cpumask_t         503 include/linux/cpumask.h extern cpumask_t cpu_present_map;
cpumask_t         504 include/linux/cpumask.h extern cpumask_t cpu_active_map;
cpumask_t          23 include/linux/cpuset.h extern void cpuset_cpus_allowed(struct task_struct *p, cpumask_t *mask);
cpumask_t          24 include/linux/cpuset.h extern void cpuset_cpus_allowed_locked(struct task_struct *p, cpumask_t *mask);
cpumask_t          89 include/linux/cpuset.h static inline void cpuset_cpus_allowed(struct task_struct *p, cpumask_t *mask)
cpumask_t          94 include/linux/cpuset.h 								cpumask_t *mask)
cpumask_t          63 include/linux/interrupt.h 	cpumask_t mask;
cpumask_t         107 include/linux/interrupt.h extern cpumask_t irq_default_affinity;
cpumask_t         109 include/linux/interrupt.h extern int irq_set_affinity(unsigned int irq, cpumask_t cpumask);
cpumask_t         115 include/linux/interrupt.h static inline int irq_set_affinity(unsigned int irq, cpumask_t cpumask)
cpumask_t         114 include/linux/irq.h 	void		(*set_affinity)(unsigned int irq, cpumask_t dest);
cpumask_t         170 include/linux/irq.h 	cpumask_t		affinity;
cpumask_t         174 include/linux/irq.h 	cpumask_t		pending_mask;
cpumask_t         209 include/linux/irq.h void set_pending_irq(unsigned int irq, cpumask_t mask);
cpumask_t         227 include/linux/irq.h static inline void set_pending_irq(unsigned int irq, cpumask_t mask)
cpumask_t         241 include/linux/irq.h extern void set_balance_irq_affinity(unsigned int irq, cpumask_t mask);
cpumask_t         243 include/linux/irq.h static inline void set_balance_irq_affinity(unsigned int irq, cpumask_t mask)
cpumask_t         217 include/linux/mm_types.h 	cpumask_t cpu_vm_mask;
cpumask_t          84 include/linux/percpu.h extern void *__percpu_alloc_mask(size_t size, gfp_t gfp, cpumask_t *mask);
cpumask_t          91 include/linux/percpu.h static __always_inline void *__percpu_alloc_mask(size_t size, gfp_t gfp, cpumask_t *mask)
cpumask_t          62 include/linux/rcuclassic.h 	cpumask_t	cpumask; /* CPUs that need to switch in order    */
cpumask_t         251 include/linux/sched.h extern cpumask_t nohz_cpu_mask;
cpumask_t         737 include/linux/sched.h 	cpumask_t cpumask;
cpumask_t         775 include/linux/sched.h 	cpumask_t span;			/* span of all CPUs in this domain */
cpumask_t         832 include/linux/sched.h extern void partition_sched_domains(int ndoms_new, cpumask_t *doms_new,
cpumask_t         841 include/linux/sched.h partition_sched_domains(int ndoms_new, cpumask_t *doms_new,
cpumask_t         926 include/linux/sched.h 				 const cpumask_t *newmask);
cpumask_t        1071 include/linux/sched.h 	cpumask_t cpus_allowed;
cpumask_t        1545 include/linux/sched.h 				const cpumask_t *new_mask);
cpumask_t        1548 include/linux/sched.h 				       const cpumask_t *new_mask)
cpumask_t        1555 include/linux/sched.h static inline int set_cpus_allowed(struct task_struct *p, cpumask_t new_mask)
cpumask_t        2137 include/linux/sched.h extern long sched_setaffinity(pid_t pid, const cpumask_t *new_mask);
cpumask_t        2138 include/linux/sched.h extern long sched_getaffinity(pid_t pid, cpumask_t *mask);
cpumask_t          53 include/linux/seq_file.h static inline int seq_cpumask(struct seq_file *m, cpumask_t *mask)
cpumask_t          65 include/linux/smp.h int smp_call_function_mask(cpumask_t mask, void(*func)(void *info), void *info,
cpumask_t          26 include/linux/stop_machine.h int stop_machine(int (*fn)(void *), void *data, const cpumask_t *cpus);
cpumask_t          37 include/linux/stop_machine.h int __stop_machine(int (*fn)(void *), void *data, const cpumask_t *cpus);
cpumask_t          41 include/linux/stop_machine.h 			       const cpumask_t *cpus)
cpumask_t          87 include/linux/tick.h extern cpumask_t *tick_get_broadcast_mask(void);
cpumask_t          90 include/linux/tick.h extern cpumask_t *tick_get_broadcast_oneshot_mask(void);
cpumask_t         368 init/main.c    cpumask_t cpu_mask_all __read_mostly = CPU_MASK_ALL;
cpumask_t         424 kernel/compat.c 				    unsigned len, cpumask_t *new_mask)
cpumask_t         428 kernel/compat.c 	if (len < sizeof(cpumask_t))
cpumask_t         429 kernel/compat.c 		memset(new_mask, 0, sizeof(cpumask_t));
cpumask_t         430 kernel/compat.c 	else if (len > sizeof(cpumask_t))
cpumask_t         431 kernel/compat.c 		len = sizeof(cpumask_t);
cpumask_t         441 kernel/compat.c 	cpumask_t new_mask;
cpumask_t         455 kernel/compat.c 	cpumask_t mask;
cpumask_t         457 kernel/compat.c 	unsigned int min_length = sizeof(cpumask_t);
cpumask_t          24 kernel/cpu.c   cpumask_t cpu_present_map __read_mostly;
cpumask_t          32 kernel/cpu.c   cpumask_t cpu_online_map __read_mostly = CPU_MASK_ALL;
cpumask_t          35 kernel/cpu.c   cpumask_t cpu_possible_map __read_mostly = CPU_MASK_ALL;
cpumask_t          67 kernel/cpu.c   cpumask_t cpu_active_map;
cpumask_t         220 kernel/cpu.c   	cpumask_t old_allowed, tmp;
cpumask_t         395 kernel/cpu.c   static cpumask_t frozen_cpus;
cpumask_t          86 kernel/cpuset.c 	cpumask_t cpus_allowed;		/* CPUs allowed to tasks in cpuset */
cpumask_t         282 kernel/cpuset.c static void guarantee_online_cpus(const struct cpuset *cs, cpumask_t *pmask)
cpumask_t         577 kernel/cpuset.c static int generate_sched_domains(cpumask_t **domains,
cpumask_t         585 kernel/cpuset.c 	cpumask_t *doms;	/* resulting partition; i.e. sched domains */
cpumask_t         597 kernel/cpuset.c 		doms = kmalloc(sizeof(cpumask_t), GFP_KERNEL);
cpumask_t         676 kernel/cpuset.c 	doms = kmalloc(ndoms * sizeof(cpumask_t), GFP_KERNEL);
cpumask_t         690 kernel/cpuset.c 		cpumask_t *dp;
cpumask_t         753 kernel/cpuset.c 	cpumask_t *doms;
cpumask_t        1308 kernel/cpuset.c 		cpumask_t mask;
cpumask_t        1324 kernel/cpuset.c 	cpumask_t cpus;
cpumask_t        1475 kernel/cpuset.c 	cpumask_t mask;
cpumask_t        1983 kernel/cpuset.c 	cpumask_t *doms;
cpumask_t        2049 kernel/cpuset.c void cpuset_cpus_allowed(struct task_struct *tsk, cpumask_t *pmask)
cpumask_t        2060 kernel/cpuset.c void cpuset_cpus_allowed_locked(struct task_struct *tsk, cpumask_t *pmask)
cpumask_t          20 kernel/irq/manage.c cpumask_t irq_default_affinity = CPU_MASK_ALL;
cpumask_t          82 kernel/irq/manage.c int irq_set_affinity(unsigned int irq, cpumask_t cpumask)
cpumask_t         113 kernel/irq/manage.c 	cpumask_t mask;
cpumask_t           4 kernel/irq/migration.c void set_pending_irq(unsigned int irq, cpumask_t mask)
cpumask_t          18 kernel/irq/migration.c 	cpumask_t tmp;
cpumask_t          23 kernel/irq/proc.c 	cpumask_t *mask = &desc->affinity;
cpumask_t          43 kernel/irq/proc.c 	cpumask_t new_value;
cpumask_t          95 kernel/irq/proc.c 	cpumask_t new_value;
cpumask_t          46 kernel/profile.c static cpumask_t prof_cpu_mask = CPU_MASK_ALL;
cpumask_t         424 kernel/profile.c 	int len = cpumask_scnprintf(page, count, *(cpumask_t *)data);
cpumask_t         434 kernel/profile.c 	cpumask_t *mask = (cpumask_t *)data;
cpumask_t         436 kernel/profile.c 	cpumask_t new_value;
cpumask_t          88 kernel/rcuclassic.c 	cpumask_t cpumask;
cpumask_t         167 kernel/rcupreempt.c static cpumask_t rcu_cpu_online_map __read_mostly = CPU_MASK_NONE;
cpumask_t         846 kernel/rcutorture.c 	cpumask_t tmp_mask;
cpumask_t         484 kernel/sched.c 	cpumask_t span;
cpumask_t         485 kernel/sched.c 	cpumask_t online;
cpumask_t         491 kernel/sched.c 	cpumask_t rto_mask;
cpumask_t        2104 kernel/sched.c 		cpumask_t *tmp)
cpumask_t        2155 kernel/sched.c 		cpumask_t span, tmpmask;
cpumask_t        3090 kernel/sched.c 		   int *sd_idle, const cpumask_t *cpus, int *balance)
cpumask_t        3411 kernel/sched.c 		   unsigned long imbalance, const cpumask_t *cpus)
cpumask_t        3450 kernel/sched.c 			int *balance, cpumask_t *cpus)
cpumask_t        3611 kernel/sched.c 			cpumask_t *cpus)
cpumask_t        3700 kernel/sched.c 	cpumask_t tmpmask;
cpumask_t        3782 kernel/sched.c 	cpumask_t cpu_mask;
cpumask_t        3871 kernel/sched.c 	cpumask_t tmp;
cpumask_t        3954 kernel/sched.c 		cpumask_t cpus = nohz.cpu_mask;
cpumask_t        5388 kernel/sched.c long sched_setaffinity(pid_t pid, const cpumask_t *in_mask)
cpumask_t        5390 kernel/sched.c 	cpumask_t cpus_allowed;
cpumask_t        5391 kernel/sched.c 	cpumask_t new_mask = *in_mask;
cpumask_t        5446 kernel/sched.c 			     cpumask_t *new_mask)
cpumask_t        5448 kernel/sched.c 	if (len < sizeof(cpumask_t)) {
cpumask_t        5449 kernel/sched.c 		memset(new_mask, 0, sizeof(cpumask_t));
cpumask_t        5450 kernel/sched.c 	} else if (len > sizeof(cpumask_t)) {
cpumask_t        5451 kernel/sched.c 		len = sizeof(cpumask_t);
cpumask_t        5465 kernel/sched.c 	cpumask_t new_mask;
cpumask_t        5475 kernel/sched.c long sched_getaffinity(pid_t pid, cpumask_t *mask)
cpumask_t        5511 kernel/sched.c 	cpumask_t mask;
cpumask_t        5513 kernel/sched.c 	if (len < sizeof(cpumask_t))
cpumask_t        5520 kernel/sched.c 	if (copy_to_user(user_mask_ptr, &mask, sizeof(cpumask_t)))
cpumask_t        5523 kernel/sched.c 	return sizeof(cpumask_t);
cpumask_t        5892 kernel/sched.c cpumask_t nohz_cpu_mask = CPU_MASK_NONE;
cpumask_t        5947 kernel/sched.c int set_cpus_allowed_ptr(struct task_struct *p, const cpumask_t *new_mask)
cpumask_t        6118 kernel/sched.c 	cpumask_t mask;
cpumask_t        6134 kernel/sched.c 			cpumask_t cpus_allowed;
cpumask_t        6646 kernel/sched.c 				  cpumask_t *groupmask)
cpumask_t        6723 kernel/sched.c 	cpumask_t *groupmask;
cpumask_t        6733 kernel/sched.c 	groupmask = kmalloc(sizeof(cpumask_t), GFP_KERNEL);
cpumask_t        6900 kernel/sched.c static cpumask_t cpu_isolated_map = CPU_MASK_NONE;
cpumask_t        6929 kernel/sched.c init_sched_build_groups(const cpumask_t *span, const cpumask_t *cpu_map,
cpumask_t        6930 kernel/sched.c 			int (*group_fn)(int cpu, const cpumask_t *cpu_map,
cpumask_t        6932 kernel/sched.c 					cpumask_t *tmpmask),
cpumask_t        6933 kernel/sched.c 			cpumask_t *covered, cpumask_t *tmpmask)
cpumask_t        7020 kernel/sched.c static void sched_domain_node_span(int node, cpumask_t *span)
cpumask_t        7051 kernel/sched.c cpu_to_cpu_group(int cpu, const cpumask_t *cpu_map, struct sched_group **sg,
cpumask_t        7052 kernel/sched.c 		 cpumask_t *unused)
cpumask_t        7070 kernel/sched.c cpu_to_core_group(int cpu, const cpumask_t *cpu_map, struct sched_group **sg,
cpumask_t        7071 kernel/sched.c 		  cpumask_t *mask)
cpumask_t        7084 kernel/sched.c cpu_to_core_group(int cpu, const cpumask_t *cpu_map, struct sched_group **sg,
cpumask_t        7085 kernel/sched.c 		  cpumask_t *unused)
cpumask_t        7097 kernel/sched.c cpu_to_phys_group(int cpu, const cpumask_t *cpu_map, struct sched_group **sg,
cpumask_t        7098 kernel/sched.c 		  cpumask_t *mask)
cpumask_t        7129 kernel/sched.c static int cpu_to_allnodes_group(int cpu, const cpumask_t *cpu_map,
cpumask_t        7130 kernel/sched.c 				 struct sched_group **sg, cpumask_t *nodemask)
cpumask_t        7172 kernel/sched.c static void free_sched_groups(const cpumask_t *cpu_map, cpumask_t *nodemask)
cpumask_t        7206 kernel/sched.c static void free_sched_groups(const cpumask_t *cpu_map, cpumask_t *nodemask)
cpumask_t        7303 kernel/sched.c 	cpumask_t tmpmask;			/* make this one first */
cpumask_t        7305 kernel/sched.c 		cpumask_t nodemask;
cpumask_t        7306 kernel/sched.c 		cpumask_t this_sibling_map;
cpumask_t        7307 kernel/sched.c 		cpumask_t this_core_map;
cpumask_t        7309 kernel/sched.c 	cpumask_t send_covered;
cpumask_t        7312 kernel/sched.c 	cpumask_t domainspan;
cpumask_t        7313 kernel/sched.c 	cpumask_t covered;
cpumask_t        7314 kernel/sched.c 	cpumask_t notcovered;
cpumask_t        7328 kernel/sched.c #define	SCHED_CPUMASK_VAR(v, a) 	cpumask_t *v = (cpumask_t *) \
cpumask_t        7370 kernel/sched.c static int __build_sched_domains(const cpumask_t *cpu_map,
cpumask_t        7376 kernel/sched.c 	cpumask_t *tmpmask;
cpumask_t        7413 kernel/sched.c 	tmpmask = (cpumask_t *)allmasks;
cpumask_t        7677 kernel/sched.c static int build_sched_domains(const cpumask_t *cpu_map)
cpumask_t        7682 kernel/sched.c static cpumask_t *doms_cur;	/* current sched domains */
cpumask_t        7692 kernel/sched.c static cpumask_t fallback_doms;
cpumask_t        7703 kernel/sched.c static int arch_init_sched_domains(const cpumask_t *cpu_map)
cpumask_t        7709 kernel/sched.c 	doms_cur = kmalloc(sizeof(cpumask_t), GFP_KERNEL);
cpumask_t        7720 kernel/sched.c static void arch_destroy_sched_domains(const cpumask_t *cpu_map,
cpumask_t        7721 kernel/sched.c 				       cpumask_t *tmpmask)
cpumask_t        7730 kernel/sched.c static void detach_destroy_domains(const cpumask_t *cpu_map)
cpumask_t        7732 kernel/sched.c 	cpumask_t tmpmask;
cpumask_t        7784 kernel/sched.c void partition_sched_domains(int ndoms_new, cpumask_t *doms_new,
cpumask_t        7971 kernel/sched.c 	cpumask_t non_isolated_cpus;
cpumask_t          70 kernel/sched_cpupri.c 		cpumask_t *lowest_mask)
cpumask_t          77 kernel/sched_cpupri.c 		cpumask_t mask;
cpumask_t          17 kernel/sched_cpupri.h 	cpumask_t  mask;
cpumask_t          28 kernel/sched_cpupri.h 		 struct task_struct *p, cpumask_t *lowest_mask);
cpumask_t         958 kernel/sched_fair.c 	cpumask_t tmp;
cpumask_t         142 kernel/sched_rt.c static inline cpumask_t sched_rt_period_mask(void)
cpumask_t         147 kernel/sched_rt.c static inline cpumask_t sched_rt_period_mask(void)
cpumask_t         215 kernel/sched_rt.c static inline cpumask_t sched_rt_period_mask(void)
cpumask_t         432 kernel/sched_rt.c 	cpumask_t span;
cpumask_t         806 kernel/sched_rt.c 	cpumask_t mask;
cpumask_t         957 kernel/sched_rt.c static DEFINE_PER_CPU(cpumask_t, local_cpu_mask);
cpumask_t         959 kernel/sched_rt.c static inline int pick_optimal_cpu(int this_cpu, cpumask_t *mask)
cpumask_t         977 kernel/sched_rt.c 	cpumask_t *lowest_mask = &__get_cpu_var(local_cpu_mask);
cpumask_t        1014 kernel/sched_rt.c 			cpumask_t domain_mask;
cpumask_t        1309 kernel/sched_rt.c 				const cpumask_t *new_mask)
cpumask_t          27 kernel/smp.c   	cpumask_t cpumask;
cpumask_t         290 kernel/smp.c   static void smp_call_function_mask_quiesce_stack(cpumask_t mask)
cpumask_t         321 kernel/smp.c   int smp_call_function_mask(cpumask_t mask, void (*func)(void *), void *info,
cpumask_t         326 kernel/smp.c   	cpumask_t allbutself;
cpumask_t         313 kernel/softlockup.c 			cpumask_t temp_cpu_online_map = cpu_online_map;
cpumask_t         102 kernel/stop_machine.c int __stop_machine(int (*fn)(void *), void *data, const cpumask_t *cpus)
cpumask_t         178 kernel/stop_machine.c int stop_machine(int (*fn)(void *), void *data, const cpumask_t *cpus)
cpumask_t         293 kernel/taskstats.c static int add_del_listener(pid_t pid, cpumask_t *maskp, int isadd)
cpumask_t         298 kernel/taskstats.c 	cpumask_t mask = *maskp;
cpumask_t         338 kernel/taskstats.c static int parse(struct nlattr *na, cpumask_t *mask)
cpumask_t         435 kernel/taskstats.c 	cpumask_t mask;
cpumask_t          31 kernel/time/tick-broadcast.c static cpumask_t tick_broadcast_mask;
cpumask_t          49 kernel/time/tick-broadcast.c cpumask_t *tick_get_broadcast_mask(void)
cpumask_t         130 kernel/time/tick-broadcast.c static void tick_do_broadcast(cpumask_t mask)
cpumask_t         163 kernel/time/tick-broadcast.c 	cpumask_t mask;
cpumask_t         363 kernel/time/tick-broadcast.c static cpumask_t tick_broadcast_oneshot_mask;
cpumask_t         368 kernel/time/tick-broadcast.c cpumask_t *tick_get_broadcast_oneshot_mask(void)
cpumask_t         392 kernel/time/tick-broadcast.c 	cpumask_t mask;
cpumask_t         495 kernel/time/tick-broadcast.c static void tick_broadcast_init_next_event(cpumask_t *mask, ktime_t expires)
cpumask_t         516 kernel/time/tick-broadcast.c 		cpumask_t mask;
cpumask_t         139 kernel/time/tick-common.c 			      const cpumask_t *cpumask)
cpumask_t          41 kernel/trace/trace.c static cpumask_t __read_mostly		tracing_buffer_mask;
cpumask_t        2067 kernel/trace/trace.c static cpumask_t tracing_cpumask = CPU_MASK_ALL;
cpumask_t        2073 kernel/trace/trace.c static cpumask_t tracing_cpumask_new;
cpumask_t        2504 kernel/trace/trace.c 	static cpumask_t mask;
cpumask_t         212 kernel/trace/trace_sysprof.c 	cpumask_t saved_mask = current->cpus_allowed;
cpumask_t          75 kernel/workqueue.c static cpumask_t cpu_singlethread_map __read_mostly;
cpumask_t          83 kernel/workqueue.c static cpumask_t cpu_populated_map __read_mostly;
cpumask_t          91 kernel/workqueue.c static const cpumask_t *wq_cpu_map(struct workqueue_struct *wq)
cpumask_t         412 kernel/workqueue.c 	const cpumask_t *cpu_map = wq_cpu_map(wq);
cpumask_t         534 kernel/workqueue.c 	const cpumask_t *cpu_map;
cpumask_t         901 kernel/workqueue.c 	const cpumask_t *cpu_map = wq_cpu_map(wq);
cpumask_t           6 lib/cpumask.c  int __first_cpu(const cpumask_t *srcp)
cpumask_t          12 lib/cpumask.c  int __next_cpu(int n, const cpumask_t *srcp)
cpumask_t          19 lib/cpumask.c  int __next_cpu_nr(int n, const cpumask_t *srcp)
cpumask_t          27 lib/cpumask.c  int __any_online_cpu(const cpumask_t *mask)
cpumask_t          34 mm/allocpercpu.c static void __percpu_depopulate_mask(void *__pdata, cpumask_t *mask)
cpumask_t          83 mm/allocpercpu.c 				  cpumask_t *mask)
cpumask_t          85 mm/allocpercpu.c 	cpumask_t populated;
cpumask_t         111 mm/allocpercpu.c void *__percpu_alloc_mask(size_t size, gfp_t gfp, cpumask_t *mask)
cpumask_t         175 mm/pdflush.c   	cpumask_t cpus_allowed;
cpumask_t        1962 mm/slub.c      static cpumask_t kmem_cach_cpu_free_init_once = CPU_MASK_NONE;
cpumask_t        3435 mm/slub.c      	cpumask_t cpus;
cpumask_t          23 mm/vmstat.c    static void sum_vm_events(unsigned long *ret, cpumask_t *cpumask)
cpumask_t         172 net/core/dev.c 	cpumask_t channel_mask;
cpumask_t         101 net/iucv/iucv.c static cpumask_t iucv_buffer_cpumask = CPU_MASK_NONE;
cpumask_t         102 net/iucv/iucv.c static cpumask_t iucv_irq_cpumask = CPU_MASK_NONE;
cpumask_t         494 net/iucv/iucv.c 	cpumask_t cpumask;
cpumask_t         558 net/iucv/iucv.c 	cpumask_t cpumask;
cpumask_t          60 virt/kvm/kvm_main.c static cpumask_t cpus_hardware_enabled;
cpumask_t         109 virt/kvm/kvm_main.c 	cpumask_t cpus;
cpumask_t         135 virt/kvm/kvm_main.c 	cpumask_t cpus;