nr_cpu_ids 1707 arch/x86/kernel/apic_64.c else if (i < nr_cpu_ids) { nr_cpu_ids 583 arch/x86/kernel/cpu/mcheck/mce_64.c cpu_tsc = kmalloc(nr_cpu_ids * sizeof(long), GFP_KERNEL); nr_cpu_ids 163 arch/x86/kernel/cpu/proc.c if ((*pos) < nr_cpu_ids && cpu_online(*pos)) nr_cpu_ids 226 arch/x86/kernel/genapic_flat_64.c if ((unsigned)cpu < nr_cpu_ids) nr_cpu_ids 169 arch/x86/kernel/genx2apic_uv_x.c if ((unsigned)cpu < nr_cpu_ids) nr_cpu_ids 142 arch/x86/kernel/nmi.c prev_nmi_count = kmalloc(nr_cpu_ids * sizeof(int), GFP_KERNEL); nr_cpu_ids 111 arch/x86/kernel/setup_percpu.c unsigned long tsize = nr_cpu_ids * sizeof(void *); nr_cpu_ids 112 arch/x86/kernel/setup_percpu.c unsigned long asize = size * (nr_cpu_ids - 1); nr_cpu_ids 182 arch/x86/kernel/setup_percpu.c NR_CPUS, nr_cpu_ids, nr_node_ids); nr_cpu_ids 1060 arch/x86/kernel/smpboot.c if (def_to_bigsmp && nr_cpu_ids > 8) { nr_cpu_ids 1082 arch/x86/kernel/smpboot.c nr_cpu_ids = 8; nr_cpu_ids 1299 arch/x86/kernel/smpboot.c nr_cpu_ids = possible; nr_cpu_ids 1234 fs/ext4/ext4.h #define EXT4_FREEBLOCKS_WATERMARK (4 * (FBC_BATCH * nr_cpu_ids)) nr_cpu_ids 40 include/linux/backing-dev.h #define BDI_STAT_BATCH (8*(1+ilog2(nr_cpu_ids))) nr_cpu_ids 146 include/linux/backing-dev.h return nr_cpu_ids * BDI_STAT_BATCH; nr_cpu_ids 412 include/linux/cpumask.h extern int nr_cpu_ids; nr_cpu_ids 436 include/linux/cpumask.h #define cpus_weight_nr(cpumask) __cpus_weight(&(cpumask), nr_cpu_ids) nr_cpu_ids 440 include/linux/cpumask.h (cpu) < nr_cpu_ids; ) nr_cpu_ids 393 include/linux/netfilter/x_tables.h + nr_cpu_ids * sizeof(char *)) nr_cpu_ids 373 init/main.c int nr_cpu_ids __read_mostly = NR_CPUS; nr_cpu_ids 384 init/main.c nr_cpu_ids = highest_cpu + 1; nr_cpu_ids 4012 kernel/sched.c if (ilb < nr_cpu_ids) nr_cpu_ids 6129 kernel/sched.c if (dest_cpu >= nr_cpu_ids) nr_cpu_ids 6133 kernel/sched.c if (dest_cpu >= nr_cpu_ids) { nr_cpu_ids 7974 kernel/sched.c sched_group_nodes_bycpu = kzalloc(nr_cpu_ids * sizeof(void **), nr_cpu_ids 8123 kernel/sched.c alloc_size += 2 * nr_cpu_ids * sizeof(void **); nr_cpu_ids 8126 kernel/sched.c alloc_size += 2 * nr_cpu_ids * sizeof(void **); nr_cpu_ids 8140 kernel/sched.c ptr += nr_cpu_ids * sizeof(void **); nr_cpu_ids 8143 kernel/sched.c ptr += nr_cpu_ids * sizeof(void **); nr_cpu_ids 8147 kernel/sched.c ptr += nr_cpu_ids * sizeof(void **); nr_cpu_ids 8150 kernel/sched.c ptr += nr_cpu_ids * sizeof(void **); nr_cpu_ids 8155 kernel/sched.c ptr += nr_cpu_ids * sizeof(void **); nr_cpu_ids 8158 kernel/sched.c ptr += nr_cpu_ids * sizeof(void **); nr_cpu_ids 8162 kernel/sched.c ptr += nr_cpu_ids * sizeof(void **); nr_cpu_ids 8165 kernel/sched.c ptr += nr_cpu_ids * sizeof(void **); nr_cpu_ids 8477 kernel/sched.c tg->cfs_rq = kzalloc(sizeof(cfs_rq) * nr_cpu_ids, GFP_KERNEL); nr_cpu_ids 8480 kernel/sched.c tg->se = kzalloc(sizeof(se) * nr_cpu_ids, GFP_KERNEL); nr_cpu_ids 8565 kernel/sched.c tg->rt_rq = kzalloc(sizeof(rt_rq) * nr_cpu_ids, GFP_KERNEL); nr_cpu_ids 8568 kernel/sched.c tg->rt_se = kzalloc(sizeof(rt_se) * nr_cpu_ids, GFP_KERNEL); nr_cpu_ids 150 kernel/time/clocksource.c if (next_cpu >= nr_cpu_ids) nr_cpu_ids 21 lib/cpumask.c return min_t(int, nr_cpu_ids, nr_cpu_ids 22 lib/cpumask.c find_next_bit(srcp->bits, nr_cpu_ids, n+1)); nr_cpu_ids 187 lib/proportions.c #define PROP_BATCH (8*(1+ilog2(nr_cpu_ids))) nr_cpu_ids 239 lib/proportions.c if (val < (nr_cpu_ids * PROP_BATCH)) nr_cpu_ids 116 mm/allocpercpu.c size_t sz = roundup(nr_cpu_ids * sizeof(void *), cache_line_size()); nr_cpu_ids 1847 mm/slub.c min_objects = 4 * (fls(nr_cpu_ids) + 1); nr_cpu_ids 3022 mm/slub.c nr_cpu_ids * sizeof(struct kmem_cache_cpu *); nr_cpu_ids 3032 mm/slub.c nr_cpu_ids, nr_node_ids); nr_cpu_ids 4076 mm/slub.c int *data = kmalloc(nr_cpu_ids * sizeof(int), GFP_KERNEL); nr_cpu_ids 1929 mm/vmscan.c if (any_online_cpu(*mask) < nr_cpu_ids) nr_cpu_ids 872 net/bridge/netfilter/ebtables.c vmalloc(nr_cpu_ids * sizeof(*(newinfo->chainstack))); nr_cpu_ids 985 net/bridge/netfilter/ebtables.c countersize = COUNTER_OFFSET(tmp.nentries) * nr_cpu_ids; nr_cpu_ids 1124 net/bridge/netfilter/ebtables.c countersize = COUNTER_OFFSET(repl->nentries) * nr_cpu_ids; nr_cpu_ids 2657 net/core/dev.c while (*pos < nr_cpu_ids) nr_cpu_ids 4642 net/core/dev.c for (i = 0; i < nr_cpu_ids; i++) nr_cpu_ids 4657 net/core/dev.c for (i = 0; i < nr_cpu_ids; i++) nr_cpu_ids 4684 net/core/dev.c net_dma.channels = kzalloc(nr_cpu_ids * sizeof(struct net_dma), nr_cpu_ids 1149 net/ipv4/icmp.c kzalloc(nr_cpu_ids * sizeof(struct sock *), GFP_KERNEL); nr_cpu_ids 808 net/ipv6/icmp.c kzalloc(nr_cpu_ids * sizeof(struct sock *), GFP_KERNEL); nr_cpu_ids 179 net/sunrpc/svc.c unsigned int maxpools = nr_cpu_ids;