tag | line | file | source code |
NR_CPUS | 41 | arch/i386/kernel/irq.c | static unsigned int int_count[NR_CPUS][NR_IRQS] = {{0},}; |
NR_CPUS | 56 | arch/i386/kernel/smp.c | volatile int cpu_number_map[NR_CPUS]; /* which CPU maps to which logical number */ |
NR_CPUS | 57 | arch/i386/kernel/smp.c | volatile int cpu_logical_map[NR_CPUS]; /* which logical number maps to which CPU */ |
NR_CPUS | 58 | arch/i386/kernel/smp.c | volatile unsigned long cpu_callin_map[NR_CPUS] = {0,}; /* We always use 0 the rest is ready for parallel delivery */ |
NR_CPUS | 60 | arch/i386/kernel/smp.c | struct cpuinfo_x86 cpu_data[NR_CPUS]; /* Per cpu bogomips and other parameters */ |
NR_CPUS | 66 | arch/i386/kernel/smp.c | int apic_version[NR_CPUS]; /* APIC version number */ |
NR_CPUS | 72 | arch/i386/kernel/smp.c | unsigned char *kernel_stacks[NR_CPUS]; /* Kernel stack pointers for CPU's (debugging) */ |
NR_CPUS | 74 | arch/i386/kernel/smp.c | static volatile unsigned char smp_cpu_in_msg[NR_CPUS]; /* True if this processor is sending an IPI */ |
NR_CPUS | 86 | arch/i386/kernel/smp.c | volatile unsigned long smp_spins[NR_CPUS]={0}; /* Count interrupt spins */ |
NR_CPUS | 87 | arch/i386/kernel/smp.c | volatile unsigned long smp_spins_syscall[NR_CPUS]={0}; /* Count syscall spins */ |
NR_CPUS | 88 | arch/i386/kernel/smp.c | volatile unsigned long smp_spins_syscall_cur[NR_CPUS]={0};/* Count spins for the actual syscall */ |
NR_CPUS | 89 | arch/i386/kernel/smp.c | volatile unsigned long smp_spins_sys_idle[NR_CPUS]={0}; /* Count spins for sys_idle */ |
NR_CPUS | 90 | arch/i386/kernel/smp.c | volatile unsigned long smp_idle_count[1+NR_CPUS]={0,}; /* Count idle ticks */ |
NR_CPUS | 96 | arch/i386/kernel/smp.c | volatile unsigned long smp_proc_in_lock[NR_CPUS] = {0,};/* for computing process time */ |
NR_CPUS | 227 | arch/i386/kernel/smp.c | if(m->mpc_apicid>NR_CPUS) |
NR_CPUS | 228 | arch/i386/kernel/smp.c | printk("Processor #%d unused. (Max %d processors).\n",m->mpc_apicid, NR_CPUS); |
NR_CPUS | 593 | arch/i386/kernel/smp.c | for (i = 0; i < NR_CPUS; i++) |
NR_CPUS | 679 | arch/i386/kernel/smp.c | for(i=0;i<NR_CPUS;i++) |
NR_CPUS | 39 | arch/sparc/kernel/smp.c | unsigned char mid_xlate[NR_CPUS] = { 0, 0, 0, 0, }; |
NR_CPUS | 40 | arch/sparc/kernel/smp.c | volatile unsigned long cpu_callin_map[NR_CPUS] = {0,}; |
NR_CPUS | 41 | arch/sparc/kernel/smp.c | volatile unsigned long smp_invalidate_needed[NR_CPUS] = { 0, }; |
NR_CPUS | 42 | arch/sparc/kernel/smp.c | volatile unsigned long smp_spinning[NR_CPUS] = { 0, }; |
NR_CPUS | 43 | arch/sparc/kernel/smp.c | struct cpuinfo_sparc cpu_data[NR_CPUS]; |
NR_CPUS | 46 | arch/sparc/kernel/smp.c | static volatile unsigned char smp_cpu_in_msg[NR_CPUS]; |
NR_CPUS | 50 | arch/sparc/kernel/smp.c | volatile int cpu_number_map[NR_CPUS]; |
NR_CPUS | 51 | arch/sparc/kernel/smp.c | volatile int cpu_logical_map[NR_CPUS]; |
NR_CPUS | 66 | arch/sparc/kernel/smp.c | volatile unsigned long smp_spins[NR_CPUS]={0}; |
NR_CPUS | 67 | arch/sparc/kernel/smp.c | volatile unsigned long smp_spins_syscall[NR_CPUS]={0}; |
NR_CPUS | 68 | arch/sparc/kernel/smp.c | volatile unsigned long smp_spins_syscall_cur[NR_CPUS]={0}; |
NR_CPUS | 69 | arch/sparc/kernel/smp.c | volatile unsigned long smp_spins_sys_idle[NR_CPUS]={0}; |
NR_CPUS | 70 | arch/sparc/kernel/smp.c | volatile unsigned long smp_idle_count[1+NR_CPUS]={0,}; |
NR_CPUS | 76 | arch/sparc/kernel/smp.c | volatile unsigned long smp_proc_in_lock[NR_CPUS] = {0,}; |
NR_CPUS | 211 | arch/sparc/kernel/smp.c | for(i=0; i < NR_CPUS; i++) |
NR_CPUS | 213 | arch/sparc/kernel/smp.c | for(i=0; i < NR_CPUS; i++) |
NR_CPUS | 224 | arch/sparc/kernel/smp.c | for(i = 0; i < NR_CPUS; i++) { |
NR_CPUS | 269 | arch/sparc/kernel/smp.c | for(i = 0; i < NR_CPUS; i++) { |
NR_CPUS | 417 | arch/sparc/kernel/smp.c | unsigned long processors_in[NR_CPUS]; /* Set when ipi entered. */ |
NR_CPUS | 418 | arch/sparc/kernel/smp.c | unsigned long processors_out[NR_CPUS]; /* Set when ipi exited. */ |
NR_CPUS | 171 | include/asm-i386/smp.h | extern struct cpuinfo_x86 cpu_data[NR_CPUS]; |
NR_CPUS | 181 | include/asm-i386/smp.h | extern unsigned char *kernel_stacks[NR_CPUS]; |
NR_CPUS | 184 | include/asm-i386/smp.h | extern volatile int cpu_number_map[NR_CPUS]; |
NR_CPUS | 185 | include/asm-i386/smp.h | extern volatile int cpu_logical_map[NR_CPUS]; |
NR_CPUS | 189 | include/asm-i386/smp.h | extern volatile unsigned long cpu_callin_map[NR_CPUS]; |
NR_CPUS | 206 | include/asm-i386/smp.h | extern volatile unsigned long smp_proc_in_lock[NR_CPUS]; /* for computing process time */ |
NR_CPUS | 32 | include/asm-sparc/smp.h | extern struct cpuinfo_sparc cpu_data[NR_CPUS]; |
NR_CPUS | 47 | include/asm-sparc/smp.h | extern volatile unsigned long smp_invalidate_needed[NR_CPUS]; |
NR_CPUS | 87 | include/asm-sparc/smp.h | extern volatile int cpu_number_map[NR_CPUS]; |
NR_CPUS | 88 | include/asm-sparc/smp.h | extern volatile int cpu_logical_map[NR_CPUS]; |
NR_CPUS | 102 | include/asm-sparc/smp.h | extern volatile unsigned long smp_proc_in_lock[NR_CPUS]; /* for computing process time */ |
NR_CPUS | 320 | include/linux/sched.h | extern struct task_struct *current_set[NR_CPUS]; |
NR_CPUS | 19 | include/linux/smp.h | extern volatile unsigned long smp_spins[NR_CPUS]; /* count of interrupt spins */ |
NR_CPUS | 24 | include/linux/smp.h | extern volatile unsigned long smp_idle_count[1+NR_CPUS];/* count idle ticks */ |
NR_CPUS | 98 | kernel/sched.c | struct task_struct *current_set[NR_CPUS]; |
NR_CPUS | 1420 | kernel/sched.c | for(cpu = 0; cpu < NR_CPUS; cpu++) |