tag | line | file | source code |
cpu_callin_map | 58 | arch/i386/kernel/smp.c | volatile unsigned long cpu_callin_map[NR_CPUS] = {0,}; /* We always use 0 the rest is ready for parallel delivery */ |
cpu_callin_map | 559 | arch/i386/kernel/smp.c | set_bit(cpuid, (unsigned long *)&cpu_callin_map[0]); |
cpu_callin_map | 835 | arch/i386/kernel/smp.c | if(cpu_callin_map[0]&(1<<i)) |
cpu_callin_map | 839 | arch/i386/kernel/smp.c | if(cpu_callin_map[0]&(1<<i)) |
cpu_callin_map | 1041 | arch/i386/kernel/smp.c | cpu_callin_map[0]=(1<<smp_src_cpu); |
cpu_callin_map | 1047 | arch/i386/kernel/smp.c | cpu_callin_map[0]=0; |
cpu_callin_map | 1052 | arch/i386/kernel/smp.c | cpu_callin_map[0]=0; |
cpu_callin_map | 1068 | arch/i386/kernel/smp.c | while(cpu_callin_map[0]!=target_map); /* Spin on the pass */ |
cpu_callin_map | 1178 | arch/i386/kernel/smp.c | set_bit(i, (unsigned long *)&cpu_callin_map[0]); |
cpu_callin_map | 40 | arch/sparc/kernel/smp.c | volatile unsigned long cpu_callin_map[NR_CPUS] = {0,}; |
cpu_callin_map | 163 | arch/sparc/kernel/smp.c | swap((unsigned long *)&cpu_callin_map[cpuid], 1); |
cpu_callin_map | 245 | arch/sparc/kernel/smp.c | if(cpu_callin_map[i]) |
cpu_callin_map | 249 | arch/sparc/kernel/smp.c | if(cpu_callin_map[i]) { |
cpu_callin_map | 258 | arch/sparc/kernel/smp.c | if(!(cpu_callin_map[i])) { |
cpu_callin_map | 373 | arch/sparc/kernel/smp.c | swap((unsigned long *) &cpu_callin_map[i], 0); |
cpu_callin_map | 376 | arch/sparc/kernel/smp.c | swap((unsigned long *) &cpu_callin_map[p], 1); |
cpu_callin_map | 382 | arch/sparc/kernel/smp.c | swap((unsigned long *) &cpu_callin_map[i], 1); |
cpu_callin_map | 392 | arch/sparc/kernel/smp.c | while(!cpu_callin_map[i]) |
cpu_callin_map | 615 | arch/sparc/kernel/smp.c | while(cpu_callin_map[i]) |
cpu_callin_map | 189 | include/asm-i386/smp.h | extern volatile unsigned long cpu_callin_map[NR_CPUS]; |
cpu_callin_map | 131 | kernel/sched.c | set_bit(cpu,&cpu_callin_map[0]); |