tag | line | file | source code |
mm | 143 | arch/alpha/kernel/process.c | p->mm->context = 0; |
mm | 155 | arch/alpha/kernel/process.c | dump->start_code = current->mm->start_code; |
mm | 156 | arch/alpha/kernel/process.c | dump->start_data = current->mm->start_data; |
mm | 158 | arch/alpha/kernel/process.c | dump->u_tsize = (current->mm->end_code - dump->start_code) >> PAGE_SHIFT; |
mm | 159 | arch/alpha/kernel/process.c | dump->u_dsize = (current->mm->brk + (PAGE_SIZE - 1) - dump->start_data) >> PAGE_SHIFT; |
mm | 161 | arch/alpha/kernel/process.c | (current->mm->start_stack - dump->start_stack + PAGE_SIZE - 1) >> PAGE_SHIFT; |
mm | 31 | arch/alpha/mm/fault.c | void get_new_asn_and_reload(struct task_struct *tsk, struct mm_struct *mm) |
mm | 33 | arch/alpha/mm/fault.c | get_new_mmu_context(tsk, mm, asn_cache); |
mm | 368 | arch/i386/kernel/process.c | dump->u_tsize = ((unsigned long) current->mm->end_code) >> PAGE_SHIFT; |
mm | 369 | arch/i386/kernel/process.c | dump->u_dsize = ((unsigned long) (current->mm->brk + (PAGE_SIZE-1))) >> PAGE_SHIFT; |
mm | 143 | arch/i386/kernel/setup.c | init_task.mm->start_code = TASK_SIZE; |
mm | 144 | arch/i386/kernel/setup.c | init_task.mm->end_code = TASK_SIZE + (unsigned long) &_etext; |
mm | 145 | arch/i386/kernel/setup.c | init_task.mm->end_data = TASK_SIZE + (unsigned long) &_edata; |
mm | 146 | arch/i386/kernel/setup.c | init_task.mm->brk = TASK_SIZE + (unsigned long) &_end; |
mm | 77 | arch/i386/kernel/vm86.c | pgd = pgd_offset(tsk->mm, 0xA0000); |
mm | 169 | arch/m68k/kernel/process.c | dump->u_tsize = ((unsigned long) current->mm->end_code) >> PAGE_SHIFT; |
mm | 170 | arch/m68k/kernel/process.c | dump->u_dsize = ((unsigned long) (current->mm->brk + |
mm | 125 | arch/m68k/kernel/setup.c | init_task.mm->start_code = 0; |
mm | 126 | arch/m68k/kernel/setup.c | init_task.mm->end_code = (unsigned long) &_etext; |
mm | 127 | arch/m68k/kernel/setup.c | init_task.mm->end_data = (unsigned long) &_edata; |
mm | 128 | arch/m68k/kernel/setup.c | init_task.mm->brk = (unsigned long) &_end; |
mm | 64 | arch/ppc/mm/fault.c | dir = pgd_offset(current->mm, regs->dar & PAGE_MASK); |
mm | 87 | arch/ppc/mm/fault.c | printk("beforefault: pgd[0] = %x[%x]\n",current->mm->pgd,*(current->mm->pgd)); |
mm | 91 | arch/ppc/mm/fault.c | printk("handled: pgd[0] = %x[%x]\n",current->mm->pgd,*(current->mm->pgd)); |
mm | 119 | arch/ppc/mm/fault.c | dir = pgd_offset(current->mm, regs->dar & PAGE_MASK); |
mm | 135 | arch/ppc/mm/fault.c | current->mm, current->mm->pgd, |
mm | 136 | arch/ppc/mm/fault.c | current->mm->pgd+((regs->dar&PAGE_MASK) >> PGDIR_SHIFT)); |
mm | 142 | arch/ppc/mm/fault.c | current->mm->pgd,dir,*dir,pmd,*pmd,pte); |
mm | 164 | arch/ppc/mm/fault.c | regs->dar, regs->dsisr, regs->nip, current->mm); |
mm | 190 | arch/ppc/mm/fault.c | for (vma = current->mm->mmap ; ; vma = vma->vm_next) |
mm | 195 | arch/ppc/mm/fault.c | regs->nip, current,current->pid,current->mm,current->mm->mmap, address, error_code); |
mm | 213 | arch/ppc/mm/fault.c | printk("stack: gpr[1]=%x ip = %x; current=%x[%d]; mm=%x; mmap=%x; address = %x error_code = %x\n",regs->gpr[1],regs->nip, current,current->pid,current->mm,current->mm->mmap, address, error_code); |
mm | 222 | arch/ppc/mm/fault.c | printk("stack2: gpr[1]=%x ip = %x; current=%x[%d]; mm=%x; mmap=%x; address = %x error_code = %x\n",regs->gpr[1],regs->nip, current,current->pid,current->mm,current->mm->mmap, address, error_code); |
mm | 242 | arch/ppc/mm/fault.c | current,address,vma->vm_flags,current->mm,vma,vma->vm_start,vma->vm_end); |
mm | 296 | arch/ppc/mm/fault.c | current,address,vma->vm_flags,current->mm,vma,vma->vm_start,vma->vm_end); |
mm | 309 | arch/ppc/mm/fault.c | dir = pgd_offset(current->mm, address & PAGE_MASK); |
mm | 730 | arch/ppc/mm/init.c | pgd = pgd_offset(current->mm, 0); |
mm | 298 | arch/sparc/kernel/process.c | dump->u_tsize = (((unsigned long) current->mm->end_code) - |
mm | 299 | arch/sparc/kernel/process.c | ((unsigned long) current->mm->start_code)) & ~(PAGE_SIZE - 1); |
mm | 300 | arch/sparc/kernel/process.c | dump->u_dsize = ((unsigned long) (current->mm->brk + (PAGE_SIZE-1))); |
mm | 309 | arch/sparc/kernel/setup.c | init_task.mm->mmap->vm_page_prot = PAGE_SHARED; |
mm | 310 | arch/sparc/kernel/setup.c | init_task.mm->mmap->vm_start = KERNBASE; |
mm | 311 | arch/sparc/kernel/setup.c | init_task.mm->mmap->vm_end = *memory_end_p; |
mm | 172 | arch/sparc/kernel/smp.c | current->mm->mmap->vm_page_prot = PAGE_SHARED; |
mm | 173 | arch/sparc/kernel/smp.c | current->mm->mmap->vm_start = KERNBASE; |
mm | 174 | arch/sparc/kernel/smp.c | current->mm->mmap->vm_end = init_task.mm->mmap->vm_end; |
mm | 531 | arch/sparc/kernel/smp.c | void smp_flush_cache_mm(struct mm_struct *mm) |
mm | 533 | arch/sparc/kernel/smp.c | if(mm->context != NO_CONTEXT) |
mm | 534 | arch/sparc/kernel/smp.c | xc1((smpfunc_t) local_flush_cache_mm, (unsigned long) mm); |
mm | 537 | arch/sparc/kernel/smp.c | void smp_flush_tlb_mm(struct mm_struct *mm) |
mm | 539 | arch/sparc/kernel/smp.c | if(mm->context != NO_CONTEXT) |
mm | 540 | arch/sparc/kernel/smp.c | xc1((smpfunc_t) local_flush_tlb_mm, (unsigned long) mm); |
mm | 543 | arch/sparc/kernel/smp.c | void smp_flush_cache_range(struct mm_struct *mm, unsigned long start, |
mm | 546 | arch/sparc/kernel/smp.c | if(mm->context != NO_CONTEXT) |
mm | 547 | arch/sparc/kernel/smp.c | xc3((smpfunc_t) local_flush_cache_range, (unsigned long) mm, |
mm | 551 | arch/sparc/kernel/smp.c | void smp_flush_tlb_range(struct mm_struct *mm, unsigned long start, |
mm | 554 | arch/sparc/kernel/smp.c | if(mm->context != NO_CONTEXT) |
mm | 555 | arch/sparc/kernel/smp.c | xc3((smpfunc_t) local_flush_tlb_range, (unsigned long) mm, |
mm | 126 | arch/sparc/kernel/sys_sunos.c | if (brk < current->mm->end_code) |
mm | 130 | arch/sparc/kernel/sys_sunos.c | oldbrk = PAGE_ALIGN(current->mm->brk); |
mm | 132 | arch/sparc/kernel/sys_sunos.c | current->mm->brk = brk; |
mm | 139 | arch/sparc/kernel/sys_sunos.c | if (brk <= current->mm->brk) { |
mm | 140 | arch/sparc/kernel/sys_sunos.c | current->mm->brk = brk; |
mm | 150 | arch/sparc/kernel/sys_sunos.c | if (brk - current->mm->end_code > rlim) |
mm | 176 | arch/sparc/kernel/sys_sunos.c | current->mm->brk = brk; |
mm | 188 | arch/sparc/kernel/sys_sunos.c | error = sunos_brk(((int) current->mm->brk) + increment); |
mm | 192 | arch/sparc/kernel/sys_sunos.c | return current->mm->brk; |
mm | 288 | arch/sparc/kernel/sys_sunos.c | pgdp = pgd_offset(current->mm, addr); |
mm | 208 | arch/sparc/mm/fault.c | (unsigned long) current->mm->context); |
mm | 210 | arch/sparc/mm/fault.c | (unsigned long) current->mm->pgd); |
mm | 23 | arch/sparc/mm/generic.c | if (current->mm->rss <= 0) |
mm | 25 | arch/sparc/mm/generic.c | current->mm->rss--; |
mm | 87 | arch/sparc/mm/generic.c | dir = pgd_offset(current->mm, from); |
mm | 88 | arch/sparc/mm/generic.c | flush_cache_range(current->mm, beg, end); |
mm | 100 | arch/sparc/mm/generic.c | flush_tlb_range(current->mm, beg, end); |
mm | 218 | arch/sparc/mm/srmmu.c | static pgd_t *srmmu_pgd_offset(struct mm_struct * mm, unsigned long address) |
mm | 220 | arch/sparc/mm/srmmu.c | return mm->pgd + ((address >> SRMMU_PGDIR_SHIFT) & (SRMMU_PTRS_PER_PGD - 1)); |
mm | 238 | arch/sparc/mm/srmmu.c | if(tsk->mm->context != NO_CONTEXT) |
mm | 239 | arch/sparc/mm/srmmu.c | ctxd_set(&srmmu_context_table[tsk->mm->context], pgdp); |
mm | 244 | arch/sparc/mm/srmmu.c | pgd_t *pgdp = srmmu_pgd_offset(init_task.mm, addr); |
mm | 255 | arch/sparc/mm/srmmu.c | pgd_t *pgdp = srmmu_pgd_offset(init_task.mm, addr); |
mm | 429 | arch/sparc/mm/srmmu.c | static inline void alloc_context(struct mm_struct *mm) |
mm | 437 | arch/sparc/mm/srmmu.c | mm->context = ctxp->ctx_number; |
mm | 438 | arch/sparc/mm/srmmu.c | ctxp->ctx_mm = mm; |
mm | 442 | arch/sparc/mm/srmmu.c | if(ctxp->ctx_mm == current->mm) |
mm | 451 | arch/sparc/mm/srmmu.c | ctxp->ctx_mm = mm; |
mm | 452 | arch/sparc/mm/srmmu.c | mm->context = ctxp->ctx_number; |
mm | 465 | arch/sparc/mm/srmmu.c | if(tsk->mm->context == NO_CONTEXT) { |
mm | 466 | arch/sparc/mm/srmmu.c | alloc_context(tsk->mm); |
mm | 467 | arch/sparc/mm/srmmu.c | ctxd_set(&srmmu_context_table[tsk->mm->context], tsk->mm->pgd); |
mm | 469 | arch/sparc/mm/srmmu.c | srmmu_set_context(tsk->mm->context); |
mm | 481 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, virt_addr); |
mm | 559 | arch/sparc/mm/srmmu.c | static void tsunami_flush_cache_mm(struct mm_struct *mm) |
mm | 562 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 572 | arch/sparc/mm/srmmu.c | static void tsunami_flush_cache_range(struct mm_struct *mm, unsigned long start, unsigned long end) |
mm | 575 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 588 | arch/sparc/mm/srmmu.c | struct mm_struct *mm = vma->vm_mm; |
mm | 589 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 633 | arch/sparc/mm/srmmu.c | static void tsunami_flush_tlb_mm(struct mm_struct *mm) |
mm | 637 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 646 | arch/sparc/mm/srmmu.c | static void tsunami_flush_tlb_range(struct mm_struct *mm, unsigned long start, unsigned long end) |
mm | 650 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 662 | arch/sparc/mm/srmmu.c | struct mm_struct *mm = vma->vm_mm; |
mm | 665 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 669 | arch/sparc/mm/srmmu.c | srmmu_set_context(mm->context); |
mm | 695 | arch/sparc/mm/srmmu.c | static void swift_flush_cache_mm(struct mm_struct *mm) |
mm | 698 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 707 | arch/sparc/mm/srmmu.c | static void swift_flush_cache_range(struct mm_struct *mm, unsigned long start, unsigned long end) |
mm | 710 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 722 | arch/sparc/mm/srmmu.c | struct mm_struct *mm = vma->vm_mm; |
mm | 723 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 756 | arch/sparc/mm/srmmu.c | static void swift_flush_tlb_mm(struct mm_struct *mm) |
mm | 760 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) |
mm | 765 | arch/sparc/mm/srmmu.c | static void swift_flush_tlb_range(struct mm_struct *mm, unsigned long start, unsigned long end) |
mm | 769 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) |
mm | 777 | arch/sparc/mm/srmmu.c | struct mm_struct *mm = vma->vm_mm; |
mm | 778 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) |
mm | 811 | arch/sparc/mm/srmmu.c | static void viking_flush_cache_mm(struct mm_struct *mm) |
mm | 814 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 823 | arch/sparc/mm/srmmu.c | static void viking_flush_cache_range(struct mm_struct *mm, unsigned long start, unsigned long end) |
mm | 826 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 838 | arch/sparc/mm/srmmu.c | struct mm_struct *mm = vma->vm_mm; |
mm | 839 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 948 | arch/sparc/mm/srmmu.c | static void viking_flush_tlb_mm(struct mm_struct *mm) |
mm | 954 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 957 | arch/sparc/mm/srmmu.c | srmmu_set_context(mm->context); |
mm | 965 | arch/sparc/mm/srmmu.c | static void viking_flush_tlb_range(struct mm_struct *mm, unsigned long start, unsigned long end) |
mm | 971 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 974 | arch/sparc/mm/srmmu.c | srmmu_set_context(mm->context); |
mm | 989 | arch/sparc/mm/srmmu.c | struct mm_struct *mm = vma->vm_mm; |
mm | 993 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 996 | arch/sparc/mm/srmmu.c | srmmu_set_context(mm->context); |
mm | 1017 | arch/sparc/mm/srmmu.c | static void cypress_flush_tlb_mm(struct mm_struct *mm) |
mm | 1023 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 1026 | arch/sparc/mm/srmmu.c | srmmu_set_context(mm->context); |
mm | 1034 | arch/sparc/mm/srmmu.c | static void cypress_flush_tlb_range(struct mm_struct *mm, unsigned long start, unsigned long end) |
mm | 1040 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 1043 | arch/sparc/mm/srmmu.c | srmmu_set_context(mm->context); |
mm | 1058 | arch/sparc/mm/srmmu.c | struct mm_struct *mm = vma->vm_mm; |
mm | 1062 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 1065 | arch/sparc/mm/srmmu.c | srmmu_set_context(mm->context); |
mm | 1081 | arch/sparc/mm/srmmu.c | static void hypersparc_flush_cache_mm(struct mm_struct *mm) |
mm | 1084 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 1094 | arch/sparc/mm/srmmu.c | static void hypersparc_flush_cache_range(struct mm_struct *mm, unsigned long start, unsigned long end) |
mm | 1097 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 1112 | arch/sparc/mm/srmmu.c | struct mm_struct *mm = vma->vm_mm; |
mm | 1117 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 1121 | arch/sparc/mm/srmmu.c | srmmu_set_context(mm->context); |
mm | 1169 | arch/sparc/mm/srmmu.c | static void hypersparc_flush_tlb_mm(struct mm_struct *mm) |
mm | 1175 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 1179 | arch/sparc/mm/srmmu.c | srmmu_set_context(mm->context); |
mm | 1188 | arch/sparc/mm/srmmu.c | static void hypersparc_flush_tlb_range(struct mm_struct *mm, unsigned long start, unsigned long end) |
mm | 1194 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 1198 | arch/sparc/mm/srmmu.c | srmmu_set_context(mm->context); |
mm | 1213 | arch/sparc/mm/srmmu.c | struct mm_struct *mm = vma->vm_mm; |
mm | 1218 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 1222 | arch/sparc/mm/srmmu.c | srmmu_set_context(mm->context); |
mm | 1244 | arch/sparc/mm/srmmu.c | if(tsk->mm->context != NO_CONTEXT) { |
mm | 1246 | arch/sparc/mm/srmmu.c | ctxd_set(&srmmu_context_table[tsk->mm->context], pgdp); |
mm | 1275 | arch/sparc/mm/srmmu.c | if(tsk->mm->context == NO_CONTEXT) { |
mm | 1276 | arch/sparc/mm/srmmu.c | alloc_context(tsk->mm); |
mm | 1277 | arch/sparc/mm/srmmu.c | ctxd_set(&srmmu_context_table[tsk->mm->context], tsk->mm->pgd); |
mm | 1279 | arch/sparc/mm/srmmu.c | srmmu_set_context(tsk->mm->context); |
mm | 1312 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, start); |
mm | 1544 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, start); |
mm | 1594 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, start); |
mm | 1634 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, start); |
mm | 1649 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, first); |
mm | 1672 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, KERNBASE); |
mm | 1688 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, start); |
mm | 1732 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, start); |
mm | 1767 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, start); |
mm | 1886 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, ptables_start); |
mm | 1892 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, (unsigned long)swapper_pg_dir); |
mm | 1944 | arch/sparc/mm/srmmu.c | struct mm_struct *mm = current->mm; |
mm | 1946 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 1947 | arch/sparc/mm/srmmu.c | flush_cache_mm(mm); |
mm | 1948 | arch/sparc/mm/srmmu.c | ctxd_set(&srmmu_context_table[mm->context], swapper_pg_dir); |
mm | 1949 | arch/sparc/mm/srmmu.c | flush_tlb_mm(mm); |
mm | 1950 | arch/sparc/mm/srmmu.c | ctx_old = ctx_list_pool + mm->context; |
mm | 1953 | arch/sparc/mm/srmmu.c | mm->context = NO_CONTEXT; |
mm | 1960 | arch/sparc/mm/srmmu.c | alloc_context(current->mm); |
mm | 1961 | arch/sparc/mm/srmmu.c | ctxd_set(&srmmu_context_table[current->mm->context], current->mm->pgd); |
mm | 1962 | arch/sparc/mm/srmmu.c | srmmu_set_context(current->mm->context); |
mm | 1969 | arch/sparc/mm/srmmu.c | struct mm_struct *mm = current->mm; |
mm | 1971 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 1977 | arch/sparc/mm/srmmu.c | flush_cache_mm(mm); |
mm | 1978 | arch/sparc/mm/srmmu.c | ctxd_set(&srmmu_context_table[mm->context], swapper_pg_dir); |
mm | 1979 | arch/sparc/mm/srmmu.c | flush_tlb_mm(mm); |
mm | 1980 | arch/sparc/mm/srmmu.c | ctx_old = ctx_list_pool + mm->context; |
mm | 1983 | arch/sparc/mm/srmmu.c | mm->context = NO_CONTEXT; |
mm | 1990 | arch/sparc/mm/srmmu.c | alloc_context(current->mm); |
mm | 1991 | arch/sparc/mm/srmmu.c | flush_cache_mm(current->mm); |
mm | 1992 | arch/sparc/mm/srmmu.c | ctxd_set(&srmmu_context_table[current->mm->context], current->mm->pgd); |
mm | 1993 | arch/sparc/mm/srmmu.c | srmmu_set_context(current->mm->context); |
mm | 941 | arch/sparc/mm/sun4c.c | sun4c_kstack_vma.vm_mm = init_task.mm; |
mm | 968 | arch/sparc/mm/sun4c.c | static void sun4c_flush_cache_mm(struct mm_struct *mm) |
mm | 974 | arch/sparc/mm/sun4c.c | if(mm->context != NO_CONTEXT) { |
mm | 979 | arch/sparc/mm/sun4c.c | sun4c_set_context(mm->context); |
mm | 988 | arch/sparc/mm/sun4c.c | static void sun4c_flush_cache_range(struct mm_struct *mm, unsigned long start, unsigned long end) |
mm | 994 | arch/sparc/mm/sun4c.c | if(mm->context != NO_CONTEXT) { |
mm | 1005 | arch/sparc/mm/sun4c.c | sun4c_set_context(mm->context); |
mm | 1036 | arch/sparc/mm/sun4c.c | struct mm_struct *mm = vma->vm_mm; |
mm | 1042 | arch/sparc/mm/sun4c.c | if(mm->context != NO_CONTEXT) { |
mm | 1047 | arch/sparc/mm/sun4c.c | sun4c_set_context(mm->context); |
mm | 1090 | arch/sparc/mm/sun4c.c | static void sun4c_flush_tlb_mm(struct mm_struct *mm) |
mm | 1097 | arch/sparc/mm/sun4c.c | if(mm->context != NO_CONTEXT) { |
mm | 1099 | arch/sparc/mm/sun4c.c | crp = &sun4c_context_ring[mm->context]; |
mm | 1101 | arch/sparc/mm/sun4c.c | ctx = mm->context; |
mm | 1103 | arch/sparc/mm/sun4c.c | sun4c_set_context(mm->context); |
mm | 1116 | arch/sparc/mm/sun4c.c | static void sun4c_flush_tlb_range(struct mm_struct *mm, unsigned long start, unsigned long end) |
mm | 1122 | arch/sparc/mm/sun4c.c | if(mm->context == NO_CONTEXT) |
mm | 1127 | arch/sparc/mm/sun4c.c | sun4c_set_context(mm->context); |
mm | 1135 | arch/sparc/mm/sun4c.c | free_user_entry(mm->context, this_entry); |
mm | 1144 | arch/sparc/mm/sun4c.c | struct mm_struct *mm = vma->vm_mm; |
mm | 1148 | arch/sparc/mm/sun4c.c | if(mm->context != NO_CONTEXT) { |
mm | 1151 | arch/sparc/mm/sun4c.c | sun4c_set_context(mm->context); |
mm | 1188 | arch/sparc/mm/sun4c.c | static inline void sun4c_alloc_context(struct mm_struct *mm) |
mm | 1196 | arch/sparc/mm/sun4c.c | mm->context = ctxp->ctx_number; |
mm | 1197 | arch/sparc/mm/sun4c.c | ctxp->ctx_mm = mm; |
mm | 1201 | arch/sparc/mm/sun4c.c | if(ctxp->ctx_mm == current->mm) |
mm | 1208 | arch/sparc/mm/sun4c.c | ctxp->ctx_mm = mm; |
mm | 1209 | arch/sparc/mm/sun4c.c | mm->context = ctxp->ctx_number; |
mm | 1238 | arch/sparc/mm/sun4c.c | if(tsk->mm->context == NO_CONTEXT) |
mm | 1239 | arch/sparc/mm/sun4c.c | sun4c_alloc_context(tsk->mm); |
mm | 1241 | arch/sparc/mm/sun4c.c | sun4c_set_context(tsk->mm->context); |
mm | 1247 | arch/sparc/mm/sun4c.c | sun4c_alloc_context(current->mm); |
mm | 1248 | arch/sparc/mm/sun4c.c | sun4c_set_context(current->mm->context); |
mm | 1255 | arch/sparc/mm/sun4c.c | struct mm_struct *mm = current->mm; |
mm | 1257 | arch/sparc/mm/sun4c.c | if(mm->context != NO_CONTEXT) { |
mm | 1258 | arch/sparc/mm/sun4c.c | sun4c_demap_context(&sun4c_context_ring[mm->context], mm->context); |
mm | 1259 | arch/sparc/mm/sun4c.c | ctx_old = ctx_list_pool + mm->context; |
mm | 1262 | arch/sparc/mm/sun4c.c | mm->context = NO_CONTEXT; |
mm | 1384 | arch/sparc/mm/sun4c.c | static pgd_t *sun4c_pgd_offset(struct mm_struct * mm, unsigned long address) |
mm | 1386 | arch/sparc/mm/sun4c.c | return mm->pgd + (address >> SUN4C_PGDIR_SHIFT); |
mm | 4185 | drivers/block/floppy.c | for (i=current->mm->env_start; i< current->mm->env_end; i ++){ |
mm | 827 | drivers/cdrom/sbpcd.c | u_int mm; |
mm | 831 | drivers/cdrom/sbpcd.c | mm = (blk + CD_MSF_OFFSET) % (CD_SECS * CD_FRAMES); |
mm | 832 | drivers/cdrom/sbpcd.c | msf.c[1] = mm / CD_FRAMES; |
mm | 833 | drivers/cdrom/sbpcd.c | msf.c[0] = mm % CD_FRAMES; |
mm | 235 | fs/binfmt_aout.c | current->mm->arg_start = (unsigned long) p; |
mm | 241 | fs/binfmt_aout.c | current->mm->arg_end = current->mm->env_start = (unsigned long) p; |
mm | 247 | fs/binfmt_aout.c | current->mm->env_end = (unsigned long) p; |
mm | 304 | fs/binfmt_aout.c | current->mm->end_code = ex.a_text + |
mm | 305 | fs/binfmt_aout.c | (current->mm->start_code = N_TXTADDR(ex)); |
mm | 306 | fs/binfmt_aout.c | current->mm->end_data = ex.a_data + |
mm | 307 | fs/binfmt_aout.c | (current->mm->start_data = N_DATADDR(ex)); |
mm | 308 | fs/binfmt_aout.c | current->mm->brk = ex.a_bss + |
mm | 309 | fs/binfmt_aout.c | (current->mm->start_brk = N_BSSADDR(ex)); |
mm | 311 | fs/binfmt_aout.c | current->mm->rss = 0; |
mm | 312 | fs/binfmt_aout.c | current->mm->mmap = NULL; |
mm | 382 | fs/binfmt_aout.c | set_brk(current->mm->start_brk, current->mm->brk); |
mm | 387 | fs/binfmt_aout.c | current->mm->start_stack = p; |
mm | 133 | fs/binfmt_elf.c | current->mm->arg_start = (unsigned long) p; |
mm | 139 | fs/binfmt_elf.c | current->mm->arg_end = current->mm->env_start = (unsigned long) p; |
mm | 145 | fs/binfmt_elf.c | current->mm->env_end = (unsigned long) p; |
mm | 290 | fs/binfmt_elf.c | current->mm->brk = interp_ex->a_bss + |
mm | 291 | fs/binfmt_elf.c | (current->mm->end_data = interp_ex->a_data + |
mm | 292 | fs/binfmt_elf.c | (current->mm->end_code = interp_ex->a_text)); |
mm | 522 | fs/binfmt_elf.c | current->mm->end_data = 0; |
mm | 523 | fs/binfmt_elf.c | current->mm->end_code = 0; |
mm | 524 | fs/binfmt_elf.c | current->mm->start_mmap = ELF_START_MMAP; |
mm | 525 | fs/binfmt_elf.c | current->mm->mmap = NULL; |
mm | 530 | fs/binfmt_elf.c | current->mm->rss = 0; |
mm | 532 | fs/binfmt_elf.c | current->mm->start_stack = bprm->p; |
mm | 643 | fs/binfmt_elf.c | current->mm->arg_start += strlen(passed_fileno) + 1; |
mm | 644 | fs/binfmt_elf.c | current->mm->start_brk = current->mm->brk = elf_brk; |
mm | 645 | fs/binfmt_elf.c | current->mm->end_code = end_code; |
mm | 646 | fs/binfmt_elf.c | current->mm->start_code = start_code; |
mm | 647 | fs/binfmt_elf.c | current->mm->end_data = end_data; |
mm | 648 | fs/binfmt_elf.c | current->mm->start_stack = bprm->p; |
mm | 657 | fs/binfmt_elf.c | printk("(start_brk) %x\n" , current->mm->start_brk); |
mm | 658 | fs/binfmt_elf.c | printk("(end_code) %x\n" , current->mm->end_code); |
mm | 659 | fs/binfmt_elf.c | printk("(start_code) %x\n" , current->mm->start_code); |
mm | 660 | fs/binfmt_elf.c | printk("(end_data) %x\n" , current->mm->end_data); |
mm | 661 | fs/binfmt_elf.c | printk("(start_stack) %x\n" , current->mm->start_stack); |
mm | 662 | fs/binfmt_elf.c | printk("(brk) %x\n" , current->mm->brk); |
mm | 966 | fs/binfmt_elf.c | for(vma = current->mm->mmap; vma != NULL; vma = vma->vm_next) { |
mm | 1094 | fs/binfmt_elf.c | len = current->mm->arg_end - current->mm->arg_start; |
mm | 1097 | fs/binfmt_elf.c | (const char *)current->mm->arg_start, len); |
mm | 1151 | fs/binfmt_elf.c | for(vma = current->mm->mmap, i = 0; |
mm | 1183 | fs/binfmt_elf.c | for(i = 0, vma = current->mm->mmap; |
mm | 296 | fs/exec.c | mpnt->vm_mm = current->mm; |
mm | 306 | fs/exec.c | current->mm->total_vm = (mpnt->vm_end - mpnt->vm_start) >> PAGE_SHIFT; |
mm | 311 | fs/exec.c | current->mm->rss++; |
mm | 374 | fs/exec.c | if (current->mm->count > 1) { |
mm | 375 | fs/exec.c | struct mm_struct *mm = kmalloc(sizeof(*mm), GFP_KERNEL); |
mm | 376 | fs/exec.c | if (!mm) { |
mm | 381 | fs/exec.c | *mm = *current->mm; |
mm | 382 | fs/exec.c | mm->def_flags = 0; /* should future lockings be kept? */ |
mm | 383 | fs/exec.c | mm->count = 1; |
mm | 384 | fs/exec.c | mm->mmap = NULL; |
mm | 385 | fs/exec.c | mm->mmap_avl = NULL; |
mm | 386 | fs/exec.c | mm->total_vm = 0; |
mm | 387 | fs/exec.c | mm->rss = 0; |
mm | 388 | fs/exec.c | current->mm->count--; |
mm | 389 | fs/exec.c | current->mm = mm; |
mm | 393 | fs/exec.c | exit_mmap(current->mm); |
mm | 130 | fs/namei.c | if (!p->mm) |
mm | 132 | fs/namei.c | for(mpnt = p->mm->mmap; mpnt; mpnt = mpnt->vm_next) { |
mm | 338 | fs/proc/array.c | if (!p || !p->mm || ptr >= TASK_SIZE) |
mm | 340 | fs/proc/array.c | page_dir = pgd_offset(p->mm,ptr); |
mm | 399 | fs/proc/array.c | if (!p || !*p || !(*p)->mm) |
mm | 401 | fs/proc/array.c | return get_array(p, (*p)->mm->env_start, (*p)->mm->env_end, buffer); |
mm | 408 | fs/proc/array.c | if (!p || !*p || !(*p)->mm) |
mm | 410 | fs/proc/array.c | return get_array(p, (*p)->mm->arg_start, (*p)->mm->arg_end, buffer); |
mm | 544 | fs/proc/array.c | struct mm_struct * mm = p->mm; |
mm | 546 | fs/proc/array.c | if (mm && mm != &init_mm) { |
mm | 547 | fs/proc/array.c | struct vm_area_struct * vma = mm->mmap; |
mm | 551 | fs/proc/array.c | for (vma = mm->mmap; vma; vma = vma->vm_next) { |
mm | 576 | fs/proc/array.c | mm->total_vm << (PAGE_SHIFT-10), |
mm | 577 | fs/proc/array.c | mm->locked_vm << (PAGE_SHIFT-10), |
mm | 578 | fs/proc/array.c | mm->rss << (PAGE_SHIFT-10), |
mm | 649 | fs/proc/array.c | if (tsk->mm && tsk->mm != &init_mm) { |
mm | 650 | fs/proc/array.c | struct vm_area_struct *vma = tsk->mm->mmap; |
mm | 714 | fs/proc/array.c | tsk->mm ? tsk->mm->rss : 0, /* you might want to shift this left 3 */ |
mm | 716 | fs/proc/array.c | tsk->mm ? tsk->mm->start_code : 0, |
mm | 717 | fs/proc/array.c | tsk->mm ? tsk->mm->end_code : 0, |
mm | 718 | fs/proc/array.c | tsk->mm ? tsk->mm->start_stack : 0, |
mm | 810 | fs/proc/array.c | if (tsk->mm && tsk->mm != &init_mm) { |
mm | 811 | fs/proc/array.c | struct vm_area_struct * vma = tsk->mm->mmap; |
mm | 814 | fs/proc/array.c | pgd_t *pgd = pgd_offset(tsk->mm, vma->vm_start); |
mm | 874 | fs/proc/array.c | if (!(*p)->mm || (*p)->mm == &init_mm || count == 0) |
mm | 882 | fs/proc/array.c | for (map = (*p)->mm->mmap, i = 0; map && (i < lineno); map = map->vm_next, i++) |
mm | 106 | fs/proc/link.c | if (!p->mm) |
mm | 108 | fs/proc/link.c | vma = p->mm->mmap; |
mm | 84 | fs/proc/mem.c | page_dir = pgd_offset(tsk->mm,addr); |
mm | 231 | fs/proc/mem.c | src_vma = tsk->mm->mmap; |
mm | 239 | fs/proc/mem.c | src_dir = pgd_offset(tsk->mm, stmp); |
mm | 266 | fs/proc/mem.c | src_vma = tsk->mm->mmap; |
mm | 276 | fs/proc/mem.c | src_dir = pgd_offset(tsk->mm, stmp); |
mm | 280 | fs/proc/mem.c | dest_dir = pgd_offset(current->mm, dtmp); |
mm | 49 | include/asm-alpha/mmu_context.h | struct mm_struct *mm, |
mm | 61 | include/asm-alpha/mmu_context.h | mm->context = asn; /* full version + asn */ |
mm | 79 | include/asm-alpha/mmu_context.h | struct mm_struct * mm = p->mm; |
mm | 81 | include/asm-alpha/mmu_context.h | if (mm) { |
mm | 84 | include/asm-alpha/mmu_context.h | if ((mm->context ^ asn) & ASN_VERSION_MASK) |
mm | 85 | include/asm-alpha/mmu_context.h | get_new_mmu_context(p, mm, asn); |
mm | 17 | include/asm-alpha/pgtable.h | #define flush_cache_mm(mm) do { } while (0) |
mm | 18 | include/asm-alpha/pgtable.h | #define flush_cache_range(mm, start, end) do { } while (0) |
mm | 50 | include/asm-alpha/pgtable.h | #define flush_tlb_current(mm) get_new_asn_and_reload(current, mm) |
mm | 51 | include/asm-alpha/pgtable.h | #define flush_tlb_other(mm) do { (mm)->context = 0; } while (0) |
mm | 60 | include/asm-alpha/pgtable.h | static inline void flush_tlb_current_page(struct mm_struct * mm, |
mm | 68 | include/asm-alpha/pgtable.h | flush_tlb_current(mm); |
mm | 79 | include/asm-alpha/pgtable.h | flush_tlb_current(current->mm); |
mm | 94 | include/asm-alpha/pgtable.h | static inline void flush_tlb_mm(struct mm_struct *mm) |
mm | 96 | include/asm-alpha/pgtable.h | if (mm != current->mm) |
mm | 97 | include/asm-alpha/pgtable.h | flush_tlb_other(mm); |
mm | 99 | include/asm-alpha/pgtable.h | flush_tlb_current(mm); |
mm | 113 | include/asm-alpha/pgtable.h | struct mm_struct * mm = vma->vm_mm; |
mm | 115 | include/asm-alpha/pgtable.h | if (mm != current->mm) |
mm | 116 | include/asm-alpha/pgtable.h | flush_tlb_other(mm); |
mm | 118 | include/asm-alpha/pgtable.h | flush_tlb_current_page(mm, vma, addr); |
mm | 125 | include/asm-alpha/pgtable.h | static inline void flush_tlb_range(struct mm_struct *mm, |
mm | 128 | include/asm-alpha/pgtable.h | flush_tlb_mm(mm); |
mm | 350 | include/asm-alpha/pgtable.h | extern inline pgd_t * pgd_offset(struct mm_struct * mm, unsigned long address) |
mm | 352 | include/asm-alpha/pgtable.h | return mm->pgd + ((address >> PGDIR_SHIFT) & (PTRS_PER_PAGE - 1)); |
mm | 24 | include/asm-i386/pgtable.h | #define flush_cache_mm(mm) do { } while (0) |
mm | 25 | include/asm-i386/pgtable.h | #define flush_cache_range(mm, start, end) do { } while (0) |
mm | 57 | include/asm-i386/pgtable.h | static inline void flush_tlb_mm(struct mm_struct *mm) |
mm | 59 | include/asm-i386/pgtable.h | if (mm == current->mm) |
mm | 66 | include/asm-i386/pgtable.h | if (vma->vm_mm == current->mm) |
mm | 70 | include/asm-i386/pgtable.h | static inline void flush_tlb_range(struct mm_struct *mm, |
mm | 73 | include/asm-i386/pgtable.h | if (mm == current->mm) |
mm | 106 | include/asm-i386/pgtable.h | if (current->mm->count == 1) /* just one copy of this mm */ |
mm | 116 | include/asm-i386/pgtable.h | static inline void flush_tlb_mm(struct mm_struct * mm) |
mm | 118 | include/asm-i386/pgtable.h | if (mm == current->mm && mm->count == 1) |
mm | 127 | include/asm-i386/pgtable.h | if (vma->vm_mm == current->mm && current->mm->count == 1) |
mm | 133 | include/asm-i386/pgtable.h | static inline void flush_tlb_range(struct mm_struct * mm, |
mm | 136 | include/asm-i386/pgtable.h | flush_tlb_mm(mm); |
mm | 147 | include/asm-i386/pgtable.h | static inline void flush_tlb_mm(struct mm_struct *mm) |
mm | 158 | include/asm-i386/pgtable.h | static inline void flush_tlb_range(struct mm_struct *mm, |
mm | 353 | include/asm-i386/pgtable.h | extern inline pgd_t * pgd_offset(struct mm_struct * mm, unsigned long address) |
mm | 355 | include/asm-i386/pgtable.h | return mm->pgd + (address >> PGDIR_SHIFT); |
mm | 34 | include/asm-m68k/pgtable.h | static inline void flush_tlb_mm(struct mm_struct *mm) |
mm | 36 | include/asm-m68k/pgtable.h | if (mm == current->mm) |
mm | 43 | include/asm-m68k/pgtable.h | if (vma->vm_mm == current->mm) |
mm | 47 | include/asm-m68k/pgtable.h | static inline void flush_tlb_range(struct mm_struct *mm, |
mm | 50 | include/asm-m68k/pgtable.h | if (mm == current->mm) |
mm | 304 | include/asm-m68k/pgtable.h | extern inline pgd_t * pgd_offset(struct mm_struct * mm, unsigned long address) |
mm | 306 | include/asm-m68k/pgtable.h | return mm->pgd + (address >> PGDIR_SHIFT); |
mm | 533 | include/asm-m68k/pgtable.h | #define flush_cache_mm(mm) flush_cache_all() |
mm | 534 | include/asm-m68k/pgtable.h | #define flush_cache_range(mm, start, end) flush_cache_all() |
mm | 539 | include/asm-m68k/pgtable.h | #define flush_cache_mm(mm) do { } while (0) |
mm | 540 | include/asm-m68k/pgtable.h | #define flush_cache_range(mm, start, end) do { } while (0) |
mm | 42 | include/asm-mips/pgtable.h | do { if ((mm_struct) == current->mm) invalidate(); } while (0) |
mm | 44 | include/asm-mips/pgtable.h | do { if ((mm_struct) == current->mm) invalidate(); } while (0) |
mm | 46 | include/asm-mips/pgtable.h | do { if ((mm_struct) == current->mm) invalidate(); } while (0) |
mm | 344 | include/asm-mips/pgtable.h | extern inline pgd_t * pgd_offset(struct mm_struct * mm, unsigned long address) |
mm | 346 | include/asm-mips/pgtable.h | return mm->pgd + (address >> PGDIR_SHIFT); |
mm | 288 | include/asm-ppc/pgtable.h | extern inline pgd_t * pgd_offset(struct mm_struct * mm, unsigned long address) |
mm | 290 | include/asm-ppc/pgtable.h | return mm->pgd + (address >> PGDIR_SHIFT); |
mm | 12 | include/asm-sparc/mmu_context.h | #define init_new_context(mm) ((mm)->context = NO_CONTEXT) |
mm | 271 | include/asm-sparc/pgtable.h | extern void smp_flush_cache_mm(struct mm_struct *mm); |
mm | 272 | include/asm-sparc/pgtable.h | extern void smp_flush_cache_range(struct mm_struct *mm, |
mm | 278 | include/asm-sparc/pgtable.h | extern void smp_flush_tlb_mm(struct mm_struct *mm); |
mm | 279 | include/asm-sparc/pgtable.h | extern void smp_flush_tlb_range(struct mm_struct *mm, |
mm | 282 | include/asm-sparc/pgtable.h | extern void smp_flush_tlb_page(struct vm_area_struct *mm, unsigned long page); |
mm | 255 | include/linux/mm.h | extern void free_page_tables(struct mm_struct * mm); |
mm | 260 | include/linux/mm.h | extern int zap_page_range(struct mm_struct *mm, unsigned long address, unsigned long size); |
mm | 338 | include/linux/mm.h | if (task->mm) { |
mm | 339 | include/linux/mm.h | struct vm_area_struct * tree = task->mm->mmap_avl; |
mm | 246 | include/linux/sched.h | struct mm_struct *mm; |
mm | 430 | ipc/shm.c | current->mm->total_vm += (shmd->vm_end - shmd->vm_start) >> PAGE_SHIFT; |
mm | 497 | ipc/shm.c | if (addr < current->mm->start_stack && |
mm | 498 | ipc/shm.c | addr > current->mm->start_stack - PAGE_SIZE*(shp->shm_npages + 4)) |
mm | 526 | ipc/shm.c | shmd->vm_mm = current->mm; |
mm | 600 | ipc/shm.c | for (shmd = current->mm->mmap; shmd; shmd = shmdnext) { |
mm | 433 | kernel/exit.c | struct mm_struct * mm = current->mm; |
mm | 436 | kernel/exit.c | if (mm != &init_mm) { |
mm | 437 | kernel/exit.c | flush_cache_mm(mm); |
mm | 438 | kernel/exit.c | flush_tlb_mm(mm); |
mm | 439 | kernel/exit.c | current->mm = &init_mm; |
mm | 444 | kernel/exit.c | if (!--mm->count) { |
mm | 445 | kernel/exit.c | exit_mmap(mm); |
mm | 446 | kernel/exit.c | free_page_tables(mm); |
mm | 447 | kernel/exit.c | kfree(mm); |
mm | 79 | kernel/fork.c | static inline int dup_mmap(struct mm_struct * mm) |
mm | 83 | kernel/fork.c | mm->mmap = NULL; |
mm | 84 | kernel/fork.c | p = &mm->mmap; |
mm | 85 | kernel/fork.c | for (mpnt = current->mm->mmap ; mpnt ; mpnt = mpnt->vm_next) { |
mm | 88 | kernel/fork.c | exit_mmap(mm); |
mm | 93 | kernel/fork.c | tmp->vm_mm = mm; |
mm | 104 | kernel/fork.c | if (copy_page_range(mm, current->mm, tmp)) { |
mm | 105 | kernel/fork.c | exit_mmap(mm); |
mm | 111 | kernel/fork.c | build_mmap_avl(mm); |
mm | 118 | kernel/fork.c | struct mm_struct * mm = kmalloc(sizeof(*tsk->mm), GFP_KERNEL); |
mm | 119 | kernel/fork.c | if (!mm) |
mm | 121 | kernel/fork.c | *mm = *current->mm; |
mm | 122 | kernel/fork.c | mm->count = 1; |
mm | 123 | kernel/fork.c | mm->def_flags = 0; |
mm | 124 | kernel/fork.c | tsk->mm = mm; |
mm | 130 | kernel/fork.c | if (dup_mmap(mm)) { |
mm | 131 | kernel/fork.c | free_page_tables(mm); |
mm | 136 | kernel/fork.c | SET_PAGE_DIR(tsk, current->mm->pgd); |
mm | 137 | kernel/fork.c | current->mm->count++; |
mm | 552 | kernel/sys.c | if (brk < current->mm->end_code) |
mm | 553 | kernel/sys.c | return current->mm->brk; |
mm | 555 | kernel/sys.c | oldbrk = PAGE_ALIGN(current->mm->brk); |
mm | 557 | kernel/sys.c | return current->mm->brk = brk; |
mm | 562 | kernel/sys.c | if (brk <= current->mm->brk) { |
mm | 563 | kernel/sys.c | current->mm->brk = brk; |
mm | 573 | kernel/sys.c | if (brk - current->mm->end_code > rlim) |
mm | 574 | kernel/sys.c | return current->mm->brk; |
mm | 579 | kernel/sys.c | return current->mm->brk; |
mm | 593 | kernel/sys.c | return current->mm->brk; |
mm | 595 | kernel/sys.c | freepages += current->mm->rss; |
mm | 598 | kernel/sys.c | return current->mm->brk; |
mm | 603 | kernel/sys.c | current->mm->brk = brk; |
mm | 755 | mm/filemap.c | if (offset >= inode->i_size && (area->vm_flags & VM_SHARED) && area->vm_mm == current->mm) |
mm | 990 | mm/filemap.c | dir = pgd_offset(current->mm, address); |
mm | 133 | mm/memory.c | page_dir = tsk->mm->pgd; |
mm | 138 | mm/memory.c | flush_cache_mm(tsk->mm); |
mm | 141 | mm/memory.c | flush_tlb_mm(tsk->mm); |
mm | 150 | mm/memory.c | void free_page_tables(struct mm_struct * mm) |
mm | 155 | mm/memory.c | page_dir = mm->pgd; |
mm | 172 | mm/memory.c | flush_cache_mm(tsk->mm); |
mm | 175 | mm/memory.c | flush_tlb_mm(tsk->mm); |
mm | 177 | mm/memory.c | tsk->mm->pgd = new_pg; |
mm | 310 | mm/memory.c | if (current->mm->rss <= 0) |
mm | 312 | mm/memory.c | current->mm->rss--; |
mm | 371 | mm/memory.c | int zap_page_range(struct mm_struct *mm, unsigned long address, unsigned long size) |
mm | 376 | mm/memory.c | dir = pgd_offset(mm, address); |
mm | 377 | mm/memory.c | flush_cache_range(mm, end - size, end); |
mm | 383 | mm/memory.c | flush_tlb_range(mm, end - size, end); |
mm | 432 | mm/memory.c | dir = pgd_offset(current->mm, address); |
mm | 433 | mm/memory.c | flush_cache_range(current->mm, beg, end); |
mm | 445 | mm/memory.c | flush_tlb_range(current->mm, beg, end); |
mm | 504 | mm/memory.c | dir = pgd_offset(current->mm, from); |
mm | 505 | mm/memory.c | flush_cache_range(current->mm, beg, end); |
mm | 517 | mm/memory.c | flush_tlb_range(current->mm, beg, end); |
mm | 549 | mm/memory.c | pgd = pgd_offset(tsk->mm,address); |
mm | 857 | mm/memory.c | pgd = pgd_offset(tsk->mm, address); |
mm | 194 | mm/mlock.c | locked += current->mm->locked_vm; |
mm | 230 | mm/mlock.c | current->mm->def_flags = def_flags; |
mm | 233 | mm/mlock.c | for (vma = current->mm->mmap; vma ; vma = vma->vm_next) { |
mm | 257 | mm/mlock.c | if (current->mm->total_vm > lock_limit) |
mm | 262 | mm/mlock.c | if (current->mm->total_vm > MAP_NR(high_memory)/2) |
mm | 58 | mm/mmap.c | if (current->mm->def_flags & VM_LOCKED) { |
mm | 59 | mm/mmap.c | unsigned long locked = current->mm->locked_vm << PAGE_SHIFT; |
mm | 126 | mm/mmap.c | vma->vm_mm = current->mm; |
mm | 131 | mm/mmap.c | vma->vm_flags |= current->mm->def_flags; |
mm | 175 | mm/mmap.c | current->mm->total_vm += len >> PAGE_SHIFT; |
mm | 178 | mm/mmap.c | current->mm->locked_vm += len >> PAGE_SHIFT; |
mm | 575 | mm/mmap.c | avl_checkheights(task->mm->mmap_avl); |
mm | 576 | mm/mmap.c | avl_checkorder(task->mm->mmap_avl); |
mm | 701 | mm/mmap.c | avl_neighbours(mpnt, current->mm->mmap_avl, &prev, &next); |
mm | 705 | mm/mmap.c | npp = (prev ? &prev->vm_next : ¤t->mm->mmap); |
mm | 711 | mm/mmap.c | avl_remove(mpnt, ¤t->mm->mmap_avl); |
mm | 737 | mm/mmap.c | zap_page_range(current->mm, st, end-st); |
mm | 742 | mm/mmap.c | zap_page_range(current->mm, addr, len); |
mm | 747 | mm/mmap.c | void build_mmap_avl(struct mm_struct * mm) |
mm | 751 | mm/mmap.c | mm->mmap_avl = NULL; |
mm | 752 | mm/mmap.c | for (vma = mm->mmap; vma; vma = vma->vm_next) |
mm | 753 | mm/mmap.c | avl_insert(vma, &mm->mmap_avl); |
mm | 757 | mm/mmap.c | void exit_mmap(struct mm_struct * mm) |
mm | 761 | mm/mmap.c | mpnt = mm->mmap; |
mm | 762 | mm/mmap.c | mm->mmap = NULL; |
mm | 763 | mm/mmap.c | mm->mmap_avl = NULL; |
mm | 764 | mm/mmap.c | mm->rss = 0; |
mm | 765 | mm/mmap.c | mm->total_vm = 0; |
mm | 766 | mm/mmap.c | mm->locked_vm = 0; |
mm | 776 | mm/mmap.c | zap_page_range(mm, mpnt->vm_start, mpnt->vm_end-mpnt->vm_start); |
mm | 796 | mm/mmap.c | p = &t->mm->mmap; |
mm | 809 | mm/mmap.c | avl_insert_neighbours(vmp, &t->mm->mmap_avl, &prev, &next); |
mm | 810 | mm/mmap.c | if ((prev ? prev->vm_next : t->mm->mmap) != next) |
mm | 815 | mm/mmap.c | t->mm->mmap = vmp; |
mm | 871 | mm/mmap.c | avl_neighbours(mpnt, task->mm->mmap_avl, &prev, &next); |
mm | 915 | mm/mmap.c | avl_remove(mpnt, &task->mm->mmap_avl); |
mm | 77 | mm/mprotect.c | dir = pgd_offset(current->mm, start); |
mm | 78 | mm/mprotect.c | flush_cache_range(current->mm, beg, end); |
mm | 84 | mm/mprotect.c | flush_tlb_range(current->mm, beg, end); |
mm | 22 | mm/mremap.c | static inline pte_t *get_one_pte(struct mm_struct *mm, unsigned long addr) |
mm | 28 | mm/mremap.c | pgd = pgd_offset(mm, addr); |
mm | 53 | mm/mremap.c | static inline pte_t *alloc_one_pte(struct mm_struct *mm, unsigned long addr) |
mm | 58 | mm/mremap.c | pmd = pmd_alloc(pgd_offset(mm, addr), addr); |
mm | 80 | mm/mremap.c | static int move_one_page(struct mm_struct *mm, unsigned long old_addr, unsigned long new_addr) |
mm | 85 | mm/mremap.c | src = get_one_pte(mm, old_addr); |
mm | 87 | mm/mremap.c | error = copy_one_pte(src, alloc_one_pte(mm, new_addr)); |
mm | 91 | mm/mremap.c | static int move_page_tables(struct mm_struct * mm, |
mm | 96 | mm/mremap.c | flush_cache_range(mm, old_addr, old_addr + len); |
mm | 97 | mm/mremap.c | flush_tlb_range(mm, old_addr, old_addr + len); |
mm | 106 | mm/mremap.c | if (move_one_page(mm, old_addr + offset, new_addr + offset)) |
mm | 119 | mm/mremap.c | flush_cache_range(mm, new_addr, new_addr + len); |
mm | 121 | mm/mremap.c | move_one_page(mm, new_addr + offset, old_addr + offset); |
mm | 122 | mm/mremap.c | flush_tlb_range(mm, new_addr, new_addr + len); |
mm | 123 | mm/mremap.c | zap_page_range(mm, new_addr, new_addr + len); |
mm | 137 | mm/mremap.c | if (new_addr && !move_page_tables(current->mm, new_addr, addr, old_len)) { |
mm | 192 | mm/mremap.c | unsigned long locked = current->mm->locked_vm << PAGE_SHIFT; |
mm | 207 | mm/mremap.c | current->mm->total_vm += pages; |
mm | 209 | mm/mremap.c | current->mm->locked_vm += pages; |
mm | 266 | mm/swapfile.c | static int unuse_process(struct mm_struct * mm, unsigned int type, unsigned long page) |
mm | 273 | mm/swapfile.c | if (!mm || mm == &init_mm) |
mm | 275 | mm/swapfile.c | vma = mm->mmap; |
mm | 277 | mm/swapfile.c | pgd_t * pgd = pgd_offset(mm, vma->vm_start); |
mm | 301 | mm/swapfile.c | if (unuse_process(p->mm, type, page)) { |
mm | 35 | mm/vmalloc.c | if (!p->mm) |
mm | 37 | mm/vmalloc.c | *pgd_offset(p->mm,address) = entry; |
mm | 263 | mm/vmscan.c | int result = swap_out_vma(p, vma, pgd_offset(p->mm, address), address, dma, wait); |
mm | 298 | mm/vmscan.c | if (p && p->swappable && p->mm->rss) |
mm | 310 | mm/vmscan.c | p->swap_cnt = AGE_CLUSTER_SIZE(p->mm->rss); |