tag | line | file | source code |
mm | 143 | arch/alpha/kernel/process.c | p->mm->context = 0; |
mm | 155 | arch/alpha/kernel/process.c | dump->start_code = current->mm->start_code; |
mm | 156 | arch/alpha/kernel/process.c | dump->start_data = current->mm->start_data; |
mm | 158 | arch/alpha/kernel/process.c | dump->u_tsize = (current->mm->end_code - dump->start_code) >> PAGE_SHIFT; |
mm | 159 | arch/alpha/kernel/process.c | dump->u_dsize = (current->mm->brk + (PAGE_SIZE - 1) - dump->start_data) >> PAGE_SHIFT; |
mm | 161 | arch/alpha/kernel/process.c | (current->mm->start_stack - dump->start_stack + PAGE_SIZE - 1) >> PAGE_SHIFT; |
mm | 31 | arch/alpha/mm/fault.c | void get_new_asn_and_reload(struct task_struct *tsk, struct mm_struct *mm) |
mm | 33 | arch/alpha/mm/fault.c | get_new_mmu_context(tsk, mm, asn_cache); |
mm | 368 | arch/i386/kernel/process.c | dump->u_tsize = ((unsigned long) current->mm->end_code) >> PAGE_SHIFT; |
mm | 369 | arch/i386/kernel/process.c | dump->u_dsize = ((unsigned long) (current->mm->brk + (PAGE_SIZE-1))) >> PAGE_SHIFT; |
mm | 143 | arch/i386/kernel/setup.c | init_task.mm->start_code = TASK_SIZE; |
mm | 144 | arch/i386/kernel/setup.c | init_task.mm->end_code = TASK_SIZE + (unsigned long) &_etext; |
mm | 145 | arch/i386/kernel/setup.c | init_task.mm->end_data = TASK_SIZE + (unsigned long) &_edata; |
mm | 146 | arch/i386/kernel/setup.c | init_task.mm->brk = TASK_SIZE + (unsigned long) &_end; |
mm | 77 | arch/i386/kernel/vm86.c | pgd = pgd_offset(tsk->mm, 0xA0000); |
mm | 169 | arch/m68k/kernel/process.c | dump->u_tsize = ((unsigned long) current->mm->end_code) >> PAGE_SHIFT; |
mm | 170 | arch/m68k/kernel/process.c | dump->u_dsize = ((unsigned long) (current->mm->brk + |
mm | 124 | arch/m68k/kernel/setup.c | init_task.mm->start_code = 0; |
mm | 125 | arch/m68k/kernel/setup.c | init_task.mm->end_code = (unsigned long) &_etext; |
mm | 126 | arch/m68k/kernel/setup.c | init_task.mm->end_data = (unsigned long) &_edata; |
mm | 127 | arch/m68k/kernel/setup.c | init_task.mm->brk = (unsigned long) &_end; |
mm | 522 | arch/m68k/mm/memory.c | void flush_cache_mm(struct mm_struct *mm){ |
mm | 524 | arch/m68k/mm/memory.c | if (mm == current->mm) |
mm | 528 | arch/m68k/mm/memory.c | void flush_cache_range(struct mm_struct *mm, unsigned long start, |
mm | 530 | arch/m68k/mm/memory.c | if (mm == current->mm) |
mm | 64 | arch/ppc/mm/fault.c | dir = pgd_offset(current->mm, regs->dar & PAGE_MASK); |
mm | 87 | arch/ppc/mm/fault.c | printk("beforefault: pgd[0] = %x[%x]\n",current->mm->pgd,*(current->mm->pgd)); |
mm | 91 | arch/ppc/mm/fault.c | printk("handled: pgd[0] = %x[%x]\n",current->mm->pgd,*(current->mm->pgd)); |
mm | 119 | arch/ppc/mm/fault.c | dir = pgd_offset(current->mm, regs->dar & PAGE_MASK); |
mm | 135 | arch/ppc/mm/fault.c | current->mm, current->mm->pgd, |
mm | 136 | arch/ppc/mm/fault.c | current->mm->pgd+((regs->dar&PAGE_MASK) >> PGDIR_SHIFT)); |
mm | 142 | arch/ppc/mm/fault.c | current->mm->pgd,dir,*dir,pmd,*pmd,pte); |
mm | 164 | arch/ppc/mm/fault.c | regs->dar, regs->dsisr, regs->nip, current->mm); |
mm | 190 | arch/ppc/mm/fault.c | for (vma = current->mm->mmap ; ; vma = vma->vm_next) |
mm | 195 | arch/ppc/mm/fault.c | regs->nip, current,current->pid,current->mm,current->mm->mmap, address, error_code); |
mm | 213 | arch/ppc/mm/fault.c | printk("stack: gpr[1]=%x ip = %x; current=%x[%d]; mm=%x; mmap=%x; address = %x error_code = %x\n",regs->gpr[1],regs->nip, current,current->pid,current->mm,current->mm->mmap, address, error_code); |
mm | 222 | arch/ppc/mm/fault.c | printk("stack2: gpr[1]=%x ip = %x; current=%x[%d]; mm=%x; mmap=%x; address = %x error_code = %x\n",regs->gpr[1],regs->nip, current,current->pid,current->mm,current->mm->mmap, address, error_code); |
mm | 242 | arch/ppc/mm/fault.c | current,address,vma->vm_flags,current->mm,vma,vma->vm_start,vma->vm_end); |
mm | 296 | arch/ppc/mm/fault.c | current,address,vma->vm_flags,current->mm,vma,vma->vm_start,vma->vm_end); |
mm | 309 | arch/ppc/mm/fault.c | dir = pgd_offset(current->mm, address & PAGE_MASK); |
mm | 730 | arch/ppc/mm/init.c | pgd = pgd_offset(current->mm, 0); |
mm | 295 | arch/sparc/kernel/process.c | dump->u_tsize = (((unsigned long) current->mm->end_code) - |
mm | 296 | arch/sparc/kernel/process.c | ((unsigned long) current->mm->start_code)) & ~(PAGE_SIZE - 1); |
mm | 297 | arch/sparc/kernel/process.c | dump->u_dsize = ((unsigned long) (current->mm->brk + (PAGE_SIZE-1))); |
mm | 300 | arch/sparc/kernel/setup.c | init_task.mm->mmap->vm_page_prot = PAGE_SHARED; |
mm | 301 | arch/sparc/kernel/setup.c | init_task.mm->mmap->vm_start = KERNBASE; |
mm | 302 | arch/sparc/kernel/setup.c | init_task.mm->mmap->vm_end = *memory_end_p; |
mm | 172 | arch/sparc/kernel/smp.c | current->mm->mmap->vm_page_prot = PAGE_SHARED; |
mm | 173 | arch/sparc/kernel/smp.c | current->mm->mmap->vm_start = KERNBASE; |
mm | 174 | arch/sparc/kernel/smp.c | current->mm->mmap->vm_end = init_task.mm->mmap->vm_end; |
mm | 529 | arch/sparc/kernel/smp.c | void smp_flush_cache_mm(struct mm_struct *mm) |
mm | 531 | arch/sparc/kernel/smp.c | if(mm->context != NO_CONTEXT) |
mm | 532 | arch/sparc/kernel/smp.c | xc1((smpfunc_t) local_flush_cache_mm, (unsigned long) mm); |
mm | 535 | arch/sparc/kernel/smp.c | void smp_flush_tlb_mm(struct mm_struct *mm) |
mm | 537 | arch/sparc/kernel/smp.c | if(mm->context != NO_CONTEXT) |
mm | 538 | arch/sparc/kernel/smp.c | xc1((smpfunc_t) local_flush_tlb_mm, (unsigned long) mm); |
mm | 541 | arch/sparc/kernel/smp.c | void smp_flush_cache_range(struct mm_struct *mm, unsigned long start, |
mm | 544 | arch/sparc/kernel/smp.c | if(mm->context != NO_CONTEXT) |
mm | 545 | arch/sparc/kernel/smp.c | xc3((smpfunc_t) local_flush_cache_range, (unsigned long) mm, |
mm | 549 | arch/sparc/kernel/smp.c | void smp_flush_tlb_range(struct mm_struct *mm, unsigned long start, |
mm | 552 | arch/sparc/kernel/smp.c | if(mm->context != NO_CONTEXT) |
mm | 553 | arch/sparc/kernel/smp.c | xc3((smpfunc_t) local_flush_tlb_range, (unsigned long) mm, |
mm | 128 | arch/sparc/kernel/sys_sunos.c | if (brk < current->mm->end_code) |
mm | 132 | arch/sparc/kernel/sys_sunos.c | oldbrk = PAGE_ALIGN(current->mm->brk); |
mm | 134 | arch/sparc/kernel/sys_sunos.c | current->mm->brk = brk; |
mm | 141 | arch/sparc/kernel/sys_sunos.c | if (brk <= current->mm->brk) { |
mm | 142 | arch/sparc/kernel/sys_sunos.c | current->mm->brk = brk; |
mm | 152 | arch/sparc/kernel/sys_sunos.c | if (brk - current->mm->end_code > rlim) |
mm | 178 | arch/sparc/kernel/sys_sunos.c | current->mm->brk = brk; |
mm | 190 | arch/sparc/kernel/sys_sunos.c | error = sunos_brk(((int) current->mm->brk) + increment); |
mm | 194 | arch/sparc/kernel/sys_sunos.c | return current->mm->brk; |
mm | 290 | arch/sparc/kernel/sys_sunos.c | pgdp = pgd_offset(current->mm, addr); |
mm | 208 | arch/sparc/mm/fault.c | (unsigned long) current->mm->context); |
mm | 210 | arch/sparc/mm/fault.c | (unsigned long) current->mm->pgd); |
mm | 24 | arch/sparc/mm/generic.c | if (current->mm->rss <= 0) |
mm | 26 | arch/sparc/mm/generic.c | current->mm->rss--; |
mm | 88 | arch/sparc/mm/generic.c | dir = pgd_offset(current->mm, from); |
mm | 89 | arch/sparc/mm/generic.c | flush_cache_range(current->mm, beg, end); |
mm | 101 | arch/sparc/mm/generic.c | flush_tlb_range(current->mm, beg, end); |
mm | 218 | arch/sparc/mm/srmmu.c | static pgd_t *srmmu_pgd_offset(struct mm_struct * mm, unsigned long address) |
mm | 220 | arch/sparc/mm/srmmu.c | return mm->pgd + ((address >> SRMMU_PGDIR_SHIFT) & (SRMMU_PTRS_PER_PGD - 1)); |
mm | 238 | arch/sparc/mm/srmmu.c | if(tsk->mm->context != NO_CONTEXT) |
mm | 239 | arch/sparc/mm/srmmu.c | ctxd_set(&srmmu_context_table[tsk->mm->context], pgdp); |
mm | 244 | arch/sparc/mm/srmmu.c | pgd_t *pgdp = srmmu_pgd_offset(init_task.mm, addr); |
mm | 255 | arch/sparc/mm/srmmu.c | pgd_t *pgdp = srmmu_pgd_offset(init_task.mm, addr); |
mm | 429 | arch/sparc/mm/srmmu.c | static inline void alloc_context(struct mm_struct *mm) |
mm | 437 | arch/sparc/mm/srmmu.c | mm->context = ctxp->ctx_number; |
mm | 438 | arch/sparc/mm/srmmu.c | ctxp->ctx_mm = mm; |
mm | 442 | arch/sparc/mm/srmmu.c | if(ctxp->ctx_mm == current->mm) |
mm | 451 | arch/sparc/mm/srmmu.c | ctxp->ctx_mm = mm; |
mm | 452 | arch/sparc/mm/srmmu.c | mm->context = ctxp->ctx_number; |
mm | 465 | arch/sparc/mm/srmmu.c | if(tsk->mm->context == NO_CONTEXT) { |
mm | 466 | arch/sparc/mm/srmmu.c | alloc_context(tsk->mm); |
mm | 467 | arch/sparc/mm/srmmu.c | ctxd_set(&srmmu_context_table[tsk->mm->context], tsk->mm->pgd); |
mm | 469 | arch/sparc/mm/srmmu.c | srmmu_set_context(tsk->mm->context); |
mm | 481 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, virt_addr); |
mm | 559 | arch/sparc/mm/srmmu.c | static void tsunami_flush_cache_mm(struct mm_struct *mm) |
mm | 562 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 572 | arch/sparc/mm/srmmu.c | static void tsunami_flush_cache_range(struct mm_struct *mm, unsigned long start, unsigned long end) |
mm | 575 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 588 | arch/sparc/mm/srmmu.c | struct mm_struct *mm = vma->vm_mm; |
mm | 589 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 636 | arch/sparc/mm/srmmu.c | static void tsunami_flush_tlb_mm(struct mm_struct *mm) |
mm | 640 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 649 | arch/sparc/mm/srmmu.c | static void tsunami_flush_tlb_range(struct mm_struct *mm, unsigned long start, unsigned long end) |
mm | 653 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 665 | arch/sparc/mm/srmmu.c | struct mm_struct *mm = vma->vm_mm; |
mm | 668 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 672 | arch/sparc/mm/srmmu.c | srmmu_set_context(mm->context); |
mm | 698 | arch/sparc/mm/srmmu.c | static void swift_flush_cache_mm(struct mm_struct *mm) |
mm | 701 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 710 | arch/sparc/mm/srmmu.c | static void swift_flush_cache_range(struct mm_struct *mm, unsigned long start, unsigned long end) |
mm | 713 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 725 | arch/sparc/mm/srmmu.c | struct mm_struct *mm = vma->vm_mm; |
mm | 726 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 759 | arch/sparc/mm/srmmu.c | static void swift_flush_tlb_mm(struct mm_struct *mm) |
mm | 763 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) |
mm | 768 | arch/sparc/mm/srmmu.c | static void swift_flush_tlb_range(struct mm_struct *mm, unsigned long start, unsigned long end) |
mm | 772 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) |
mm | 780 | arch/sparc/mm/srmmu.c | struct mm_struct *mm = vma->vm_mm; |
mm | 781 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) |
mm | 813 | arch/sparc/mm/srmmu.c | static void viking_flush_cache_mm(struct mm_struct *mm) |
mm | 816 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 824 | arch/sparc/mm/srmmu.c | static void viking_flush_cache_range(struct mm_struct *mm, unsigned long start, unsigned long end) |
mm | 827 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 838 | arch/sparc/mm/srmmu.c | struct mm_struct *mm = vma->vm_mm; |
mm | 839 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 946 | arch/sparc/mm/srmmu.c | static void viking_flush_tlb_mm(struct mm_struct *mm) |
mm | 952 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 955 | arch/sparc/mm/srmmu.c | srmmu_set_context(mm->context); |
mm | 963 | arch/sparc/mm/srmmu.c | static void viking_flush_tlb_range(struct mm_struct *mm, unsigned long start, unsigned long end) |
mm | 969 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 972 | arch/sparc/mm/srmmu.c | srmmu_set_context(mm->context); |
mm | 987 | arch/sparc/mm/srmmu.c | struct mm_struct *mm = vma->vm_mm; |
mm | 991 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 994 | arch/sparc/mm/srmmu.c | srmmu_set_context(mm->context); |
mm | 1015 | arch/sparc/mm/srmmu.c | static void cypress_flush_tlb_mm(struct mm_struct *mm) |
mm | 1021 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 1024 | arch/sparc/mm/srmmu.c | srmmu_set_context(mm->context); |
mm | 1032 | arch/sparc/mm/srmmu.c | static void cypress_flush_tlb_range(struct mm_struct *mm, unsigned long start, unsigned long end) |
mm | 1038 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 1041 | arch/sparc/mm/srmmu.c | srmmu_set_context(mm->context); |
mm | 1056 | arch/sparc/mm/srmmu.c | struct mm_struct *mm = vma->vm_mm; |
mm | 1060 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 1063 | arch/sparc/mm/srmmu.c | srmmu_set_context(mm->context); |
mm | 1079 | arch/sparc/mm/srmmu.c | static void hypersparc_flush_cache_mm(struct mm_struct *mm) |
mm | 1082 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 1092 | arch/sparc/mm/srmmu.c | static void hypersparc_flush_cache_range(struct mm_struct *mm, unsigned long start, unsigned long end) |
mm | 1095 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 1110 | arch/sparc/mm/srmmu.c | struct mm_struct *mm = vma->vm_mm; |
mm | 1115 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 1119 | arch/sparc/mm/srmmu.c | srmmu_set_context(mm->context); |
mm | 1167 | arch/sparc/mm/srmmu.c | static void hypersparc_flush_tlb_mm(struct mm_struct *mm) |
mm | 1173 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 1177 | arch/sparc/mm/srmmu.c | srmmu_set_context(mm->context); |
mm | 1186 | arch/sparc/mm/srmmu.c | static void hypersparc_flush_tlb_range(struct mm_struct *mm, unsigned long start, unsigned long end) |
mm | 1192 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 1196 | arch/sparc/mm/srmmu.c | srmmu_set_context(mm->context); |
mm | 1211 | arch/sparc/mm/srmmu.c | struct mm_struct *mm = vma->vm_mm; |
mm | 1216 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 1220 | arch/sparc/mm/srmmu.c | srmmu_set_context(mm->context); |
mm | 1242 | arch/sparc/mm/srmmu.c | if(tsk->mm->context != NO_CONTEXT) { |
mm | 1244 | arch/sparc/mm/srmmu.c | ctxd_set(&srmmu_context_table[tsk->mm->context], pgdp); |
mm | 1273 | arch/sparc/mm/srmmu.c | if(tsk->mm->context == NO_CONTEXT) { |
mm | 1274 | arch/sparc/mm/srmmu.c | alloc_context(tsk->mm); |
mm | 1275 | arch/sparc/mm/srmmu.c | ctxd_set(&srmmu_context_table[tsk->mm->context], tsk->mm->pgd); |
mm | 1277 | arch/sparc/mm/srmmu.c | srmmu_set_context(tsk->mm->context); |
mm | 1310 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, start); |
mm | 1542 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, start); |
mm | 1592 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, start); |
mm | 1632 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, start); |
mm | 1647 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, first); |
mm | 1670 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, KERNBASE); |
mm | 1686 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, start); |
mm | 1730 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, start); |
mm | 1765 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, start); |
mm | 1884 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, ptables_start); |
mm | 1890 | arch/sparc/mm/srmmu.c | pgdp = srmmu_pgd_offset(init_task.mm, (unsigned long)swapper_pg_dir); |
mm | 1942 | arch/sparc/mm/srmmu.c | struct mm_struct *mm = current->mm; |
mm | 1944 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 1945 | arch/sparc/mm/srmmu.c | flush_cache_mm(mm); |
mm | 1946 | arch/sparc/mm/srmmu.c | ctxd_set(&srmmu_context_table[mm->context], swapper_pg_dir); |
mm | 1947 | arch/sparc/mm/srmmu.c | flush_tlb_mm(mm); |
mm | 1948 | arch/sparc/mm/srmmu.c | ctx_old = ctx_list_pool + mm->context; |
mm | 1951 | arch/sparc/mm/srmmu.c | mm->context = NO_CONTEXT; |
mm | 1958 | arch/sparc/mm/srmmu.c | alloc_context(current->mm); |
mm | 1959 | arch/sparc/mm/srmmu.c | ctxd_set(&srmmu_context_table[current->mm->context], current->mm->pgd); |
mm | 1960 | arch/sparc/mm/srmmu.c | srmmu_set_context(current->mm->context); |
mm | 1967 | arch/sparc/mm/srmmu.c | struct mm_struct *mm = current->mm; |
mm | 1969 | arch/sparc/mm/srmmu.c | if(mm->context != NO_CONTEXT) { |
mm | 1975 | arch/sparc/mm/srmmu.c | flush_cache_mm(mm); |
mm | 1976 | arch/sparc/mm/srmmu.c | ctxd_set(&srmmu_context_table[mm->context], swapper_pg_dir); |
mm | 1977 | arch/sparc/mm/srmmu.c | flush_tlb_mm(mm); |
mm | 1978 | arch/sparc/mm/srmmu.c | ctx_old = ctx_list_pool + mm->context; |
mm | 1981 | arch/sparc/mm/srmmu.c | mm->context = NO_CONTEXT; |
mm | 1988 | arch/sparc/mm/srmmu.c | alloc_context(current->mm); |
mm | 1989 | arch/sparc/mm/srmmu.c | flush_cache_mm(current->mm); |
mm | 1990 | arch/sparc/mm/srmmu.c | ctxd_set(&srmmu_context_table[current->mm->context], current->mm->pgd); |
mm | 1991 | arch/sparc/mm/srmmu.c | srmmu_set_context(current->mm->context); |
mm | 941 | arch/sparc/mm/sun4c.c | sun4c_kstack_vma.vm_mm = init_task.mm; |
mm | 968 | arch/sparc/mm/sun4c.c | static void sun4c_flush_cache_mm(struct mm_struct *mm) |
mm | 974 | arch/sparc/mm/sun4c.c | if(mm->context != NO_CONTEXT) { |
mm | 979 | arch/sparc/mm/sun4c.c | sun4c_set_context(mm->context); |
mm | 988 | arch/sparc/mm/sun4c.c | static void sun4c_flush_cache_range(struct mm_struct *mm, unsigned long start, unsigned long end) |
mm | 994 | arch/sparc/mm/sun4c.c | if(mm->context != NO_CONTEXT) { |
mm | 1005 | arch/sparc/mm/sun4c.c | sun4c_set_context(mm->context); |
mm | 1036 | arch/sparc/mm/sun4c.c | struct mm_struct *mm = vma->vm_mm; |
mm | 1042 | arch/sparc/mm/sun4c.c | if(mm->context != NO_CONTEXT) { |
mm | 1047 | arch/sparc/mm/sun4c.c | sun4c_set_context(mm->context); |
mm | 1090 | arch/sparc/mm/sun4c.c | static void sun4c_flush_tlb_mm(struct mm_struct *mm) |
mm | 1097 | arch/sparc/mm/sun4c.c | if(mm->context != NO_CONTEXT) { |
mm | 1099 | arch/sparc/mm/sun4c.c | crp = &sun4c_context_ring[mm->context]; |
mm | 1101 | arch/sparc/mm/sun4c.c | ctx = mm->context; |
mm | 1103 | arch/sparc/mm/sun4c.c | sun4c_set_context(mm->context); |
mm | 1116 | arch/sparc/mm/sun4c.c | static void sun4c_flush_tlb_range(struct mm_struct *mm, unsigned long start, unsigned long end) |
mm | 1122 | arch/sparc/mm/sun4c.c | if(mm->context == NO_CONTEXT) |
mm | 1127 | arch/sparc/mm/sun4c.c | sun4c_set_context(mm->context); |
mm | 1135 | arch/sparc/mm/sun4c.c | free_user_entry(mm->context, this_entry); |
mm | 1144 | arch/sparc/mm/sun4c.c | struct mm_struct *mm = vma->vm_mm; |
mm | 1148 | arch/sparc/mm/sun4c.c | if(mm->context != NO_CONTEXT) { |
mm | 1151 | arch/sparc/mm/sun4c.c | sun4c_set_context(mm->context); |
mm | 1188 | arch/sparc/mm/sun4c.c | static inline void sun4c_alloc_context(struct mm_struct *mm) |
mm | 1196 | arch/sparc/mm/sun4c.c | mm->context = ctxp->ctx_number; |
mm | 1197 | arch/sparc/mm/sun4c.c | ctxp->ctx_mm = mm; |
mm | 1201 | arch/sparc/mm/sun4c.c | if(ctxp->ctx_mm == current->mm) |
mm | 1208 | arch/sparc/mm/sun4c.c | ctxp->ctx_mm = mm; |
mm | 1209 | arch/sparc/mm/sun4c.c | mm->context = ctxp->ctx_number; |
mm | 1238 | arch/sparc/mm/sun4c.c | if(tsk->mm->context == NO_CONTEXT) |
mm | 1239 | arch/sparc/mm/sun4c.c | sun4c_alloc_context(tsk->mm); |
mm | 1241 | arch/sparc/mm/sun4c.c | sun4c_set_context(tsk->mm->context); |
mm | 1247 | arch/sparc/mm/sun4c.c | sun4c_alloc_context(current->mm); |
mm | 1248 | arch/sparc/mm/sun4c.c | sun4c_set_context(current->mm->context); |
mm | 1255 | arch/sparc/mm/sun4c.c | struct mm_struct *mm = current->mm; |
mm | 1257 | arch/sparc/mm/sun4c.c | if(mm->context != NO_CONTEXT) { |
mm | 1258 | arch/sparc/mm/sun4c.c | sun4c_demap_context(&sun4c_context_ring[mm->context], mm->context); |
mm | 1259 | arch/sparc/mm/sun4c.c | ctx_old = ctx_list_pool + mm->context; |
mm | 1262 | arch/sparc/mm/sun4c.c | mm->context = NO_CONTEXT; |
mm | 1384 | arch/sparc/mm/sun4c.c | static pgd_t *sun4c_pgd_offset(struct mm_struct * mm, unsigned long address) |
mm | 1386 | arch/sparc/mm/sun4c.c | return mm->pgd + (address >> SUN4C_PGDIR_SHIFT); |
mm | 4197 | drivers/block/floppy.c | for (i=current->mm->env_start; i< current->mm->env_end; i ++){ |
mm | 827 | drivers/cdrom/sbpcd.c | u_int mm; |
mm | 831 | drivers/cdrom/sbpcd.c | mm = (blk + CD_MSF_OFFSET) % (CD_SECS * CD_FRAMES); |
mm | 832 | drivers/cdrom/sbpcd.c | msf.c[1] = mm / CD_FRAMES; |
mm | 833 | drivers/cdrom/sbpcd.c | msf.c[0] = mm % CD_FRAMES; |
mm | 235 | fs/binfmt_aout.c | current->mm->arg_start = (unsigned long) p; |
mm | 241 | fs/binfmt_aout.c | current->mm->arg_end = current->mm->env_start = (unsigned long) p; |
mm | 247 | fs/binfmt_aout.c | current->mm->env_end = (unsigned long) p; |
mm | 304 | fs/binfmt_aout.c | current->mm->end_code = ex.a_text + |
mm | 305 | fs/binfmt_aout.c | (current->mm->start_code = N_TXTADDR(ex)); |
mm | 306 | fs/binfmt_aout.c | current->mm->end_data = ex.a_data + |
mm | 307 | fs/binfmt_aout.c | (current->mm->start_data = N_DATADDR(ex)); |
mm | 308 | fs/binfmt_aout.c | current->mm->brk = ex.a_bss + |
mm | 309 | fs/binfmt_aout.c | (current->mm->start_brk = N_BSSADDR(ex)); |
mm | 311 | fs/binfmt_aout.c | current->mm->rss = 0; |
mm | 312 | fs/binfmt_aout.c | current->mm->mmap = NULL; |
mm | 382 | fs/binfmt_aout.c | set_brk(current->mm->start_brk, current->mm->brk); |
mm | 387 | fs/binfmt_aout.c | current->mm->start_stack = p; |
mm | 133 | fs/binfmt_elf.c | current->mm->arg_start = (unsigned long) p; |
mm | 139 | fs/binfmt_elf.c | current->mm->arg_end = current->mm->env_start = (unsigned long) p; |
mm | 145 | fs/binfmt_elf.c | current->mm->env_end = (unsigned long) p; |
mm | 290 | fs/binfmt_elf.c | current->mm->brk = interp_ex->a_bss + |
mm | 291 | fs/binfmt_elf.c | (current->mm->end_data = interp_ex->a_data + |
mm | 292 | fs/binfmt_elf.c | (current->mm->end_code = interp_ex->a_text)); |
mm | 522 | fs/binfmt_elf.c | current->mm->end_data = 0; |
mm | 523 | fs/binfmt_elf.c | current->mm->end_code = 0; |
mm | 524 | fs/binfmt_elf.c | current->mm->start_mmap = ELF_START_MMAP; |
mm | 525 | fs/binfmt_elf.c | current->mm->mmap = NULL; |
mm | 530 | fs/binfmt_elf.c | current->mm->rss = 0; |
mm | 532 | fs/binfmt_elf.c | current->mm->start_stack = bprm->p; |
mm | 643 | fs/binfmt_elf.c | current->mm->arg_start += strlen(passed_fileno) + 1; |
mm | 644 | fs/binfmt_elf.c | current->mm->start_brk = current->mm->brk = elf_brk; |
mm | 645 | fs/binfmt_elf.c | current->mm->end_code = end_code; |
mm | 646 | fs/binfmt_elf.c | current->mm->start_code = start_code; |
mm | 647 | fs/binfmt_elf.c | current->mm->end_data = end_data; |
mm | 648 | fs/binfmt_elf.c | current->mm->start_stack = bprm->p; |
mm | 657 | fs/binfmt_elf.c | printk("(start_brk) %x\n" , current->mm->start_brk); |
mm | 658 | fs/binfmt_elf.c | printk("(end_code) %x\n" , current->mm->end_code); |
mm | 659 | fs/binfmt_elf.c | printk("(start_code) %x\n" , current->mm->start_code); |
mm | 660 | fs/binfmt_elf.c | printk("(end_data) %x\n" , current->mm->end_data); |
mm | 661 | fs/binfmt_elf.c | printk("(start_stack) %x\n" , current->mm->start_stack); |
mm | 662 | fs/binfmt_elf.c | printk("(brk) %x\n" , current->mm->brk); |
mm | 966 | fs/binfmt_elf.c | for(vma = current->mm->mmap; vma != NULL; vma = vma->vm_next) { |
mm | 1094 | fs/binfmt_elf.c | len = current->mm->arg_end - current->mm->arg_start; |
mm | 1097 | fs/binfmt_elf.c | (const char *)current->mm->arg_start, len); |
mm | 1151 | fs/binfmt_elf.c | for(vma = current->mm->mmap, i = 0; |
mm | 1183 | fs/binfmt_elf.c | for(i = 0, vma = current->mm->mmap; |
mm | 296 | fs/exec.c | mpnt->vm_mm = current->mm; |
mm | 306 | fs/exec.c | current->mm->total_vm = (mpnt->vm_end - mpnt->vm_start) >> PAGE_SHIFT; |
mm | 311 | fs/exec.c | current->mm->rss++; |
mm | 374 | fs/exec.c | if (current->mm->count > 1) { |
mm | 375 | fs/exec.c | struct mm_struct *mm = kmalloc(sizeof(*mm), GFP_KERNEL); |
mm | 376 | fs/exec.c | if (!mm) { |
mm | 381 | fs/exec.c | *mm = *current->mm; |
mm | 382 | fs/exec.c | mm->def_flags = 0; /* should future lockings be kept? */ |
mm | 383 | fs/exec.c | mm->count = 1; |
mm | 384 | fs/exec.c | mm->mmap = NULL; |
mm | 385 | fs/exec.c | mm->mmap_avl = NULL; |
mm | 386 | fs/exec.c | mm->total_vm = 0; |
mm | 387 | fs/exec.c | mm->rss = 0; |
mm | 388 | fs/exec.c | current->mm->count--; |
mm | 389 | fs/exec.c | current->mm = mm; |
mm | 393 | fs/exec.c | exit_mmap(current->mm); |
mm | 130 | fs/namei.c | if (!p->mm) |
mm | 132 | fs/namei.c | for(mpnt = p->mm->mmap; mpnt; mpnt = mpnt->vm_next) { |
mm | 338 | fs/proc/array.c | if (!p || !p->mm || ptr >= TASK_SIZE) |
mm | 340 | fs/proc/array.c | page_dir = pgd_offset(p->mm,ptr); |
mm | 399 | fs/proc/array.c | if (!p || !*p || !(*p)->mm) |
mm | 401 | fs/proc/array.c | return get_array(p, (*p)->mm->env_start, (*p)->mm->env_end, buffer); |
mm | 408 | fs/proc/array.c | if (!p || !*p || !(*p)->mm) |
mm | 410 | fs/proc/array.c | return get_array(p, (*p)->mm->arg_start, (*p)->mm->arg_end, buffer); |
mm | 549 | fs/proc/array.c | struct mm_struct * mm = p->mm; |
mm | 551 | fs/proc/array.c | if (mm && mm != &init_mm) { |
mm | 552 | fs/proc/array.c | struct vm_area_struct * vma = mm->mmap; |
mm | 556 | fs/proc/array.c | for (vma = mm->mmap; vma; vma = vma->vm_next) { |
mm | 581 | fs/proc/array.c | mm->total_vm << (PAGE_SHIFT-10), |
mm | 582 | fs/proc/array.c | mm->locked_vm << (PAGE_SHIFT-10), |
mm | 583 | fs/proc/array.c | mm->rss << (PAGE_SHIFT-10), |
mm | 654 | fs/proc/array.c | if (tsk->mm && tsk->mm != &init_mm) { |
mm | 655 | fs/proc/array.c | struct vm_area_struct *vma = tsk->mm->mmap; |
mm | 719 | fs/proc/array.c | tsk->mm ? tsk->mm->rss : 0, /* you might want to shift this left 3 */ |
mm | 721 | fs/proc/array.c | tsk->mm ? tsk->mm->start_code : 0, |
mm | 722 | fs/proc/array.c | tsk->mm ? tsk->mm->end_code : 0, |
mm | 723 | fs/proc/array.c | tsk->mm ? tsk->mm->start_stack : 0, |
mm | 815 | fs/proc/array.c | if (tsk->mm && tsk->mm != &init_mm) { |
mm | 816 | fs/proc/array.c | struct vm_area_struct * vma = tsk->mm->mmap; |
mm | 819 | fs/proc/array.c | pgd_t *pgd = pgd_offset(tsk->mm, vma->vm_start); |
mm | 879 | fs/proc/array.c | if (!(*p)->mm || (*p)->mm == &init_mm || count == 0) |
mm | 887 | fs/proc/array.c | for (map = (*p)->mm->mmap, i = 0; map && (i < lineno); map = map->vm_next, i++) |
mm | 106 | fs/proc/link.c | if (!p->mm) |
mm | 108 | fs/proc/link.c | vma = p->mm->mmap; |
mm | 103 | fs/proc/mem.c | page_dir = pgd_offset(tsk->mm,addr); |
mm | 237 | fs/proc/mem.c | src_vma = tsk->mm->mmap; |
mm | 245 | fs/proc/mem.c | src_dir = pgd_offset(tsk->mm, stmp); |
mm | 272 | fs/proc/mem.c | src_vma = tsk->mm->mmap; |
mm | 282 | fs/proc/mem.c | src_dir = pgd_offset(tsk->mm, stmp); |
mm | 286 | fs/proc/mem.c | dest_dir = pgd_offset(current->mm, dtmp); |
mm | 49 | include/asm-alpha/mmu_context.h | struct mm_struct *mm, |
mm | 61 | include/asm-alpha/mmu_context.h | mm->context = asn; /* full version + asn */ |
mm | 79 | include/asm-alpha/mmu_context.h | struct mm_struct * mm = p->mm; |
mm | 81 | include/asm-alpha/mmu_context.h | if (mm) { |
mm | 84 | include/asm-alpha/mmu_context.h | if ((mm->context ^ asn) & ASN_VERSION_MASK) |
mm | 85 | include/asm-alpha/mmu_context.h | get_new_mmu_context(p, mm, asn); |
mm | 17 | include/asm-alpha/pgtable.h | #define flush_cache_mm(mm) do { } while (0) |
mm | 18 | include/asm-alpha/pgtable.h | #define flush_cache_range(mm, start, end) do { } while (0) |
mm | 50 | include/asm-alpha/pgtable.h | #define flush_tlb_current(mm) get_new_asn_and_reload(current, mm) |
mm | 51 | include/asm-alpha/pgtable.h | #define flush_tlb_other(mm) do { (mm)->context = 0; } while (0) |
mm | 60 | include/asm-alpha/pgtable.h | static inline void flush_tlb_current_page(struct mm_struct * mm, |
mm | 68 | include/asm-alpha/pgtable.h | flush_tlb_current(mm); |
mm | 79 | include/asm-alpha/pgtable.h | flush_tlb_current(current->mm); |
mm | 94 | include/asm-alpha/pgtable.h | static inline void flush_tlb_mm(struct mm_struct *mm) |
mm | 96 | include/asm-alpha/pgtable.h | if (mm != current->mm) |
mm | 97 | include/asm-alpha/pgtable.h | flush_tlb_other(mm); |
mm | 99 | include/asm-alpha/pgtable.h | flush_tlb_current(mm); |
mm | 113 | include/asm-alpha/pgtable.h | struct mm_struct * mm = vma->vm_mm; |
mm | 115 | include/asm-alpha/pgtable.h | if (mm != current->mm) |
mm | 116 | include/asm-alpha/pgtable.h | flush_tlb_other(mm); |
mm | 118 | include/asm-alpha/pgtable.h | flush_tlb_current_page(mm, vma, addr); |
mm | 125 | include/asm-alpha/pgtable.h | static inline void flush_tlb_range(struct mm_struct *mm, |
mm | 128 | include/asm-alpha/pgtable.h | flush_tlb_mm(mm); |
mm | 350 | include/asm-alpha/pgtable.h | extern inline pgd_t * pgd_offset(struct mm_struct * mm, unsigned long address) |
mm | 352 | include/asm-alpha/pgtable.h | return mm->pgd + ((address >> PGDIR_SHIFT) & (PTRS_PER_PAGE - 1)); |
mm | 24 | include/asm-i386/pgtable.h | #define flush_cache_mm(mm) do { } while (0) |
mm | 25 | include/asm-i386/pgtable.h | #define flush_cache_range(mm, start, end) do { } while (0) |
mm | 57 | include/asm-i386/pgtable.h | static inline void flush_tlb_mm(struct mm_struct *mm) |
mm | 59 | include/asm-i386/pgtable.h | if (mm == current->mm) |
mm | 66 | include/asm-i386/pgtable.h | if (vma->vm_mm == current->mm) |
mm | 70 | include/asm-i386/pgtable.h | static inline void flush_tlb_range(struct mm_struct *mm, |
mm | 73 | include/asm-i386/pgtable.h | if (mm == current->mm) |
mm | 106 | include/asm-i386/pgtable.h | if (current->mm->count == 1) /* just one copy of this mm */ |
mm | 116 | include/asm-i386/pgtable.h | static inline void flush_tlb_mm(struct mm_struct * mm) |
mm | 118 | include/asm-i386/pgtable.h | if (mm == current->mm && mm->count == 1) |
mm | 127 | include/asm-i386/pgtable.h | if (vma->vm_mm == current->mm && current->mm->count == 1) |
mm | 133 | include/asm-i386/pgtable.h | static inline void flush_tlb_range(struct mm_struct * mm, |
mm | 136 | include/asm-i386/pgtable.h | flush_tlb_mm(mm); |
mm | 147 | include/asm-i386/pgtable.h | static inline void flush_tlb_mm(struct mm_struct *mm) |
mm | 158 | include/asm-i386/pgtable.h | static inline void flush_tlb_range(struct mm_struct *mm, |
mm | 353 | include/asm-i386/pgtable.h | extern inline pgd_t * pgd_offset(struct mm_struct * mm, unsigned long address) |
mm | 355 | include/asm-i386/pgtable.h | return mm->pgd + (address >> PGDIR_SHIFT); |
mm | 30 | include/asm-m68k/pgtable.h | static inline void flush_tlb_mm(struct mm_struct *mm) |
mm | 32 | include/asm-m68k/pgtable.h | if (mm == current->mm) |
mm | 39 | include/asm-m68k/pgtable.h | if (vma->vm_mm == current->mm) |
mm | 43 | include/asm-m68k/pgtable.h | static inline void flush_tlb_range(struct mm_struct *mm, |
mm | 46 | include/asm-m68k/pgtable.h | if (mm == current->mm) |
mm | 300 | include/asm-m68k/pgtable.h | extern inline pgd_t * pgd_offset(struct mm_struct * mm, unsigned long address) |
mm | 302 | include/asm-m68k/pgtable.h | return mm->pgd + (address >> PGDIR_SHIFT); |
mm | 528 | include/asm-m68k/pgtable.h | #define flush_cache_mm(mm) flush_cache_all() |
mm | 529 | include/asm-m68k/pgtable.h | #define flush_cache_range(mm, start, end) flush_cache_all() |
mm | 42 | include/asm-mips/pgtable.h | do { if ((mm_struct) == current->mm) invalidate(); } while (0) |
mm | 44 | include/asm-mips/pgtable.h | do { if ((mm_struct) == current->mm) invalidate(); } while (0) |
mm | 46 | include/asm-mips/pgtable.h | do { if ((mm_struct) == current->mm) invalidate(); } while (0) |
mm | 344 | include/asm-mips/pgtable.h | extern inline pgd_t * pgd_offset(struct mm_struct * mm, unsigned long address) |
mm | 346 | include/asm-mips/pgtable.h | return mm->pgd + (address >> PGDIR_SHIFT); |
mm | 288 | include/asm-ppc/pgtable.h | extern inline pgd_t * pgd_offset(struct mm_struct * mm, unsigned long address) |
mm | 290 | include/asm-ppc/pgtable.h | return mm->pgd + (address >> PGDIR_SHIFT); |
mm | 12 | include/asm-sparc/mmu_context.h | #define init_new_context(mm) ((mm)->context = NO_CONTEXT) |
mm | 271 | include/asm-sparc/pgtable.h | extern void smp_flush_cache_mm(struct mm_struct *mm); |
mm | 272 | include/asm-sparc/pgtable.h | extern void smp_flush_cache_range(struct mm_struct *mm, |
mm | 278 | include/asm-sparc/pgtable.h | extern void smp_flush_tlb_mm(struct mm_struct *mm); |
mm | 279 | include/asm-sparc/pgtable.h | extern void smp_flush_tlb_range(struct mm_struct *mm, |
mm | 282 | include/asm-sparc/pgtable.h | extern void smp_flush_tlb_page(struct vm_area_struct *mm, unsigned long page); |
mm | 256 | include/linux/mm.h | extern void free_page_tables(struct mm_struct * mm); |
mm | 261 | include/linux/mm.h | extern int zap_page_range(struct mm_struct *mm, unsigned long address, unsigned long size); |
mm | 339 | include/linux/mm.h | if (task->mm) { |
mm | 340 | include/linux/mm.h | struct vm_area_struct * tree = task->mm->mmap_avl; |
mm | 246 | include/linux/sched.h | struct mm_struct *mm; |
mm | 430 | ipc/shm.c | current->mm->total_vm += (shmd->vm_end - shmd->vm_start) >> PAGE_SHIFT; |
mm | 497 | ipc/shm.c | if (addr < current->mm->start_stack && |
mm | 498 | ipc/shm.c | addr > current->mm->start_stack - PAGE_SIZE*(shp->shm_npages + 4)) |
mm | 526 | ipc/shm.c | shmd->vm_mm = current->mm; |
mm | 600 | ipc/shm.c | for (shmd = current->mm->mmap; shmd; shmd = shmdnext) { |
mm | 453 | kernel/exit.c | struct mm_struct * mm = tsk->mm; |
mm | 456 | kernel/exit.c | if (mm != &init_mm) { |
mm | 457 | kernel/exit.c | flush_cache_mm(mm); |
mm | 458 | kernel/exit.c | flush_tlb_mm(mm); |
mm | 459 | kernel/exit.c | tsk->mm = &init_mm; |
mm | 464 | kernel/exit.c | if (!--mm->count) { |
mm | 465 | kernel/exit.c | exit_mmap(mm); |
mm | 466 | kernel/exit.c | free_page_tables(mm); |
mm | 467 | kernel/exit.c | kfree(mm); |
mm | 79 | kernel/fork.c | static inline int dup_mmap(struct mm_struct * mm) |
mm | 83 | kernel/fork.c | mm->mmap = NULL; |
mm | 84 | kernel/fork.c | p = &mm->mmap; |
mm | 85 | kernel/fork.c | for (mpnt = current->mm->mmap ; mpnt ; mpnt = mpnt->vm_next) { |
mm | 88 | kernel/fork.c | exit_mmap(mm); |
mm | 93 | kernel/fork.c | tmp->vm_mm = mm; |
mm | 104 | kernel/fork.c | if (copy_page_range(mm, current->mm, tmp)) { |
mm | 105 | kernel/fork.c | exit_mmap(mm); |
mm | 111 | kernel/fork.c | build_mmap_avl(mm); |
mm | 118 | kernel/fork.c | struct mm_struct * mm = kmalloc(sizeof(*tsk->mm), GFP_KERNEL); |
mm | 119 | kernel/fork.c | if (!mm) |
mm | 121 | kernel/fork.c | *mm = *current->mm; |
mm | 122 | kernel/fork.c | mm->count = 1; |
mm | 123 | kernel/fork.c | mm->def_flags = 0; |
mm | 124 | kernel/fork.c | tsk->mm = mm; |
mm | 130 | kernel/fork.c | if (dup_mmap(mm)) { |
mm | 131 | kernel/fork.c | free_page_tables(mm); |
mm | 136 | kernel/fork.c | SET_PAGE_DIR(tsk, current->mm->pgd); |
mm | 137 | kernel/fork.c | current->mm->count++; |
mm | 559 | kernel/sys.c | if (brk < current->mm->end_code) |
mm | 560 | kernel/sys.c | return current->mm->brk; |
mm | 562 | kernel/sys.c | oldbrk = PAGE_ALIGN(current->mm->brk); |
mm | 564 | kernel/sys.c | return current->mm->brk = brk; |
mm | 569 | kernel/sys.c | if (brk <= current->mm->brk) { |
mm | 570 | kernel/sys.c | current->mm->brk = brk; |
mm | 580 | kernel/sys.c | if (brk - current->mm->end_code > rlim) |
mm | 581 | kernel/sys.c | return current->mm->brk; |
mm | 586 | kernel/sys.c | return current->mm->brk; |
mm | 600 | kernel/sys.c | return current->mm->brk; |
mm | 602 | kernel/sys.c | freepages += current->mm->rss; |
mm | 605 | kernel/sys.c | return current->mm->brk; |
mm | 610 | kernel/sys.c | current->mm->brk = brk; |
mm | 769 | mm/filemap.c | if (offset >= inode->i_size && (area->vm_flags & VM_SHARED) && area->vm_mm == current->mm) |
mm | 133 | mm/memory.c | page_dir = tsk->mm->pgd; |
mm | 138 | mm/memory.c | flush_cache_mm(tsk->mm); |
mm | 141 | mm/memory.c | flush_tlb_mm(tsk->mm); |
mm | 150 | mm/memory.c | void free_page_tables(struct mm_struct * mm) |
mm | 155 | mm/memory.c | page_dir = mm->pgd; |
mm | 172 | mm/memory.c | flush_cache_mm(tsk->mm); |
mm | 175 | mm/memory.c | flush_tlb_mm(tsk->mm); |
mm | 177 | mm/memory.c | tsk->mm->pgd = new_pg; |
mm | 310 | mm/memory.c | if (current->mm->rss <= 0) |
mm | 312 | mm/memory.c | current->mm->rss--; |
mm | 371 | mm/memory.c | int zap_page_range(struct mm_struct *mm, unsigned long address, unsigned long size) |
mm | 376 | mm/memory.c | dir = pgd_offset(mm, address); |
mm | 377 | mm/memory.c | flush_cache_range(mm, end - size, end); |
mm | 383 | mm/memory.c | flush_tlb_range(mm, end - size, end); |
mm | 432 | mm/memory.c | dir = pgd_offset(current->mm, address); |
mm | 433 | mm/memory.c | flush_cache_range(current->mm, beg, end); |
mm | 445 | mm/memory.c | flush_tlb_range(current->mm, beg, end); |
mm | 504 | mm/memory.c | dir = pgd_offset(current->mm, from); |
mm | 505 | mm/memory.c | flush_cache_range(current->mm, beg, end); |
mm | 517 | mm/memory.c | flush_tlb_range(current->mm, beg, end); |
mm | 549 | mm/memory.c | pgd = pgd_offset(tsk->mm,address); |
mm | 857 | mm/memory.c | pgd = pgd_offset(tsk->mm, address); |
mm | 194 | mm/mlock.c | locked += current->mm->locked_vm; |
mm | 230 | mm/mlock.c | current->mm->def_flags = def_flags; |
mm | 233 | mm/mlock.c | for (vma = current->mm->mmap; vma ; vma = vma->vm_next) { |
mm | 257 | mm/mlock.c | if (current->mm->total_vm > lock_limit) |
mm | 262 | mm/mlock.c | if (current->mm->total_vm > MAP_NR(high_memory)/2) |
mm | 58 | mm/mmap.c | if (current->mm->def_flags & VM_LOCKED) { |
mm | 59 | mm/mmap.c | unsigned long locked = current->mm->locked_vm << PAGE_SHIFT; |
mm | 126 | mm/mmap.c | vma->vm_mm = current->mm; |
mm | 131 | mm/mmap.c | vma->vm_flags |= current->mm->def_flags; |
mm | 175 | mm/mmap.c | current->mm->total_vm += len >> PAGE_SHIFT; |
mm | 178 | mm/mmap.c | current->mm->locked_vm += len >> PAGE_SHIFT; |
mm | 575 | mm/mmap.c | avl_checkheights(task->mm->mmap_avl); |
mm | 576 | mm/mmap.c | avl_checkorder(task->mm->mmap_avl); |
mm | 701 | mm/mmap.c | avl_neighbours(mpnt, current->mm->mmap_avl, &prev, &next); |
mm | 705 | mm/mmap.c | npp = (prev ? &prev->vm_next : ¤t->mm->mmap); |
mm | 711 | mm/mmap.c | avl_remove(mpnt, ¤t->mm->mmap_avl); |
mm | 737 | mm/mmap.c | zap_page_range(current->mm, st, end-st); |
mm | 742 | mm/mmap.c | zap_page_range(current->mm, addr, len); |
mm | 747 | mm/mmap.c | void build_mmap_avl(struct mm_struct * mm) |
mm | 751 | mm/mmap.c | mm->mmap_avl = NULL; |
mm | 752 | mm/mmap.c | for (vma = mm->mmap; vma; vma = vma->vm_next) |
mm | 753 | mm/mmap.c | avl_insert(vma, &mm->mmap_avl); |
mm | 757 | mm/mmap.c | void exit_mmap(struct mm_struct * mm) |
mm | 761 | mm/mmap.c | mpnt = mm->mmap; |
mm | 762 | mm/mmap.c | mm->mmap = NULL; |
mm | 763 | mm/mmap.c | mm->mmap_avl = NULL; |
mm | 764 | mm/mmap.c | mm->rss = 0; |
mm | 765 | mm/mmap.c | mm->total_vm = 0; |
mm | 766 | mm/mmap.c | mm->locked_vm = 0; |
mm | 776 | mm/mmap.c | zap_page_range(mm, mpnt->vm_start, mpnt->vm_end-mpnt->vm_start); |
mm | 796 | mm/mmap.c | p = &t->mm->mmap; |
mm | 809 | mm/mmap.c | avl_insert_neighbours(vmp, &t->mm->mmap_avl, &prev, &next); |
mm | 810 | mm/mmap.c | if ((prev ? prev->vm_next : t->mm->mmap) != next) |
mm | 815 | mm/mmap.c | t->mm->mmap = vmp; |
mm | 871 | mm/mmap.c | avl_neighbours(mpnt, task->mm->mmap_avl, &prev, &next); |
mm | 915 | mm/mmap.c | avl_remove(mpnt, &task->mm->mmap_avl); |
mm | 77 | mm/mprotect.c | dir = pgd_offset(current->mm, start); |
mm | 78 | mm/mprotect.c | flush_cache_range(current->mm, beg, end); |
mm | 84 | mm/mprotect.c | flush_tlb_range(current->mm, beg, end); |
mm | 22 | mm/mremap.c | static inline pte_t *get_one_pte(struct mm_struct *mm, unsigned long addr) |
mm | 28 | mm/mremap.c | pgd = pgd_offset(mm, addr); |
mm | 53 | mm/mremap.c | static inline pte_t *alloc_one_pte(struct mm_struct *mm, unsigned long addr) |
mm | 58 | mm/mremap.c | pmd = pmd_alloc(pgd_offset(mm, addr), addr); |
mm | 80 | mm/mremap.c | static int move_one_page(struct mm_struct *mm, unsigned long old_addr, unsigned long new_addr) |
mm | 85 | mm/mremap.c | src = get_one_pte(mm, old_addr); |
mm | 87 | mm/mremap.c | error = copy_one_pte(src, alloc_one_pte(mm, new_addr)); |
mm | 91 | mm/mremap.c | static int move_page_tables(struct mm_struct * mm, |
mm | 96 | mm/mremap.c | flush_cache_range(mm, old_addr, old_addr + len); |
mm | 97 | mm/mremap.c | flush_tlb_range(mm, old_addr, old_addr + len); |
mm | 106 | mm/mremap.c | if (move_one_page(mm, old_addr + offset, new_addr + offset)) |
mm | 119 | mm/mremap.c | flush_cache_range(mm, new_addr, new_addr + len); |
mm | 121 | mm/mremap.c | move_one_page(mm, new_addr + offset, old_addr + offset); |
mm | 122 | mm/mremap.c | flush_tlb_range(mm, new_addr, new_addr + len); |
mm | 123 | mm/mremap.c | zap_page_range(mm, new_addr, new_addr + len); |
mm | 137 | mm/mremap.c | if (new_addr && !move_page_tables(current->mm, new_addr, addr, old_len)) { |
mm | 192 | mm/mremap.c | unsigned long locked = current->mm->locked_vm << PAGE_SHIFT; |
mm | 207 | mm/mremap.c | current->mm->total_vm += pages; |
mm | 209 | mm/mremap.c | current->mm->locked_vm += pages; |
mm | 267 | mm/swapfile.c | static int unuse_process(struct mm_struct * mm, unsigned int type, unsigned long page) |
mm | 274 | mm/swapfile.c | if (!mm || mm == &init_mm) |
mm | 276 | mm/swapfile.c | vma = mm->mmap; |
mm | 278 | mm/swapfile.c | pgd_t * pgd = pgd_offset(mm, vma->vm_start); |
mm | 302 | mm/swapfile.c | if (unuse_process(p->mm, type, page)) { |
mm | 35 | mm/vmalloc.c | if (!p->mm) |
mm | 37 | mm/vmalloc.c | *pgd_offset(p->mm,address) = entry; |
mm | 263 | mm/vmscan.c | int result = swap_out_vma(p, vma, pgd_offset(p->mm, address), address, dma, wait); |
mm | 298 | mm/vmscan.c | if (p && p->swappable && p->mm->rss) |
mm | 310 | mm/vmscan.c | p->swap_cnt = AGE_CLUSTER_SIZE(p->mm->rss); |