| tag | line | file | source code |
| mpnt | 80 | fs/binfmt_elf.c | struct vm_area_struct *mpnt; |
| mpnt | 82 | fs/binfmt_elf.c | mpnt = (struct vm_area_struct *)kmalloc(sizeof(*mpnt), GFP_KERNEL); |
| mpnt | 83 | fs/binfmt_elf.c | if (mpnt) { |
| mpnt | 84 | fs/binfmt_elf.c | mpnt->vm_task = current; |
| mpnt | 85 | fs/binfmt_elf.c | mpnt->vm_start = PAGE_MASK & (unsigned long) p; |
| mpnt | 86 | fs/binfmt_elf.c | mpnt->vm_end = TASK_SIZE; |
| mpnt | 87 | fs/binfmt_elf.c | mpnt->vm_page_prot = PAGE_PRIVATE|PAGE_DIRTY; |
| mpnt | 89 | fs/binfmt_elf.c | mpnt->vm_flags = VM_STACK_FLAGS; |
| mpnt | 90 | fs/binfmt_elf.c | mpnt->vm_pte = 0; |
| mpnt | 93 | fs/binfmt_elf.c | mpnt->vm_flags = VM_GROWSDOWN; |
| mpnt | 96 | fs/binfmt_elf.c | mpnt->vm_share = NULL; |
| mpnt | 97 | fs/binfmt_elf.c | mpnt->vm_inode = NULL; |
| mpnt | 98 | fs/binfmt_elf.c | mpnt->vm_offset = 0; |
| mpnt | 99 | fs/binfmt_elf.c | mpnt->vm_ops = NULL; |
| mpnt | 100 | fs/binfmt_elf.c | insert_vm_struct(current, mpnt); |
| mpnt | 102 | fs/binfmt_elf.c | current->mm->stk_vma = mpnt; |
| mpnt | 324 | fs/exec.c | struct vm_area_struct *mpnt; |
| mpnt | 326 | fs/exec.c | mpnt = (struct vm_area_struct *)kmalloc(sizeof(*mpnt), GFP_KERNEL); |
| mpnt | 327 | fs/exec.c | if (mpnt) { |
| mpnt | 328 | fs/exec.c | mpnt->vm_task = current; |
| mpnt | 329 | fs/exec.c | mpnt->vm_start = PAGE_MASK & (unsigned long) p; |
| mpnt | 330 | fs/exec.c | mpnt->vm_end = TASK_SIZE; |
| mpnt | 331 | fs/exec.c | mpnt->vm_page_prot = PAGE_PRIVATE|PAGE_DIRTY; |
| mpnt | 332 | fs/exec.c | mpnt->vm_flags = VM_STACK_FLAGS; |
| mpnt | 333 | fs/exec.c | mpnt->vm_share = NULL; |
| mpnt | 334 | fs/exec.c | mpnt->vm_ops = NULL; |
| mpnt | 335 | fs/exec.c | mpnt->vm_offset = 0; |
| mpnt | 336 | fs/exec.c | mpnt->vm_inode = NULL; |
| mpnt | 337 | fs/exec.c | mpnt->vm_pte = 0; |
| mpnt | 338 | fs/exec.c | insert_vm_struct(current, mpnt); |
| mpnt | 532 | fs/exec.c | struct vm_area_struct * mpnt, *mpnt1; |
| mpnt | 546 | fs/exec.c | mpnt = current->mm->mmap; |
| mpnt | 548 | fs/exec.c | while (mpnt) { |
| mpnt | 549 | fs/exec.c | mpnt1 = mpnt->vm_next; |
| mpnt | 550 | fs/exec.c | if (mpnt->vm_ops && mpnt->vm_ops->close) |
| mpnt | 551 | fs/exec.c | mpnt->vm_ops->close(mpnt); |
| mpnt | 552 | fs/exec.c | if (mpnt->vm_inode) |
| mpnt | 553 | fs/exec.c | iput(mpnt->vm_inode); |
| mpnt | 554 | fs/exec.c | kfree(mpnt); |
| mpnt | 555 | fs/exec.c | mpnt = mpnt1; |
| mpnt | 133 | fs/namei.c | struct vm_area_struct * mpnt; |
| mpnt | 136 | fs/namei.c | for(mpnt = (*p)->mm->mmap; mpnt; mpnt = mpnt->vm_next) { |
| mpnt | 137 | fs/namei.c | if (inode != mpnt->vm_inode) |
| mpnt | 139 | fs/namei.c | if (mpnt->vm_flags & VM_DENYWRITE) |
| mpnt | 61 | ibcs/binfmt_elf.c | struct vm_area_struct *mpnt; |
| mpnt | 63 | ibcs/binfmt_elf.c | mpnt = (struct vm_area_struct *)kmalloc(sizeof(*mpnt), GFP_KERNEL); |
| mpnt | 64 | ibcs/binfmt_elf.c | if (mpnt) { |
| mpnt | 65 | ibcs/binfmt_elf.c | mpnt->vm_task = current; |
| mpnt | 66 | ibcs/binfmt_elf.c | mpnt->vm_start = PAGE_MASK & (unsigned long) p; |
| mpnt | 67 | ibcs/binfmt_elf.c | mpnt->vm_end = TASK_SIZE; |
| mpnt | 68 | ibcs/binfmt_elf.c | mpnt->vm_page_prot = PAGE_PRIVATE|PAGE_DIRTY; |
| mpnt | 69 | ibcs/binfmt_elf.c | mpnt->vm_flags = VM_STACK_FLAGS; |
| mpnt | 70 | ibcs/binfmt_elf.c | mpnt->vm_share = NULL; |
| mpnt | 71 | ibcs/binfmt_elf.c | mpnt->vm_ops = NULL; |
| mpnt | 72 | ibcs/binfmt_elf.c | mpnt->vm_inode = NULL; |
| mpnt | 73 | ibcs/binfmt_elf.c | mpnt->vm_offset = 0; |
| mpnt | 74 | ibcs/binfmt_elf.c | mpnt->vm_pte = 0; |
| mpnt | 75 | ibcs/binfmt_elf.c | insert_vm_struct(current, mpnt); |
| mpnt | 359 | kernel/exit.c | struct vm_area_struct * mpnt; |
| mpnt | 361 | kernel/exit.c | mpnt = current->mm->mmap; |
| mpnt | 363 | kernel/exit.c | while (mpnt) { |
| mpnt | 364 | kernel/exit.c | struct vm_area_struct * next = mpnt->vm_next; |
| mpnt | 365 | kernel/exit.c | if (mpnt->vm_ops && mpnt->vm_ops->close) |
| mpnt | 366 | kernel/exit.c | mpnt->vm_ops->close(mpnt); |
| mpnt | 367 | kernel/exit.c | if (mpnt->vm_inode) |
| mpnt | 368 | kernel/exit.c | iput(mpnt->vm_inode); |
| mpnt | 369 | kernel/exit.c | kfree(mpnt); |
| mpnt | 370 | kernel/exit.c | mpnt = next; |
| mpnt | 93 | kernel/fork.c | struct vm_area_struct * mpnt, **p, *tmp; |
| mpnt | 97 | kernel/fork.c | for (mpnt = current->mm->mmap ; mpnt ; mpnt = mpnt->vm_next) { |
| mpnt | 101 | kernel/fork.c | *tmp = *mpnt; |
| mpnt | 824 | mm/memory.c | struct vm_area_struct * mpnt; |
| mpnt | 831 | mm/memory.c | for (mpnt = (*p)->mm->mmap; mpnt; mpnt = mpnt->vm_next) { |
| mpnt | 833 | mm/memory.c | if (mpnt->vm_inode != inode) |
| mpnt | 836 | mm/memory.c | if ((mpnt->vm_offset ^ area->vm_offset) & ~PAGE_MASK) |
| mpnt | 839 | mm/memory.c | from_address = offset + mpnt->vm_start - mpnt->vm_offset; |
| mpnt | 840 | mm/memory.c | if (from_address < mpnt->vm_start || from_address >= mpnt->vm_end) |
| mpnt | 843 | mm/memory.c | if (!try_to_share(address, area, from_address, mpnt, give_page)) |
| mpnt | 223 | mm/mmap.c | struct vm_area_struct *mpnt; |
| mpnt | 256 | mm/mmap.c | mpnt = (struct vm_area_struct *)kmalloc(sizeof(*mpnt), GFP_KERNEL); |
| mpnt | 258 | mm/mmap.c | if (!mpnt) |
| mpnt | 260 | mm/mmap.c | *mpnt = *area; |
| mpnt | 261 | mm/mmap.c | mpnt->vm_offset += (end - area->vm_start); |
| mpnt | 262 | mm/mmap.c | mpnt->vm_start = end; |
| mpnt | 263 | mm/mmap.c | if (mpnt->vm_inode) |
| mpnt | 264 | mm/mmap.c | mpnt->vm_inode->i_count++; |
| mpnt | 265 | mm/mmap.c | if (mpnt->vm_ops && mpnt->vm_ops->open) |
| mpnt | 266 | mm/mmap.c | mpnt->vm_ops->open(mpnt); |
| mpnt | 268 | mm/mmap.c | insert_vm_struct(current, mpnt); |
| mpnt | 272 | mm/mmap.c | mpnt = (struct vm_area_struct *)kmalloc(sizeof(*mpnt), GFP_KERNEL); |
| mpnt | 273 | mm/mmap.c | if (!mpnt) |
| mpnt | 275 | mm/mmap.c | *mpnt = *area; |
| mpnt | 276 | mm/mmap.c | if (mpnt->vm_ops && mpnt->vm_ops->open) |
| mpnt | 277 | mm/mmap.c | mpnt->vm_ops->open(mpnt); |
| mpnt | 282 | mm/mmap.c | insert_vm_struct(current, mpnt); |
| mpnt | 298 | mm/mmap.c | struct vm_area_struct *mpnt, **npp, *free; |
| mpnt | 314 | mm/mmap.c | for (mpnt = *npp; mpnt != NULL; mpnt = *npp) { |
| mpnt | 317 | mm/mmap.c | if ((addr < mpnt->vm_start && end <= mpnt->vm_start) || |
| mpnt | 318 | mm/mmap.c | (addr >= mpnt->vm_end && end > mpnt->vm_end)) |
| mpnt | 320 | mm/mmap.c | npp = &mpnt->vm_next; |
| mpnt | 324 | mm/mmap.c | *npp = mpnt->vm_next; |
| mpnt | 325 | mm/mmap.c | mpnt->vm_next = free; |
| mpnt | 326 | mm/mmap.c | free = mpnt; |
| mpnt | 341 | mm/mmap.c | mpnt = free; |
| mpnt | 344 | mm/mmap.c | st = addr < mpnt->vm_start ? mpnt->vm_start : addr; |
| mpnt | 346 | mm/mmap.c | end = end > mpnt->vm_end ? mpnt->vm_end : end; |
| mpnt | 348 | mm/mmap.c | if (mpnt->vm_ops && mpnt->vm_ops->unmap) |
| mpnt | 349 | mm/mmap.c | mpnt->vm_ops->unmap(mpnt, st, end-st); |
| mpnt | 351 | mm/mmap.c | unmap_fixup(mpnt, st, end-st); |
| mpnt | 353 | mm/mmap.c | kfree(mpnt); |
| mpnt | 388 | mm/mmap.c | struct vm_area_struct **p, *mpnt; |
| mpnt | 391 | mm/mmap.c | while ((mpnt = *p) != NULL) { |
| mpnt | 392 | mm/mmap.c | if (mpnt->vm_start > vmp->vm_start) |
| mpnt | 394 | mm/mmap.c | if (mpnt->vm_end > vmp->vm_start) |
| mpnt | 396 | mm/mmap.c | p = &mpnt->vm_next; |
| mpnt | 398 | mm/mmap.c | vmp->vm_next = mpnt; |
| mpnt | 407 | mm/mmap.c | void merge_segments(struct vm_area_struct *mpnt) |
| mpnt | 411 | mm/mmap.c | if (mpnt == NULL) |
| mpnt | 414 | mm/mmap.c | for(prev = mpnt, mpnt = mpnt->vm_next; |
| mpnt | 415 | mm/mmap.c | mpnt != NULL; |
| mpnt | 416 | mm/mmap.c | prev = mpnt, mpnt = next) |
| mpnt | 418 | mm/mmap.c | next = mpnt->vm_next; |
| mpnt | 423 | mm/mmap.c | if (mpnt->vm_inode != prev->vm_inode) |
| mpnt | 425 | mm/mmap.c | if (mpnt->vm_pte != prev->vm_pte) |
| mpnt | 427 | mm/mmap.c | if (mpnt->vm_ops != prev->vm_ops) |
| mpnt | 429 | mm/mmap.c | if (mpnt->vm_page_prot != prev->vm_page_prot || |
| mpnt | 430 | mm/mmap.c | mpnt->vm_flags != prev->vm_flags) |
| mpnt | 432 | mm/mmap.c | if (prev->vm_end != mpnt->vm_start) |
| mpnt | 437 | mm/mmap.c | if ((mpnt->vm_inode != NULL) || (mpnt->vm_flags & VM_SHM)) { |
| mpnt | 438 | mm/mmap.c | if (prev->vm_offset + prev->vm_end - prev->vm_start != mpnt->vm_offset) |
| mpnt | 447 | mm/mmap.c | prev->vm_end = mpnt->vm_end; |
| mpnt | 448 | mm/mmap.c | prev->vm_next = mpnt->vm_next; |
| mpnt | 449 | mm/mmap.c | if (mpnt->vm_ops && mpnt->vm_ops->close) { |
| mpnt | 450 | mm/mmap.c | mpnt->vm_offset += mpnt->vm_end - mpnt->vm_start; |
| mpnt | 451 | mm/mmap.c | mpnt->vm_start = mpnt->vm_end; |
| mpnt | 452 | mm/mmap.c | mpnt->vm_ops->close(mpnt); |
| mpnt | 454 | mm/mmap.c | if (mpnt->vm_inode) |
| mpnt | 455 | mm/mmap.c | mpnt->vm_inode->i_count--; |
| mpnt | 456 | mm/mmap.c | kfree_s(mpnt, sizeof(*mpnt)); |
| mpnt | 457 | mm/mmap.c | mpnt = prev; |