| tag | line | file | source code |
| area | 30 | fs/msdos/mmap.c | struct vm_area_struct * area, |
| area | 35 | fs/msdos/mmap.c | struct inode * inode = area->vm_inode; |
| area | 41 | fs/msdos/mmap.c | pos = address - area->vm_start + area->vm_offset; |
| area | 33 | fs/nfs/mmap.c | static unsigned long nfs_file_mmap_nopage(struct vm_area_struct * area, |
| area | 36 | fs/nfs/mmap.c | struct inode * inode = area->vm_inode; |
| area | 45 | fs/nfs/mmap.c | pos = address - area->vm_start + area->vm_offset; |
| area | 48 | fs/nfs/mmap.c | if (address + PAGE_SIZE > area->vm_end) { |
| area | 49 | fs/nfs/mmap.c | clear = address + PAGE_SIZE - area->vm_end; |
| area | 29 | fs/smbfs/mmap.c | smb_file_mmap_nopage(struct vm_area_struct * area, |
| area | 32 | fs/smbfs/mmap.c | struct inode * inode = area->vm_inode; |
| area | 40 | fs/smbfs/mmap.c | pos = address - area->vm_start + area->vm_offset; |
| area | 43 | fs/smbfs/mmap.c | if (address + PAGE_SIZE > area->vm_end) { |
| area | 44 | fs/smbfs/mmap.c | clear = address + PAGE_SIZE - area->vm_end; |
| area | 96 | include/linux/mm.h | void (*open)(struct vm_area_struct * area); |
| area | 97 | include/linux/mm.h | void (*close)(struct vm_area_struct * area); |
| area | 98 | include/linux/mm.h | void (*unmap)(struct vm_area_struct *area, unsigned long, size_t); |
| area | 99 | include/linux/mm.h | void (*protect)(struct vm_area_struct *area, unsigned long, size_t, unsigned int newprot); |
| area | 100 | include/linux/mm.h | int (*sync)(struct vm_area_struct *area, unsigned long, size_t, unsigned int flags); |
| area | 101 | include/linux/mm.h | void (*advise)(struct vm_area_struct *area, unsigned long, size_t, unsigned int advise); |
| area | 102 | include/linux/mm.h | unsigned long (*nopage)(struct vm_area_struct * area, unsigned long address, |
| area | 104 | include/linux/mm.h | unsigned long (*wppage)(struct vm_area_struct * area, unsigned long address, |
| area | 56 | mm/filemap.c | static unsigned long filemap_nopage(struct vm_area_struct * area, unsigned long address, |
| area | 59 | mm/filemap.c | struct inode * inode = area->vm_inode; |
| area | 62 | mm/filemap.c | address = (address & PAGE_MASK) - area->vm_start + area->vm_offset; |
| area | 63 | mm/filemap.c | if (address >= inode->i_size && (area->vm_flags & VM_SHARED) && area->vm_mm == current->mm) |
| area | 884 | mm/memory.c | static int share_page(struct vm_area_struct * area, unsigned long address, |
| area | 893 | mm/memory.c | if (!area || !(inode = area->vm_inode) || inode->i_count < 2) |
| area | 897 | mm/memory.c | if (write_access && !(area->vm_flags & VM_SHARED)) { |
| area | 902 | mm/memory.c | offset = address - area->vm_start + area->vm_offset; |
| area | 905 | mm/memory.c | for (mpnt = area->vm_next_share; mpnt != area; mpnt = mpnt->vm_next_share) { |
| area | 912 | mm/memory.c | if ((mpnt->vm_offset ^ area->vm_offset) & ~PAGE_MASK) |
| area | 919 | mm/memory.c | if (!try_to_share(address, area, from_address, mpnt, give_page)) |
| area | 641 | mm/mmap.c | static void unmap_fixup(struct vm_area_struct *area, |
| area | 647 | mm/mmap.c | if (addr < area->vm_start || addr >= area->vm_end || |
| area | 648 | mm/mmap.c | end <= area->vm_start || end > area->vm_end || |
| area | 652 | mm/mmap.c | area->vm_start, area->vm_end, addr, end); |
| area | 657 | mm/mmap.c | if (addr == area->vm_start && end == area->vm_end) { |
| area | 658 | mm/mmap.c | if (area->vm_ops && area->vm_ops->close) |
| area | 659 | mm/mmap.c | area->vm_ops->close(area); |
| area | 660 | mm/mmap.c | if (area->vm_inode) |
| area | 661 | mm/mmap.c | iput(area->vm_inode); |
| area | 666 | mm/mmap.c | if (end == area->vm_end) |
| area | 667 | mm/mmap.c | area->vm_end = addr; |
| area | 669 | mm/mmap.c | if (addr == area->vm_start) { |
| area | 670 | mm/mmap.c | area->vm_offset += (end - area->vm_start); |
| area | 671 | mm/mmap.c | area->vm_start = end; |
| area | 680 | mm/mmap.c | *mpnt = *area; |
| area | 681 | mm/mmap.c | mpnt->vm_offset += (end - area->vm_start); |
| area | 687 | mm/mmap.c | area->vm_end = addr; /* Truncate area */ |
| area | 695 | mm/mmap.c | *mpnt = *area; |
| area | 698 | mm/mmap.c | if (area->vm_ops && area->vm_ops->close) { |
| area | 699 | mm/mmap.c | area->vm_end = area->vm_start; |
| area | 700 | mm/mmap.c | area->vm_ops->close(area); |
| area | 237 | mm/vmalloc.c | struct vm_struct **p, *tmp, *area; |
| area | 239 | mm/vmalloc.c | area = (struct vm_struct *) kmalloc(sizeof(*area), GFP_KERNEL); |
| area | 240 | mm/vmalloc.c | if (!area) |
| area | 243 | mm/vmalloc.c | area->size = size + PAGE_SIZE; |
| area | 244 | mm/vmalloc.c | area->next = NULL; |
| area | 250 | mm/vmalloc.c | area->addr = addr; |
| area | 251 | mm/vmalloc.c | area->next = *p; |
| area | 252 | mm/vmalloc.c | *p = area; |
| area | 253 | mm/vmalloc.c | return area; |
| area | 280 | mm/vmalloc.c | struct vm_struct *area; |
| area | 285 | mm/vmalloc.c | area = get_vm_area(size); |
| area | 286 | mm/vmalloc.c | if (!area) |
| area | 288 | mm/vmalloc.c | addr = area->addr; |
| area | 304 | mm/vmalloc.c | struct vm_struct * area; |
| area | 313 | mm/vmalloc.c | area = get_vm_area(size); |
| area | 314 | mm/vmalloc.c | if (!area) |
| area | 316 | mm/vmalloc.c | addr = area->addr; |