tag | line | file | source code |
area | 29 | arch/m68k/fpsp040/fpsp.h | | | FPU fsave area | |
area | 46 | arch/m68k/fpsp040/fpsp.h | | offsets refer to the Local Variable area and the fsave area. |
area | 63 | arch/m68k/fpsp040/fpsp.h | | Sometimes the exception handler will transform the fsave area |
area | 71 | arch/m68k/fpsp040/fpsp.h | | restored from the "local variable" area and can be used as |
area | 28 | fs/fat/mmap.c | struct vm_area_struct * area, |
area | 32 | fs/fat/mmap.c | struct inode * inode = area->vm_inode; |
area | 42 | fs/fat/mmap.c | pos = address - area->vm_start + area->vm_offset; |
area | 33 | fs/ncpfs/mmap.c | ncp_file_mmap_nopage(struct vm_area_struct * area, |
area | 36 | fs/ncpfs/mmap.c | struct inode * inode = area->vm_inode; |
area | 48 | fs/ncpfs/mmap.c | pos = address - area->vm_start + area->vm_offset; |
area | 51 | fs/ncpfs/mmap.c | if (address + PAGE_SIZE > area->vm_end) |
area | 53 | fs/ncpfs/mmap.c | clear = address + PAGE_SIZE - area->vm_end; |
area | 27 | fs/smbfs/mmap.c | smb_file_mmap_nopage(struct vm_area_struct * area, |
area | 30 | fs/smbfs/mmap.c | struct inode * inode = area->vm_inode; |
area | 42 | fs/smbfs/mmap.c | pos = address - area->vm_start + area->vm_offset; |
area | 45 | fs/smbfs/mmap.c | if (address + PAGE_SIZE > area->vm_end) { |
area | 46 | fs/smbfs/mmap.c | clear = address + PAGE_SIZE - area->vm_end; |
area | 97 | include/linux/mm.h | void (*open)(struct vm_area_struct * area); |
area | 98 | include/linux/mm.h | void (*close)(struct vm_area_struct * area); |
area | 99 | include/linux/mm.h | void (*unmap)(struct vm_area_struct *area, unsigned long, size_t); |
area | 100 | include/linux/mm.h | void (*protect)(struct vm_area_struct *area, unsigned long, size_t, unsigned int newprot); |
area | 101 | include/linux/mm.h | int (*sync)(struct vm_area_struct *area, unsigned long, size_t, unsigned int flags); |
area | 102 | include/linux/mm.h | void (*advise)(struct vm_area_struct *area, unsigned long, size_t, unsigned int advise); |
area | 103 | include/linux/mm.h | unsigned long (*nopage)(struct vm_area_struct * area, unsigned long address, int write_access); |
area | 104 | include/linux/mm.h | unsigned long (*wppage)(struct vm_area_struct * area, unsigned long address, |
area | 762 | mm/filemap.c | static unsigned long filemap_nopage(struct vm_area_struct * area, unsigned long address, int no_share) |
area | 765 | mm/filemap.c | struct inode * inode = area->vm_inode; |
area | 768 | mm/filemap.c | offset = (address & PAGE_MASK) - area->vm_start + area->vm_offset; |
area | 769 | mm/filemap.c | if (offset >= inode->i_size && (area->vm_flags & VM_SHARED) && area->vm_mm == current->mm) |
area | 604 | mm/mmap.c | static void unmap_fixup(struct vm_area_struct *area, |
area | 610 | mm/mmap.c | if (addr < area->vm_start || addr >= area->vm_end || |
area | 611 | mm/mmap.c | end <= area->vm_start || end > area->vm_end || |
area | 615 | mm/mmap.c | area->vm_start, area->vm_end, addr, end); |
area | 618 | mm/mmap.c | area->vm_mm->total_vm -= len >> PAGE_SHIFT; |
area | 619 | mm/mmap.c | if (area->vm_flags & VM_LOCKED) |
area | 620 | mm/mmap.c | area->vm_mm->locked_vm -= len >> PAGE_SHIFT; |
area | 623 | mm/mmap.c | if (addr == area->vm_start && end == area->vm_end) { |
area | 624 | mm/mmap.c | if (area->vm_ops && area->vm_ops->close) |
area | 625 | mm/mmap.c | area->vm_ops->close(area); |
area | 626 | mm/mmap.c | if (area->vm_inode) |
area | 627 | mm/mmap.c | iput(area->vm_inode); |
area | 632 | mm/mmap.c | if (end == area->vm_end) |
area | 633 | mm/mmap.c | area->vm_end = addr; |
area | 635 | mm/mmap.c | if (addr == area->vm_start) { |
area | 636 | mm/mmap.c | area->vm_offset += (end - area->vm_start); |
area | 637 | mm/mmap.c | area->vm_start = end; |
area | 646 | mm/mmap.c | *mpnt = *area; |
area | 647 | mm/mmap.c | mpnt->vm_offset += (end - area->vm_start); |
area | 653 | mm/mmap.c | area->vm_end = addr; /* Truncate area */ |
area | 661 | mm/mmap.c | *mpnt = *area; |
area | 664 | mm/mmap.c | if (area->vm_ops && area->vm_ops->close) { |
area | 665 | mm/mmap.c | area->vm_end = area->vm_start; |
area | 666 | mm/mmap.c | area->vm_ops->close(area); |
area | 136 | mm/page_alloc.c | #define MARK_USED(index, order, area) \ |
area | 137 | mm/page_alloc.c | change_bit((index) >> (1+(order)), (area)->map) |
area | 141 | mm/page_alloc.c | do { struct free_area_struct * area = free_area+order; \ |
area | 143 | mm/page_alloc.c | do { struct page *prev = &area->list, *ret; \ |
area | 144 | mm/page_alloc.c | while (&area->list != (ret = prev->next)) { \ |
area | 148 | mm/page_alloc.c | MARK_USED(map_nr, new_order, area); \ |
area | 150 | mm/page_alloc.c | EXPAND(ret, map_nr, order, new_order, area); \ |
area | 156 | mm/page_alloc.c | new_order++; area++; \ |
area | 160 | mm/page_alloc.c | #define EXPAND(map,index,low,high,area) \ |
area | 163 | mm/page_alloc.c | area--; high--; size >>= 1; \ |
area | 164 | mm/page_alloc.c | add_mem_queue(&area->list, map); \ |
area | 165 | mm/page_alloc.c | MARK_USED(index, high, area); \ |
area | 240 | mm/vmalloc.c | struct vm_struct **p, *tmp, *area; |
area | 242 | mm/vmalloc.c | area = (struct vm_struct *) kmalloc(sizeof(*area), GFP_KERNEL); |
area | 243 | mm/vmalloc.c | if (!area) |
area | 246 | mm/vmalloc.c | area->size = size + PAGE_SIZE; |
area | 247 | mm/vmalloc.c | area->next = NULL; |
area | 253 | mm/vmalloc.c | area->addr = addr; |
area | 254 | mm/vmalloc.c | area->next = *p; |
area | 255 | mm/vmalloc.c | *p = area; |
area | 256 | mm/vmalloc.c | return area; |
area | 283 | mm/vmalloc.c | struct vm_struct *area; |
area | 288 | mm/vmalloc.c | area = get_vm_area(size); |
area | 289 | mm/vmalloc.c | if (!area) |
area | 291 | mm/vmalloc.c | addr = area->addr; |
area | 307 | mm/vmalloc.c | struct vm_struct * area; |
area | 316 | mm/vmalloc.c | area = get_vm_area(size); |
area | 317 | mm/vmalloc.c | if (!area) |
area | 319 | mm/vmalloc.c | addr = area->addr; |