tag | line | file | source code |
order | 169 | drivers/block/floppy.c | int order; |
order | 173 | drivers/block/floppy.c | : "=r" (order) |
order | 176 | drivers/block/floppy.c | for (order = 0; order < NR_MEM_LISTS; ++order) |
order | 177 | drivers/block/floppy.c | if (size <= (PAGE_SIZE << order)) |
order | 178 | drivers/block/floppy.c | return order; |
order | 185 | drivers/block/floppy.c | int order = __get_order(size); |
order | 187 | drivers/block/floppy.c | if (order >= NR_MEM_LISTS) |
order | 189 | drivers/block/floppy.c | return __get_dma_pages(GFP_KERNEL,order); |
order | 153 | include/linux/mm.h | #define __get_dma_pages(priority, order) __get_free_pages((priority),(order),MAX_DMA_ADDRESS) |
order | 169 | include/linux/mm.h | extern void free_pages(unsigned long addr, unsigned long order); |
order | 73 | mm/kmalloc.c | int order; |
order | 142 | mm/kmalloc.c | #define NBLOCKS(order) (sizes[order].nblocks) |
order | 143 | mm/kmalloc.c | #define BLOCKSIZE(order) (sizes[order].size) |
order | 144 | mm/kmalloc.c | #define AREASIZE(order) (PAGE_SIZE<<(sizes[order].gfporder)) |
order | 149 | mm/kmalloc.c | int order; |
order | 155 | mm/kmalloc.c | for (order = 0; BLOCKSIZE(order); order++) { |
order | 156 | mm/kmalloc.c | if ((NBLOCKS(order) * BLOCKSIZE(order) + sizeof(struct page_descriptor)) > |
order | 157 | mm/kmalloc.c | AREASIZE(order)) { |
order | 159 | mm/kmalloc.c | (int) (NBLOCKS(order) * BLOCKSIZE(order) + |
order | 161 | mm/kmalloc.c | (int) AREASIZE(order), |
order | 162 | mm/kmalloc.c | BLOCKSIZE(order)); |
order | 173 | mm/kmalloc.c | int order; |
order | 177 | mm/kmalloc.c | for (order = 0; BLOCKSIZE(order); order++) |
order | 178 | mm/kmalloc.c | if (size <= BLOCKSIZE(order)) |
order | 179 | mm/kmalloc.c | return order; |
order | 187 | mm/kmalloc.c | int order, i, sz; |
order | 191 | mm/kmalloc.c | order = get_order(size); |
order | 192 | mm/kmalloc.c | if (order < 0) { |
order | 199 | mm/kmalloc.c | pg = &sizes[order].firstfree; |
order | 203 | mm/kmalloc.c | pg = &sizes[order].dmafree; |
order | 234 | mm/kmalloc.c | sz = BLOCKSIZE(order); |
order | 238 | mm/kmalloc.c | sizes[order].gfporder, max_addr); |
order | 248 | mm/kmalloc.c | sizes[order].npages++; |
order | 251 | mm/kmalloc.c | for (i = NBLOCKS(order), p = BH(page + 1); i > 1; i--, p = p->bh_next) { |
order | 259 | mm/kmalloc.c | page->order = order; |
order | 260 | mm/kmalloc.c | page->nfree = NBLOCKS(order); |
order | 277 | mm/kmalloc.c | sizes[order].nmallocs++; |
order | 278 | mm/kmalloc.c | sizes[order].nbytesmalloced += size; |
order | 288 | mm/kmalloc.c | int order; |
order | 296 | mm/kmalloc.c | order = page->order; |
order | 297 | mm/kmalloc.c | pg = &sizes[order].firstfree; |
order | 300 | mm/kmalloc.c | pg = &sizes[order].dmafree; |
order | 303 | mm/kmalloc.c | if ((order < 0) || |
order | 304 | mm/kmalloc.c | (order > sizeof(sizes) / sizeof(sizes[0])) || |
order | 308 | mm/kmalloc.c | p, page->next, page->order); |
order | 325 | mm/kmalloc.c | if (page->nfree == NBLOCKS(order)) { |
order | 338 | mm/kmalloc.c | sizes[order].npages--; |
order | 339 | mm/kmalloc.c | free_pages((long) page, sizes[order].gfporder); |
order | 341 | mm/kmalloc.c | sizes[order].nfrees++; |
order | 342 | mm/kmalloc.c | sizes[order].nbytesmalloced -= size; |
order | 682 | mm/swap.c | static inline void free_pages_ok(unsigned long addr, unsigned long order) |
order | 684 | mm/swap.c | unsigned long index = MAP_NR(addr) >> (1 + order); |
order | 685 | mm/swap.c | unsigned long mask = PAGE_MASK << order; |
order | 688 | mm/swap.c | nr_free_pages += 1 << order; |
order | 689 | mm/swap.c | while (order < NR_MEM_LISTS-1) { |
order | 690 | mm/swap.c | if (!change_bit(index, free_area_map[order])) |
order | 692 | mm/swap.c | remove_mem_queue(free_area_list+order, (struct mem_list *) (addr ^ (1+~mask))); |
order | 693 | mm/swap.c | order++; |
order | 698 | mm/swap.c | add_mem_queue(free_area_list+order, (struct mem_list *) addr); |
order | 716 | mm/swap.c | void free_pages(unsigned long addr, unsigned long order) |
order | 726 | mm/swap.c | free_pages_ok(addr, order); |
order | 744 | mm/swap.c | #define RMQUEUE(order, limit) \ |
order | 745 | mm/swap.c | do { struct mem_list * queue = free_area_list+order; \ |
order | 746 | mm/swap.c | unsigned long new_order = order; \ |
order | 752 | mm/swap.c | nr_free_pages -= 1 << order; \ |
order | 754 | mm/swap.c | EXPAND(ret, order, new_order); \ |
order | 763 | mm/swap.c | static inline int mark_used(unsigned long addr, unsigned long order) |
order | 765 | mm/swap.c | return change_bit(MAP_NR(addr) >> (1+order), free_area_map[order]); |
order | 779 | mm/swap.c | unsigned long __get_free_pages(int priority, unsigned long order, unsigned long limit) |
order | 799 | mm/swap.c | RMQUEUE(order, limit); |
order | 816 | mm/swap.c | unsigned long order, flags; |
order | 822 | mm/swap.c | for (order=0 ; order < NR_MEM_LISTS; order++) { |
order | 825 | mm/swap.c | for (tmp = free_area_list[order].next ; tmp != free_area_list + order ; tmp = tmp->next) { |
order | 828 | mm/swap.c | total += nr * ((PAGE_SIZE>>10) << order); |
order | 829 | mm/swap.c | printk("%lu*%lukB ", nr, (PAGE_SIZE>>10) << order); |