tag | line | file | source code |
order | 169 | drivers/block/floppy.c | int order; |
order | 173 | drivers/block/floppy.c | : "=r" (order) |
order | 176 | drivers/block/floppy.c | for (order = 0; order < NR_MEM_LISTS; ++order) |
order | 177 | drivers/block/floppy.c | if (size <= (PAGE_SIZE << order)) |
order | 178 | drivers/block/floppy.c | return order; |
order | 185 | drivers/block/floppy.c | int order = __get_order(size); |
order | 187 | drivers/block/floppy.c | if (order >= NR_MEM_LISTS) |
order | 189 | drivers/block/floppy.c | return __get_dma_pages(GFP_KERNEL,order); |
order | 153 | include/linux/mm.h | #define __get_dma_pages(priority, order) __get_free_pages((priority),(order),MAX_DMA_ADDRESS) |
order | 169 | include/linux/mm.h | extern void free_pages(unsigned long addr, unsigned long order); |
order | 72 | mm/kmalloc.c | int order; |
order | 139 | mm/kmalloc.c | #define NBLOCKS(order) (sizes[order].nblocks) |
order | 140 | mm/kmalloc.c | #define BLOCKSIZE(order) (sizes[order].size) |
order | 141 | mm/kmalloc.c | #define AREASIZE(order) (PAGE_SIZE<<(sizes[order].gfporder)) |
order | 146 | mm/kmalloc.c | int order; |
order | 152 | mm/kmalloc.c | for (order = 0;BLOCKSIZE(order);order++) |
order | 154 | mm/kmalloc.c | if ((NBLOCKS (order)*BLOCKSIZE(order) + sizeof (struct page_descriptor)) > |
order | 155 | mm/kmalloc.c | AREASIZE(order)) |
order | 158 | mm/kmalloc.c | (int) (NBLOCKS (order) * BLOCKSIZE(order) + |
order | 160 | mm/kmalloc.c | (int) AREASIZE(order), |
order | 161 | mm/kmalloc.c | BLOCKSIZE (order)); |
order | 172 | mm/kmalloc.c | int order; |
order | 176 | mm/kmalloc.c | for (order = 0;BLOCKSIZE(order);order++) |
order | 177 | mm/kmalloc.c | if (size <= BLOCKSIZE (order)) |
order | 178 | mm/kmalloc.c | return order; |
order | 185 | mm/kmalloc.c | int order,tries,i,sz; |
order | 203 | mm/kmalloc.c | order = get_order (size); |
order | 204 | mm/kmalloc.c | if (order < 0) |
order | 219 | mm/kmalloc.c | if ((page = (dma_flag ? sizes[order].dmafree : sizes[order].firstfree)) && |
order | 229 | mm/kmalloc.c | sizes[order].dmafree = page->next; |
order | 231 | mm/kmalloc.c | sizes[order].firstfree = page->next; |
order | 236 | mm/kmalloc.c | sizes [order].nmallocs++; |
order | 237 | mm/kmalloc.c | sizes [order].nbytesmalloced += size; |
order | 250 | mm/kmalloc.c | sz = BLOCKSIZE(order); /* sz is the size of the blocks we're dealing with */ |
order | 258 | mm/kmalloc.c | sizes[order].gfporder, max_addr); |
order | 272 | mm/kmalloc.c | sizes[order].npages++; |
order | 275 | mm/kmalloc.c | for (i=NBLOCKS(order),p=BH (page+1);i > 1;i--,p=p->bh_next) |
order | 284 | mm/kmalloc.c | page->order = order; |
order | 285 | mm/kmalloc.c | page->nfree = NBLOCKS(order); |
order | 298 | mm/kmalloc.c | page->next = sizes[order].dmafree; |
order | 299 | mm/kmalloc.c | sizes[order].dmafree = page; |
order | 301 | mm/kmalloc.c | page->next = sizes[order].firstfree; |
order | 302 | mm/kmalloc.c | sizes[order].firstfree = page; |
order | 324 | mm/kmalloc.c | int order; |
order | 329 | mm/kmalloc.c | order = page->order; |
order | 330 | mm/kmalloc.c | if ((order < 0) || |
order | 331 | mm/kmalloc.c | (order > sizeof (sizes)/sizeof (sizes[0])) || |
order | 336 | mm/kmalloc.c | p, page->next, page->order); |
order | 363 | mm/kmalloc.c | page->next = sizes[order].firstfree; |
order | 364 | mm/kmalloc.c | sizes[order].firstfree = page; |
order | 369 | mm/kmalloc.c | if (page->nfree == NBLOCKS (page->order)) |
order | 374 | mm/kmalloc.c | if (sizes[order].firstfree == page) |
order | 376 | mm/kmalloc.c | sizes[order].firstfree = page->next; |
order | 378 | mm/kmalloc.c | else if (sizes[order].dmafree == page) |
order | 380 | mm/kmalloc.c | sizes[order].dmafree = page->next; |
order | 384 | mm/kmalloc.c | for (pg2=sizes[order].firstfree; |
order | 389 | mm/kmalloc.c | for (pg2=sizes[order].dmafree; |
order | 399 | mm/kmalloc.c | free_pages ((long)page, sizes[order].gfporder); |
order | 407 | mm/kmalloc.c | sizes[order].nfrees++; /* Noncritical (monitoring) admin stuff */ |
order | 408 | mm/kmalloc.c | sizes[order].nbytesmalloced -= size; |
order | 673 | mm/swap.c | static inline void free_pages_ok(unsigned long addr, unsigned long order) |
order | 675 | mm/swap.c | unsigned long index = MAP_NR(addr) >> (1 + order); |
order | 676 | mm/swap.c | unsigned long mask = PAGE_MASK << order; |
order | 679 | mm/swap.c | nr_free_pages += 1 << order; |
order | 680 | mm/swap.c | while (order < NR_MEM_LISTS-1) { |
order | 681 | mm/swap.c | if (!change_bit(index, free_area_map[order])) |
order | 683 | mm/swap.c | remove_mem_queue(free_area_list+order, (struct mem_list *) (addr ^ (1+~mask))); |
order | 684 | mm/swap.c | order++; |
order | 689 | mm/swap.c | add_mem_queue(free_area_list+order, (struct mem_list *) addr); |
order | 707 | mm/swap.c | void free_pages(unsigned long addr, unsigned long order) |
order | 717 | mm/swap.c | free_pages_ok(addr, order); |
order | 735 | mm/swap.c | #define RMQUEUE(order, limit) \ |
order | 736 | mm/swap.c | do { struct mem_list * queue = free_area_list+order; \ |
order | 737 | mm/swap.c | unsigned long new_order = order; \ |
order | 743 | mm/swap.c | nr_free_pages -= 1 << order; \ |
order | 745 | mm/swap.c | EXPAND(ret, order, new_order); \ |
order | 754 | mm/swap.c | static inline int mark_used(unsigned long addr, unsigned long order) |
order | 756 | mm/swap.c | return change_bit(MAP_NR(addr) >> (1+order), free_area_map[order]); |
order | 770 | mm/swap.c | unsigned long __get_free_pages(int priority, unsigned long order, unsigned long limit) |
order | 790 | mm/swap.c | RMQUEUE(order, limit); |
order | 807 | mm/swap.c | unsigned long order, flags; |
order | 813 | mm/swap.c | for (order=0 ; order < NR_MEM_LISTS; order++) { |
order | 816 | mm/swap.c | for (tmp = free_area_list[order].next ; tmp != free_area_list + order ; tmp = tmp->next) { |
order | 819 | mm/swap.c | total += nr * ((PAGE_SIZE>>10) << order); |
order | 820 | mm/swap.c | printk("%lu*%lukB ", nr, (PAGE_SIZE>>10) << order); |