tag | line | file | source code |
order | 176 | drivers/block/floppy.c | int order; |
order | 179 | drivers/block/floppy.c | order = -1; |
order | 182 | drivers/block/floppy.c | order++; |
order | 184 | drivers/block/floppy.c | return order; |
order | 189 | drivers/block/floppy.c | int order = __get_order(size); |
order | 191 | drivers/block/floppy.c | return __get_dma_pages(GFP_KERNEL,order); |
order | 90 | drivers/char/ftape/kernel-interface.c | int order; |
order | 93 | drivers/char/ftape/kernel-interface.c | order = -1; |
order | 96 | drivers/char/ftape/kernel-interface.c | order++; |
order | 98 | drivers/char/ftape/kernel-interface.c | return order; |
order | 102 | drivers/char/ftape/kernel-interface.c | void *dmaalloc(int order) |
order | 104 | drivers/char/ftape/kernel-interface.c | return (void *) __get_dma_pages(GFP_KERNEL, order); |
order | 108 | drivers/char/ftape/kernel-interface.c | void dmafree(void *addr, int order) |
order | 110 | drivers/char/ftape/kernel-interface.c | free_pages((unsigned long) addr, order); |
order | 125 | drivers/char/ftape/kernel-interface.c | int order; |
order | 151 | drivers/char/ftape/kernel-interface.c | order = __get_order(BUFF_SIZE); |
order | 153 | drivers/char/ftape/kernel-interface.c | tape_buffer[n] = (byte *) dmaalloc(order); |
order | 158 | drivers/char/ftape/kernel-interface.c | dmafree(tape_buffer[n], order); |
order | 191 | drivers/char/ftape/kernel-interface.c | int order; |
order | 199 | drivers/char/ftape/kernel-interface.c | order = __get_order(BUFF_SIZE); |
order | 202 | drivers/char/ftape/kernel-interface.c | dmafree(tape_buffer[n], order); |
order | 8 | drivers/isdn/teles/buffers.c | BufPoolInit(struct BufPool *bp, int order, int bpps, |
order | 22 | drivers/isdn/teles/buffers.c | bp->pageorder = order; |
order | 25 | drivers/isdn/teles/buffers.c | bp->bufsize = BUFFER_SIZE(order, bpps); |
order | 354 | drivers/isdn/teles/teles.h | void BufPoolInit(struct BufPool *bp, int order, int bpps, |
order | 461 | drivers/isdn/teles/teles.h | #define PART_SIZE(order,bpps) (( (PAGE_SIZE<<order) -\ |
order | 463 | drivers/isdn/teles/teles.h | #define BUFFER_SIZE(order,bpps) (PART_SIZE(order,bpps)-\ |
order | 330 | drivers/pci/pci.c | long order; |
order | 332 | drivers/pci/pci.c | order = dev_info[i].vendor - (long) vendor; |
order | 333 | drivers/pci/pci.c | if (!order) |
order | 334 | drivers/pci/pci.c | order = dev_info[i].device - (long) dev; |
order | 336 | drivers/pci/pci.c | if (order < 0) |
order | 344 | drivers/pci/pci.c | if (order > 0) |
order | 2290 | drivers/scsi/scsi.c | int order, a_size; |
order | 2291 | drivers/scsi/scsi.c | for (order = 0, a_size = PAGE_SIZE; |
order | 2292 | drivers/scsi/scsi.c | a_size < size; order++, a_size <<= 1) |
order | 2295 | drivers/scsi/scsi.c | order); |
order | 2313 | drivers/scsi/scsi.c | int order, a_size; |
order | 2315 | drivers/scsi/scsi.c | for (order = 0, a_size = PAGE_SIZE; |
order | 2316 | drivers/scsi/scsi.c | a_size < size; order++, a_size <<= 1) |
order | 2318 | drivers/scsi/scsi.c | free_pages((unsigned long)ptr, order); |
order | 282 | include/asm-mips/dma.h | #define __get_dma_pages(priority, order) __get_free_pages((priority),(order), 1) |
order | 150 | include/linux/mm.h | #define __get_dma_pages(priority, order) __get_free_pages((priority),(order),1) |
order | 166 | include/linux/mm.h | extern void free_pages(unsigned long addr, unsigned long order); |
order | 66 | mm/kmalloc.c | int order; |
order | 135 | mm/kmalloc.c | #define NBLOCKS(order) (sizes[order].nblocks) |
order | 136 | mm/kmalloc.c | #define BLOCKSIZE(order) (sizes[order].size) |
order | 137 | mm/kmalloc.c | #define AREASIZE(order) (PAGE_SIZE<<(sizes[order].gfporder)) |
order | 142 | mm/kmalloc.c | int order; |
order | 148 | mm/kmalloc.c | for (order = 0; BLOCKSIZE(order); order++) { |
order | 149 | mm/kmalloc.c | if ((NBLOCKS(order) * BLOCKSIZE(order) + sizeof(struct page_descriptor)) > |
order | 150 | mm/kmalloc.c | AREASIZE(order)) { |
order | 152 | mm/kmalloc.c | (int) (NBLOCKS(order) * BLOCKSIZE(order) + |
order | 154 | mm/kmalloc.c | (int) AREASIZE(order), |
order | 155 | mm/kmalloc.c | BLOCKSIZE(order)); |
order | 166 | mm/kmalloc.c | int order; |
order | 170 | mm/kmalloc.c | for (order = 0; BLOCKSIZE(order); order++) |
order | 171 | mm/kmalloc.c | if (size <= BLOCKSIZE(order)) |
order | 172 | mm/kmalloc.c | return order; |
order | 180 | mm/kmalloc.c | int order, i, sz, dma; |
order | 184 | mm/kmalloc.c | order = get_order(size); |
order | 185 | mm/kmalloc.c | if (order < 0) { |
order | 192 | mm/kmalloc.c | pg = &sizes[order].firstfree; |
order | 196 | mm/kmalloc.c | pg = &sizes[order].dmafree; |
order | 229 | mm/kmalloc.c | sz = BLOCKSIZE(order); |
order | 232 | mm/kmalloc.c | sizes[order].gfporder, dma); |
order | 242 | mm/kmalloc.c | sizes[order].npages++; |
order | 245 | mm/kmalloc.c | for (i = NBLOCKS(order), p = BH(page + 1); i > 1; i--, p = p->bh_next) { |
order | 253 | mm/kmalloc.c | page->order = order; |
order | 254 | mm/kmalloc.c | page->nfree = NBLOCKS(order); |
order | 271 | mm/kmalloc.c | sizes[order].nmallocs++; |
order | 272 | mm/kmalloc.c | sizes[order].nbytesmalloced += size; |
order | 285 | mm/kmalloc.c | int order; |
order | 293 | mm/kmalloc.c | order = page->order; |
order | 294 | mm/kmalloc.c | pg = &sizes[order].firstfree; |
order | 297 | mm/kmalloc.c | pg = &sizes[order].dmafree; |
order | 300 | mm/kmalloc.c | if ((order < 0) || |
order | 301 | mm/kmalloc.c | (order >= sizeof(sizes) / sizeof(sizes[0])) || |
order | 305 | mm/kmalloc.c | p, page->next, page->order); |
order | 325 | mm/kmalloc.c | if (page->nfree == NBLOCKS(order)) { |
order | 338 | mm/kmalloc.c | sizes[order].npages--; |
order | 339 | mm/kmalloc.c | free_pages((long) page, sizes[order].gfporder); |
order | 341 | mm/kmalloc.c | sizes[order].nfrees++; |
order | 342 | mm/kmalloc.c | sizes[order].nbytesmalloced -= size; |
order | 88 | mm/page_alloc.c | static inline void free_pages_ok(unsigned long map_nr, unsigned long order) |
order | 90 | mm/page_alloc.c | unsigned long index = map_nr >> (1 + order); |
order | 91 | mm/page_alloc.c | unsigned long mask = (~0UL) << order; |
order | 96 | mm/page_alloc.c | nr_free_pages += 1 << order; |
order | 97 | mm/page_alloc.c | while (order < NR_MEM_LISTS-1) { |
order | 98 | mm/page_alloc.c | if (!change_bit(index, free_area[order].map)) |
order | 100 | mm/page_alloc.c | remove_mem_queue(&free_area[order].list, list(map_nr ^ (1+~mask))); |
order | 102 | mm/page_alloc.c | order++; |
order | 106 | mm/page_alloc.c | add_mem_queue(&free_area[order].list, list(map_nr)); |
order | 126 | mm/page_alloc.c | void free_pages(unsigned long addr, unsigned long order) |
order | 139 | mm/page_alloc.c | free_pages_ok(map_nr, order); |
order | 156 | mm/page_alloc.c | #define MARK_USED(index, order, area) \ |
order | 157 | mm/page_alloc.c | change_bit((index) >> (1+(order)), (area)->map) |
order | 160 | mm/page_alloc.c | #define RMQUEUE(order, dma) \ |
order | 161 | mm/page_alloc.c | do { struct free_area_struct * area = free_area+order; \ |
order | 162 | mm/page_alloc.c | unsigned long new_order = order; \ |
order | 169 | mm/page_alloc.c | nr_free_pages -= 1 << order; \ |
order | 170 | mm/page_alloc.c | EXPAND(ret, map_nr, order, new_order, area); \ |
order | 193 | mm/page_alloc.c | unsigned long __get_free_pages(int priority, unsigned long order, int dma) |
order | 198 | mm/page_alloc.c | if (order >= NR_MEM_LISTS) |
order | 215 | mm/page_alloc.c | RMQUEUE(order, dma); |
order | 232 | mm/page_alloc.c | unsigned long order, flags; |
order | 238 | mm/page_alloc.c | for (order=0 ; order < NR_MEM_LISTS; order++) { |
order | 241 | mm/page_alloc.c | for (tmp = free_area[order].list.next ; tmp != &free_area[order].list ; tmp = tmp->next) { |
order | 244 | mm/page_alloc.c | total += nr * ((PAGE_SIZE>>10) << order); |
order | 245 | mm/page_alloc.c | printk("%lu*%lukB ", nr, (PAGE_SIZE>>10) << order); |