tag | line | file | source code |
order | 176 | drivers/block/floppy.c | int order; |
order | 179 | drivers/block/floppy.c | order = -1; |
order | 182 | drivers/block/floppy.c | order++; |
order | 184 | drivers/block/floppy.c | return order; |
order | 189 | drivers/block/floppy.c | int order = __get_order(size); |
order | 191 | drivers/block/floppy.c | return __get_dma_pages(GFP_KERNEL,order); |
order | 100 | drivers/char/ftape/kernel-interface.c | int order; |
order | 102 | drivers/char/ftape/kernel-interface.c | for (order = 0; order < NR_MEM_LISTS; ++order) |
order | 103 | drivers/char/ftape/kernel-interface.c | if (size <= (PAGE_SIZE << order)) |
order | 104 | drivers/char/ftape/kernel-interface.c | return order; |
order | 109 | drivers/char/ftape/kernel-interface.c | void *dmaalloc(int order) |
order | 111 | drivers/char/ftape/kernel-interface.c | return (void *) __get_dma_pages(GFP_KERNEL, order); |
order | 115 | drivers/char/ftape/kernel-interface.c | void dmafree(void *addr, int order) |
order | 117 | drivers/char/ftape/kernel-interface.c | free_pages((unsigned long) addr, order); |
order | 132 | drivers/char/ftape/kernel-interface.c | int order; |
order | 158 | drivers/char/ftape/kernel-interface.c | order = __get_order(BUFF_SIZE); |
order | 159 | drivers/char/ftape/kernel-interface.c | if (order < 0) { |
order | 167 | drivers/char/ftape/kernel-interface.c | tape_buffer[n] = (byte *) dmaalloc(order); |
order | 172 | drivers/char/ftape/kernel-interface.c | dmafree(tape_buffer[n], order); |
order | 205 | drivers/char/ftape/kernel-interface.c | int order; |
order | 213 | drivers/char/ftape/kernel-interface.c | order = __get_order(BUFF_SIZE); |
order | 214 | drivers/char/ftape/kernel-interface.c | if (order < 0) { |
order | 219 | drivers/char/ftape/kernel-interface.c | dmafree(tape_buffer[n], order); |
order | 8 | drivers/isdn/teles/buffers.c | BufPoolInit(struct BufPool *bp, int order, int bpps, |
order | 22 | drivers/isdn/teles/buffers.c | bp->pageorder = order; |
order | 25 | drivers/isdn/teles/buffers.c | bp->bufsize = BUFFER_SIZE(order, bpps); |
order | 354 | drivers/isdn/teles/teles.h | void BufPoolInit(struct BufPool *bp, int order, int bpps, |
order | 461 | drivers/isdn/teles/teles.h | #define PART_SIZE(order,bpps) (( (PAGE_SIZE<<order) -\ |
order | 463 | drivers/isdn/teles/teles.h | #define BUFFER_SIZE(order,bpps) (PART_SIZE(order,bpps)-\ |
order | 313 | drivers/pci/pci.c | long order; |
order | 315 | drivers/pci/pci.c | order = dev_info[i].vendor - (long) vendor; |
order | 316 | drivers/pci/pci.c | if (!order) |
order | 317 | drivers/pci/pci.c | order = dev_info[i].device - (long) dev; |
order | 319 | drivers/pci/pci.c | if (order < 0) |
order | 327 | drivers/pci/pci.c | if (order > 0) |
order | 2254 | drivers/scsi/scsi.c | int order, a_size; |
order | 2255 | drivers/scsi/scsi.c | for (order = 0, a_size = PAGE_SIZE; |
order | 2256 | drivers/scsi/scsi.c | a_size < size; order++, a_size <<= 1) |
order | 2259 | drivers/scsi/scsi.c | order); |
order | 2277 | drivers/scsi/scsi.c | int order, a_size; |
order | 2279 | drivers/scsi/scsi.c | for (order = 0, a_size = PAGE_SIZE; |
order | 2280 | drivers/scsi/scsi.c | a_size < size; order++, a_size <<= 1) |
order | 2282 | drivers/scsi/scsi.c | free_pages((unsigned long)ptr, order); |
order | 282 | include/asm-mips/dma.h | #define __get_dma_pages(priority, order) __get_free_pages((priority),(order), 1) |
order | 150 | include/linux/mm.h | #define __get_dma_pages(priority, order) __get_free_pages((priority),(order),1) |
order | 166 | include/linux/mm.h | extern void free_pages(unsigned long addr, unsigned long order); |
order | 66 | mm/kmalloc.c | int order; |
order | 135 | mm/kmalloc.c | #define NBLOCKS(order) (sizes[order].nblocks) |
order | 136 | mm/kmalloc.c | #define BLOCKSIZE(order) (sizes[order].size) |
order | 137 | mm/kmalloc.c | #define AREASIZE(order) (PAGE_SIZE<<(sizes[order].gfporder)) |
order | 142 | mm/kmalloc.c | int order; |
order | 148 | mm/kmalloc.c | for (order = 0; BLOCKSIZE(order); order++) { |
order | 149 | mm/kmalloc.c | if ((NBLOCKS(order) * BLOCKSIZE(order) + sizeof(struct page_descriptor)) > |
order | 150 | mm/kmalloc.c | AREASIZE(order)) { |
order | 152 | mm/kmalloc.c | (int) (NBLOCKS(order) * BLOCKSIZE(order) + |
order | 154 | mm/kmalloc.c | (int) AREASIZE(order), |
order | 155 | mm/kmalloc.c | BLOCKSIZE(order)); |
order | 166 | mm/kmalloc.c | int order; |
order | 170 | mm/kmalloc.c | for (order = 0; BLOCKSIZE(order); order++) |
order | 171 | mm/kmalloc.c | if (size <= BLOCKSIZE(order)) |
order | 172 | mm/kmalloc.c | return order; |
order | 180 | mm/kmalloc.c | int order, i, sz; |
order | 184 | mm/kmalloc.c | order = get_order(size); |
order | 185 | mm/kmalloc.c | if (order < 0) { |
order | 191 | mm/kmalloc.c | pg = &sizes[order].firstfree; |
order | 194 | mm/kmalloc.c | pg = &sizes[order].dmafree; |
order | 227 | mm/kmalloc.c | sz = BLOCKSIZE(order); |
order | 230 | mm/kmalloc.c | sizes[order].gfporder, priority & GFP_DMA); |
order | 240 | mm/kmalloc.c | sizes[order].npages++; |
order | 243 | mm/kmalloc.c | for (i = NBLOCKS(order), p = BH(page + 1); i > 1; i--, p = p->bh_next) { |
order | 251 | mm/kmalloc.c | page->order = order; |
order | 252 | mm/kmalloc.c | page->nfree = NBLOCKS(order); |
order | 269 | mm/kmalloc.c | sizes[order].nmallocs++; |
order | 270 | mm/kmalloc.c | sizes[order].nbytesmalloced += size; |
order | 283 | mm/kmalloc.c | int order; |
order | 291 | mm/kmalloc.c | order = page->order; |
order | 292 | mm/kmalloc.c | pg = &sizes[order].firstfree; |
order | 295 | mm/kmalloc.c | pg = &sizes[order].dmafree; |
order | 298 | mm/kmalloc.c | if ((order < 0) || |
order | 299 | mm/kmalloc.c | (order >= sizeof(sizes) / sizeof(sizes[0])) || |
order | 303 | mm/kmalloc.c | p, page->next, page->order); |
order | 323 | mm/kmalloc.c | if (page->nfree == NBLOCKS(order)) { |
order | 336 | mm/kmalloc.c | sizes[order].npages--; |
order | 337 | mm/kmalloc.c | free_pages((long) page, sizes[order].gfporder); |
order | 339 | mm/kmalloc.c | sizes[order].nfrees++; |
order | 340 | mm/kmalloc.c | sizes[order].nbytesmalloced -= size; |
order | 88 | mm/page_alloc.c | static inline void free_pages_ok(unsigned long map_nr, unsigned long order) |
order | 90 | mm/page_alloc.c | unsigned long index = map_nr >> (1 + order); |
order | 91 | mm/page_alloc.c | unsigned long mask = (~0UL) << order; |
order | 96 | mm/page_alloc.c | nr_free_pages += 1 << order; |
order | 97 | mm/page_alloc.c | while (order < NR_MEM_LISTS-1) { |
order | 98 | mm/page_alloc.c | if (!change_bit(index, free_area[order].map)) |
order | 100 | mm/page_alloc.c | remove_mem_queue(&free_area[order].list, list(map_nr ^ (1+~mask))); |
order | 102 | mm/page_alloc.c | order++; |
order | 106 | mm/page_alloc.c | add_mem_queue(&free_area[order].list, list(map_nr)); |
order | 126 | mm/page_alloc.c | void free_pages(unsigned long addr, unsigned long order) |
order | 139 | mm/page_alloc.c | free_pages_ok(map_nr, order); |
order | 156 | mm/page_alloc.c | #define MARK_USED(index, order, area) \ |
order | 157 | mm/page_alloc.c | change_bit((index) >> (1+(order)), (area)->map) |
order | 160 | mm/page_alloc.c | #define RMQUEUE(order, dma) \ |
order | 161 | mm/page_alloc.c | do { struct free_area_struct * area = free_area+order; \ |
order | 162 | mm/page_alloc.c | unsigned long new_order = order; \ |
order | 169 | mm/page_alloc.c | nr_free_pages -= 1 << order; \ |
order | 170 | mm/page_alloc.c | EXPAND(ret, map_nr, order, new_order, area); \ |
order | 193 | mm/page_alloc.c | unsigned long __get_free_pages(int priority, unsigned long order, int dma) |
order | 198 | mm/page_alloc.c | if (order >= NR_MEM_LISTS) |
order | 215 | mm/page_alloc.c | RMQUEUE(order, dma); |
order | 232 | mm/page_alloc.c | unsigned long order, flags; |
order | 238 | mm/page_alloc.c | for (order=0 ; order < NR_MEM_LISTS; order++) { |
order | 241 | mm/page_alloc.c | for (tmp = free_area[order].list.next ; tmp != &free_area[order].list ; tmp = tmp->next) { |
order | 244 | mm/page_alloc.c | total += nr * ((PAGE_SIZE>>10) << order); |
order | 245 | mm/page_alloc.c | printk("%lu*%lukB ", nr, (PAGE_SIZE>>10) << order); |