tag | line | file | source code |
priority | 111 | drivers/net/atp.c | #define alloc_skb(size, priority) (struct sk_buff *) kmalloc(size,priority) |
priority | 58 | fs/buffer.c | static int shrink_specific_buffers(unsigned int priority, int size); |
priority | 1391 | fs/buffer.c | int shrink_buffers(unsigned int priority) |
priority | 1393 | fs/buffer.c | if (priority < 2) { |
priority | 1397 | fs/buffer.c | if(priority == 2) wakeup_bdflush(1); |
priority | 1402 | fs/buffer.c | return shrink_specific_buffers(priority, 0); |
priority | 1405 | fs/buffer.c | static int shrink_specific_buffers(unsigned int priority, int size) |
priority | 1436 | fs/buffer.c | if(priority > 3 && nlist == BUF_SHARED) continue; |
priority | 1439 | fs/buffer.c | i = nr_buffers_type[nlist] >> priority; |
priority | 1447 | fs/buffer.c | if (priority) |
priority | 336 | fs/proc/array.c | (*p)->priority, /* this is the nice value --- |
priority | 365 | include/linux/fs.h | extern int shrink_buffers(unsigned int priority); |
priority | 10 | include/linux/malloc.h | void *deb_kmalloc(const char *deb_file, unsigned short deb_line,unsigned int size, int priority); |
priority | 20 | include/linux/malloc.h | void * kmalloc(unsigned int size, int priority); |
priority | 118 | include/linux/mm.h | #define __get_free_page(priority) __get_free_pages((priority),0) |
priority | 119 | include/linux/mm.h | extern unsigned long __get_free_pages(int priority, unsigned long gfporder); |
priority | 120 | include/linux/mm.h | extern inline unsigned long get_free_page(int priority) |
priority | 124 | include/linux/mm.h | page = __get_free_page(priority); |
priority | 179 | include/linux/sched.h | long priority; |
priority | 103 | include/linux/skbuff.h | extern struct sk_buff * alloc_skb(unsigned int size, int priority); |
priority | 105 | include/linux/skbuff.h | extern struct sk_buff * skb_clone(struct sk_buff *skb, int priority); |
priority | 221 | kernel/sched.c | current->counter < current->priority*2) { |
priority | 236 | kernel/sched.c | p->counter = (p->counter >> 1) + p->priority; |
priority | 614 | kernel/sched.c | if (current->priority < 15) |
priority | 719 | kernel/sched.c | newprio = current->priority - increment; |
priority | 724 | kernel/sched.c | current->priority = newprio; |
priority | 61 | kernel/sys.c | int priority; |
priority | 66 | kernel/sys.c | if ((priority = PZERO - niceval) <= 0) |
priority | 67 | kernel/sys.c | priority = 1; |
priority | 79 | kernel/sys.c | if (priority > (*p)->priority && !suser()) |
priority | 82 | kernel/sys.c | (*p)->priority = priority; |
priority | 98 | kernel/sys.c | if ((*p)->priority > max_prio) |
priority | 99 | kernel/sys.c | max_prio = (*p)->priority; |
priority | 163 | mm/kmalloc.c | void * kmalloc (size_t size, int priority) |
priority | 171 | mm/kmalloc.c | if (intr_count && priority != GFP_ATOMIC) { |
priority | 176 | mm/kmalloc.c | priority = GFP_ATOMIC; |
priority | 227 | mm/kmalloc.c | page = (struct page_descriptor *) __get_free_pages (priority & GFP_LEVEL_MASK, sizes[order].gfporder); |
priority | 304 | mm/swap.c | static int swap_out(unsigned int priority) |
priority | 311 | mm/swap.c | int counter = NR_TASKS * 2 >> priority; |
priority | 314 | mm/swap.c | counter = NR_TASKS * 2 >> priority; |
priority | 415 | mm/swap.c | static int swap_out(unsigned int priority) |
priority | 424 | mm/swap.c | counter >>= priority; |
priority | 473 | mm/swap.c | static int try_to_free_page(int priority) |
priority | 478 | mm/swap.c | if (priority != GFP_NOBUFFER && shrink_buffers(i)) |
priority | 603 | mm/swap.c | unsigned long __get_free_pages(int priority, unsigned long order) |
priority | 607 | mm/swap.c | if (intr_count && priority != GFP_ATOMIC) { |
priority | 611 | mm/swap.c | ((unsigned long *)&priority)[-1]); |
priority | 612 | mm/swap.c | priority = GFP_ATOMIC; |
priority | 618 | mm/swap.c | if ((priority==GFP_ATOMIC) || nr_free_pages > MAX_SECONDARY_PAGES) { |
priority | 624 | mm/swap.c | if (priority != GFP_BUFFER && try_to_free_page(priority)) |
priority | 414 | net/inet/arp.c | dev_queue_xmit(skb,skb->dev,skb->sk->priority); |
priority | 1899 | net/inet/ip.c | dev_queue_xmit(skb, dev, sk->priority); |
priority | 1977 | net/inet/ip.c | dev_queue_xmit(skb, dev, sk->priority); |
priority | 215 | net/inet/packet.c | dev_queue_xmit(skb, dev, sk->priority); |
priority | 434 | net/inet/skbuff.c | struct sk_buff *alloc_skb(unsigned int size,int priority) |
priority | 438 | net/inet/skbuff.c | if (intr_count && priority!=GFP_ATOMIC) { |
priority | 443 | net/inet/skbuff.c | priority = GFP_ATOMIC; |
priority | 448 | net/inet/skbuff.c | skb=(struct sk_buff *)kmalloc(size,priority); |
priority | 499 | net/inet/skbuff.c | struct sk_buff *skb_clone(struct sk_buff *skb, int priority) |
priority | 504 | net/inet/skbuff.c | n=alloc_skb(skb->mem_len-sizeof(struct sk_buff),priority); |
priority | 557 | net/inet/sock.c | sk->priority = val; |
priority | 642 | net/inet/sock.c | val = sk->priority; |
priority | 819 | net/inet/sock.c | sk->priority = 1; |
priority | 1519 | net/inet/sock.c | int priority) |
priority | 1523 | net/inet/sock.c | struct sk_buff * c = alloc_skb(size, priority); |
priority | 1532 | net/inet/sock.c | sk, size, force, priority)); |
priority | 1535 | net/inet/sock.c | return(alloc_skb(size, priority)); |
priority | 1540 | net/inet/sock.c | sock_rmalloc(struct sock *sk, unsigned long size, int force, int priority) |
priority | 1544 | net/inet/sock.c | struct sk_buff *c = alloc_skb(size, priority); |
priority | 1553 | net/inet/sock.c | sk,size,force, priority)); |
priority | 1556 | net/inet/sock.c | return(alloc_skb(size, priority)); |
priority | 130 | net/inet/sock.h | unsigned char priority; |
priority | 177 | net/inet/sock.h | int priority); |
priority | 180 | net/inet/sock.h | int priority); |
priority | 259 | net/inet/sock.h | int priority); |
priority | 262 | net/inet/sock.h | int priority); |