tag | line | file | source code |
cache | 126 | arch/alpha/kernel/setup.c | set_hae(hae.cache); /* sync HAE register w/hae_cache */ |
cache | 500 | drivers/net/de4x5.c | } cache; |
cache | 754 | drivers/net/de4x5.c | lp->cache.priv = tmp; |
cache | 794 | drivers/net/de4x5.c | lp->cache.buf = tmp; |
cache | 1070 | drivers/net/de4x5.c | if (lp->cache.skb && !dev->interrupt) { |
cache | 1159 | drivers/net/de4x5.c | while (lp->cache.skb && !dev->tbusy && lp->tx_enable) { |
cache | 2510 | drivers/net/de4x5.c | if (!lp->cache.save_cnt) { |
cache | 2529 | drivers/net/de4x5.c | lp->cache.save_cnt++; |
cache | 2544 | drivers/net/de4x5.c | if (lp->cache.save_cnt) { |
cache | 2551 | drivers/net/de4x5.c | for (i=0; TX_BUFFS_AVAIL && lp->cache.skb; i++) { |
cache | 2559 | drivers/net/de4x5.c | lp->cache.save_cnt--; |
cache | 2574 | drivers/net/de4x5.c | lp->cache.csr0 = inl(DE4X5_BMR); |
cache | 2575 | drivers/net/de4x5.c | lp->cache.csr6 = (inl(DE4X5_OMR) & ~(OMR_ST | OMR_SR)); |
cache | 2576 | drivers/net/de4x5.c | lp->cache.csr7 = inl(DE4X5_IMR); |
cache | 2578 | drivers/net/de4x5.c | lp->cache.csr13 = inl(DE4X5_SICR); |
cache | 2579 | drivers/net/de4x5.c | lp->cache.csr14 = inl(DE4X5_STRR); |
cache | 2580 | drivers/net/de4x5.c | lp->cache.csr15 = inl(DE4X5_SIGR); |
cache | 2585 | drivers/net/de4x5.c | outl(lp->cache.csr0, DE4X5_BMR); |
cache | 2586 | drivers/net/de4x5.c | outl(lp->cache.csr6, DE4X5_OMR); |
cache | 2587 | drivers/net/de4x5.c | outl(lp->cache.csr7, DE4X5_IMR); |
cache | 2596 | drivers/net/de4x5.c | reset_init_sia(dev, lp->cache.csr13, lp->cache.csr14, |
cache | 2597 | drivers/net/de4x5.c | lp->cache.csr15); |
cache | 2610 | drivers/net/de4x5.c | if (lp->cache.skb) { |
cache | 2611 | drivers/net/de4x5.c | for (p=lp->cache.skb; p->next; p=p->next); |
cache | 2614 | drivers/net/de4x5.c | lp->cache.skb = skb; |
cache | 2624 | drivers/net/de4x5.c | struct sk_buff *p = lp->cache.skb; |
cache | 2626 | drivers/net/de4x5.c | lp->cache.skb = skb; |
cache | 2635 | drivers/net/de4x5.c | struct sk_buff *p = lp->cache.skb; |
cache | 2638 | drivers/net/de4x5.c | lp->cache.skb = p->next; |
cache | 3822 | drivers/net/de4x5.c | if (lp->cache.buf) { /* MAC buffers allocated? */ |
cache | 3823 | drivers/net/de4x5.c | kfree(lp->cache.buf); /* Free the MAC buffers */ |
cache | 3828 | drivers/net/de4x5.c | kfree(lp->cache.priv); /* Free the private area */ |
cache | 3835 | drivers/net/de4x5.c | if (lp->cache.buf) { /* Are MAC buffers allocated */ |
cache | 3836 | drivers/net/de4x5.c | kfree(lp->cache.buf); |
cache | 3842 | drivers/net/de4x5.c | kfree(lp->cache.priv); |
cache | 15 | fs/fat/cache.c | static struct fat_cache *fat_cache,cache[FAT_CACHE]; |
cache | 109 | fs/fat/cache.c | fat_cache = &cache[0]; |
cache | 111 | fs/fat/cache.c | cache[count].device = 0; |
cache | 112 | fs/fat/cache.c | cache[count].next = count == FAT_CACHE-1 ? NULL : |
cache | 113 | fs/fat/cache.c | &cache[count+1]; |
cache | 108 | fs/xiafs/bitmap.c | int bznr[], u_char cache, int first_zone, |
cache | 119 | fs/xiafs/bitmap.c | if (!cache) |
cache | 150 | fs/xiafs/bitmap.c | #define xiafs_unlock_super(sb, cache) if (cache) unlock_super(sb); |
cache | 169 | fs/xiafs/bitmap.c | int bznr[], u_char cache, int first_zone, int bmap_zones, |
cache | 195 | fs/xiafs/bitmap.c | bh = get__map_zone(sb, bmap_buf, bznr, cache, first_zone, |
cache | 208 | fs/xiafs/bitmap.c | if (cache) |
cache | 210 | fs/xiafs/bitmap.c | xiafs_unlock_super(sb, cache); |
cache | 283 | include/asm-alpha/alcor.h | if (msb != hae.cache) { |
cache | 298 | include/asm-alpha/alcor.h | if (msb != hae.cache) { |
cache | 317 | include/asm-alpha/alcor.h | if (msb != hae.cache) { |
cache | 329 | include/asm-alpha/alcor.h | if (msb != hae.cache) { |
cache | 253 | include/asm-alpha/apecs.h | if (msb != hae.cache) { |
cache | 270 | include/asm-alpha/apecs.h | if (msb != hae.cache) { |
cache | 291 | include/asm-alpha/apecs.h | if (msb != hae.cache) { |
cache | 305 | include/asm-alpha/apecs.h | if (msb != hae.cache) { |
cache | 17 | include/asm-alpha/io.h | unsigned long cache; |
cache | 38 | include/asm-alpha/io.h | hae.cache = new_hae; |
cache | 94 | include/asm-alpha/jensen.h | if (addr != hae.cache) |
cache | 226 | include/asm-alpha/lca.h | if (msb != hae.cache) { |
cache | 243 | include/asm-alpha/lca.h | if (msb != hae.cache) { |
cache | 265 | include/asm-alpha/lca.h | if (msb != hae.cache) { |
cache | 281 | include/asm-alpha/lca.h | if (msb != hae.cache) { |
cache | 33 | include/asm-mips/cachectl.h | extern int sys_cacheflush(void *addr, int nbytes, int cache); |
cache | 97 | net/ipv4/ipmr.c | static void ipmr_cache_delete(struct mfc_cache *cache) |
cache | 107 | net/ipv4/ipmr.c | if(cache->mfc_flags&MFC_QUEUED) |
cache | 110 | net/ipv4/ipmr.c | del_timer(&cache->mfc_timer); |
cache | 114 | net/ipv4/ipmr.c | line=MFC_HASH(cache->mfc_mcastgrp,cache->mfc_origin); |
cache | 124 | net/ipv4/ipmr.c | if(*cp==cache) |
cache | 126 | net/ipv4/ipmr.c | *cp=cache->next; |
cache | 137 | net/ipv4/ipmr.c | if(cache->mfc_flags&MFC_QUEUED) |
cache | 140 | net/ipv4/ipmr.c | while((skb=skb_dequeue(&cache->mfc_unresolved))) |
cache | 143 | net/ipv4/ipmr.c | kfree_s(cache,sizeof(cache)); |
cache | 152 | net/ipv4/ipmr.c | struct mfc_cache *cache=(struct mfc_cache *)data; |
cache | 153 | net/ipv4/ipmr.c | ipmr_cache_delete(cache); |
cache | 174 | net/ipv4/ipmr.c | struct mfc_cache *cache; |
cache | 175 | net/ipv4/ipmr.c | cache=mfc_cache_array[line]; |
cache | 176 | net/ipv4/ipmr.c | while(cache!=NULL) |
cache | 178 | net/ipv4/ipmr.c | if(cache->mfc_origin==origin && cache->mfc_mcastgrp==mcastgrp) |
cache | 179 | net/ipv4/ipmr.c | return cache; |
cache | 180 | net/ipv4/ipmr.c | cache=cache->next; |
cache | 182 | net/ipv4/ipmr.c | cache=cache_resolve_queue; |
cache | 183 | net/ipv4/ipmr.c | while(cache!=NULL) |
cache | 185 | net/ipv4/ipmr.c | if(cache->mfc_origin==origin && cache->mfc_mcastgrp==mcastgrp) |
cache | 186 | net/ipv4/ipmr.c | return cache; |
cache | 187 | net/ipv4/ipmr.c | cache=cache->next; |
cache | 213 | net/ipv4/ipmr.c | static void ipmr_cache_resolve(struct mfc_cache *cache) |
cache | 220 | net/ipv4/ipmr.c | del_timer(&cache->mfc_timer); |
cache | 221 | net/ipv4/ipmr.c | cache->mfc_flags&=~MFC_QUEUED; |
cache | 228 | net/ipv4/ipmr.c | if((*p)==cache) |
cache | 230 | net/ipv4/ipmr.c | *p=cache->next; |
cache | 240 | net/ipv4/ipmr.c | ipmr_cache_insert(cache); |
cache | 244 | net/ipv4/ipmr.c | while((skb=skb_dequeue(&cache->mfc_unresolved))) |
cache | 296 | net/ipv4/ipmr.c | static void ipmr_cache_unresolved(struct mfc_cache *cache, vifi_t vifi, struct sk_buff *skb, int is_frag) |
cache | 298 | net/ipv4/ipmr.c | if(cache==NULL) |
cache | 303 | net/ipv4/ipmr.c | if(cache_resolve_queue_len>=10 || (cache=ipmr_cache_alloc(GFP_ATOMIC))==NULL) |
cache | 311 | net/ipv4/ipmr.c | cache->mfc_parent=vifi; |
cache | 312 | net/ipv4/ipmr.c | cache->mfc_origin=skb->ip_hdr->saddr; |
cache | 313 | net/ipv4/ipmr.c | cache->mfc_mcastgrp=skb->ip_hdr->daddr; |
cache | 314 | net/ipv4/ipmr.c | cache->mfc_flags=MFC_QUEUED; |
cache | 318 | net/ipv4/ipmr.c | cache->next=cache_resolve_queue; |
cache | 319 | net/ipv4/ipmr.c | cache_resolve_queue=cache; |
cache | 324 | net/ipv4/ipmr.c | cache->mfc_timer.expires=jiffies+10*HZ; |
cache | 325 | net/ipv4/ipmr.c | add_timer(&cache->mfc_timer); |
cache | 335 | net/ipv4/ipmr.c | if(cache->mfc_queuelen>3) |
cache | 344 | net/ipv4/ipmr.c | cache->mfc_queuelen++; |
cache | 346 | net/ipv4/ipmr.c | skb_queue_tail(&cache->mfc_unresolved,skb); |
cache | 355 | net/ipv4/ipmr.c | struct mfc_cache *cache; |
cache | 364 | net/ipv4/ipmr.c | cache=ipmr_cache_find(mfc->mfcc_origin.s_addr,mfc->mfcc_mcastgrp.s_addr); |
cache | 371 | net/ipv4/ipmr.c | if(cache) |
cache | 373 | net/ipv4/ipmr.c | ipmr_cache_delete(cache); |
cache | 380 | net/ipv4/ipmr.c | if(cache) |
cache | 386 | net/ipv4/ipmr.c | cache->mfc_flags|=MFC_RESOLVED; |
cache | 387 | net/ipv4/ipmr.c | memcpy(cache->mfc_ttls, mfc->mfcc_ttls,sizeof(cache->mfc_ttls)); |
cache | 394 | net/ipv4/ipmr.c | if(cache->mfc_flags&MFC_QUEUED) |
cache | 395 | net/ipv4/ipmr.c | ipmr_cache_resolve(cache); /* Unhook & send the frames */ |
cache | 404 | net/ipv4/ipmr.c | cache=ipmr_cache_alloc(GFP_ATOMIC); |
cache | 405 | net/ipv4/ipmr.c | if(cache==NULL) |
cache | 410 | net/ipv4/ipmr.c | cache->mfc_flags=MFC_RESOLVED; |
cache | 411 | net/ipv4/ipmr.c | cache->mfc_origin=mfc->mfcc_origin.s_addr; |
cache | 412 | net/ipv4/ipmr.c | cache->mfc_mcastgrp=mfc->mfcc_mcastgrp.s_addr; |
cache | 413 | net/ipv4/ipmr.c | cache->mfc_parent=mfc->mfcc_parent; |
cache | 414 | net/ipv4/ipmr.c | memcpy(cache->mfc_ttls, mfc->mfcc_ttls,sizeof(cache->mfc_ttls)); |
cache | 415 | net/ipv4/ipmr.c | ipmr_cache_insert(cache); |
cache | 729 | net/ipv4/ipmr.c | struct mfc_cache *cache; |
cache | 742 | net/ipv4/ipmr.c | cache=ipmr_cache_find(skb->ip_hdr->saddr,skb->ip_hdr->daddr); |
cache | 748 | net/ipv4/ipmr.c | if(cache==NULL || (cache->mfc_flags&MFC_QUEUED)) |
cache | 749 | net/ipv4/ipmr.c | ipmr_cache_unresolved(cache,vif,skb, is_frag); |
cache | 761 | net/ipv4/ipmr.c | if(skb->ip_hdr->ttl > cache->mfc_ttls[ct] && cache->mfc_ttls[ct]>0) |