tag | line | file | source code |
cache | 126 | arch/alpha/kernel/setup.c | set_hae(hae.cache); /* sync HAE register w/hae_cache */ |
cache | 522 | drivers/net/de4x5.c | } cache; |
cache | 784 | drivers/net/de4x5.c | lp->cache.priv = tmp; |
cache | 835 | drivers/net/de4x5.c | kfree(lp->cache.priv); |
cache | 839 | drivers/net/de4x5.c | lp->cache.buf = tmp; |
cache | 1105 | drivers/net/de4x5.c | if (lp->cache.skb && !dev->interrupt) { |
cache | 1198 | drivers/net/de4x5.c | while (lp->cache.skb && !dev->tbusy && lp->tx_enable) { |
cache | 2701 | drivers/net/de4x5.c | while (lp->cache.skb) { |
cache | 2722 | drivers/net/de4x5.c | if (!lp->cache.save_cnt) { |
cache | 2730 | drivers/net/de4x5.c | lp->cache.save_cnt++; |
cache | 2744 | drivers/net/de4x5.c | if (lp->cache.save_cnt) { |
cache | 2750 | drivers/net/de4x5.c | lp->cache.save_cnt--; |
cache | 2766 | drivers/net/de4x5.c | lp->cache.csr0 = inl(DE4X5_BMR); |
cache | 2767 | drivers/net/de4x5.c | lp->cache.csr6 = (inl(DE4X5_OMR) & ~(OMR_ST | OMR_SR)); |
cache | 2768 | drivers/net/de4x5.c | lp->cache.csr7 = inl(DE4X5_IMR); |
cache | 2770 | drivers/net/de4x5.c | lp->cache.csr13 = inl(DE4X5_SICR); |
cache | 2771 | drivers/net/de4x5.c | lp->cache.csr14 = inl(DE4X5_STRR); |
cache | 2772 | drivers/net/de4x5.c | lp->cache.csr15 = inl(DE4X5_SIGR); |
cache | 2777 | drivers/net/de4x5.c | outl(lp->cache.csr0, DE4X5_BMR); |
cache | 2778 | drivers/net/de4x5.c | outl(lp->cache.csr6, DE4X5_OMR); |
cache | 2779 | drivers/net/de4x5.c | outl(lp->cache.csr7, DE4X5_IMR); |
cache | 2788 | drivers/net/de4x5.c | reset_init_sia(dev, lp->cache.csr13, lp->cache.csr14, |
cache | 2789 | drivers/net/de4x5.c | lp->cache.csr15); |
cache | 2803 | drivers/net/de4x5.c | if (lp->cache.skb) { |
cache | 2804 | drivers/net/de4x5.c | for (p=lp->cache.skb; p->next; p=p->next); |
cache | 2807 | drivers/net/de4x5.c | lp->cache.skb = skb; |
cache | 2818 | drivers/net/de4x5.c | struct sk_buff *p = lp->cache.skb; |
cache | 2820 | drivers/net/de4x5.c | lp->cache.skb = skb; |
cache | 2830 | drivers/net/de4x5.c | struct sk_buff *p = lp->cache.skb; |
cache | 2833 | drivers/net/de4x5.c | lp->cache.skb = p->next; |
cache | 4092 | drivers/net/de4x5.c | if (lp->cache.buf) { /* MAC buffers allocated? */ |
cache | 4093 | drivers/net/de4x5.c | kfree(lp->cache.buf); /* Free the MAC buffers */ |
cache | 4098 | drivers/net/de4x5.c | kfree(lp->cache.priv); /* Free the private area */ |
cache | 4105 | drivers/net/de4x5.c | if (lp->cache.buf) { /* Are MAC buffers allocated */ |
cache | 4106 | drivers/net/de4x5.c | kfree(lp->cache.buf); |
cache | 4112 | drivers/net/de4x5.c | kfree(lp->cache.priv); |
cache | 16 | fs/fat/cache.c | static struct fat_cache *fat_cache,cache[FAT_CACHE]; |
cache | 110 | fs/fat/cache.c | fat_cache = &cache[0]; |
cache | 112 | fs/fat/cache.c | cache[count].device = 0; |
cache | 113 | fs/fat/cache.c | cache[count].next = count == FAT_CACHE-1 ? NULL : |
cache | 114 | fs/fat/cache.c | &cache[count+1]; |
cache | 108 | fs/xiafs/bitmap.c | int bznr[], u_char cache, int first_zone, |
cache | 119 | fs/xiafs/bitmap.c | if (!cache) |
cache | 150 | fs/xiafs/bitmap.c | #define xiafs_unlock_super(sb, cache) if (cache) unlock_super(sb); |
cache | 169 | fs/xiafs/bitmap.c | int bznr[], u_char cache, int first_zone, int bmap_zones, |
cache | 195 | fs/xiafs/bitmap.c | bh = get__map_zone(sb, bmap_buf, bznr, cache, first_zone, |
cache | 208 | fs/xiafs/bitmap.c | if (cache) |
cache | 210 | fs/xiafs/bitmap.c | xiafs_unlock_super(sb, cache); |
cache | 283 | include/asm-alpha/alcor.h | if (msb != hae.cache) { |
cache | 298 | include/asm-alpha/alcor.h | if (msb != hae.cache) { |
cache | 317 | include/asm-alpha/alcor.h | if (msb != hae.cache) { |
cache | 329 | include/asm-alpha/alcor.h | if (msb != hae.cache) { |
cache | 321 | include/asm-alpha/apecs.h | if (msb != hae.cache) { |
cache | 338 | include/asm-alpha/apecs.h | if (msb != hae.cache) { |
cache | 359 | include/asm-alpha/apecs.h | if (msb != hae.cache) { |
cache | 373 | include/asm-alpha/apecs.h | if (msb != hae.cache) { |
cache | 21 | include/asm-alpha/io.h | unsigned long cache; |
cache | 42 | include/asm-alpha/io.h | hae.cache = new_hae; |
cache | 94 | include/asm-alpha/jensen.h | if (addr != hae.cache) |
cache | 226 | include/asm-alpha/lca.h | if (msb != hae.cache) { |
cache | 243 | include/asm-alpha/lca.h | if (msb != hae.cache) { |
cache | 265 | include/asm-alpha/lca.h | if (msb != hae.cache) { |
cache | 281 | include/asm-alpha/lca.h | if (msb != hae.cache) { |
cache | 33 | include/asm-mips/cachectl.h | extern int sys_cacheflush(void *addr, int nbytes, int cache); |
cache | 300 | net/ipv4/icmp.c | struct icmp_xrl_cache cache[XRLIM_CACHE_SIZE]; /* Rate cache */ |
cache | 360 | net/ipv4/icmp.c | xr->cache[entry].daddr = INADDR_NONE; |
cache | 384 | net/ipv4/icmp.c | for (c = r->cache; c < &r->cache[XRLIM_CACHE_SIZE]; c++) /* Cache lookup */ |
cache | 390 | net/ipv4/icmp.c | if (c == &r->cache[XRLIM_CACHE_SIZE]) { /* Cache miss */ |
cache | 393 | net/ipv4/icmp.c | c = r->cache; |
cache | 394 | net/ipv4/icmp.c | for (d = r->cache; d < &r->cache[XRLIM_CACHE_SIZE]; d++) |
cache | 97 | net/ipv4/ipmr.c | static void ipmr_cache_delete(struct mfc_cache *cache) |
cache | 107 | net/ipv4/ipmr.c | if(cache->mfc_flags&MFC_QUEUED) |
cache | 110 | net/ipv4/ipmr.c | del_timer(&cache->mfc_timer); |
cache | 114 | net/ipv4/ipmr.c | line=MFC_HASH(cache->mfc_mcastgrp,cache->mfc_origin); |
cache | 124 | net/ipv4/ipmr.c | if(*cp==cache) |
cache | 126 | net/ipv4/ipmr.c | *cp=cache->next; |
cache | 137 | net/ipv4/ipmr.c | if(cache->mfc_flags&MFC_QUEUED) |
cache | 140 | net/ipv4/ipmr.c | while((skb=skb_dequeue(&cache->mfc_unresolved))) |
cache | 143 | net/ipv4/ipmr.c | kfree_s(cache,sizeof(cache)); |
cache | 152 | net/ipv4/ipmr.c | struct mfc_cache *cache=(struct mfc_cache *)data; |
cache | 153 | net/ipv4/ipmr.c | ipmr_cache_delete(cache); |
cache | 174 | net/ipv4/ipmr.c | struct mfc_cache *cache; |
cache | 175 | net/ipv4/ipmr.c | cache=mfc_cache_array[line]; |
cache | 176 | net/ipv4/ipmr.c | while(cache!=NULL) |
cache | 178 | net/ipv4/ipmr.c | if(cache->mfc_origin==origin && cache->mfc_mcastgrp==mcastgrp) |
cache | 179 | net/ipv4/ipmr.c | return cache; |
cache | 180 | net/ipv4/ipmr.c | cache=cache->next; |
cache | 182 | net/ipv4/ipmr.c | cache=cache_resolve_queue; |
cache | 183 | net/ipv4/ipmr.c | while(cache!=NULL) |
cache | 185 | net/ipv4/ipmr.c | if(cache->mfc_origin==origin && cache->mfc_mcastgrp==mcastgrp) |
cache | 186 | net/ipv4/ipmr.c | return cache; |
cache | 187 | net/ipv4/ipmr.c | cache=cache->next; |
cache | 213 | net/ipv4/ipmr.c | static void ipmr_cache_resolve(struct mfc_cache *cache) |
cache | 220 | net/ipv4/ipmr.c | del_timer(&cache->mfc_timer); |
cache | 221 | net/ipv4/ipmr.c | cache->mfc_flags&=~MFC_QUEUED; |
cache | 228 | net/ipv4/ipmr.c | if((*p)==cache) |
cache | 230 | net/ipv4/ipmr.c | *p=cache->next; |
cache | 240 | net/ipv4/ipmr.c | ipmr_cache_insert(cache); |
cache | 244 | net/ipv4/ipmr.c | while((skb=skb_dequeue(&cache->mfc_unresolved))) |
cache | 296 | net/ipv4/ipmr.c | static void ipmr_cache_unresolved(struct mfc_cache *cache, vifi_t vifi, struct sk_buff *skb, int is_frag) |
cache | 298 | net/ipv4/ipmr.c | if(cache==NULL) |
cache | 303 | net/ipv4/ipmr.c | if(cache_resolve_queue_len>=10 || (cache=ipmr_cache_alloc(GFP_ATOMIC))==NULL) |
cache | 311 | net/ipv4/ipmr.c | cache->mfc_parent=vifi; |
cache | 312 | net/ipv4/ipmr.c | cache->mfc_origin=skb->ip_hdr->saddr; |
cache | 313 | net/ipv4/ipmr.c | cache->mfc_mcastgrp=skb->ip_hdr->daddr; |
cache | 314 | net/ipv4/ipmr.c | cache->mfc_flags=MFC_QUEUED; |
cache | 318 | net/ipv4/ipmr.c | cache->next=cache_resolve_queue; |
cache | 319 | net/ipv4/ipmr.c | cache_resolve_queue=cache; |
cache | 324 | net/ipv4/ipmr.c | cache->mfc_timer.expires=jiffies+10*HZ; |
cache | 325 | net/ipv4/ipmr.c | add_timer(&cache->mfc_timer); |
cache | 335 | net/ipv4/ipmr.c | if(cache->mfc_queuelen>3) |
cache | 344 | net/ipv4/ipmr.c | cache->mfc_queuelen++; |
cache | 346 | net/ipv4/ipmr.c | skb_queue_tail(&cache->mfc_unresolved,skb); |
cache | 355 | net/ipv4/ipmr.c | struct mfc_cache *cache; |
cache | 364 | net/ipv4/ipmr.c | cache=ipmr_cache_find(mfc->mfcc_origin.s_addr,mfc->mfcc_mcastgrp.s_addr); |
cache | 371 | net/ipv4/ipmr.c | if(cache) |
cache | 373 | net/ipv4/ipmr.c | ipmr_cache_delete(cache); |
cache | 380 | net/ipv4/ipmr.c | if(cache) |
cache | 386 | net/ipv4/ipmr.c | cache->mfc_flags|=MFC_RESOLVED; |
cache | 387 | net/ipv4/ipmr.c | memcpy(cache->mfc_ttls, mfc->mfcc_ttls,sizeof(cache->mfc_ttls)); |
cache | 394 | net/ipv4/ipmr.c | if(cache->mfc_flags&MFC_QUEUED) |
cache | 395 | net/ipv4/ipmr.c | ipmr_cache_resolve(cache); /* Unhook & send the frames */ |
cache | 404 | net/ipv4/ipmr.c | cache=ipmr_cache_alloc(GFP_ATOMIC); |
cache | 405 | net/ipv4/ipmr.c | if(cache==NULL) |
cache | 410 | net/ipv4/ipmr.c | cache->mfc_flags=MFC_RESOLVED; |
cache | 411 | net/ipv4/ipmr.c | cache->mfc_origin=mfc->mfcc_origin.s_addr; |
cache | 412 | net/ipv4/ipmr.c | cache->mfc_mcastgrp=mfc->mfcc_mcastgrp.s_addr; |
cache | 413 | net/ipv4/ipmr.c | cache->mfc_parent=mfc->mfcc_parent; |
cache | 414 | net/ipv4/ipmr.c | memcpy(cache->mfc_ttls, mfc->mfcc_ttls,sizeof(cache->mfc_ttls)); |
cache | 415 | net/ipv4/ipmr.c | ipmr_cache_insert(cache); |
cache | 729 | net/ipv4/ipmr.c | struct mfc_cache *cache; |
cache | 742 | net/ipv4/ipmr.c | cache=ipmr_cache_find(skb->ip_hdr->saddr,skb->ip_hdr->daddr); |
cache | 748 | net/ipv4/ipmr.c | if(cache==NULL || (cache->mfc_flags&MFC_QUEUED)) |
cache | 749 | net/ipv4/ipmr.c | ipmr_cache_unresolved(cache,vif,skb, is_frag); |
cache | 761 | net/ipv4/ipmr.c | if(skb->ip_hdr->ttl > cache->mfc_ttls[ct] && cache->mfc_ttls[ct]>0) |