tag | line | file | source code |
skb2 | 78 | drivers/net/loopback.c | struct sk_buff *skb2=skb; |
skb2 | 82 | drivers/net/loopback.c | dev_kfree_skb(skb2, FREE_READ); |
skb2 | 125 | drivers/net/tunnel.c | struct sk_buff *skb2; /* The output packet */ |
skb2 | 209 | drivers/net/tunnel.c | if ( !(skb2 = dev_alloc_skb(newlen)) ) |
skb2 | 219 | drivers/net/tunnel.c | skb2->free=1; |
skb2 | 220 | drivers/net/tunnel.c | skb_put(skb2,newlen); |
skb2 | 221 | drivers/net/tunnel.c | iph=skb2->h.iph=(struct iphdr *)skb2->data; |
skb2 | 222 | drivers/net/tunnel.c | skb2->ip_hdr=iph; |
skb2 | 223 | drivers/net/tunnel.c | memcpy(skb2->h.iph, skb->data, ip_header_len ); |
skb2 | 224 | drivers/net/tunnel.c | memcpy(skb2->data + ip_header_len, skb->data, skb->len); |
skb2 | 234 | drivers/net/tunnel.c | iph->tot_len = htons(skb2->len); |
skb2 | 250 | drivers/net/tunnel.c | if(ip_forward(skb2, dev, 0, iph->daddr)) |
skb2 | 251 | drivers/net/tunnel.c | kfree_skb(skb2, FREE_WRITE); |
skb2 | 1652 | net/appletalk/ddp.c | struct sk_buff *skb2=skb_clone(skb, GFP_KERNEL); |
skb2 | 1653 | net/appletalk/ddp.c | if(skb2) |
skb2 | 1658 | net/appletalk/ddp.c | if(aarp_send_ddp(dev,skb2,&usat->sat_addr, NULL)==-1) |
skb2 | 1659 | net/appletalk/ddp.c | kfree_skb(skb2, FREE_WRITE); |
skb2 | 410 | net/core/dev.c | struct sk_buff *skb2; |
skb2 | 411 | net/core/dev.c | if ((skb2 = skb_clone(skb, GFP_ATOMIC)) == NULL) |
skb2 | 413 | net/core/dev.c | skb2->h.raw = skb2->data + dev->hard_header_len; |
skb2 | 414 | net/core/dev.c | skb2->mac.raw = skb2->data; |
skb2 | 415 | net/core/dev.c | ptype->func(skb2, skb->dev, ptype); |
skb2 | 699 | net/core/dev.c | struct sk_buff *skb2=skb_clone(skb, GFP_ATOMIC); |
skb2 | 700 | net/core/dev.c | if(skb2) |
skb2 | 701 | net/core/dev.c | pt_prev->func(skb2,skb->dev, pt_prev); |
skb2 | 716 | net/core/dev.c | struct sk_buff *skb2; |
skb2 | 718 | net/core/dev.c | skb2=skb_clone(skb, GFP_ATOMIC); |
skb2 | 725 | net/core/dev.c | if(skb2) |
skb2 | 726 | net/core/dev.c | pt_prev->func(skb2, skb->dev, pt_prev); |
skb2 | 103 | net/core/skbuff.c | struct sk_buff *skb2 = skb->next; |
skb2 | 105 | net/core/skbuff.c | while (skb2 != skb && i < 5) { |
skb2 | 106 | net/core/skbuff.c | if (skb_check(skb2, 0, line, file) < 0) { |
skb2 | 111 | net/core/skbuff.c | skb2 = skb2->next; |
skb2 | 358 | net/ipv4/af_inet.c | struct sk_buff *skb2; |
skb2 | 370 | net/ipv4/af_inet.c | skb2 = skb->link3; |
skb2 | 372 | net/ipv4/af_inet.c | skb = skb2; |
skb2 | 89 | net/ipv4/ip_forward.c | struct sk_buff *skb2; /* Output packet */ |
skb2 | 258 | net/ipv4/ip_forward.c | skb2 = alloc_skb(dev2->hard_header_len + skb->len + encap + 15, GFP_ATOMIC); |
skb2 | 262 | net/ipv4/ip_forward.c | skb2 = alloc_skb(dev2->hard_header_len + skb->len + 15, GFP_ATOMIC); |
skb2 | 269 | net/ipv4/ip_forward.c | if (skb2 == NULL) |
skb2 | 277 | net/ipv4/ip_forward.c | IS_SKB(skb2); |
skb2 | 281 | net/ipv4/ip_forward.c | skb2->protocol=htons(ETH_P_IP); |
skb2 | 286 | net/ipv4/ip_forward.c | ip_encap(skb2,skb->len, dev2, raddr); |
skb2 | 290 | net/ipv4/ip_forward.c | ip_send(rt,skb2,raddr,skb->len,dev2,dev2->pa_addr); |
skb2 | 297 | net/ipv4/ip_forward.c | ptr = skb_put(skb2,skb->len); |
skb2 | 298 | net/ipv4/ip_forward.c | skb2->free = 1; |
skb2 | 299 | net/ipv4/ip_forward.c | skb2->h.raw = ptr; |
skb2 | 305 | net/ipv4/ip_forward.c | memcpy(skb2->proto_priv, skb->proto_priv, sizeof(skb->proto_priv)); |
skb2 | 306 | net/ipv4/ip_forward.c | iph = skb2->ip_hdr = skb2->h.iph; |
skb2 | 314 | net/ipv4/ip_forward.c | skb2 = skb; |
skb2 | 315 | net/ipv4/ip_forward.c | skb2->dev=dev2; |
skb2 | 404 | net/ipv4/ip_forward.c | if(skb2->len > dev2->mtu + dev2->hard_header_len) |
skb2 | 406 | net/ipv4/ip_forward.c | ip_fragment(NULL,skb2,dev2, is_frag); |
skb2 | 407 | net/ipv4/ip_forward.c | kfree_skb(skb2,FREE_WRITE); |
skb2 | 425 | net/ipv4/ip_forward.c | dev_queue_xmit(skb2, dev2, SOPRI_INTERACTIVE); |
skb2 | 427 | net/ipv4/ip_forward.c | dev_queue_xmit(skb2, dev2, SOPRI_BACKGROUND); |
skb2 | 429 | net/ipv4/ip_forward.c | dev_queue_xmit(skb2, dev2, SOPRI_NORMAL); |
skb2 | 445 | net/ipv4/ip_forward.c | if(skb==skb2) |
skb2 | 414 | net/ipv4/ip_fragment.c | struct sk_buff *skb2; |
skb2 | 599 | net/ipv4/ip_fragment.c | skb2 = ip_glue(qp); /* glue together the fragments */ |
skb2 | 600 | net/ipv4/ip_fragment.c | return(skb2); |
skb2 | 623 | net/ipv4/ip_fragment.c | struct sk_buff *skb2; |
skb2 | 711 | net/ipv4/ip_fragment.c | if ((skb2 = alloc_skb(len + hlen+15,GFP_ATOMIC)) == NULL) |
skb2 | 722 | net/ipv4/ip_fragment.c | skb2->arp = skb->arp; |
skb2 | 725 | net/ipv4/ip_fragment.c | skb2->free = 1; |
skb2 | 726 | net/ipv4/ip_fragment.c | skb_put(skb2,len + hlen); |
skb2 | 727 | net/ipv4/ip_fragment.c | skb2->h.raw=(char *) skb2->data; |
skb2 | 737 | net/ipv4/ip_fragment.c | sk->wmem_alloc += skb2->truesize; |
skb2 | 738 | net/ipv4/ip_fragment.c | skb2->sk=sk; |
skb2 | 741 | net/ipv4/ip_fragment.c | skb2->raddr = skb->raddr; /* For rebuild_header - must be here */ |
skb2 | 747 | net/ipv4/ip_fragment.c | memcpy(skb2->h.raw, raw, hlen); |
skb2 | 752 | net/ipv4/ip_fragment.c | memcpy(skb2->h.raw + hlen, ptr, len); |
skb2 | 755 | net/ipv4/ip_fragment.c | skb2->h.raw+=dev->hard_header_len; |
skb2 | 760 | net/ipv4/ip_fragment.c | iph = (struct iphdr *)(skb2->h.raw/*+dev->hard_header_len*/); |
skb2 | 762 | net/ipv4/ip_fragment.c | skb2->ip_hdr = iph; |
skb2 | 788 | net/ipv4/ip_fragment.c | ip_queue_xmit(sk, dev, skb2, 2); |
skb2 | 598 | net/ipv4/ip_fw.c | struct sk_buff *skb2; |
skb2 | 737 | net/ipv4/ip_fw.c | skb2 = alloc_skb(MAX_HEADER + skb->len+diff, GFP_ATOMIC); |
skb2 | 738 | net/ipv4/ip_fw.c | if (skb2 == NULL) { |
skb2 | 742 | net/ipv4/ip_fw.c | skb2->free = skb->free; |
skb2 | 743 | net/ipv4/ip_fw.c | skb_reserve(skb2,MAX_HEADER); |
skb2 | 744 | net/ipv4/ip_fw.c | skb_put(skb2,skb->len + diff); |
skb2 | 745 | net/ipv4/ip_fw.c | skb2->h.raw = skb2->data + (skb->h.raw - skb->data); |
skb2 | 746 | net/ipv4/ip_fw.c | iph=skb2->h.iph; |
skb2 | 758 | net/ipv4/ip_fw.c | memcpy(skb2->data, skb->data, (p - (char *)skb->data)); |
skb2 | 759 | net/ipv4/ip_fw.c | memcpy(&skb2->data[(p - (char *)skb->data)], buf, strlen(buf)); |
skb2 | 760 | net/ipv4/ip_fw.c | memcpy(&skb2->data[(p - (char *)skb->data) + strlen(buf)], data, |
skb2 | 775 | net/ipv4/ip_fw.c | return skb2; |
skb2 | 513 | net/ipv4/ip_input.c | struct sk_buff *skb2; |
skb2 | 530 | net/ipv4/ip_input.c | skb2 = skb_clone(skb, GFP_ATOMIC); |
skb2 | 531 | net/ipv4/ip_input.c | if(skb2==NULL) |
skb2 | 536 | net/ipv4/ip_input.c | skb2 = skb; |
skb2 | 546 | net/ipv4/ip_input.c | ipprot->handler(skb2, dev, opt, daddr, |
skb2 | 572 | net/ipv4/ip_input.c | struct sk_buff *skb2=skb_clone(skb, GFP_ATOMIC); |
skb2 | 573 | net/ipv4/ip_input.c | if(skb2) |
skb2 | 575 | net/ipv4/ip_input.c | skb2->free=1; |
skb2 | 576 | net/ipv4/ip_input.c | ipmr_forward(skb2, is_frag); |
skb2 | 725 | net/ipv4/ipmr.c | struct sk_buff *skb2; |
skb2 | 760 | net/ipv4/ipmr.c | skb2=skb_clone(skb, GFP_ATOMIC); |
skb2 | 761 | net/ipv4/ipmr.c | if(skb2) |
skb2 | 763 | net/ipv4/ipmr.c | skb2->free=1; |
skb2 | 764 | net/ipv4/ipmr.c | ipmr_queue_xmit(skb2, &vif_table[psend], skb->dev, is_frag); |
skb2 | 3498 | net/ipv4/tcp.c | struct sk_buff *skb2; |
skb2 | 3501 | net/ipv4/tcp.c | skb2 = sk->send_head; |
skb2 | 3514 | net/ipv4/tcp.c | while (skb2 != NULL) |
skb2 | 3516 | net/ipv4/tcp.c | skb = skb2; |
skb2 | 3517 | net/ipv4/tcp.c | skb2 = skb->link3; |
skb2 | 4106 | net/ipv4/tcp.c | struct sk_buff *skb1, *skb2; |
skb2 | 4302 | net/ipv4/tcp.c | for(skb2 = skb->next; |
skb2 | 4303 | net/ipv4/tcp.c | skb2 != (struct sk_buff *)&sk->receive_queue; |
skb2 | 4304 | net/ipv4/tcp.c | skb2 = skb2->next) |
skb2 | 4306 | net/ipv4/tcp.c | if (before(skb2->seq, sk->acked_seq+1)) |
skb2 | 4308 | net/ipv4/tcp.c | if (after(skb2->end_seq, sk->acked_seq)) |
skb2 | 4311 | net/ipv4/tcp.c | (skb2->end_seq - sk->acked_seq); |
skb2 | 4315 | net/ipv4/tcp.c | sk->acked_seq = skb2->end_seq; |
skb2 | 4317 | net/ipv4/tcp.c | skb2->acked = 1; |
skb2 | 4322 | net/ipv4/tcp.c | if (skb2->h.th->fin) |
skb2 | 356 | net/ipx/af_ipx.c | struct sk_buff *skb1 = NULL, *skb2 = NULL; |
skb2 | 429 | net/ipx/af_ipx.c | skb2 = skb_clone(skb1, GFP_ATOMIC); |
skb2 | 430 | net/ipx/af_ipx.c | if (skb2 != NULL) |
skb2 | 431 | net/ipx/af_ipx.c | skb2->arp = skb2->free = 1; |
skb2 | 434 | net/ipx/af_ipx.c | skb2 = skb1; |
skb2 | 439 | net/ipx/af_ipx.c | if (skb2 == NULL) |
skb2 | 443 | net/ipx/af_ipx.c | (void) ipxitf_def_skb_handler(sock2, skb2); |
skb2 | 451 | net/ipx/af_ipx.c | struct sk_buff *skb2; |
skb2 | 464 | net/ipx/af_ipx.c | skb2 = alloc_skb(len, GFP_ATOMIC); |
skb2 | 465 | net/ipx/af_ipx.c | if (skb2 != NULL) { |
skb2 | 466 | net/ipx/af_ipx.c | skb_reserve(skb2,out_offset); |
skb2 | 467 | net/ipx/af_ipx.c | skb2->h.raw=skb_put(skb2,skb->len); |
skb2 | 468 | net/ipx/af_ipx.c | skb2->free=1; |
skb2 | 469 | net/ipx/af_ipx.c | skb2->arp=1; |
skb2 | 470 | net/ipx/af_ipx.c | memcpy(skb2->h.raw, skb->h.raw, skb->len); |
skb2 | 473 | net/ipx/af_ipx.c | return skb2; |