taglinefilesource code
buff47arch/alpha/lib/checksum.cstatic inline unsigned long do_csum(unsigned char * buff, int len)
buff54arch/alpha/lib/checksum.codd = 1 & (unsigned long) buff;
buff56arch/alpha/lib/checksum.cresult = *buff << 8;
buff58arch/alpha/lib/checksum.cbuff++;
buff62arch/alpha/lib/checksum.cif (2 & (unsigned long) buff) {
buff63arch/alpha/lib/checksum.cresult += *(unsigned short *) buff;
buff66arch/alpha/lib/checksum.cbuff += 2;
buff70arch/alpha/lib/checksum.cif (4 & (unsigned long) buff) {
buff71arch/alpha/lib/checksum.cresult += *(unsigned int *) buff;
buff74arch/alpha/lib/checksum.cbuff += 4;
buff80arch/alpha/lib/checksum.cunsigned long w = *(unsigned long *) buff;
buff82arch/alpha/lib/checksum.cbuff += 8;
buff91arch/alpha/lib/checksum.cresult += *(unsigned int *) buff;
buff92arch/alpha/lib/checksum.cbuff += 4;
buff96arch/alpha/lib/checksum.cresult += *(unsigned short *) buff;
buff97arch/alpha/lib/checksum.cbuff += 2;
buff101arch/alpha/lib/checksum.cresult += *buff;
buff130arch/alpha/lib/checksum.cunsigned int csum_partial(unsigned char * buff, int len, unsigned int sum)
buff132arch/alpha/lib/checksum.cunsigned long result = do_csum(buff, len);
buff167arch/alpha/lib/checksum.cunsigned short ip_compute_csum(unsigned char * buff, int len)
buff169arch/alpha/lib/checksum.creturn ~from64to16(do_csum(buff,len));
buff26arch/i386/lib/checksum.cunsigned int csum_partial(const unsigned char * buff, int len, unsigned int sum) {
buff93arch/i386/lib/checksum.c: "0"(sum), "c"(len), "S"(buff)
buff101arch/ppc/kernel/stubs.cunsigned int csum_partial(unsigned char * buff, int len, unsigned int sum)
buff592drivers/cdrom/sonycd535.cByte **buff, int buf_size)
buff626drivers/cdrom/sonycd535.cdata_buff = buff[sector_count++];
buff49drivers/net/apricot.c#define kfree_skbmem(buff, size) kfree_s(buff,size)
buff582drivers/net/arcnet.cint arcnetA_header(unsigned char *buff,struct device *dev,
buff585drivers/net/arcnet.cint arcnetS_header(unsigned char *buff,struct device *dev,
buff2712drivers/net/arcnet.cint arcnetA_header(unsigned char *buff,struct device *dev,
buff2722drivers/net/arcnet.cbuff;
buff2796drivers/net/arcnet.cint arcnetS_header(unsigned char *buff,struct device *dev,
buff2806drivers/net/arcnet.cbuff;
buff2862drivers/net/arcnet.cint arcnetA_rebuild_header(void *buff,struct device *dev,unsigned long dst,
buff2865drivers/net/arcnet.cstruct ClientData *head = (struct ClientData *)buff;
buff2900drivers/net/arcnet.cint arcnetS_rebuild_header(void *buff,struct device *dev,unsigned long dst,
buff2903drivers/net/arcnet.cstruct S_ClientData *head = (struct S_ClientData *)buff;
buff151drivers/net/eql.cstatic int eql_rebuild_header(void *buff, struct device *dev, 
buff419drivers/net/eql.ceql_rebuild_header(void *buff, struct device *dev, 
buff1084drivers/net/pi2.cstatic int pi_rebuild_header(void *buff, struct device *dev, unsigned long raddr,
buff1087drivers/net/pi2.creturn ax25_rebuild_header(buff, dev, raddr, skb);
buff145drivers/net/plip.cstatic int plip_rebuild_header(void *buff, struct device *dev,
buff857drivers/net/plip.cplip_rebuild_header(void *buff, struct device *dev, unsigned long dst,
buff861drivers/net/plip.cstruct ethhdr *eth = (struct ethhdr *)buff;
buff865drivers/net/plip.creturn nl->orig_rebuild_header(buff, dev, dst, skb);
buff202drivers/net/ppp.cstatic int ppp_dev_header (unsigned char *buff, struct device *dev,
buff3282drivers/net/ppp.cstatic int ppp_dev_header (unsigned char *buff, struct device *dev,
buff3295drivers/net/ppp.cppp_dev_rebuild (void *buff, struct device *dev, unsigned long raddr,
buff548drivers/net/slip.csl_rebuild_header(void *buff, struct device *dev, unsigned long raddr,
buff556drivers/net/slip.creturn ax25_rebuild_header(buff, dev, raddr, skb);
buff521drivers/scsi/aha1542.cvoid *buff = SCpnt->request_buffer;
buff648drivers/scsi/aha1542.cif(((unsigned int) buff & 0xff000000)) goto baddma;
buff649drivers/scsi/aha1542.cany2scsi(ccb[mbo].dataptr, buff);
buff258drivers/scsi/aha1740.cvoid *buff = SCpnt->request_buffer;
buff354drivers/scsi/aha1740.cecb[ecbno].dataptr = (long) buff;
buff79drivers/scsi/aic7xxx_proc.cstatic u8 buff[512];
buff125drivers/scsi/aic7xxx_proc.cmemset(buff, 0, sizeof(buff));
buff696drivers/scsi/eata_dma.cstatic char *buff;
buff705drivers/scsi/eata_dma.cbuff = dma_scratch;
buff709drivers/scsi/eata_dma.cmemset(buff, 0, 256);
buff717drivers/scsi/eata_dma.ccp->cp_dataDMA = htonl(virt_to_bus(buff));
buff744drivers/scsi/eata_dma.c(u32) sp->scsi_stat, buff, sp));
buff758drivers/scsi/eata_dma.creturn (buff);
buff853drivers/scsi/eata_dma.cchar *buff = 0;
buff907drivers/scsi/eata_dma.cbuff = get_board_data(base, gc->IRQ, gc->scsi_id[3]);
buff909drivers/scsi/eata_dma.cif (buff == NULL) {
buff933drivers/scsi/eata_dma.cif (gc->DMA_support == FALSE && buff != NULL)  
buff935drivers/scsi/eata_dma.c&buff[16], base);
buff979drivers/scsi/eata_dma.cstrncpy(SD(sh)->vendor, &buff[8], 8);
buff981drivers/scsi/eata_dma.cstrncpy(SD(sh)->name, &buff[16], 17);
buff983drivers/scsi/eata_dma.cSD(sh)->revision[0] = buff[32];
buff984drivers/scsi/eata_dma.cSD(sh)->revision[1] = buff[33];
buff985drivers/scsi/eata_dma.cSD(sh)->revision[2] = buff[34];
buff987drivers/scsi/eata_dma.cSD(sh)->revision[4] = buff[35];
buff1015drivers/scsi/eata_dma.cif (buff[21] == '4')
buff1017drivers/scsi/eata_dma.celse if (buff[21] == '2')
buff75drivers/scsi/eata_dma_proc.cstatic u8 buff[512];
buff106drivers/scsi/eata_dma_proc.cmemset(buff, 0, sizeof(buff));
buff108drivers/scsi/eata_dma_proc.ccc = (coco *)     (buff + 0x148);
buff109drivers/scsi/eata_dma_proc.cst = (scsitrans *)(buff + 0x164); 
buff110drivers/scsi/eata_dma_proc.csm = (scsimod *)  (buff + 0x16c);
buff111drivers/scsi/eata_dma_proc.chb = (hobu *)     (buff + 0x172);
buff112drivers/scsi/eata_dma_proc.csb = (scbu *)     (buff + 0x178);
buff113drivers/scsi/eata_dma_proc.cbt = (boty *)     (buff + 0x17e);
buff114drivers/scsi/eata_dma_proc.cmc = (memco *)    (buff + 0x186);
buff115drivers/scsi/eata_dma_proc.cfm = (firm *)     (buff + 0x18e);
buff116drivers/scsi/eata_dma_proc.csi = (subinf *)   (buff + 0x196); 
buff117drivers/scsi/eata_dma_proc.cpi = (pcinf *)    (buff + 0x19c);
buff118drivers/scsi/eata_dma_proc.cal = (arrlim *)   (buff + 0x1a2);
buff191drivers/scsi/eata_dma_proc.cscsi_do_cmd (&scmd, cmnd, buff + 0x144, 0x66,  
buff539drivers/scsi/eata_pio.cstatic char buff[256];
buff543drivers/scsi/eata_pio.cmemset(buff, 0, sizeof(buff));
buff575drivers/scsi/eata_pio.cinsw(base+HA_RDATA, &buff, 127);
buff577drivers/scsi/eata_pio.creturn (buff);
buff679drivers/scsi/eata_pio.cchar *buff;
buff693drivers/scsi/eata_pio.cif ((buff = get_pio_board_data((uint)base, gc->IRQ, gc->scsi_id[3], 
buff737drivers/scsi/eata_pio.cstrncpy(SD(sh)->vendor, &buff[8], 8);
buff739drivers/scsi/eata_pio.cstrncpy(SD(sh)->name, &buff[16], 17);
buff741drivers/scsi/eata_pio.cSD(sh)->revision[0] = buff[32];
buff742drivers/scsi/eata_pio.cSD(sh)->revision[1] = buff[33];
buff743drivers/scsi/eata_pio.cSD(sh)->revision[2] = buff[34];
buff745drivers/scsi/eata_pio.cSD(sh)->revision[4] = buff[35];
buff772drivers/scsi/eata_pio.cif (buff[21] == '4')
buff774drivers/scsi/eata_pio.celse if (buff[21] == '2')
buff46drivers/scsi/eata_pio_proc.cstatic u8 buff[512];
buff63drivers/scsi/eata_pio_proc.cmemset(buff, 0, sizeof(buff));
buff423drivers/scsi/in2000.cvoid *buff = SCpnt->request_buffer;
buff505drivers/scsi/in2000.cin2000_scatter = (struct scatterlist *) buff;
buff511drivers/scsi/in2000.cin2000_dataptr = (unsigned short*) buff;
buff182drivers/scsi/scsi_debug.cunsigned char * buff;
buff195drivers/scsi/scsi_debug.cbuff = (unsigned char *) SCpnt->request_buffer;
buff218drivers/scsi/scsi_debug.cprintk("scsi_debug: Requesting sense buffer (%x %x %x %d):", SCpnt, buff, done, bufflen);
buff223drivers/scsi/scsi_debug.cmemset(buff, 0, bufflen);
buff224drivers/scsi/scsi_debug.cmemcpy(buff, sense_buffer, bufflen);
buff235drivers/scsi/scsi_debug.cprintk("Inquiry...(%x %d)\n", buff, bufflen);
buff236drivers/scsi/scsi_debug.cmemset(buff, 0, bufflen);
buff237drivers/scsi/scsi_debug.cbuff[0] = TYPE_DISK;
buff238drivers/scsi/scsi_debug.cbuff[1] = 0x80;  /* Removable disk */
buff239drivers/scsi/scsi_debug.cbuff[2] = 1;
buff240drivers/scsi/scsi_debug.cbuff[4] = 33 - 5;
buff241drivers/scsi/scsi_debug.cmemcpy(&buff[8],"Foo Inc",7);
buff242drivers/scsi/scsi_debug.cmemcpy(&buff[16],"XYZZY",5);
buff243drivers/scsi/scsi_debug.cmemcpy(&buff[32],"1",1);
buff247drivers/scsi/scsi_debug.cprintk("Test unit ready(%x %d)\n", buff, bufflen);
buff248drivers/scsi/scsi_debug.cif (buff)
buff249drivers/scsi/scsi_debug.cmemset(buff, 0, bufflen);
buff255drivers/scsi/scsi_debug.cmemset(buff, 0, bufflen);
buff256drivers/scsi/scsi_debug.cbuff[0] = (CAPACITY >> 24);
buff257drivers/scsi/scsi_debug.cbuff[1] = (CAPACITY >> 16) & 0xff;
buff258drivers/scsi/scsi_debug.cbuff[2] = (CAPACITY >> 8) & 0xff;
buff259drivers/scsi/scsi_debug.cbuff[3] = CAPACITY & 0xff;
buff260drivers/scsi/scsi_debug.cbuff[6] = 2; /* 512 byte sectors */
buff291drivers/scsi/scsi_debug.csgpnt = (struct scatterlist *) buff;
buff292drivers/scsi/scsi_debug.cbuff = sgpnt[sgcount].address;
buff301drivers/scsi/scsi_debug.cmemset(buff, 0, bufflen);
buff306drivers/scsi/scsi_debug.cmemset(buff, 0, bufflen);
buff307drivers/scsi/scsi_debug.c*((unsigned short *) (buff+510)) = 0xAA55;
buff308drivers/scsi/scsi_debug.cp = (struct partition* ) (buff + 0x1be);
buff348drivers/scsi/scsi_debug.cmemcpy(buff, &target, sizeof(target));
buff349drivers/scsi/scsi_debug.cmemcpy(buff+sizeof(target), cmd, 24);
buff350drivers/scsi/scsi_debug.cmemcpy(buff+60, &block, sizeof(block));
buff351drivers/scsi/scsi_debug.cmemcpy(buff+64, SCpnt, sizeof(Scsi_Cmnd));
buff356drivers/scsi/scsi_debug.cmemcpy(buff+128, bh, sizeof(struct buffer_head));
buff363drivers/scsi/scsi_debug.cbuff = sgpnt[sgcount].address;
buff390drivers/scsi/scsi_debug.csgpnt = (struct scatterlist *) buff;
buff391drivers/scsi/scsi_debug.cbuff = sgpnt[sgcount].address;
buff394drivers/scsi/scsi_debug.cif (block != *((unsigned long *) (buff+60))) {
buff395drivers/scsi/scsi_debug.cprintk("%x %x :",block,  *((unsigned long *) (buff+60)));
buff513drivers/scsi/sd.cchar * buff, *bounce_buffer;
buff628drivers/scsi/sd.cbuff = SCpnt->request.buffer;
buff647drivers/scsi/sd.cbuff = SCpnt->request.buffer;
buff692drivers/scsi/sd.cbuff = SCpnt->request.buffer;
buff697drivers/scsi/sd.cbuff = (char *) sgpnt;
buff738drivers/scsi/sd.cbuff = SCpnt->request.buffer;
buff819drivers/scsi/sd.cif (((long) buff) + (this_count << 9) - 1 > ISA_DMA_THRESHOLD && 
buff822drivers/scsi/sd.cbuff = bounce_buffer;
buff824drivers/scsi/sd.cbuff = (char *) scsi_malloc(this_count << 9);
buff825drivers/scsi/sd.cif(buff == NULL) {  /* Try backing off a bit if we are low on mem*/
buff827drivers/scsi/sd.cbuff = (char *) scsi_malloc(this_count << 9);
buff828drivers/scsi/sd.cif(!buff) panic("Ran out of DMA buffers.");
buff831drivers/scsi/sd.cmemcpy(buff, (char *)SCpnt->request.buffer, this_count << 9);
buff889drivers/scsi/sd.cscsi_do_cmd (SCpnt, (void *) cmd, buff, 
buff95drivers/scsi/seagate.cvoid *buff, int bufflen, int reselect);
buff618drivers/scsi/seagate.cvoid *buff, int bufflen, int reselect) {
buff56drivers/scsi/sg.cchar *buff;   /* the buffer */
buff64drivers/scsi/sg.cstatic void sg_free(char *buff,int size);
buff136drivers/scsi/sg.cif (scsi_generics[dev].buff != NULL)
buff137drivers/scsi/sg.csg_free(scsi_generics[dev].buff,scsi_generics[dev].buff_len);
buff138drivers/scsi/sg.cscsi_generics[dev].buff=NULL;
buff181drivers/scsi/sg.cstatic void sg_free(char *buff,int size) 
buff184drivers/scsi/sg.cif (buff==big_buff)
buff191drivers/scsi/sg.cscsi_free(buff,size);
buff241drivers/scsi/sg.cmemcpy_tofs(buf,device->buff,count-sizeof(struct sg_header));
buff251drivers/scsi/sg.csg_free(device->buff,device->buff_len);
buff252drivers/scsi/sg.cdevice->buff = NULL;
buff374drivers/scsi/sg.cif ((bsize<0) || !(device->buff=sg_malloc(device->buff_len=bsize)))
buff393drivers/scsi/sg.csg_free(device->buff,device->buff_len);
buff394drivers/scsi/sg.cdevice->buff = NULL;
buff427drivers/scsi/sg.csg_free( device->buff, device->buff_len );
buff428drivers/scsi/sg.cdevice->buff = NULL;
buff443drivers/scsi/sg.cif( direction == SG_SEND )  memcpy_fromfs(device->buff,buf, amt);
buff460drivers/scsi/sg.c(void *) device->buff,amt,
buff587drivers/scsi/sg.cscsi_generics[i].buff=NULL;
buff159drivers/sound/audio.ctranslate_bytes (const unsigned char *table, unsigned char *buff, int n)
buff167drivers/sound/audio.cbuff[i] = table[buff[i]];
buff172drivers/sound/audio.ctranslate_bytes (const void *table, void *buff, int n)
buff181drivers/sound/audio.c:     "b" ((long) table), "c" (n), "D" ((long) buff), "S" ((long) buff)
buff47fs/smbfs/sock.cstatic int _send(struct socket *sock, const void *buff, int len, int nonblock, unsigned flags) {
buff51fs/smbfs/sock.ciov.iov_base = (void *)buff;
buff33include/asm-alpha/checksum.hextern unsigned int csum_partial(const unsigned char * buff, int len, unsigned int sum);
buff56include/asm-alpha/checksum.hextern unsigned short ip_compute_csum(unsigned char * buff, int len);
buff28include/asm-alpha/ipsum.hextern inline unsigned short ip_compute_csum(unsigned char * buff, int len)
buff39include/asm-alpha/ipsum.hstatic inline unsigned short ip_fast_csum(unsigned char * buff, int wlen)
buff16include/asm-i386/checksum.hunsigned int csum_partial(const unsigned char * buff, int len, unsigned int sum);
buff123include/asm-i386/checksum.hstatic inline unsigned short ip_compute_csum(unsigned char * buff, int len) {
buff135include/asm-i386/checksum.h: "0" (csum_partial(buff, len, 0)));
buff33include/asm-ppc/checksum.hextern unsigned int csum_partial(const unsigned char * buff, int len, unsigned int sum);
buff56include/asm-ppc/checksum.hextern unsigned short ip_compute_csum(unsigned char * buff, int len);
buff25include/asm-sparc/checksum.hdo_csum(unsigned char * buff, int len)
buff32include/asm-sparc/checksum.hodd = 1 & (unsigned long) buff;
buff34include/asm-sparc/checksum.hresult = *buff << 8;
buff36include/asm-sparc/checksum.hbuff++;
buff40include/asm-sparc/checksum.hif (2 & (unsigned long) buff) {
buff41include/asm-sparc/checksum.hresult += *(unsigned short *) buff;
buff44include/asm-sparc/checksum.hbuff += 2;
buff50include/asm-sparc/checksum.hunsigned long w = *(unsigned long *) buff;
buff52include/asm-sparc/checksum.hbuff += 4;
buff62include/asm-sparc/checksum.hresult += *(unsigned short *) buff;
buff63include/asm-sparc/checksum.hbuff += 2;
buff67include/asm-sparc/checksum.hresult += (*buff) << 8;
buff92include/asm-sparc/checksum.hextern inline unsigned int csum_partial(unsigned char * buff, int len, unsigned int sum)
buff94include/asm-sparc/checksum.hunsigned long result = do_csum(buff, len);
buff128include/asm-sparc/checksum.hextern inline unsigned short ip_compute_csum(unsigned char * buff, int len)
buff130include/asm-sparc/checksum.hreturn ~from32to16(do_csum(buff,len));
buff34include/linux/etherdevice.hextern int    eth_rebuild_header(void *buff, struct device *dev,
buff229include/linux/netdevice.hextern int    dev_rint(unsigned char *buff, long len, int flags,
buff34include/linux/trdevice.hextern int    tr_rebuild_header(void *buff, struct device *dev,
buff36include/net/protocol.hvoid      (*err_handler)(int type, int code, unsigned char *buff,
buff29include/net/raw.hextern int  raw_read(struct sock *sk, unsigned char *buff,
buff288include/net/sock.hint      (*rcv)(struct sk_buff *buff, struct device *dev,
buff41include/net/udp.hextern int  udp_read(struct sock *sk, unsigned char *buff,
buff71net/802/tr.cint tr_rebuild_header(void *buff, struct device *dev, unsigned long dest,
buff74net/802/tr.cstruct trh_hdr *trh=(struct trh_hdr *)buff;
buff75net/802/tr.cstruct trllc *trllc=(struct trllc *)(buff+sizeof(struct trh_hdr));
buff2212net/ax25/af_ax25.cunsigned char *buff = skb_push(skb, AX25_HEADER_LEN);
buff2214net/ax25/af_ax25.c*buff++ = 0;  /* KISS DATA */
buff2217net/ax25/af_ax25.cmemcpy(buff, daddr, dev->addr_len);  /* Address specified */
buff2219net/ax25/af_ax25.cbuff[6] &= ~LAPB_C;
buff2220net/ax25/af_ax25.cbuff[6] &= ~LAPB_E;
buff2221net/ax25/af_ax25.cbuff[6] |= SSSID_SPARE;
buff2222net/ax25/af_ax25.cbuff += AX25_ADDR_LEN;
buff2225net/ax25/af_ax25.cmemcpy(buff, saddr, dev->addr_len);
buff2227net/ax25/af_ax25.cmemcpy(buff, dev->dev_addr, dev->addr_len);
buff2229net/ax25/af_ax25.cbuff[6] &= ~LAPB_C;
buff2230net/ax25/af_ax25.cbuff[6] |= LAPB_E;
buff2231net/ax25/af_ax25.cbuff[6] |= SSSID_SPARE;
buff2232net/ax25/af_ax25.cbuff   += AX25_ADDR_LEN;
buff2234net/ax25/af_ax25.c*buff++ = LAPB_UI;  /* UI */
buff2239net/ax25/af_ax25.c*buff++ = AX25_P_IP;
buff2243net/ax25/af_ax25.c*buff++ = AX25_P_ARP;
buff2246net/ax25/af_ax25.c*buff++ = 0;
buff484net/core/dev.cint dev_rint(unsigned char *buff, long len, int flags, struct device *dev)
buff492net/core/dev.cif (dev == NULL || buff == NULL || len <= 0) 
buff497net/core/dev.cskb = (struct sk_buff *) buff;
buff530net/core/dev.c(unsigned long) buff);
buff531net/core/dev.cmemcpy(to, buff, amount);
buff534net/core/dev.cbuff += amount;
buff536net/core/dev.cif ((unsigned long) buff == dev->rmem_end)
buff537net/core/dev.cbuff = (unsigned char *) dev->rmem_start;
buff139net/ethernet/eth.cint eth_rebuild_header(void *buff, struct device *dev, unsigned long dst,
buff142net/ethernet/eth.cstruct ethhdr *eth = (struct ethhdr *)buff;
buff1551net/ipv4/tcp.cstruct sk_buff *buff;
buff1564net/ipv4/tcp.cbuff = sock_wmalloc(sk, MAX_ACK_SIZE, 1, GFP_ATOMIC);
buff1565net/ipv4/tcp.cif (buff == NULL) 
buff1586net/ipv4/tcp.cbuff->sk = sk;
buff1587net/ipv4/tcp.cbuff->localroute = sk->localroute;
buff1593net/ipv4/tcp.ctmp = sk->prot->build_header(buff, sk->saddr, daddr, &dev,
buff1597net/ipv4/tcp.cbuff->free = 1;
buff1598net/ipv4/tcp.csock_wfree(sk, buff);
buff1601net/ipv4/tcp.ct1 =(struct tcphdr *)skb_put(buff,sizeof(struct tcphdr));
buff1655net/ipv4/tcp.csk->prot->queue_xmit(sk, dev, buff, 1);
buff2060net/ipv4/tcp.cstruct sk_buff *buff;
buff2083net/ipv4/tcp.cbuff = sock_wmalloc(sk,MAX_ACK_SIZE,1, GFP_ATOMIC);
buff2084net/ipv4/tcp.cif (buff == NULL) 
buff2091net/ipv4/tcp.cbuff->sk = sk;
buff2092net/ipv4/tcp.cbuff->localroute = sk->localroute;
buff2098net/ipv4/tcp.ctmp = sk->prot->build_header(buff, sk->saddr, sk->daddr, &dev,
buff2102net/ipv4/tcp.cbuff->free = 1;
buff2103net/ipv4/tcp.csock_wfree(sk, buff);
buff2107net/ipv4/tcp.ct1 =(struct tcphdr *)skb_put(buff,sizeof(struct tcphdr));
buff2125net/ipv4/tcp.csk->prot->queue_xmit(sk, dev, buff, 1);
buff2592net/ipv4/tcp.cstruct sk_buff *buff;
buff2598net/ipv4/tcp.cbuff = sock_wmalloc(sk, MAX_RESET_SIZE,1 , GFP_KERNEL);
buff2601net/ipv4/tcp.cif (buff == NULL)
buff2612net/ipv4/tcp.cbuff->sk = sk;
buff2613net/ipv4/tcp.cbuff->localroute = sk->localroute;
buff2619net/ipv4/tcp.ctmp = prot->build_header(buff,sk->saddr, sk->daddr, &dev,
buff2630net/ipv4/tcp.cbuff->free = 1;
buff2631net/ipv4/tcp.csock_wfree(sk,buff);
buff2646net/ipv4/tcp.ct1 =(struct tcphdr *)skb_put(buff,sizeof(struct tcphdr));
buff2647net/ipv4/tcp.cbuff->dev = dev;
buff2649net/ipv4/tcp.cbuff->seq = sk->write_seq;
buff2651net/ipv4/tcp.cbuff->end_seq = sk->write_seq;
buff2652net/ipv4/tcp.ct1->seq = htonl(buff->seq);
buff2668net/ipv4/tcp.cbuff->free = 0;
buff2669net/ipv4/tcp.cif (buff->next != NULL) 
buff2672net/ipv4/tcp.cskb_unlink(buff);
buff2674net/ipv4/tcp.cskb_queue_tail(&sk->write_queue, buff);
buff2679net/ipv4/tcp.csk->prot->queue_xmit(sk, dev, buff, 0);
buff2747net/ipv4/tcp.cstruct sk_buff *buff;
buff2764net/ipv4/tcp.cbuff = sock_wmalloc(NULL, MAX_RESET_SIZE, 1, GFP_ATOMIC);
buff2765net/ipv4/tcp.cif (buff == NULL) 
buff2768net/ipv4/tcp.cbuff->sk = NULL;
buff2769net/ipv4/tcp.cbuff->dev = dev;
buff2770net/ipv4/tcp.cbuff->localroute = 0;
buff2776net/ipv4/tcp.ctmp = prot->build_header(buff, saddr, daddr, &ndev, IPPROTO_TCP, opt,
buff2780net/ipv4/tcp.cbuff->free = 1;
buff2781net/ipv4/tcp.csock_wfree(NULL, buff);
buff2785net/ipv4/tcp.ct1 =(struct tcphdr *)skb_put(buff,sizeof(struct tcphdr));
buff2819net/ipv4/tcp.cprot->queue_xmit(NULL, ndev, buff, 1);
buff2924net/ipv4/tcp.cstruct sk_buff *buff;
buff3128net/ipv4/tcp.cbuff = sock_wmalloc(newsk, MAX_SYN_SIZE, 1, GFP_ATOMIC);
buff3129net/ipv4/tcp.cif (buff == NULL) 
buff3141net/ipv4/tcp.cbuff->sk = newsk;
buff3142net/ipv4/tcp.cbuff->localroute = newsk->localroute;
buff3148net/ipv4/tcp.ctmp = sk->prot->build_header(buff, newsk->saddr, newsk->daddr, &ndev,
buff3158net/ipv4/tcp.cbuff->free = 1;
buff3159net/ipv4/tcp.ckfree_skb(buff,FREE_WRITE);
buff3169net/ipv4/tcp.ct1 =(struct tcphdr *)skb_put(buff,sizeof(struct tcphdr));
buff3172net/ipv4/tcp.cbuff->seq = newsk->write_seq++;
buff3173net/ipv4/tcp.cbuff->end_seq = newsk->write_seq;
buff3179net/ipv4/tcp.ct1->seq = ntohl(buff->seq);
buff3192net/ipv4/tcp.cptr = skb_put(buff,4);
buff3199net/ipv4/tcp.cnewsk->prot->queue_xmit(newsk, ndev, buff, 0);
buff4531net/ipv4/tcp.cstruct sk_buff *buff;
buff4578net/ipv4/tcp.cbuff = sock_wmalloc(sk,MAX_SYN_SIZE,0, GFP_KERNEL);
buff4579net/ipv4/tcp.cif (buff == NULL) 
buff4584net/ipv4/tcp.cbuff->sk = sk;
buff4585net/ipv4/tcp.cbuff->free = 0;
buff4586net/ipv4/tcp.cbuff->localroute = sk->localroute;
buff4593net/ipv4/tcp.ctmp = sk->prot->build_header(buff, sk->saddr, sk->daddr, &dev,
buff4597net/ipv4/tcp.csock_wfree(sk, buff);
buff4605net/ipv4/tcp.ct1 = (struct tcphdr *) skb_put(buff,sizeof(struct tcphdr));
buff4608net/ipv4/tcp.cbuff->seq = sk->write_seq++;
buff4609net/ipv4/tcp.ct1->seq = htonl(buff->seq);
buff4611net/ipv4/tcp.cbuff->end_seq = sk->write_seq;
buff4664net/ipv4/tcp.cptr = skb_put(buff,4);
buff4688net/ipv4/tcp.csk->prot->queue_xmit(sk, dev, buff, 0);  
buff5211net/ipv4/tcp.cstruct sk_buff *buff,*skb;
buff5269net/ipv4/tcp.cbuff = sock_wmalloc(sk, win_size + th->doff * 4 + 
buff5273net/ipv4/tcp.cif ( buff == NULL )
buff5281net/ipv4/tcp.cbuff->free = /*0*/1;
buff5283net/ipv4/tcp.cbuff->sk = sk;
buff5284net/ipv4/tcp.cbuff->localroute = sk->localroute;
buff5290net/ipv4/tcp.ctmp = sk->prot->build_header(buff, sk->saddr, sk->daddr, &dev,
buff5291net/ipv4/tcp.cIPPROTO_TCP, sk->opt, buff->truesize,
buff5295net/ipv4/tcp.csock_wfree(sk, buff);
buff5303net/ipv4/tcp.cbuff->dev = dev;
buff5305net/ipv4/tcp.cnth = (struct tcphdr *) skb_put(buff,th->doff*4);
buff5328net/ipv4/tcp.cmemcpy(skb_put(buff,win_size), tcp_data_start, win_size);
buff5334net/ipv4/tcp.cbuff->end_seq = sk->sent_seq + win_size;
buff5335net/ipv4/tcp.csk->sent_seq = buff->end_seq;    /* Hack */
buff5378net/ipv4/tcp.cbuff = sock_wmalloc(sk,MAX_ACK_SIZE,1, GFP_ATOMIC);
buff5379net/ipv4/tcp.cif (buff == NULL) 
buff5382net/ipv4/tcp.cbuff->free = 1;
buff5383net/ipv4/tcp.cbuff->sk = sk;
buff5384net/ipv4/tcp.cbuff->localroute = sk->localroute;
buff5390net/ipv4/tcp.ctmp = sk->prot->build_header(buff, sk->saddr, sk->daddr, &dev,
buff5394net/ipv4/tcp.csock_wfree(sk, buff);
buff5398net/ipv4/tcp.ct1 = (struct tcphdr *)skb_put(buff,sizeof(struct tcphdr));
buff5426net/ipv4/tcp.csk->prot->queue_xmit(sk, dev, buff, 1);
buff51net/ipv4/utils.cstatic char buff[18];
buff55net/ipv4/utils.csprintf(buff, "%d.%d.%d.%d",
buff57net/ipv4/utils.creturn(buff);
buff79net/netrom/nr_dev.cunsigned char *buff = skb_push(skb, NR_NETWORK_LEN + NR_TRANSPORT_LEN);
buff81net/netrom/nr_dev.cmemcpy(buff, (saddr != NULL) ? saddr : dev->dev_addr, dev->addr_len);
buff82net/netrom/nr_dev.cbuff[6] &= ~LAPB_C;
buff83net/netrom/nr_dev.cbuff[6] &= ~LAPB_E;
buff84net/netrom/nr_dev.cbuff[6] |= SSSID_SPARE;
buff85net/netrom/nr_dev.cbuff    += AX25_ADDR_LEN;
buff88net/netrom/nr_dev.cmemcpy(buff, daddr, dev->addr_len);
buff89net/netrom/nr_dev.cbuff[6] &= ~LAPB_C;
buff90net/netrom/nr_dev.cbuff[6] |= LAPB_E;
buff91net/netrom/nr_dev.cbuff[6] |= SSSID_SPARE;
buff92net/netrom/nr_dev.cbuff    += AX25_ADDR_LEN;
buff94net/netrom/nr_dev.c*buff++ = nr_default.ttl;
buff96net/netrom/nr_dev.c*buff++ = NR_PROTO_IP;
buff97net/netrom/nr_dev.c*buff++ = NR_PROTO_IP;
buff98net/netrom/nr_dev.c*buff++ = 0;
buff99net/netrom/nr_dev.c*buff++ = 0;
buff100net/netrom/nr_dev.c*buff++ = NR_PROTOEXT;
buff108net/netrom/nr_dev.cstatic int nr_rebuild_header(void *buff, struct device *dev,
buff112net/netrom/nr_dev.cunsigned char *bp = (unsigned char *)buff;
buff864net/socket.casmlinkage int sys_send(int fd, void * buff, int len, unsigned flags)
buff879net/socket.cerr=verify_area(VERIFY_READ, buff, len);
buff883net/socket.ciov.iov_base=buff;
buff898net/socket.casmlinkage int sys_sendto(int fd, void * buff, int len, unsigned flags,
buff915net/socket.cerr=verify_area(VERIFY_READ,buff,len);
buff922net/socket.ciov.iov_base=buff;