taglinefilesource code
req1382drivers/block/ide-cd.cstruct request req;
req1392drivers/block/ide-cd.cide_init_drive_cmd (&req);
req1393drivers/block/ide-cd.creq.cmd = PACKET_COMMAND;
req1394drivers/block/ide-cd.creq.buffer = (char *)pc;
req1395drivers/block/ide-cd.c(void) ide_do_drive_cmd (drive, &req, ide_wait);
req2480drivers/block/ide-cd.cstruct request req;
req2481drivers/block/ide-cd.cide_init_drive_cmd (&req);
req2482drivers/block/ide-cd.creq.cmd = RESET_DRIVE_COMMAND;
req2483drivers/block/ide-cd.creturn ide_do_drive_cmd (drive, &req, ide_wait);
req108drivers/block/linear.cstatic int linear_map (int minor, struct md_dev *mddev, struct request *req)
req118drivers/block/linear.cwhile (req->nr_sectors)
req120drivers/block/linear.cblock=req->sector >> 1;
req137drivers/block/linear.cif (req->sem)        /* This is a paging request */
req139drivers/block/linear.creq->rq_dev=tmp_dev->dev;
req140drivers/block/linear.creq->sector=rblock << 1;
req141drivers/block/linear.cadd_request (blk_dev+MAJOR (tmp_dev->dev), req);
req148drivers/block/linear.cfor (nblk=0, bh=bh2=req->bh;
req159drivers/block/linear.cpending[queue].cmd=req->cmd;
req162drivers/block/linear.cpending[queue].current_nr_sectors=req->bh->b_size >> 9;
req163drivers/block/linear.cpending[queue].bh=req->bh;
req167drivers/block/linear.creq->bh=bh;
req168drivers/block/linear.creq->sector+=nblk << 1;
req169drivers/block/linear.creq->nr_sectors-=nblk << 1;
req172drivers/block/linear.creq->rq_status=RQ_INACTIVE;
req115drivers/block/ll_rw_blk.cregister struct request *req, *limit;
req125drivers/block/ll_rw_blk.creq = prev_found;
req127drivers/block/ll_rw_blk.creq = ((req > all_requests) ? req : limit) - 1;
req128drivers/block/ll_rw_blk.cif (req->rq_status == RQ_INACTIVE)
req130drivers/block/ll_rw_blk.cif (req == prev_found)
req133drivers/block/ll_rw_blk.cprev_found = req;
req134drivers/block/ll_rw_blk.creq->rq_status = RQ_ACTIVE;
req135drivers/block/ll_rw_blk.creq->rq_dev = dev;
req136drivers/block/ll_rw_blk.creturn req;
req144drivers/block/ll_rw_blk.cregister struct request *req;
req151drivers/block/ll_rw_blk.creq = get_request(n, dev);
req153drivers/block/ll_rw_blk.cif (req)
req159drivers/block/ll_rw_blk.creturn req;
req164drivers/block/ll_rw_blk.cregister struct request *req;
req167drivers/block/ll_rw_blk.creq = get_request(n, dev);
req169drivers/block/ll_rw_blk.cif (req)
req170drivers/block/ll_rw_blk.creturn req;
req224drivers/block/ll_rw_blk.cvoid add_request(struct blk_dev_struct * dev, struct request * req)
req230drivers/block/ll_rw_blk.cswitch (MAJOR(req->rq_dev)) {
req232drivers/block/ll_rw_blk.cdisk_index = (MINOR(req->rq_dev) & 0x0070) >> 4;
req234drivers/block/ll_rw_blk.cdrive_stat_acct(req->cmd, req->nr_sectors, disk_index);
req238drivers/block/ll_rw_blk.cdisk_index = (MINOR(req->rq_dev) & 0x0040) >> 6;
req239drivers/block/ll_rw_blk.cdrive_stat_acct(req->cmd, req->nr_sectors, disk_index);
req242drivers/block/ll_rw_blk.cdisk_index = ((MINOR(req->rq_dev) & 0x0040) >> 6) + 2;
req243drivers/block/ll_rw_blk.cdrive_stat_acct(req->cmd, req->nr_sectors, disk_index);
req248drivers/block/ll_rw_blk.creq->next = NULL;
req250drivers/block/ll_rw_blk.cif (req->bh && req->bh->b_dev==req->bh->b_rdev)
req251drivers/block/ll_rw_blk.cmark_buffer_clean(req->bh);
req253drivers/block/ll_rw_blk.cdev->current_request = req;
req260drivers/block/ll_rw_blk.cif ((IN_ORDER(tmp,req) ||
req262drivers/block/ll_rw_blk.cIN_ORDER(req,tmp->next))
req265drivers/block/ll_rw_blk.creq->next = tmp->next;
req266drivers/block/ll_rw_blk.ctmp->next = req;
req270drivers/block/ll_rw_blk.cif (scsi_major(MAJOR(req->rq_dev)) && MAJOR(req->rq_dev)!=MD_MAJOR)
req279drivers/block/ll_rw_blk.cstruct request * req;
req340drivers/block/ll_rw_blk.creq = blk_dev[major].current_request;
req341drivers/block/ll_rw_blk.cif (!req) {
req358drivers/block/ll_rw_blk.creq = req->next;
req359drivers/block/ll_rw_blk.cif (!req)
req368drivers/block/ll_rw_blk.cif (req->sem)
req370drivers/block/ll_rw_blk.cif (req->cmd != rw)
req372drivers/block/ll_rw_blk.cif (req->nr_sectors >= 244)
req374drivers/block/ll_rw_blk.cif (req->rq_dev != bh->b_dev)
req377drivers/block/ll_rw_blk.cif (req->sector + req->nr_sectors == sector) {
req378drivers/block/ll_rw_blk.creq->bhtail->b_reqnext = bh;
req379drivers/block/ll_rw_blk.creq->bhtail = bh;
req381drivers/block/ll_rw_blk.c} else if (req->sector - count == sector) {
req382drivers/block/ll_rw_blk.cbh->b_reqnext = req->bh;
req383drivers/block/ll_rw_blk.creq->bh = bh;
req384drivers/block/ll_rw_blk.creq->buffer = bh->b_data;
req385drivers/block/ll_rw_blk.creq->current_nr_sectors = count;
req386drivers/block/ll_rw_blk.creq->sector = sector;
req390drivers/block/ll_rw_blk.creq->nr_sectors += count;
req395drivers/block/ll_rw_blk.c} while ((req = req->next) != NULL);
req401drivers/block/ll_rw_blk.creq = get_request(max_req, bh->b_dev);
req405drivers/block/ll_rw_blk.cif (!req) {
req410drivers/block/ll_rw_blk.creq = __get_request_wait(max_req, bh->b_dev);
req414drivers/block/ll_rw_blk.creq->cmd = rw;
req415drivers/block/ll_rw_blk.creq->errors = 0;
req416drivers/block/ll_rw_blk.creq->sector = sector;
req417drivers/block/ll_rw_blk.creq->nr_sectors = count;
req418drivers/block/ll_rw_blk.creq->current_nr_sectors = count;
req419drivers/block/ll_rw_blk.creq->buffer = bh->b_data;
req420drivers/block/ll_rw_blk.creq->sem = NULL;
req421drivers/block/ll_rw_blk.creq->bh = bh;
req422drivers/block/ll_rw_blk.creq->bhtail = bh;
req423drivers/block/ll_rw_blk.creq->next = NULL;
req424drivers/block/ll_rw_blk.cadd_request(major+blk_dev,req);
req549drivers/block/ll_rw_blk.cstruct request * req[8];
req577drivers/block/ll_rw_blk.creq[j] = get_request_wait(NR_REQUEST, dev);
req580drivers/block/ll_rw_blk.creq[j] = get_request(NR_REQUEST, dev);
req582drivers/block/ll_rw_blk.cif (req[j] == NULL)
req585drivers/block/ll_rw_blk.creq[j]->cmd = rw;
req586drivers/block/ll_rw_blk.creq[j]->errors = 0;
req587drivers/block/ll_rw_blk.creq[j]->sector = (b[i] * buffersize) >> 9;
req588drivers/block/ll_rw_blk.creq[j]->nr_sectors = buffersize >> 9;
req589drivers/block/ll_rw_blk.creq[j]->current_nr_sectors = buffersize >> 9;
req590drivers/block/ll_rw_blk.creq[j]->buffer = buf;
req591drivers/block/ll_rw_blk.creq[j]->sem = &sem;
req592drivers/block/ll_rw_blk.creq[j]->bh = NULL;
req593drivers/block/ll_rw_blk.creq[j]->next = NULL;
req594drivers/block/ll_rw_blk.cadd_request(major+blk_dev,req[j]);
req605drivers/block/ll_rw_blk.cstruct request * req;
req618drivers/block/ll_rw_blk.creq = all_requests + NR_REQUEST;
req619drivers/block/ll_rw_blk.cwhile (--req >= all_requests) {
req620drivers/block/ll_rw_blk.creq->rq_status = RQ_INACTIVE;
req621drivers/block/ll_rw_blk.creq->next = NULL;
req378drivers/block/md.cstatic inline int remap_request (int minor, struct request *req)
req386drivers/block/md.creturn (md_dev[minor].pers->map(minor, md_dev+minor, req));
req392drivers/block/md.cstruct request *req;
req401drivers/block/md.creq = blk_dev[MD_MAJOR].current_request;
req402drivers/block/md.cif (!req || (req->rq_status == RQ_INACTIVE))
req409drivers/block/md.creqsize=req->nr_sectors>>1;
req410drivers/block/md.cchunksize=1 << FACTOR_SHIFT(FACTOR(md_dev+MINOR(req->rq_dev)));
req411drivers/block/md.cif (reqsize==chunksize) (md_dev+MINOR(req->rq_dev))->equal_count++;
req412drivers/block/md.cif (reqsize<chunksize) (md_dev+MINOR(req->rq_dev))->smallest_count++;
req413drivers/block/md.cif (reqsize>chunksize) (md_dev+MINOR(req->rq_dev))->biggest_count++;
req416drivers/block/md.cblk_dev[MD_MAJOR].current_request = req->next;
req419drivers/block/md.cminor = MINOR(req->rq_dev);
req420drivers/block/md.cif ((MAJOR(req->rq_dev) != MD_MAJOR) || (minor >= MAX_REAL))
req422drivers/block/md.cprintk("md: bad device: %s\n", kdevname(req->rq_dev));
req423drivers/block/md.cend_request(0, req);
req427drivers/block/md.cswitch (remap_request (minor, req))
req430drivers/block/md.creq->rq_status=RQ_INACTIVE;
req438drivers/block/md.cend_request (0, req);
req454drivers/block/md.cstruct request *req;
req478drivers/block/md.c&& (req = blk_dev[major].current_request))
req485drivers/block/md.creq = req->next;
req487drivers/block/md.cwhile (req && !found)
req489drivers/block/md.cif (req->rq_status!=RQ_ACTIVE)
req492drivers/block/md.cif (req->rq_dev == dev &&
req493drivers/block/md.c!req->sem &&
req494drivers/block/md.creq->cmd == rw &&
req495drivers/block/md.creq->sector + req->nr_sectors == pending[i].sector &&
req496drivers/block/md.c(req->nr_sectors + pending[i].nr_sectors) < 245)
req498drivers/block/md.creq->bhtail->b_reqnext = bh;
req499drivers/block/md.creq->bhtail = pending[i].bhtail;
req500drivers/block/md.creq->nr_sectors += pending[i].nr_sectors;
req506drivers/block/md.creq->rq_dev == dev &&
req507drivers/block/md.c!req->sem &&
req508drivers/block/md.creq->cmd == rw &&
req509drivers/block/md.creq->sector - pending[i].nr_sectors == pending[i].sector &&
req510drivers/block/md.c(req->nr_sectors + pending[i].nr_sectors) < 245)
req512drivers/block/md.creq->nr_sectors += pending[i].nr_sectors;
req513drivers/block/md.cpending[i].bhtail->b_reqnext = req->bh;
req514drivers/block/md.creq->buffer = bh->b_data;
req515drivers/block/md.creq->current_nr_sectors = bh->b_size >> 9;
req516drivers/block/md.creq->sector = pending[i].sector;
req517drivers/block/md.creq->bh = bh;
req522drivers/block/md.creq = req->next;
req531drivers/block/md.creq=get_md_request (max_req, dev);
req534drivers/block/md.creq->cmd = rw;
req535drivers/block/md.creq->errors = 0;
req537drivers/block/md.creq->shared_count = 0;
req539drivers/block/md.creq->sector = pending[i].sector;
req540drivers/block/md.creq->nr_sectors = pending[i].nr_sectors;
req541drivers/block/md.creq->current_nr_sectors = bh->b_size >> 9;
req542drivers/block/md.creq->buffer = bh->b_data;
req543drivers/block/md.creq->sem = NULL;
req544drivers/block/md.creq->bh = bh;
req545drivers/block/md.creq->bhtail = pending[i].bhtail;
req546drivers/block/md.creq->next = NULL;
req548drivers/block/md.cadd_request (blk_dev + MAJOR(dev), req);
req182drivers/block/raid0.cstatic int raid0_map (int minor, struct md_dev *mddev, struct request *req)
req195drivers/block/raid0.cwhile (req->bh || req->sem)
req197drivers/block/raid0.cblock=req->sector >> 1;
req215drivers/block/raid0.cif (req->sem)    /* This is a paging request */
req217drivers/block/raid0.creq->rq_dev=tmp_dev->dev;
req218drivers/block/raid0.creq->sector=rblock << 1;
req219drivers/block/raid0.cadd_request (blk_dev+MAJOR (tmp_dev->dev), req);
req228drivers/block/raid0.ci<(1UL << FACTOR_SHIFT(factor)) && req->bh;
req231drivers/block/raid0.cbh=req->bh;
req246drivers/block/raid0.cpending[queue].cmd=req->cmd;
req257drivers/block/raid0.cend_redirect (req);  /* Separate bh from the request */
req261drivers/block/raid0.creq->rq_status=RQ_INACTIVE;
req2799drivers/char/serial.cint register_serial(struct serial_struct *req)
req2808drivers/char/serial.cif (rs_table[i].port == req->port)
req2825drivers/char/serial.c"device already open\n", i, req->port, req->irq);
req2828drivers/char/serial.cinfo->irq = req->irq;
req2829drivers/char/serial.cinfo->port = req->port;
req2830drivers/char/serial.cinfo->flags = req->flags;
req847drivers/scsi/scsi.cScsi_Cmnd * request_queueable (struct request * req, Scsi_Device * device)
req857drivers/scsi/scsi.cif (req && req->rq_status == RQ_INACTIVE)
req905drivers/scsi/scsi.cif (req) {
req906drivers/scsi/scsi.cmemcpy(&SCpnt->request, req, sizeof(struct request));
req908drivers/scsi/scsi.cbhp = bh = req->bh;
req914drivers/scsi/scsi.cwhile(req->nr_sectors && bh){
req917drivers/scsi/scsi.creq->nr_sectors -= bh->b_size >> 9;
req918drivers/scsi/scsi.creq->sector += bh->b_size >> 9;
req922drivers/scsi/scsi.cif(req->nr_sectors && bh && bh->b_reqnext){  /* Any leftovers? */
req924drivers/scsi/scsi.creq->bh = bh->b_reqnext; /* Divide request */
req926drivers/scsi/scsi.cbh = req->bh;
req929drivers/scsi/scsi.cSCpnt->request.nr_sectors -= req->nr_sectors;
req930drivers/scsi/scsi.creq->current_nr_sectors = bh->b_size >> 9;
req931drivers/scsi/scsi.creq->buffer = bh->b_data;
req934drivers/scsi/scsi.creq->rq_status = RQ_INACTIVE;
req973drivers/scsi/scsi.cstruct request * req = NULL;
req985drivers/scsi/scsi.cif (reqp) req = *reqp;
req988drivers/scsi/scsi.cif (req) {
req989drivers/scsi/scsi.cif(req->rq_status == RQ_INACTIVE) return NULL;
req990drivers/scsi/scsi.cdev = req->rq_dev;
req1039drivers/scsi/scsi.cif (req && (req->rq_status == RQ_INACTIVE || req->rq_dev != dev)) {
req1070drivers/scsi/scsi.cif (req) {
req1071drivers/scsi/scsi.cmemcpy(&SCpnt->request, req, sizeof(struct request));
req1073drivers/scsi/scsi.cbhp = bh = req->bh;
req1079drivers/scsi/scsi.cwhile(req->nr_sectors && bh){
req1082drivers/scsi/scsi.creq->nr_sectors -= bh->b_size >> 9;
req1083drivers/scsi/scsi.creq->sector += bh->b_size >> 9;
req1087drivers/scsi/scsi.cif(req->nr_sectors && bh && bh->b_reqnext){/* Any leftovers? */
req1089drivers/scsi/scsi.creq->bh = bh->b_reqnext; /* Divide request */
req1091drivers/scsi/scsi.cbh = req->bh;
req1093drivers/scsi/scsi.cSCpnt->request.nr_sectors -= req->nr_sectors;
req1094drivers/scsi/scsi.creq->current_nr_sectors = bh->b_size >> 9;
req1095drivers/scsi/scsi.creq->buffer = bh->b_data;
req1100drivers/scsi/scsi.creq->rq_status = RQ_INACTIVE;
req1101drivers/scsi/scsi.c*reqp = req->next;
req3135drivers/scsi/scsi.cstruct request * req;
req3137drivers/scsi/scsi.creq = blk_dev[i].current_request;
req3138drivers/scsi/scsi.cwhile(req) {
req3140drivers/scsi/scsi.ckdevname(req->rq_dev),
req3141drivers/scsi/scsi.creq->cmd,
req3142drivers/scsi/scsi.creq->sector,
req3143drivers/scsi/scsi.creq->nr_sectors,
req3144drivers/scsi/scsi.creq->current_nr_sectors);
req3145drivers/scsi/scsi.creq = req->next;
req460drivers/scsi/scsi.hstruct request * req;
req463drivers/scsi/scsi.hreq = &SCpnt->request;
req464drivers/scsi/scsi.hreq->errors = 0;
req467drivers/scsi/scsi.hkdevname(req->rq_dev), req->sector);
req471drivers/scsi/scsi.hif ((bh = req->bh) != NULL) {
req472drivers/scsi/scsi.hreq->bh = bh->b_reqnext;
req473drivers/scsi/scsi.hreq->nr_sectors -= bh->b_size >> 9;
req474drivers/scsi/scsi.hreq->sector += bh->b_size >> 9;
req479drivers/scsi/scsi.hif ((bh = req->bh) != NULL) {
req480drivers/scsi/scsi.hreq->current_nr_sectors = bh->b_size >> 9;
req481drivers/scsi/scsi.hif (req->nr_sectors < req->current_nr_sectors) {
req482drivers/scsi/scsi.hreq->nr_sectors = req->current_nr_sectors;
req488drivers/scsi/scsi.hif (req->bh){
req489drivers/scsi/scsi.hreq->buffer = bh->b_data;
req492drivers/scsi/scsi.hDEVICE_OFF(req->rq_dev);
req493drivers/scsi/scsi.hif (req->sem != NULL) {
req494drivers/scsi/scsi.hup(req->sem);
req505drivers/scsi/scsi.hreq->rq_status = RQ_INACTIVE;
req94drivers/scsi/scsi_ioctl.cstruct request * req;
req96drivers/scsi/scsi_ioctl.creq = &SCpnt->request;
req97drivers/scsi/scsi_ioctl.creq->rq_status = RQ_SCSI_DONE; /* Busy, but indicate request done */
req99drivers/scsi/scsi_ioctl.cif (req->sem != NULL) {
req100drivers/scsi/scsi_ioctl.cup(req->sem);
req410drivers/scsi/sd.cstruct request * req = NULL;
req483drivers/scsi/sd.creq = CURRENT;
req484drivers/scsi/sd.cwhile(req){
req485drivers/scsi/sd.cSCpnt = request_queueable(req, rscsi_disks[DEVICE_NR(req->rq_dev)].device);
req487drivers/scsi/sd.creq1 = req;
req488drivers/scsi/sd.creq = req->next;
req490drivers/scsi/sd.cif (SCpnt && req->rq_status == RQ_INACTIVE) {
req491drivers/scsi/sd.cif (req == CURRENT) 
req494drivers/scsi/sd.creq1->next = req->next;
req942drivers/scsi/sd.cstruct request * req;
req944drivers/scsi/sd.creq = &SCpnt->request;
req945drivers/scsi/sd.creq->rq_status = RQ_SCSI_DONE; /* Busy, but indicate request done */
req947drivers/scsi/sd.cif (req->sem != NULL) {
req948drivers/scsi/sd.cup(req->sem);
req558drivers/scsi/sr.cstruct request * req = NULL;
req613drivers/scsi/sr.creq = CURRENT;
req614drivers/scsi/sr.cwhile(req){
req615drivers/scsi/sr.cSCpnt = request_queueable(req,
req616drivers/scsi/sr.cscsi_CDs[DEVICE_NR(req->rq_dev)].device);
req618drivers/scsi/sr.creq1 = req;
req619drivers/scsi/sr.creq = req->next;
req621drivers/scsi/sr.cif (SCpnt && req->rq_status == RQ_INACTIVE) {
req622drivers/scsi/sr.cif (req == CURRENT) 
req625drivers/scsi/sr.creq1->next = req->next;
req965drivers/scsi/sr.cstruct request * req;
req967drivers/scsi/sr.creq = &SCpnt->request;
req968drivers/scsi/sr.creq->rq_status = RQ_SCSI_DONE; /* Busy, but indicate request done */
req970drivers/scsi/sr.cif (req->sem != NULL) {
req971drivers/scsi/sr.cup(req->sem);
req23drivers/scsi/sr_ioctl.cstruct request * req;
req25drivers/scsi/sr_ioctl.creq = &SCpnt->request;
req26drivers/scsi/sr_ioctl.creq->rq_status = RQ_SCSI_DONE; /* Busy, but indicate request done */
req28drivers/scsi/sr_ioctl.cif (req->sem != NULL) {
req29drivers/scsi/sr_ioctl.cup(req->sem);
req383include/linux/blk.hstruct request *req = hwgroup->rq;
req385include/linux/blk.hstatic void end_request (int uptodate, struct request * req) {
req388include/linux/blk.hstruct request *req = CURRENT;
req392include/linux/blk.hreq->errors = 0;
req395include/linux/blk.hkdevname(req->rq_dev), req->sector);
req396include/linux/blk.hreq->nr_sectors--;
req397include/linux/blk.hreq->nr_sectors &= ~SECTOR_MASK;
req398include/linux/blk.hreq->sector += (BLOCK_SIZE / 512);
req399include/linux/blk.hreq->sector &= ~SECTOR_MASK;    
req402include/linux/blk.hif ((bh = req->bh) != NULL) {
req403include/linux/blk.hreq->bh = bh->b_reqnext;
req407include/linux/blk.hif ((bh = req->bh) != NULL) {
req408include/linux/blk.hreq->current_nr_sectors = bh->b_size >> 9;
req409include/linux/blk.hif (req->nr_sectors < req->current_nr_sectors) {
req410include/linux/blk.hreq->nr_sectors = req->current_nr_sectors;
req413include/linux/blk.hreq->buffer = bh->b_data;
req418include/linux/blk.hadd_blkdev_randomness(MAJOR(req->rq_dev));
req421include/linux/blk.hblk_dev[MAJOR(req->rq_dev)].current_request = req->next;
req424include/linux/blk.hDEVICE_OFF(req->rq_dev);
req425include/linux/blk.hCURRENT = req->next;
req427include/linux/blk.hif (req->sem != NULL)
req428include/linux/blk.hup(req->sem);
req429include/linux/blk.hreq->rq_status = RQ_INACTIVE;
req436include/linux/blk.hextern inline void end_redirect (struct request *req)
req440include/linux/blk.hreq->errors = 0;
req442include/linux/blk.hif ((bh = req->bh) != NULL)
req444include/linux/blk.hreq->bh = bh->b_reqnext;
req447include/linux/blk.hif ((bh = req->bh) != NULL)
req449include/linux/blk.hreq->sector += req->current_nr_sectors;
req450include/linux/blk.hreq->current_nr_sectors = bh->b_size >> 9;
req452include/linux/blk.hif (req->nr_sectors < req->current_nr_sectors)
req454include/linux/blk.hreq->nr_sectors = req->current_nr_sectors;
req458include/linux/blk.hreq->buffer = bh->b_data;
req54include/linux/blkdev.hextern void add_request(struct blk_dev_struct * dev, struct request * req);
req115include/linux/if_arp.hunsigned short  req;      /* request type */
req110include/linux/md.hint (*map)(int minor, struct md_dev *md_dev, struct request *req);
req204include/linux/serial.hextern int register_serial(struct serial_struct *req);
req447net/ipv4/arp.carpreq.req = ARPD_UPDATE;
req503net/ipv4/arp.carpreq.req = ARPD_LOOKUP;
req284net/ipv4/rarp.cstatic int rarp_req_set(struct arpreq *req)
req294net/ipv4/rarp.cmemcpy_fromfs(&r, req, sizeof(r));
req388net/ipv4/rarp.cstatic int rarp_req_get(struct arpreq *req)
req399net/ipv4/rarp.cmemcpy_fromfs(&r, req, sizeof(r));
req434net/ipv4/rarp.cmemcpy_tofs(req, &r, sizeof(r));