taglinefilesource code
req1382drivers/block/ide-cd.cstruct request req;
req1392drivers/block/ide-cd.cide_init_drive_cmd (&req);
req1393drivers/block/ide-cd.creq.cmd = PACKET_COMMAND;
req1394drivers/block/ide-cd.creq.buffer = (char *)pc;
req1395drivers/block/ide-cd.c(void) ide_do_drive_cmd (drive, &req, ide_wait);
req2480drivers/block/ide-cd.cstruct request req;
req2481drivers/block/ide-cd.cide_init_drive_cmd (&req);
req2482drivers/block/ide-cd.creq.cmd = RESET_DRIVE_COMMAND;
req2483drivers/block/ide-cd.creturn ide_do_drive_cmd (drive, &req, ide_wait);
req108drivers/block/linear.cstatic int linear_map (int minor, struct md_dev *mddev, struct request *req)
req118drivers/block/linear.cwhile (req->nr_sectors)
req120drivers/block/linear.cblock=req->sector >> 1;
req137drivers/block/linear.cif (req->sem)        /* This is a paging request */
req139drivers/block/linear.creq->rq_dev=tmp_dev->dev;
req140drivers/block/linear.creq->sector=rblock << 1;
req141drivers/block/linear.cadd_request (blk_dev+MAJOR (tmp_dev->dev), req);
req148drivers/block/linear.cfor (nblk=0, bh=bh2=req->bh;
req159drivers/block/linear.cpending[queue].cmd=req->cmd;
req162drivers/block/linear.cpending[queue].current_nr_sectors=req->bh->b_size >> 9;
req163drivers/block/linear.cpending[queue].bh=req->bh;
req167drivers/block/linear.creq->bh=bh;
req168drivers/block/linear.creq->sector+=nblk << 1;
req169drivers/block/linear.creq->nr_sectors-=nblk << 1;
req172drivers/block/linear.creq->rq_status=RQ_INACTIVE;
req102drivers/block/ll_rw_blk.cstruct request * req;
req107drivers/block/ll_rw_blk.creq = dev->current_request;
req108drivers/block/ll_rw_blk.cif (req && req->rq_status == RQ_INACTIVE && req->cmd == -1) {
req109drivers/block/ll_rw_blk.cdev->current_request = req->next;
req123drivers/block/ll_rw_blk.cregister struct request *req, *limit;
req133drivers/block/ll_rw_blk.creq = prev_found;
req135drivers/block/ll_rw_blk.creq = ((req > all_requests) ? req : limit) - 1;
req136drivers/block/ll_rw_blk.cif (req->rq_status == RQ_INACTIVE)
req138drivers/block/ll_rw_blk.cif (req == prev_found)
req141drivers/block/ll_rw_blk.cprev_found = req;
req142drivers/block/ll_rw_blk.creq->rq_status = RQ_ACTIVE;
req143drivers/block/ll_rw_blk.creq->rq_dev = dev;
req144drivers/block/ll_rw_blk.creturn req;
req152drivers/block/ll_rw_blk.cregister struct request *req;
req160drivers/block/ll_rw_blk.creq = get_request(n, dev);
req162drivers/block/ll_rw_blk.cif (req)
req168drivers/block/ll_rw_blk.creturn req;
req173drivers/block/ll_rw_blk.cregister struct request *req;
req176drivers/block/ll_rw_blk.creq = get_request(n, dev);
req178drivers/block/ll_rw_blk.cif (req)
req179drivers/block/ll_rw_blk.creturn req;
req233drivers/block/ll_rw_blk.cvoid add_request(struct blk_dev_struct * dev, struct request * req)
req239drivers/block/ll_rw_blk.cswitch (MAJOR(req->rq_dev)) {
req241drivers/block/ll_rw_blk.cdisk_index = (MINOR(req->rq_dev) & 0x0070) >> 4;
req243drivers/block/ll_rw_blk.cdrive_stat_acct(req->cmd, req->nr_sectors, disk_index);
req247drivers/block/ll_rw_blk.cdisk_index = (MINOR(req->rq_dev) & 0x0040) >> 6;
req248drivers/block/ll_rw_blk.cdrive_stat_acct(req->cmd, req->nr_sectors, disk_index);
req251drivers/block/ll_rw_blk.cdisk_index = ((MINOR(req->rq_dev) & 0x0040) >> 6) + 2;
req252drivers/block/ll_rw_blk.cdrive_stat_acct(req->cmd, req->nr_sectors, disk_index);
req257drivers/block/ll_rw_blk.creq->next = NULL;
req259drivers/block/ll_rw_blk.cif (req->bh && req->bh->b_dev==req->bh->b_rdev)
req260drivers/block/ll_rw_blk.cmark_buffer_clean(req->bh);
req262drivers/block/ll_rw_blk.cdev->current_request = req;
req269drivers/block/ll_rw_blk.cif ((IN_ORDER(tmp,req) ||
req271drivers/block/ll_rw_blk.cIN_ORDER(req,tmp->next))
req274drivers/block/ll_rw_blk.creq->next = tmp->next;
req275drivers/block/ll_rw_blk.ctmp->next = req;
req279drivers/block/ll_rw_blk.cif (scsi_major(MAJOR(req->rq_dev)) && MAJOR(req->rq_dev)!=MD_MAJOR)
req288drivers/block/ll_rw_blk.cstruct request * req;
req358drivers/block/ll_rw_blk.c&& (req = blk_dev[major].current_request))
req363drivers/block/ll_rw_blk.creq = req->next;
req365drivers/block/ll_rw_blk.cwhile (req) {
req366drivers/block/ll_rw_blk.cif (req->rq_dev == bh->b_dev &&
req367drivers/block/ll_rw_blk.c!req->sem &&
req368drivers/block/ll_rw_blk.creq->cmd == rw &&
req369drivers/block/ll_rw_blk.creq->sector + req->nr_sectors == sector &&
req370drivers/block/ll_rw_blk.creq->nr_sectors < 244)
req372drivers/block/ll_rw_blk.creq->bhtail->b_reqnext = bh;
req373drivers/block/ll_rw_blk.creq->bhtail = bh;
req374drivers/block/ll_rw_blk.creq->nr_sectors += count;
req381drivers/block/ll_rw_blk.cif (req->rq_dev == bh->b_dev &&
req382drivers/block/ll_rw_blk.c!req->sem &&
req383drivers/block/ll_rw_blk.creq->cmd == rw &&
req384drivers/block/ll_rw_blk.creq->sector - count == sector &&
req385drivers/block/ll_rw_blk.creq->nr_sectors < 244)
req387drivers/block/ll_rw_blk.creq->nr_sectors += count;
req388drivers/block/ll_rw_blk.cbh->b_reqnext = req->bh;
req389drivers/block/ll_rw_blk.creq->buffer = bh->b_data;
req390drivers/block/ll_rw_blk.creq->current_nr_sectors = count;
req391drivers/block/ll_rw_blk.creq->sector = sector;
req393drivers/block/ll_rw_blk.creq->bh = bh;
req399drivers/block/ll_rw_blk.creq = req->next;
req406drivers/block/ll_rw_blk.creq = get_request(max_req, bh->b_dev);
req410drivers/block/ll_rw_blk.cif (!req) {
req415drivers/block/ll_rw_blk.creq = __get_request_wait(max_req, bh->b_dev);
req419drivers/block/ll_rw_blk.creq->cmd = rw;
req420drivers/block/ll_rw_blk.creq->errors = 0;
req421drivers/block/ll_rw_blk.creq->sector = sector;
req422drivers/block/ll_rw_blk.creq->nr_sectors = count;
req423drivers/block/ll_rw_blk.creq->current_nr_sectors = count;
req424drivers/block/ll_rw_blk.creq->buffer = bh->b_data;
req425drivers/block/ll_rw_blk.creq->sem = NULL;
req426drivers/block/ll_rw_blk.creq->bh = bh;
req427drivers/block/ll_rw_blk.creq->bhtail = bh;
req428drivers/block/ll_rw_blk.creq->next = NULL;
req429drivers/block/ll_rw_blk.cadd_request(major+blk_dev,req);
req563drivers/block/ll_rw_blk.cstruct request * req[8];
req591drivers/block/ll_rw_blk.creq[j] = get_request_wait(NR_REQUEST, dev);
req594drivers/block/ll_rw_blk.creq[j] = get_request(NR_REQUEST, dev);
req596drivers/block/ll_rw_blk.cif (req[j] == NULL)
req599drivers/block/ll_rw_blk.creq[j]->cmd = rw;
req600drivers/block/ll_rw_blk.creq[j]->errors = 0;
req601drivers/block/ll_rw_blk.creq[j]->sector = (b[i] * buffersize) >> 9;
req602drivers/block/ll_rw_blk.creq[j]->nr_sectors = buffersize >> 9;
req603drivers/block/ll_rw_blk.creq[j]->current_nr_sectors = buffersize >> 9;
req604drivers/block/ll_rw_blk.creq[j]->buffer = buf;
req605drivers/block/ll_rw_blk.creq[j]->sem = &sem;
req606drivers/block/ll_rw_blk.creq[j]->bh = NULL;
req607drivers/block/ll_rw_blk.creq[j]->next = NULL;
req608drivers/block/ll_rw_blk.cadd_request(major+blk_dev,req[j]);
req619drivers/block/ll_rw_blk.cstruct request * req;
req627drivers/block/ll_rw_blk.creq = all_requests + NR_REQUEST;
req628drivers/block/ll_rw_blk.cwhile (--req >= all_requests) {
req629drivers/block/ll_rw_blk.creq->rq_status = RQ_INACTIVE;
req630drivers/block/ll_rw_blk.creq->next = NULL;
req363drivers/block/md.cstatic inline int remap_request (int minor, struct request *req)
req371drivers/block/md.creturn (md_dev[minor].pers->map(minor, md_dev+minor, req));
req377drivers/block/md.cstruct request *req;
req386drivers/block/md.creq = blk_dev[MD_MAJOR].current_request;
req387drivers/block/md.cif (!req || (req->rq_status == RQ_INACTIVE))
req394drivers/block/md.creqsize=req->nr_sectors>>1;
req395drivers/block/md.cchunksize=1 << FACTOR_SHIFT(FACTOR(md_dev+MINOR(req->rq_dev)));
req396drivers/block/md.cif (reqsize==chunksize) (md_dev+MINOR(req->rq_dev))->equal_count++;
req397drivers/block/md.cif (reqsize<chunksize) (md_dev+MINOR(req->rq_dev))->smallest_count++;
req398drivers/block/md.cif (reqsize>chunksize) (md_dev+MINOR(req->rq_dev))->biggest_count++;
req401drivers/block/md.cblk_dev[MD_MAJOR].current_request = req->next;
req404drivers/block/md.cminor = MINOR(req->rq_dev);
req405drivers/block/md.cif ((MAJOR(req->rq_dev) != MD_MAJOR) || (minor >= MAX_REAL))
req407drivers/block/md.cprintk("md: bad device number: 0x%04x\n", req->rq_dev);
req408drivers/block/md.cend_request(0, req);
req412drivers/block/md.cswitch (remap_request (minor, req))
req415drivers/block/md.creq->rq_status=RQ_INACTIVE;
req423drivers/block/md.cend_request (0, req);
req439drivers/block/md.cstruct request *req;
req462drivers/block/md.c&& (req = blk_dev[major].current_request))
req469drivers/block/md.creq = req->next;
req471drivers/block/md.cwhile (req && !found)
req473drivers/block/md.cif (req->rq_status!=RQ_INACTIVE && req->rq_status!=RQ_ACTIVE)
req476drivers/block/md.cif (req->rq_dev == dev &&
req477drivers/block/md.c!req->sem &&
req478drivers/block/md.creq->cmd == rw &&
req479drivers/block/md.creq->sector + req->nr_sectors == pending[i].sector &&
req480drivers/block/md.c(req->nr_sectors + pending[i].nr_sectors) < 245)
req482drivers/block/md.creq->bhtail->b_reqnext = bh;
req483drivers/block/md.creq->bhtail = pending[i].bhtail;
req484drivers/block/md.creq->nr_sectors += pending[i].nr_sectors;
req490drivers/block/md.creq->rq_dev == dev &&
req491drivers/block/md.c!req->sem &&
req492drivers/block/md.creq->cmd == rw &&
req493drivers/block/md.creq->sector - pending[i].nr_sectors == pending[i].sector &&
req494drivers/block/md.c(req->nr_sectors + pending[i].nr_sectors) < 245)
req496drivers/block/md.creq->nr_sectors += pending[i].nr_sectors;
req497drivers/block/md.cbh->b_reqnext = req->bh;
req498drivers/block/md.creq->buffer = bh->b_data;
req499drivers/block/md.creq->current_nr_sectors = bh->b_size >> 9;
req500drivers/block/md.creq->sector = pending[i].sector;
req501drivers/block/md.creq->bh = bh;
req506drivers/block/md.creq = req->next;
req514drivers/block/md.creq=get_md_request (max_req, dev);
req517drivers/block/md.creq->cmd = rw;
req518drivers/block/md.creq->errors = 0;
req520drivers/block/md.creq->shared_count = 0;
req522drivers/block/md.creq->sector = pending[i].sector;
req523drivers/block/md.creq->nr_sectors = pending[i].nr_sectors;
req524drivers/block/md.creq->current_nr_sectors = bh->b_size >> 9;
req525drivers/block/md.creq->buffer = bh->b_data;
req526drivers/block/md.creq->sem = NULL;
req527drivers/block/md.creq->bh = bh;
req528drivers/block/md.creq->bhtail = pending[i].bhtail;
req529drivers/block/md.creq->next = NULL;
req531drivers/block/md.cadd_request (blk_dev + MAJOR(dev), req);
req182drivers/block/raid0.cstatic int raid0_map (int minor, struct md_dev *mddev, struct request *req)
req195drivers/block/raid0.cwhile (req->bh || req->sem)
req197drivers/block/raid0.cblock=req->sector >> 1;
req215drivers/block/raid0.cif (req->sem)    /* This is a paging request */
req217drivers/block/raid0.creq->rq_dev=tmp_dev->dev;
req218drivers/block/raid0.creq->sector=rblock << 1;
req219drivers/block/raid0.cadd_request (blk_dev+MAJOR (tmp_dev->dev), req);
req228drivers/block/raid0.ci<(1UL << FACTOR_SHIFT(factor)) && req->bh;
req231drivers/block/raid0.cbh=req->bh;
req246drivers/block/raid0.cpending[queue].cmd=req->cmd;
req257drivers/block/raid0.cend_redirect (req);  /* Separate bh from the request */
req261drivers/block/raid0.creq->rq_status=RQ_INACTIVE;
req2799drivers/char/serial.cint register_serial(struct serial_struct *req)
req2808drivers/char/serial.cif (rs_table[i].port == req->port)
req2825drivers/char/serial.c"device already open\n", i, req->port, req->irq);
req2828drivers/char/serial.cinfo->irq = req->irq;
req2829drivers/char/serial.cinfo->port = req->port;
req2830drivers/char/serial.cinfo->flags = req->flags;
req847drivers/scsi/scsi.cScsi_Cmnd * request_queueable (struct request * req, Scsi_Device * device)
req857drivers/scsi/scsi.cif (req && req->rq_status == RQ_INACTIVE)
req905drivers/scsi/scsi.cif (req) {
req906drivers/scsi/scsi.cmemcpy(&SCpnt->request, req, sizeof(struct request));
req908drivers/scsi/scsi.cbhp = bh = req->bh;
req914drivers/scsi/scsi.cwhile(req->nr_sectors && bh){
req917drivers/scsi/scsi.creq->nr_sectors -= bh->b_size >> 9;
req918drivers/scsi/scsi.creq->sector += bh->b_size >> 9;
req922drivers/scsi/scsi.cif(req->nr_sectors && bh && bh->b_reqnext){  /* Any leftovers? */
req924drivers/scsi/scsi.creq->bh = bh->b_reqnext; /* Divide request */
req926drivers/scsi/scsi.cbh = req->bh;
req929drivers/scsi/scsi.cSCpnt->request.nr_sectors -= req->nr_sectors;
req930drivers/scsi/scsi.creq->current_nr_sectors = bh->b_size >> 9;
req931drivers/scsi/scsi.creq->buffer = bh->b_data;
req934drivers/scsi/scsi.creq->rq_status = RQ_INACTIVE;
req973drivers/scsi/scsi.cstruct request * req = NULL;
req985drivers/scsi/scsi.cif (reqp) req = *reqp;
req988drivers/scsi/scsi.cif (req) {
req989drivers/scsi/scsi.cif(req->rq_status == RQ_INACTIVE) return NULL;
req990drivers/scsi/scsi.cdev = req->rq_dev;
req1039drivers/scsi/scsi.cif (req && (req->rq_status == RQ_INACTIVE || req->rq_dev != dev)) {
req1070drivers/scsi/scsi.cif (req) {
req1071drivers/scsi/scsi.cmemcpy(&SCpnt->request, req, sizeof(struct request));
req1073drivers/scsi/scsi.cbhp = bh = req->bh;
req1079drivers/scsi/scsi.cwhile(req->nr_sectors && bh){
req1082drivers/scsi/scsi.creq->nr_sectors -= bh->b_size >> 9;
req1083drivers/scsi/scsi.creq->sector += bh->b_size >> 9;
req1087drivers/scsi/scsi.cif(req->nr_sectors && bh && bh->b_reqnext){/* Any leftovers? */
req1089drivers/scsi/scsi.creq->bh = bh->b_reqnext; /* Divide request */
req1091drivers/scsi/scsi.cbh = req->bh;
req1093drivers/scsi/scsi.cSCpnt->request.nr_sectors -= req->nr_sectors;
req1094drivers/scsi/scsi.creq->current_nr_sectors = bh->b_size >> 9;
req1095drivers/scsi/scsi.creq->buffer = bh->b_data;
req1100drivers/scsi/scsi.creq->rq_status = RQ_INACTIVE;
req1101drivers/scsi/scsi.c*reqp = req->next;
req3134drivers/scsi/scsi.cstruct request * req;
req3136drivers/scsi/scsi.creq = blk_dev[i].current_request;
req3137drivers/scsi/scsi.cwhile(req) {
req3139drivers/scsi/scsi.ckdevname(req->rq_dev),
req3140drivers/scsi/scsi.creq->cmd,
req3141drivers/scsi/scsi.creq->sector,
req3142drivers/scsi/scsi.creq->nr_sectors,
req3143drivers/scsi/scsi.creq->current_nr_sectors);
req3144drivers/scsi/scsi.creq = req->next;
req460drivers/scsi/scsi.hstruct request * req;
req463drivers/scsi/scsi.hreq = &SCpnt->request;
req464drivers/scsi/scsi.hreq->errors = 0;
req467drivers/scsi/scsi.hkdevname(req->rq_dev), req->sector);
req471drivers/scsi/scsi.hif ((bh = req->bh) != NULL) {
req472drivers/scsi/scsi.hreq->bh = bh->b_reqnext;
req473drivers/scsi/scsi.hreq->nr_sectors -= bh->b_size >> 9;
req474drivers/scsi/scsi.hreq->sector += bh->b_size >> 9;
req479drivers/scsi/scsi.hif ((bh = req->bh) != NULL) {
req480drivers/scsi/scsi.hreq->current_nr_sectors = bh->b_size >> 9;
req481drivers/scsi/scsi.hif (req->nr_sectors < req->current_nr_sectors) {
req482drivers/scsi/scsi.hreq->nr_sectors = req->current_nr_sectors;
req488drivers/scsi/scsi.hif (req->bh){
req489drivers/scsi/scsi.hreq->buffer = bh->b_data;
req492drivers/scsi/scsi.hDEVICE_OFF(req->rq_dev);
req493drivers/scsi/scsi.hif (req->sem != NULL) {
req494drivers/scsi/scsi.hup(req->sem);
req505drivers/scsi/scsi.hreq->rq_status = RQ_INACTIVE;
req94drivers/scsi/scsi_ioctl.cstruct request * req;
req96drivers/scsi/scsi_ioctl.creq = &SCpnt->request;
req97drivers/scsi/scsi_ioctl.creq->rq_status = RQ_SCSI_DONE; /* Busy, but indicate request done */
req99drivers/scsi/scsi_ioctl.cif (req->sem != NULL) {
req100drivers/scsi/scsi_ioctl.cup(req->sem);
req410drivers/scsi/sd.cstruct request * req = NULL;
req483drivers/scsi/sd.creq = CURRENT;
req484drivers/scsi/sd.cwhile(req){
req485drivers/scsi/sd.cSCpnt = request_queueable(req, rscsi_disks[DEVICE_NR(req->rq_dev)].device);
req487drivers/scsi/sd.creq1 = req;
req488drivers/scsi/sd.creq = req->next;
req490drivers/scsi/sd.cif (SCpnt && req->rq_status == RQ_INACTIVE) {
req491drivers/scsi/sd.cif (req == CURRENT) 
req494drivers/scsi/sd.creq1->next = req->next;
req942drivers/scsi/sd.cstruct request * req;
req944drivers/scsi/sd.creq = &SCpnt->request;
req945drivers/scsi/sd.creq->rq_status = RQ_SCSI_DONE; /* Busy, but indicate request done */
req947drivers/scsi/sd.cif (req->sem != NULL) {
req948drivers/scsi/sd.cup(req->sem);
req558drivers/scsi/sr.cstruct request * req = NULL;
req613drivers/scsi/sr.creq = CURRENT;
req614drivers/scsi/sr.cwhile(req){
req615drivers/scsi/sr.cSCpnt = request_queueable(req,
req616drivers/scsi/sr.cscsi_CDs[DEVICE_NR(req->rq_dev)].device);
req618drivers/scsi/sr.creq1 = req;
req619drivers/scsi/sr.creq = req->next;
req621drivers/scsi/sr.cif (SCpnt && req->rq_status == RQ_INACTIVE) {
req622drivers/scsi/sr.cif (req == CURRENT) 
req625drivers/scsi/sr.creq1->next = req->next;
req965drivers/scsi/sr.cstruct request * req;
req967drivers/scsi/sr.creq = &SCpnt->request;
req968drivers/scsi/sr.creq->rq_status = RQ_SCSI_DONE; /* Busy, but indicate request done */
req970drivers/scsi/sr.cif (req->sem != NULL) {
req971drivers/scsi/sr.cup(req->sem);
req23drivers/scsi/sr_ioctl.cstruct request * req;
req25drivers/scsi/sr_ioctl.creq = &SCpnt->request;
req26drivers/scsi/sr_ioctl.creq->rq_status = RQ_SCSI_DONE; /* Busy, but indicate request done */
req28drivers/scsi/sr_ioctl.cif (req->sem != NULL) {
req29drivers/scsi/sr_ioctl.cup(req->sem);
req383include/linux/blk.hstruct request *req = hwgroup->rq;
req385include/linux/blk.hstatic void end_request (int uptodate, struct request * req) {
req388include/linux/blk.hstruct request *req = CURRENT;
req392include/linux/blk.hreq->errors = 0;
req395include/linux/blk.hkdevname(req->rq_dev), req->sector);
req396include/linux/blk.hreq->nr_sectors--;
req397include/linux/blk.hreq->nr_sectors &= ~SECTOR_MASK;
req398include/linux/blk.hreq->sector += (BLOCK_SIZE / 512);
req399include/linux/blk.hreq->sector &= ~SECTOR_MASK;    
req402include/linux/blk.hif ((bh = req->bh) != NULL) {
req403include/linux/blk.hreq->bh = bh->b_reqnext;
req407include/linux/blk.hif ((bh = req->bh) != NULL) {
req408include/linux/blk.hreq->current_nr_sectors = bh->b_size >> 9;
req409include/linux/blk.hif (req->nr_sectors < req->current_nr_sectors) {
req410include/linux/blk.hreq->nr_sectors = req->current_nr_sectors;
req413include/linux/blk.hreq->buffer = bh->b_data;
req418include/linux/blk.hadd_blkdev_randomness(MAJOR(req->rq_dev));
req421include/linux/blk.hblk_dev[MAJOR(req->rq_dev)].current_request = req->next;
req424include/linux/blk.hDEVICE_OFF(req->rq_dev);
req425include/linux/blk.hCURRENT = req->next;
req427include/linux/blk.hif (req->sem != NULL)
req428include/linux/blk.hup(req->sem);
req429include/linux/blk.hreq->rq_status = RQ_INACTIVE;
req436include/linux/blk.hextern inline void end_redirect (struct request *req)
req440include/linux/blk.hreq->errors = 0;
req442include/linux/blk.hif ((bh = req->bh) != NULL)
req444include/linux/blk.hreq->bh = bh->b_reqnext;
req447include/linux/blk.hif ((bh = req->bh) != NULL)
req449include/linux/blk.hreq->sector += req->current_nr_sectors;
req450include/linux/blk.hreq->current_nr_sectors = bh->b_size >> 9;
req452include/linux/blk.hif (req->nr_sectors < req->current_nr_sectors)
req454include/linux/blk.hreq->nr_sectors = req->current_nr_sectors;
req458include/linux/blk.hreq->buffer = bh->b_data;
req51include/linux/blkdev.hextern void add_request(struct blk_dev_struct * dev, struct request * req);
req110include/linux/md.hint (*map)(int minor, struct md_dev *md_dev, struct request *req);
req204include/linux/serial.hextern int register_serial(struct serial_struct *req);
req284net/ipv4/rarp.cstatic int rarp_req_set(struct arpreq *req)
req294net/ipv4/rarp.cmemcpy_fromfs(&r, req, sizeof(r));
req388net/ipv4/rarp.cstatic int rarp_req_get(struct arpreq *req)
req399net/ipv4/rarp.cmemcpy_fromfs(&r, req, sizeof(r));
req434net/ipv4/rarp.cmemcpy_tofs(req, &r, sizeof(r));