1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432143314341435143614371438143914401441144214431444144514461447144814491450145114521453145414551456145714581459146014611462146314641465146614671468146914701471147214731474147514761477147814791480148114821483148414851486148714881489149014911492149314941495149614971498149915001501150215031504150515061507150815091510151115121513151415151516151715181519152015211522152315241525152615271528152915301531153215331534153515361537153815391540154115421543154415451546154715481549155015511552155315541555155615571558155915601561156215631564156515661567156815691570157115721573157415751576157715781579158015811582158315841585158615871588158915901591159215931594159515961597159815991600160116021603160416051606160716081609161016111612161316141615161616171618161916201621162216231624162516261627162816291630163116321633163416351636163716381639164016411642164316441645164616471648164916501651165216531654165516561657165816591660166116621663166416651666166716681669167016711672167316741675167616771678167916801681168216831684168516861687168816891690169116921693169416951696169716981699170017011702170317041705170617071708170917101711171217131714171517161717171817191720172117221723172417251726172717281729173017311732173317341735173617371738173917401741174217431744174517461747174817491750175117521753175417551756175717581759176017611762176317641765176617671768176917701771177217731774177517761777177817791780178117821783178417851786178717881789179017911792179317941795179617971798179918001801180218031804180518061807180818091810181118121813181418151816181718181819182018211822182318241825182618271828182918301831183218331834183518361837183818391840184118421843184418451846184718481849185018511852185318541855185618571858185918601861186218631864186518661867186818691870187118721873187418751876187718781879188018811882188318841885188618871888188918901891189218931894189518961897189818991900190119021903190419051906190719081909191019111912191319141915191619171918191919201921192219231924192519261927192819291930193119321933193419351936193719381939194019411942194319441945194619471948194919501951195219531954195519561957195819591960196119621963196419651966196719681969197019711972197319741975197619771978197919801981198219831984198519861987198819891990199119921993199419951996199719981999200020012002200320042005200620072008200920102011201220132014201520162017201820192020202120222023202420252026202720282029203020312032203320342035203620372038203920402041204220432044204520462047204820492050205120522053205420552056205720582059206020612062206320642065206620672068206920702071207220732074207520762077207820792080208120822083208420852086208720882089209020912092209320942095209620972098209921002101210221032104210521062107210821092110211121122113211421152116211721182119212021212122212321242125212621272128212921302131213221332134213521362137213821392140214121422143214421452146214721482149215021512152215321542155215621572158215921602161216221632164216521662167216821692170217121722173217421752176217721782179218021812182218321842185218621872188218921902191219221932194219521962197219821992200220122022203220422052206220722082209221022112212221322142215221622172218221922202221222222232224222522262227222822292230223122322233223422352236223722382239224022412242224322442245224622472248224922502251225222532254225522562257225822592260226122622263226422652266226722682269227022712272227322742275227622772278227922802281228222832284228522862287228822892290229122922293229422952296229722982299230023012302230323042305230623072308230923102311231223132314231523162317231823192320232123222323232423252326232723282329233023312332233323342335233623372338233923402341234223432344234523462347234823492350235123522353235423552356235723582359236023612362236323642365236623672368236923702371237223732374237523762377237823792380238123822383238423852386238723882389239023912392239323942395239623972398239924002401240224032404240524062407240824092410241124122413241424152416241724182419242024212422242324242425242624272428242924302431243224332434243524362437243824392440244124422443244424452446244724482449245024512452245324542455245624572458245924602461246224632464246524662467246824692470247124722473247424752476247724782479248024812482248324842485248624872488248924902491249224932494249524962497249824992500250125022503250425052506250725082509251025112512251325142515251625172518251925202521252225232524252525262527 |
- /*
- * This file is part of the Chelsio FCoE driver for Linux.
- *
- * Copyright (c) 2008-2012 Chelsio Communications, Inc. All rights reserved.
- *
- * This software is available to you under a choice of one of two
- * licenses. You may choose to be licensed under the terms of the GNU
- * General Public License (GPL) Version 2, available from the file
- * COPYING in the main directory of this source tree, or the
- * OpenIB.org BSD license below:
- *
- * Redistribution and use in source and binary forms, with or
- * without modification, are permitted provided that the following
- * conditions are met:
- *
- * - Redistributions of source code must retain the above
- * copyright notice, this list of conditions and the following
- * disclaimer.
- *
- * - Redistributions in binary form must reproduce the above
- * copyright notice, this list of conditions and the following
- * disclaimer in the documentation and/or other materials
- * provided with the distribution.
- *
- * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
- * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
- * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
- * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
- * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
- * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
- * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
- * SOFTWARE.
- */
- #include <linux/device.h>
- #include <linux/delay.h>
- #include <linux/ctype.h>
- #include <linux/kernel.h>
- #include <linux/slab.h>
- #include <linux/string.h>
- #include <linux/compiler.h>
- #include <linux/export.h>
- #include <linux/module.h>
- #include <asm/unaligned.h>
- #include <asm/page.h>
- #include <scsi/scsi.h>
- #include <scsi/scsi_device.h>
- #include <scsi/scsi_transport_fc.h>
- #include "csio_hw.h"
- #include "csio_lnode.h"
- #include "csio_rnode.h"
- #include "csio_scsi.h"
- #include "csio_init.h"
- int csio_scsi_eqsize = 65536;
- int csio_scsi_iqlen = 128;
- int csio_scsi_ioreqs = 2048;
- uint32_t csio_max_scan_tmo;
- uint32_t csio_delta_scan_tmo = 5;
- int csio_lun_qdepth = 32;
- static int csio_ddp_descs = 128;
- static int csio_do_abrt_cls(struct csio_hw *,
- struct csio_ioreq *, bool);
- static void csio_scsis_uninit(struct csio_ioreq *, enum csio_scsi_ev);
- static void csio_scsis_io_active(struct csio_ioreq *, enum csio_scsi_ev);
- static void csio_scsis_tm_active(struct csio_ioreq *, enum csio_scsi_ev);
- static void csio_scsis_aborting(struct csio_ioreq *, enum csio_scsi_ev);
- static void csio_scsis_closing(struct csio_ioreq *, enum csio_scsi_ev);
- static void csio_scsis_shost_cmpl_await(struct csio_ioreq *, enum csio_scsi_ev);
- /*
- * csio_scsi_match_io - Match an ioreq with the given SCSI level data.
- * @ioreq: The I/O request
- * @sld: Level information
- *
- * Should be called with lock held.
- *
- */
- static bool
- csio_scsi_match_io(struct csio_ioreq *ioreq, struct csio_scsi_level_data *sld)
- {
- struct scsi_cmnd *scmnd = csio_scsi_cmnd(ioreq);
- switch (sld->level) {
- case CSIO_LEV_LUN:
- if (scmnd == NULL)
- return false;
- return ((ioreq->lnode == sld->lnode) &&
- (ioreq->rnode == sld->rnode) &&
- ((uint64_t)scmnd->device->lun == sld->oslun));
- case CSIO_LEV_RNODE:
- return ((ioreq->lnode == sld->lnode) &&
- (ioreq->rnode == sld->rnode));
- case CSIO_LEV_LNODE:
- return (ioreq->lnode == sld->lnode);
- case CSIO_LEV_ALL:
- return true;
- default:
- return false;
- }
- }
- /*
- * csio_scsi_gather_active_ios - Gather active I/Os based on level
- * @scm: SCSI module
- * @sld: Level information
- * @dest: The queue where these I/Os have to be gathered.
- *
- * Should be called with lock held.
- */
- static void
- csio_scsi_gather_active_ios(struct csio_scsim *scm,
- struct csio_scsi_level_data *sld,
- struct list_head *dest)
- {
- struct list_head *tmp, *next;
- if (list_empty(&scm->active_q))
- return;
- /* Just splice the entire active_q into dest */
- if (sld->level == CSIO_LEV_ALL) {
- list_splice_tail_init(&scm->active_q, dest);
- return;
- }
- list_for_each_safe(tmp, next, &scm->active_q) {
- if (csio_scsi_match_io((struct csio_ioreq *)tmp, sld)) {
- list_del_init(tmp);
- list_add_tail(tmp, dest);
- }
- }
- }
- static inline bool
- csio_scsi_itnexus_loss_error(uint16_t error)
- {
- switch (error) {
- case FW_ERR_LINK_DOWN:
- case FW_RDEV_NOT_READY:
- case FW_ERR_RDEV_LOST:
- case FW_ERR_RDEV_LOGO:
- case FW_ERR_RDEV_IMPL_LOGO:
- return 1;
- }
- return 0;
- }
- /*
- * csio_scsi_fcp_cmnd - Frame the SCSI FCP command paylod.
- * @req: IO req structure.
- * @addr: DMA location to place the payload.
- *
- * This routine is shared between FCP_WRITE, FCP_READ and FCP_CMD requests.
- */
- static inline void
- csio_scsi_fcp_cmnd(struct csio_ioreq *req, void *addr)
- {
- struct fcp_cmnd *fcp_cmnd = (struct fcp_cmnd *)addr;
- struct scsi_cmnd *scmnd = csio_scsi_cmnd(req);
- /* Check for Task Management */
- if (likely(scmnd->SCp.Message == 0)) {
- int_to_scsilun(scmnd->device->lun, &fcp_cmnd->fc_lun);
- fcp_cmnd->fc_tm_flags = 0;
- fcp_cmnd->fc_cmdref = 0;
- memcpy(fcp_cmnd->fc_cdb, scmnd->cmnd, 16);
- fcp_cmnd->fc_pri_ta = FCP_PTA_SIMPLE;
- fcp_cmnd->fc_dl = cpu_to_be32(scsi_bufflen(scmnd));
- if (req->nsge)
- if (req->datadir == DMA_TO_DEVICE)
- fcp_cmnd->fc_flags = FCP_CFL_WRDATA;
- else
- fcp_cmnd->fc_flags = FCP_CFL_RDDATA;
- else
- fcp_cmnd->fc_flags = 0;
- } else {
- memset(fcp_cmnd, 0, sizeof(*fcp_cmnd));
- int_to_scsilun(scmnd->device->lun, &fcp_cmnd->fc_lun);
- fcp_cmnd->fc_tm_flags = (uint8_t)scmnd->SCp.Message;
- }
- }
- /*
- * csio_scsi_init_cmd_wr - Initialize the SCSI CMD WR.
- * @req: IO req structure.
- * @addr: DMA location to place the payload.
- * @size: Size of WR (including FW WR + immed data + rsp SG entry
- *
- * Wrapper for populating fw_scsi_cmd_wr.
- */
- static inline void
- csio_scsi_init_cmd_wr(struct csio_ioreq *req, void *addr, uint32_t size)
- {
- struct csio_hw *hw = req->lnode->hwp;
- struct csio_rnode *rn = req->rnode;
- struct fw_scsi_cmd_wr *wr = (struct fw_scsi_cmd_wr *)addr;
- struct csio_dma_buf *dma_buf;
- uint8_t imm = csio_hw_to_scsim(hw)->proto_cmd_len;
- wr->op_immdlen = cpu_to_be32(FW_WR_OP_V(FW_SCSI_CMD_WR) |
- FW_SCSI_CMD_WR_IMMDLEN(imm));
- wr->flowid_len16 = cpu_to_be32(FW_WR_FLOWID_V(rn->flowid) |
- FW_WR_LEN16_V(
- DIV_ROUND_UP(size, 16)));
- wr->cookie = (uintptr_t) req;
- wr->iqid = cpu_to_be16(csio_q_physiqid(hw, req->iq_idx));
- wr->tmo_val = (uint8_t) req->tmo;
- wr->r3 = 0;
- memset(&wr->r5, 0, 8);
- /* Get RSP DMA buffer */
- dma_buf = &req->dma_buf;
- /* Prepare RSP SGL */
- wr->rsp_dmalen = cpu_to_be32(dma_buf->len);
- wr->rsp_dmaaddr = cpu_to_be64(dma_buf->paddr);
- wr->r6 = 0;
- wr->u.fcoe.ctl_pri = 0;
- wr->u.fcoe.cp_en_class = 0;
- wr->u.fcoe.r4_lo[0] = 0;
- wr->u.fcoe.r4_lo[1] = 0;
- /* Frame a FCP command */
- csio_scsi_fcp_cmnd(req, (void *)((uintptr_t)addr +
- sizeof(struct fw_scsi_cmd_wr)));
- }
- #define CSIO_SCSI_CMD_WR_SZ(_imm) \
- (sizeof(struct fw_scsi_cmd_wr) + /* WR size */ \
- ALIGN((_imm), 16)) /* Immed data */
- #define CSIO_SCSI_CMD_WR_SZ_16(_imm) \
- (ALIGN(CSIO_SCSI_CMD_WR_SZ((_imm)), 16))
- /*
- * csio_scsi_cmd - Create a SCSI CMD WR.
- * @req: IO req structure.
- *
- * Gets a WR slot in the ingress queue and initializes it with SCSI CMD WR.
- *
- */
- static inline void
- csio_scsi_cmd(struct csio_ioreq *req)
- {
- struct csio_wr_pair wrp;
- struct csio_hw *hw = req->lnode->hwp;
- struct csio_scsim *scsim = csio_hw_to_scsim(hw);
- uint32_t size = CSIO_SCSI_CMD_WR_SZ_16(scsim->proto_cmd_len);
- req->drv_status = csio_wr_get(hw, req->eq_idx, size, &wrp);
- if (unlikely(req->drv_status != 0))
- return;
- if (wrp.size1 >= size) {
- /* Initialize WR in one shot */
- csio_scsi_init_cmd_wr(req, wrp.addr1, size);
- } else {
- uint8_t *tmpwr = csio_q_eq_wrap(hw, req->eq_idx);
- /*
- * Make a temporary copy of the WR and write back
- * the copy into the WR pair.
- */
- csio_scsi_init_cmd_wr(req, (void *)tmpwr, size);
- memcpy(wrp.addr1, tmpwr, wrp.size1);
- memcpy(wrp.addr2, tmpwr + wrp.size1, size - wrp.size1);
- }
- }
- /*
- * csio_scsi_init_ulptx_dsgl - Fill in a ULP_TX_SC_DSGL
- * @hw: HW module
- * @req: IO request
- * @sgl: ULP TX SGL pointer.
- *
- */
- static inline void
- csio_scsi_init_ultptx_dsgl(struct csio_hw *hw, struct csio_ioreq *req,
- struct ulptx_sgl *sgl)
- {
- struct ulptx_sge_pair *sge_pair = NULL;
- struct scatterlist *sgel;
- uint32_t i = 0;
- uint32_t xfer_len;
- struct list_head *tmp;
- struct csio_dma_buf *dma_buf;
- struct scsi_cmnd *scmnd = csio_scsi_cmnd(req);
- sgl->cmd_nsge = htonl(ULPTX_CMD_V(ULP_TX_SC_DSGL) | ULPTX_MORE_F |
- ULPTX_NSGE_V(req->nsge));
- /* Now add the data SGLs */
- if (likely(!req->dcopy)) {
- scsi_for_each_sg(scmnd, sgel, req->nsge, i) {
- if (i == 0) {
- sgl->addr0 = cpu_to_be64(sg_dma_address(sgel));
- sgl->len0 = cpu_to_be32(sg_dma_len(sgel));
- sge_pair = (struct ulptx_sge_pair *)(sgl + 1);
- continue;
- }
- if ((i - 1) & 0x1) {
- sge_pair->addr[1] = cpu_to_be64(
- sg_dma_address(sgel));
- sge_pair->len[1] = cpu_to_be32(
- sg_dma_len(sgel));
- sge_pair++;
- } else {
- sge_pair->addr[0] = cpu_to_be64(
- sg_dma_address(sgel));
- sge_pair->len[0] = cpu_to_be32(
- sg_dma_len(sgel));
- }
- }
- } else {
- /* Program sg elements with driver's DDP buffer */
- xfer_len = scsi_bufflen(scmnd);
- list_for_each(tmp, &req->gen_list) {
- dma_buf = (struct csio_dma_buf *)tmp;
- if (i == 0) {
- sgl->addr0 = cpu_to_be64(dma_buf->paddr);
- sgl->len0 = cpu_to_be32(
- min(xfer_len, dma_buf->len));
- sge_pair = (struct ulptx_sge_pair *)(sgl + 1);
- } else if ((i - 1) & 0x1) {
- sge_pair->addr[1] = cpu_to_be64(dma_buf->paddr);
- sge_pair->len[1] = cpu_to_be32(
- min(xfer_len, dma_buf->len));
- sge_pair++;
- } else {
- sge_pair->addr[0] = cpu_to_be64(dma_buf->paddr);
- sge_pair->len[0] = cpu_to_be32(
- min(xfer_len, dma_buf->len));
- }
- xfer_len -= min(xfer_len, dma_buf->len);
- i++;
- }
- }
- }
- /*
- * csio_scsi_init_read_wr - Initialize the READ SCSI WR.
- * @req: IO req structure.
- * @wrp: DMA location to place the payload.
- * @size: Size of WR (including FW WR + immed data + rsp SG entry + data SGL
- *
- * Wrapper for populating fw_scsi_read_wr.
- */
- static inline void
- csio_scsi_init_read_wr(struct csio_ioreq *req, void *wrp, uint32_t size)
- {
- struct csio_hw *hw = req->lnode->hwp;
- struct csio_rnode *rn = req->rnode;
- struct fw_scsi_read_wr *wr = (struct fw_scsi_read_wr *)wrp;
- struct ulptx_sgl *sgl;
- struct csio_dma_buf *dma_buf;
- uint8_t imm = csio_hw_to_scsim(hw)->proto_cmd_len;
- struct scsi_cmnd *scmnd = csio_scsi_cmnd(req);
- wr->op_immdlen = cpu_to_be32(FW_WR_OP_V(FW_SCSI_READ_WR) |
- FW_SCSI_READ_WR_IMMDLEN(imm));
- wr->flowid_len16 = cpu_to_be32(FW_WR_FLOWID_V(rn->flowid) |
- FW_WR_LEN16_V(DIV_ROUND_UP(size, 16)));
- wr->cookie = (uintptr_t)req;
- wr->iqid = cpu_to_be16(csio_q_physiqid(hw, req->iq_idx));
- wr->tmo_val = (uint8_t)(req->tmo);
- wr->use_xfer_cnt = 1;
- wr->xfer_cnt = cpu_to_be32(scsi_bufflen(scmnd));
- wr->ini_xfer_cnt = cpu_to_be32(scsi_bufflen(scmnd));
- /* Get RSP DMA buffer */
- dma_buf = &req->dma_buf;
- /* Prepare RSP SGL */
- wr->rsp_dmalen = cpu_to_be32(dma_buf->len);
- wr->rsp_dmaaddr = cpu_to_be64(dma_buf->paddr);
- wr->r4 = 0;
- wr->u.fcoe.ctl_pri = 0;
- wr->u.fcoe.cp_en_class = 0;
- wr->u.fcoe.r3_lo[0] = 0;
- wr->u.fcoe.r3_lo[1] = 0;
- csio_scsi_fcp_cmnd(req, (void *)((uintptr_t)wrp +
- sizeof(struct fw_scsi_read_wr)));
- /* Move WR pointer past command and immediate data */
- sgl = (struct ulptx_sgl *)((uintptr_t)wrp +
- sizeof(struct fw_scsi_read_wr) + ALIGN(imm, 16));
- /* Fill in the DSGL */
- csio_scsi_init_ultptx_dsgl(hw, req, sgl);
- }
- /*
- * csio_scsi_init_write_wr - Initialize the WRITE SCSI WR.
- * @req: IO req structure.
- * @wrp: DMA location to place the payload.
- * @size: Size of WR (including FW WR + immed data + rsp SG entry + data SGL
- *
- * Wrapper for populating fw_scsi_write_wr.
- */
- static inline void
- csio_scsi_init_write_wr(struct csio_ioreq *req, void *wrp, uint32_t size)
- {
- struct csio_hw *hw = req->lnode->hwp;
- struct csio_rnode *rn = req->rnode;
- struct fw_scsi_write_wr *wr = (struct fw_scsi_write_wr *)wrp;
- struct ulptx_sgl *sgl;
- struct csio_dma_buf *dma_buf;
- uint8_t imm = csio_hw_to_scsim(hw)->proto_cmd_len;
- struct scsi_cmnd *scmnd = csio_scsi_cmnd(req);
- wr->op_immdlen = cpu_to_be32(FW_WR_OP_V(FW_SCSI_WRITE_WR) |
- FW_SCSI_WRITE_WR_IMMDLEN(imm));
- wr->flowid_len16 = cpu_to_be32(FW_WR_FLOWID_V(rn->flowid) |
- FW_WR_LEN16_V(DIV_ROUND_UP(size, 16)));
- wr->cookie = (uintptr_t)req;
- wr->iqid = cpu_to_be16(csio_q_physiqid(hw, req->iq_idx));
- wr->tmo_val = (uint8_t)(req->tmo);
- wr->use_xfer_cnt = 1;
- wr->xfer_cnt = cpu_to_be32(scsi_bufflen(scmnd));
- wr->ini_xfer_cnt = cpu_to_be32(scsi_bufflen(scmnd));
- /* Get RSP DMA buffer */
- dma_buf = &req->dma_buf;
- /* Prepare RSP SGL */
- wr->rsp_dmalen = cpu_to_be32(dma_buf->len);
- wr->rsp_dmaaddr = cpu_to_be64(dma_buf->paddr);
- wr->r4 = 0;
- wr->u.fcoe.ctl_pri = 0;
- wr->u.fcoe.cp_en_class = 0;
- wr->u.fcoe.r3_lo[0] = 0;
- wr->u.fcoe.r3_lo[1] = 0;
- csio_scsi_fcp_cmnd(req, (void *)((uintptr_t)wrp +
- sizeof(struct fw_scsi_write_wr)));
- /* Move WR pointer past command and immediate data */
- sgl = (struct ulptx_sgl *)((uintptr_t)wrp +
- sizeof(struct fw_scsi_write_wr) + ALIGN(imm, 16));
- /* Fill in the DSGL */
- csio_scsi_init_ultptx_dsgl(hw, req, sgl);
- }
- /* Calculate WR size needed for fw_scsi_read_wr/fw_scsi_write_wr */
- #define CSIO_SCSI_DATA_WRSZ(req, oper, sz, imm) \
- do { \
- (sz) = sizeof(struct fw_scsi_##oper##_wr) + /* WR size */ \
- ALIGN((imm), 16) + /* Immed data */ \
- sizeof(struct ulptx_sgl); /* ulptx_sgl */ \
- \
- if (unlikely((req)->nsge > 1)) \
- (sz) += (sizeof(struct ulptx_sge_pair) * \
- (ALIGN(((req)->nsge - 1), 2) / 2)); \
- /* Data SGE */ \
- } while (0)
- /*
- * csio_scsi_read - Create a SCSI READ WR.
- * @req: IO req structure.
- *
- * Gets a WR slot in the ingress queue and initializes it with
- * SCSI READ WR.
- *
- */
- static inline void
- csio_scsi_read(struct csio_ioreq *req)
- {
- struct csio_wr_pair wrp;
- uint32_t size;
- struct csio_hw *hw = req->lnode->hwp;
- struct csio_scsim *scsim = csio_hw_to_scsim(hw);
- CSIO_SCSI_DATA_WRSZ(req, read, size, scsim->proto_cmd_len);
- size = ALIGN(size, 16);
- req->drv_status = csio_wr_get(hw, req->eq_idx, size, &wrp);
- if (likely(req->drv_status == 0)) {
- if (likely(wrp.size1 >= size)) {
- /* Initialize WR in one shot */
- csio_scsi_init_read_wr(req, wrp.addr1, size);
- } else {
- uint8_t *tmpwr = csio_q_eq_wrap(hw, req->eq_idx);
- /*
- * Make a temporary copy of the WR and write back
- * the copy into the WR pair.
- */
- csio_scsi_init_read_wr(req, (void *)tmpwr, size);
- memcpy(wrp.addr1, tmpwr, wrp.size1);
- memcpy(wrp.addr2, tmpwr + wrp.size1, size - wrp.size1);
- }
- }
- }
- /*
- * csio_scsi_write - Create a SCSI WRITE WR.
- * @req: IO req structure.
- *
- * Gets a WR slot in the ingress queue and initializes it with
- * SCSI WRITE WR.
- *
- */
- static inline void
- csio_scsi_write(struct csio_ioreq *req)
- {
- struct csio_wr_pair wrp;
- uint32_t size;
- struct csio_hw *hw = req->lnode->hwp;
- struct csio_scsim *scsim = csio_hw_to_scsim(hw);
- CSIO_SCSI_DATA_WRSZ(req, write, size, scsim->proto_cmd_len);
- size = ALIGN(size, 16);
- req->drv_status = csio_wr_get(hw, req->eq_idx, size, &wrp);
- if (likely(req->drv_status == 0)) {
- if (likely(wrp.size1 >= size)) {
- /* Initialize WR in one shot */
- csio_scsi_init_write_wr(req, wrp.addr1, size);
- } else {
- uint8_t *tmpwr = csio_q_eq_wrap(hw, req->eq_idx);
- /*
- * Make a temporary copy of the WR and write back
- * the copy into the WR pair.
- */
- csio_scsi_init_write_wr(req, (void *)tmpwr, size);
- memcpy(wrp.addr1, tmpwr, wrp.size1);
- memcpy(wrp.addr2, tmpwr + wrp.size1, size - wrp.size1);
- }
- }
- }
- /*
- * csio_setup_ddp - Setup DDP buffers for Read request.
- * @req: IO req structure.
- *
- * Checks SGLs/Data buffers are virtually contiguous required for DDP.
- * If contiguous,driver posts SGLs in the WR otherwise post internal
- * buffers for such request for DDP.
- */
- static inline void
- csio_setup_ddp(struct csio_scsim *scsim, struct csio_ioreq *req)
- {
- #ifdef __CSIO_DEBUG__
- struct csio_hw *hw = req->lnode->hwp;
- #endif
- struct scatterlist *sgel = NULL;
- struct scsi_cmnd *scmnd = csio_scsi_cmnd(req);
- uint64_t sg_addr = 0;
- uint32_t ddp_pagesz = 4096;
- uint32_t buf_off;
- struct csio_dma_buf *dma_buf = NULL;
- uint32_t alloc_len = 0;
- uint32_t xfer_len = 0;
- uint32_t sg_len = 0;
- uint32_t i;
- scsi_for_each_sg(scmnd, sgel, req->nsge, i) {
- sg_addr = sg_dma_address(sgel);
- sg_len = sg_dma_len(sgel);
- buf_off = sg_addr & (ddp_pagesz - 1);
- /* Except 1st buffer,all buffer addr have to be Page aligned */
- if (i != 0 && buf_off) {
- csio_dbg(hw, "SGL addr not DDP aligned (%llx:%d)\n",
- sg_addr, sg_len);
- goto unaligned;
- }
- /* Except last buffer,all buffer must end on page boundary */
- if ((i != (req->nsge - 1)) &&
- ((buf_off + sg_len) & (ddp_pagesz - 1))) {
- csio_dbg(hw,
- "SGL addr not ending on page boundary"
- "(%llx:%d)\n", sg_addr, sg_len);
- goto unaligned;
- }
- }
- /* SGL's are virtually contiguous. HW will DDP to SGLs */
- req->dcopy = 0;
- csio_scsi_read(req);
- return;
- unaligned:
- CSIO_INC_STATS(scsim, n_unaligned);
- /*
- * For unaligned SGLs, driver will allocate internal DDP buffer.
- * Once command is completed data from DDP buffer copied to SGLs
- */
- req->dcopy = 1;
- /* Use gen_list to store the DDP buffers */
- INIT_LIST_HEAD(&req->gen_list);
- xfer_len = scsi_bufflen(scmnd);
- i = 0;
- /* Allocate ddp buffers for this request */
- while (alloc_len < xfer_len) {
- dma_buf = csio_get_scsi_ddp(scsim);
- if (dma_buf == NULL || i > scsim->max_sge) {
- req->drv_status = -EBUSY;
- break;
- }
- alloc_len += dma_buf->len;
- /* Added to IO req */
- list_add_tail(&dma_buf->list, &req->gen_list);
- i++;
- }
- if (!req->drv_status) {
- /* set number of ddp bufs used */
- req->nsge = i;
- csio_scsi_read(req);
- return;
- }
- /* release dma descs */
- if (i > 0)
- csio_put_scsi_ddp_list(scsim, &req->gen_list, i);
- }
- /*
- * csio_scsi_init_abrt_cls_wr - Initialize an ABORT/CLOSE WR.
- * @req: IO req structure.
- * @addr: DMA location to place the payload.
- * @size: Size of WR
- * @abort: abort OR close
- *
- * Wrapper for populating fw_scsi_cmd_wr.
- */
- static inline void
- csio_scsi_init_abrt_cls_wr(struct csio_ioreq *req, void *addr, uint32_t size,
- bool abort)
- {
- struct csio_hw *hw = req->lnode->hwp;
- struct csio_rnode *rn = req->rnode;
- struct fw_scsi_abrt_cls_wr *wr = (struct fw_scsi_abrt_cls_wr *)addr;
- wr->op_immdlen = cpu_to_be32(FW_WR_OP_V(FW_SCSI_ABRT_CLS_WR));
- wr->flowid_len16 = cpu_to_be32(FW_WR_FLOWID_V(rn->flowid) |
- FW_WR_LEN16_V(
- DIV_ROUND_UP(size, 16)));
- wr->cookie = (uintptr_t) req;
- wr->iqid = cpu_to_be16(csio_q_physiqid(hw, req->iq_idx));
- wr->tmo_val = (uint8_t) req->tmo;
- /* 0 for CHK_ALL_IO tells FW to look up t_cookie */
- wr->sub_opcode_to_chk_all_io =
- (FW_SCSI_ABRT_CLS_WR_SUB_OPCODE(abort) |
- FW_SCSI_ABRT_CLS_WR_CHK_ALL_IO(0));
- wr->r3[0] = 0;
- wr->r3[1] = 0;
- wr->r3[2] = 0;
- wr->r3[3] = 0;
- /* Since we re-use the same ioreq for abort as well */
- wr->t_cookie = (uintptr_t) req;
- }
- static inline void
- csio_scsi_abrt_cls(struct csio_ioreq *req, bool abort)
- {
- struct csio_wr_pair wrp;
- struct csio_hw *hw = req->lnode->hwp;
- uint32_t size = ALIGN(sizeof(struct fw_scsi_abrt_cls_wr), 16);
- req->drv_status = csio_wr_get(hw, req->eq_idx, size, &wrp);
- if (req->drv_status != 0)
- return;
- if (wrp.size1 >= size) {
- /* Initialize WR in one shot */
- csio_scsi_init_abrt_cls_wr(req, wrp.addr1, size, abort);
- } else {
- uint8_t *tmpwr = csio_q_eq_wrap(hw, req->eq_idx);
- /*
- * Make a temporary copy of the WR and write back
- * the copy into the WR pair.
- */
- csio_scsi_init_abrt_cls_wr(req, (void *)tmpwr, size, abort);
- memcpy(wrp.addr1, tmpwr, wrp.size1);
- memcpy(wrp.addr2, tmpwr + wrp.size1, size - wrp.size1);
- }
- }
- /*****************************************************************************/
- /* START: SCSI SM */
- /*****************************************************************************/
- static void
- csio_scsis_uninit(struct csio_ioreq *req, enum csio_scsi_ev evt)
- {
- struct csio_hw *hw = req->lnode->hwp;
- struct csio_scsim *scsim = csio_hw_to_scsim(hw);
- switch (evt) {
- case CSIO_SCSIE_START_IO:
- if (req->nsge) {
- if (req->datadir == DMA_TO_DEVICE) {
- req->dcopy = 0;
- csio_scsi_write(req);
- } else
- csio_setup_ddp(scsim, req);
- } else {
- csio_scsi_cmd(req);
- }
- if (likely(req->drv_status == 0)) {
- /* change state and enqueue on active_q */
- csio_set_state(&req->sm, csio_scsis_io_active);
- list_add_tail(&req->sm.sm_list, &scsim->active_q);
- csio_wr_issue(hw, req->eq_idx, false);
- CSIO_INC_STATS(scsim, n_active);
- return;
- }
- break;
- case CSIO_SCSIE_START_TM:
- csio_scsi_cmd(req);
- if (req->drv_status == 0) {
- /*
- * NOTE: We collect the affected I/Os prior to issuing
- * LUN reset, and not after it. This is to prevent
- * aborting I/Os that get issued after the LUN reset,
- * but prior to LUN reset completion (in the event that
- * the host stack has not blocked I/Os to a LUN that is
- * being reset.
- */
- csio_set_state(&req->sm, csio_scsis_tm_active);
- list_add_tail(&req->sm.sm_list, &scsim->active_q);
- csio_wr_issue(hw, req->eq_idx, false);
- CSIO_INC_STATS(scsim, n_tm_active);
- }
- return;
- case CSIO_SCSIE_ABORT:
- case CSIO_SCSIE_CLOSE:
- /*
- * NOTE:
- * We could get here due to :
- * - a window in the cleanup path of the SCSI module
- * (csio_scsi_abort_io()). Please see NOTE in this function.
- * - a window in the time we tried to issue an abort/close
- * of a request to FW, and the FW completed the request
- * itself.
- * Print a message for now, and return INVAL either way.
- */
- req->drv_status = -EINVAL;
- csio_warn(hw, "Trying to abort/close completed IO:%p!\n", req);
- break;
- default:
- csio_dbg(hw, "Unhandled event:%d sent to req:%p\n", evt, req);
- CSIO_DB_ASSERT(0);
- }
- }
- static void
- csio_scsis_io_active(struct csio_ioreq *req, enum csio_scsi_ev evt)
- {
- struct csio_hw *hw = req->lnode->hwp;
- struct csio_scsim *scm = csio_hw_to_scsim(hw);
- struct csio_rnode *rn;
- switch (evt) {
- case CSIO_SCSIE_COMPLETED:
- CSIO_DEC_STATS(scm, n_active);
- list_del_init(&req->sm.sm_list);
- csio_set_state(&req->sm, csio_scsis_uninit);
- /*
- * In MSIX mode, with multiple queues, the SCSI compeltions
- * could reach us sooner than the FW events sent to indicate
- * I-T nexus loss (link down, remote device logo etc). We
- * dont want to be returning such I/Os to the upper layer
- * immediately, since we wouldnt have reported the I-T nexus
- * loss itself. This forces us to serialize such completions
- * with the reporting of the I-T nexus loss. Therefore, we
- * internally queue up such up such completions in the rnode.
- * The reporting of I-T nexus loss to the upper layer is then
- * followed by the returning of I/Os in this internal queue.
- * Having another state alongwith another queue helps us take
- * actions for events such as ABORT received while we are
- * in this rnode queue.
- */
- if (unlikely(req->wr_status != FW_SUCCESS)) {
- rn = req->rnode;
- /*
- * FW says remote device is lost, but rnode
- * doesnt reflect it.
- */
- if (csio_scsi_itnexus_loss_error(req->wr_status) &&
- csio_is_rnode_ready(rn)) {
- csio_set_state(&req->sm,
- csio_scsis_shost_cmpl_await);
- list_add_tail(&req->sm.sm_list,
- &rn->host_cmpl_q);
- }
- }
- break;
- case CSIO_SCSIE_ABORT:
- csio_scsi_abrt_cls(req, SCSI_ABORT);
- if (req->drv_status == 0) {
- csio_wr_issue(hw, req->eq_idx, false);
- csio_set_state(&req->sm, csio_scsis_aborting);
- }
- break;
- case CSIO_SCSIE_CLOSE:
- csio_scsi_abrt_cls(req, SCSI_CLOSE);
- if (req->drv_status == 0) {
- csio_wr_issue(hw, req->eq_idx, false);
- csio_set_state(&req->sm, csio_scsis_closing);
- }
- break;
- case CSIO_SCSIE_DRVCLEANUP:
- req->wr_status = FW_HOSTERROR;
- CSIO_DEC_STATS(scm, n_active);
- csio_set_state(&req->sm, csio_scsis_uninit);
- break;
- default:
- csio_dbg(hw, "Unhandled event:%d sent to req:%p\n", evt, req);
- CSIO_DB_ASSERT(0);
- }
- }
- static void
- csio_scsis_tm_active(struct csio_ioreq *req, enum csio_scsi_ev evt)
- {
- struct csio_hw *hw = req->lnode->hwp;
- struct csio_scsim *scm = csio_hw_to_scsim(hw);
- switch (evt) {
- case CSIO_SCSIE_COMPLETED:
- CSIO_DEC_STATS(scm, n_tm_active);
- list_del_init(&req->sm.sm_list);
- csio_set_state(&req->sm, csio_scsis_uninit);
- break;
- case CSIO_SCSIE_ABORT:
- csio_scsi_abrt_cls(req, SCSI_ABORT);
- if (req->drv_status == 0) {
- csio_wr_issue(hw, req->eq_idx, false);
- csio_set_state(&req->sm, csio_scsis_aborting);
- }
- break;
- case CSIO_SCSIE_CLOSE:
- csio_scsi_abrt_cls(req, SCSI_CLOSE);
- if (req->drv_status == 0) {
- csio_wr_issue(hw, req->eq_idx, false);
- csio_set_state(&req->sm, csio_scsis_closing);
- }
- break;
- case CSIO_SCSIE_DRVCLEANUP:
- req->wr_status = FW_HOSTERROR;
- CSIO_DEC_STATS(scm, n_tm_active);
- csio_set_state(&req->sm, csio_scsis_uninit);
- break;
- default:
- csio_dbg(hw, "Unhandled event:%d sent to req:%p\n", evt, req);
- CSIO_DB_ASSERT(0);
- }
- }
- static void
- csio_scsis_aborting(struct csio_ioreq *req, enum csio_scsi_ev evt)
- {
- struct csio_hw *hw = req->lnode->hwp;
- struct csio_scsim *scm = csio_hw_to_scsim(hw);
- switch (evt) {
- case CSIO_SCSIE_COMPLETED:
- csio_dbg(hw,
- "ioreq %p recvd cmpltd (wr_status:%d) "
- "in aborting st\n", req, req->wr_status);
- /*
- * Use -ECANCELED to explicitly tell the ABORTED event that
- * the original I/O was returned to driver by FW.
- * We dont really care if the I/O was returned with success by
- * FW (because the ABORT and completion of the I/O crossed each
- * other), or any other return value. Once we are in aborting
- * state, the success or failure of the I/O is unimportant to
- * us.
- */
- req->drv_status = -ECANCELED;
- break;
- case CSIO_SCSIE_ABORT:
- CSIO_INC_STATS(scm, n_abrt_dups);
- break;
- case CSIO_SCSIE_ABORTED:
- csio_dbg(hw, "abort of %p return status:0x%x drv_status:%x\n",
- req, req->wr_status, req->drv_status);
- /*
- * Check if original I/O WR completed before the Abort
- * completion.
- */
- if (req->drv_status != -ECANCELED) {
- csio_warn(hw,
- "Abort completed before original I/O,"
- " req:%p\n", req);
- CSIO_DB_ASSERT(0);
- }
- /*
- * There are the following possible scenarios:
- * 1. The abort completed successfully, FW returned FW_SUCCESS.
- * 2. The completion of an I/O and the receipt of
- * abort for that I/O by the FW crossed each other.
- * The FW returned FW_EINVAL. The original I/O would have
- * returned with FW_SUCCESS or any other SCSI error.
- * 3. The FW couldnt sent the abort out on the wire, as there
- * was an I-T nexus loss (link down, remote device logged
- * out etc). FW sent back an appropriate IT nexus loss status
- * for the abort.
- * 4. FW sent an abort, but abort timed out (remote device
- * didnt respond). FW replied back with
- * FW_SCSI_ABORT_TIMEDOUT.
- * 5. FW couldnt genuinely abort the request for some reason,
- * and sent us an error.
- *
- * The first 3 scenarios are treated as succesful abort
- * operations by the host, while the last 2 are failed attempts
- * to abort. Manipulate the return value of the request
- * appropriately, so that host can convey these results
- * back to the upper layer.
- */
- if ((req->wr_status == FW_SUCCESS) ||
- (req->wr_status == FW_EINVAL) ||
- csio_scsi_itnexus_loss_error(req->wr_status))
- req->wr_status = FW_SCSI_ABORT_REQUESTED;
- CSIO_DEC_STATS(scm, n_active);
- list_del_init(&req->sm.sm_list);
- csio_set_state(&req->sm, csio_scsis_uninit);
- break;
- case CSIO_SCSIE_DRVCLEANUP:
- req->wr_status = FW_HOSTERROR;
- CSIO_DEC_STATS(scm, n_active);
- csio_set_state(&req->sm, csio_scsis_uninit);
- break;
- case CSIO_SCSIE_CLOSE:
- /*
- * We can receive this event from the module
- * cleanup paths, if the FW forgot to reply to the ABORT WR
- * and left this ioreq in this state. For now, just ignore
- * the event. The CLOSE event is sent to this state, as
- * the LINK may have already gone down.
- */
- break;
- default:
- csio_dbg(hw, "Unhandled event:%d sent to req:%p\n", evt, req);
- CSIO_DB_ASSERT(0);
- }
- }
- static void
- csio_scsis_closing(struct csio_ioreq *req, enum csio_scsi_ev evt)
- {
- struct csio_hw *hw = req->lnode->hwp;
- struct csio_scsim *scm = csio_hw_to_scsim(hw);
- switch (evt) {
- case CSIO_SCSIE_COMPLETED:
- csio_dbg(hw,
- "ioreq %p recvd cmpltd (wr_status:%d) "
- "in closing st\n", req, req->wr_status);
- /*
- * Use -ECANCELED to explicitly tell the CLOSED event that
- * the original I/O was returned to driver by FW.
- * We dont really care if the I/O was returned with success by
- * FW (because the CLOSE and completion of the I/O crossed each
- * other), or any other return value. Once we are in aborting
- * state, the success or failure of the I/O is unimportant to
- * us.
- */
- req->drv_status = -ECANCELED;
- break;
- case CSIO_SCSIE_CLOSED:
- /*
- * Check if original I/O WR completed before the Close
- * completion.
- */
- if (req->drv_status != -ECANCELED) {
- csio_fatal(hw,
- "Close completed before original I/O,"
- " req:%p\n", req);
- CSIO_DB_ASSERT(0);
- }
- /*
- * Either close succeeded, or we issued close to FW at the
- * same time FW compelted it to us. Either way, the I/O
- * is closed.
- */
- CSIO_DB_ASSERT((req->wr_status == FW_SUCCESS) ||
- (req->wr_status == FW_EINVAL));
- req->wr_status = FW_SCSI_CLOSE_REQUESTED;
- CSIO_DEC_STATS(scm, n_active);
- list_del_init(&req->sm.sm_list);
- csio_set_state(&req->sm, csio_scsis_uninit);
- break;
- case CSIO_SCSIE_CLOSE:
- break;
- case CSIO_SCSIE_DRVCLEANUP:
- req->wr_status = FW_HOSTERROR;
- CSIO_DEC_STATS(scm, n_active);
- csio_set_state(&req->sm, csio_scsis_uninit);
- break;
- default:
- csio_dbg(hw, "Unhandled event:%d sent to req:%p\n", evt, req);
- CSIO_DB_ASSERT(0);
- }
- }
- static void
- csio_scsis_shost_cmpl_await(struct csio_ioreq *req, enum csio_scsi_ev evt)
- {
- switch (evt) {
- case CSIO_SCSIE_ABORT:
- case CSIO_SCSIE_CLOSE:
- /*
- * Just succeed the abort request, and hope that
- * the remote device unregister path will cleanup
- * this I/O to the upper layer within a sane
- * amount of time.
- */
- /*
- * A close can come in during a LINK DOWN. The FW would have
- * returned us the I/O back, but not the remote device lost
- * FW event. In this interval, if the I/O times out at the upper
- * layer, a close can come in. Take the same action as abort:
- * return success, and hope that the remote device unregister
- * path will cleanup this I/O. If the FW still doesnt send
- * the msg, the close times out, and the upper layer resorts
- * to the next level of error recovery.
- */
- req->drv_status = 0;
- break;
- case CSIO_SCSIE_DRVCLEANUP:
- csio_set_state(&req->sm, csio_scsis_uninit);
- break;
- default:
- csio_dbg(req->lnode->hwp, "Unhandled event:%d sent to req:%p\n",
- evt, req);
- CSIO_DB_ASSERT(0);
- }
- }
- /*
- * csio_scsi_cmpl_handler - WR completion handler for SCSI.
- * @hw: HW module.
- * @wr: The completed WR from the ingress queue.
- * @len: Length of the WR.
- * @flb: Freelist buffer array.
- * @priv: Private object
- * @scsiwr: Pointer to SCSI WR.
- *
- * This is the WR completion handler called per completion from the
- * ISR. It is called with lock held. It walks past the RSS and CPL message
- * header where the actual WR is present.
- * It then gets the status, WR handle (ioreq pointer) and the len of
- * the WR, based on WR opcode. Only on a non-good status is the entire
- * WR copied into the WR cache (ioreq->fw_wr).
- * The ioreq corresponding to the WR is returned to the caller.
- * NOTE: The SCSI queue doesnt allocate a freelist today, hence
- * no freelist buffer is expected.
- */
- struct csio_ioreq *
- csio_scsi_cmpl_handler(struct csio_hw *hw, void *wr, uint32_t len,
- struct csio_fl_dma_buf *flb, void *priv, uint8_t **scsiwr)
- {
- struct csio_ioreq *ioreq = NULL;
- struct cpl_fw6_msg *cpl;
- uint8_t *tempwr;
- uint8_t status;
- struct csio_scsim *scm = csio_hw_to_scsim(hw);
- /* skip RSS header */
- cpl = (struct cpl_fw6_msg *)((uintptr_t)wr + sizeof(__be64));
- if (unlikely(cpl->opcode != CPL_FW6_MSG)) {
- csio_warn(hw, "Error: Invalid CPL msg %x recvd on SCSI q\n",
- cpl->opcode);
- CSIO_INC_STATS(scm, n_inval_cplop);
- return NULL;
- }
- tempwr = (uint8_t *)(cpl->data);
- status = csio_wr_status(tempwr);
- *scsiwr = tempwr;
- if (likely((*tempwr == FW_SCSI_READ_WR) ||
- (*tempwr == FW_SCSI_WRITE_WR) ||
- (*tempwr == FW_SCSI_CMD_WR))) {
- ioreq = (struct csio_ioreq *)((uintptr_t)
- (((struct fw_scsi_read_wr *)tempwr)->cookie));
- CSIO_DB_ASSERT(virt_addr_valid(ioreq));
- ioreq->wr_status = status;
- return ioreq;
- }
- if (*tempwr == FW_SCSI_ABRT_CLS_WR) {
- ioreq = (struct csio_ioreq *)((uintptr_t)
- (((struct fw_scsi_abrt_cls_wr *)tempwr)->cookie));
- CSIO_DB_ASSERT(virt_addr_valid(ioreq));
- ioreq->wr_status = status;
- return ioreq;
- }
- csio_warn(hw, "WR with invalid opcode in SCSI IQ: %x\n", *tempwr);
- CSIO_INC_STATS(scm, n_inval_scsiop);
- return NULL;
- }
- /*
- * csio_scsi_cleanup_io_q - Cleanup the given queue.
- * @scm: SCSI module.
- * @q: Queue to be cleaned up.
- *
- * Called with lock held. Has to exit with lock held.
- */
- void
- csio_scsi_cleanup_io_q(struct csio_scsim *scm, struct list_head *q)
- {
- struct csio_hw *hw = scm->hw;
- struct csio_ioreq *ioreq;
- struct list_head *tmp, *next;
- struct scsi_cmnd *scmnd;
- /* Call back the completion routines of the active_q */
- list_for_each_safe(tmp, next, q) {
- ioreq = (struct csio_ioreq *)tmp;
- csio_scsi_drvcleanup(ioreq);
- list_del_init(&ioreq->sm.sm_list);
- scmnd = csio_scsi_cmnd(ioreq);
- spin_unlock_irq(&hw->lock);
- /*
- * Upper layers may have cleared this command, hence this
- * check to avoid accessing stale references.
- */
- if (scmnd != NULL)
- ioreq->io_cbfn(hw, ioreq);
- spin_lock_irq(&scm->freelist_lock);
- csio_put_scsi_ioreq(scm, ioreq);
- spin_unlock_irq(&scm->freelist_lock);
- spin_lock_irq(&hw->lock);
- }
- }
- #define CSIO_SCSI_ABORT_Q_POLL_MS 2000
- static void
- csio_abrt_cls(struct csio_ioreq *ioreq, struct scsi_cmnd *scmnd)
- {
- struct csio_lnode *ln = ioreq->lnode;
- struct csio_hw *hw = ln->hwp;
- int ready = 0;
- struct csio_scsim *scsim = csio_hw_to_scsim(hw);
- int rv;
- if (csio_scsi_cmnd(ioreq) != scmnd) {
- CSIO_INC_STATS(scsim, n_abrt_race_comp);
- return;
- }
- ready = csio_is_lnode_ready(ln);
- rv = csio_do_abrt_cls(hw, ioreq, (ready ? SCSI_ABORT : SCSI_CLOSE));
- if (rv != 0) {
- if (ready)
- CSIO_INC_STATS(scsim, n_abrt_busy_error);
- else
- CSIO_INC_STATS(scsim, n_cls_busy_error);
- }
- }
- /*
- * csio_scsi_abort_io_q - Abort all I/Os on given queue
- * @scm: SCSI module.
- * @q: Queue to abort.
- * @tmo: Timeout in ms
- *
- * Attempt to abort all I/Os on given queue, and wait for a max
- * of tmo milliseconds for them to complete. Returns success
- * if all I/Os are aborted. Else returns -ETIMEDOUT.
- * Should be entered with lock held. Exits with lock held.
- * NOTE:
- * Lock has to be held across the loop that aborts I/Os, since dropping the lock
- * in between can cause the list to be corrupted. As a result, the caller
- * of this function has to ensure that the number of I/os to be aborted
- * is finite enough to not cause lock-held-for-too-long issues.
- */
- static int
- csio_scsi_abort_io_q(struct csio_scsim *scm, struct list_head *q, uint32_t tmo)
- {
- struct csio_hw *hw = scm->hw;
- struct list_head *tmp, *next;
- int count = DIV_ROUND_UP(tmo, CSIO_SCSI_ABORT_Q_POLL_MS);
- struct scsi_cmnd *scmnd;
- if (list_empty(q))
- return 0;
- csio_dbg(hw, "Aborting SCSI I/Os\n");
- /* Now abort/close I/Os in the queue passed */
- list_for_each_safe(tmp, next, q) {
- scmnd = csio_scsi_cmnd((struct csio_ioreq *)tmp);
- csio_abrt_cls((struct csio_ioreq *)tmp, scmnd);
- }
- /* Wait till all active I/Os are completed/aborted/closed */
- while (!list_empty(q) && count--) {
- spin_unlock_irq(&hw->lock);
- msleep(CSIO_SCSI_ABORT_Q_POLL_MS);
- spin_lock_irq(&hw->lock);
- }
- /* all aborts completed */
- if (list_empty(q))
- return 0;
- return -ETIMEDOUT;
- }
- /*
- * csio_scsim_cleanup_io - Cleanup all I/Os in SCSI module.
- * @scm: SCSI module.
- * @abort: abort required.
- * Called with lock held, should exit with lock held.
- * Can sleep when waiting for I/Os to complete.
- */
- int
- csio_scsim_cleanup_io(struct csio_scsim *scm, bool abort)
- {
- struct csio_hw *hw = scm->hw;
- int rv = 0;
- int count = DIV_ROUND_UP(60 * 1000, CSIO_SCSI_ABORT_Q_POLL_MS);
- /* No I/Os pending */
- if (list_empty(&scm->active_q))
- return 0;
- /* Wait until all active I/Os are completed */
- while (!list_empty(&scm->active_q) && count--) {
- spin_unlock_irq(&hw->lock);
- msleep(CSIO_SCSI_ABORT_Q_POLL_MS);
- spin_lock_irq(&hw->lock);
- }
- /* all I/Os completed */
- if (list_empty(&scm->active_q))
- return 0;
- /* Else abort */
- if (abort) {
- rv = csio_scsi_abort_io_q(scm, &scm->active_q, 30000);
- if (rv == 0)
- return rv;
- csio_dbg(hw, "Some I/O aborts timed out, cleaning up..\n");
- }
- csio_scsi_cleanup_io_q(scm, &scm->active_q);
- CSIO_DB_ASSERT(list_empty(&scm->active_q));
- return rv;
- }
- /*
- * csio_scsim_cleanup_io_lnode - Cleanup all I/Os of given lnode.
- * @scm: SCSI module.
- * @lnode: lnode
- *
- * Called with lock held, should exit with lock held.
- * Can sleep (with dropped lock) when waiting for I/Os to complete.
- */
- int
- csio_scsim_cleanup_io_lnode(struct csio_scsim *scm, struct csio_lnode *ln)
- {
- struct csio_hw *hw = scm->hw;
- struct csio_scsi_level_data sld;
- int rv;
- int count = DIV_ROUND_UP(60 * 1000, CSIO_SCSI_ABORT_Q_POLL_MS);
- csio_dbg(hw, "Gathering all SCSI I/Os on lnode %p\n", ln);
- sld.level = CSIO_LEV_LNODE;
- sld.lnode = ln;
- INIT_LIST_HEAD(&ln->cmpl_q);
- csio_scsi_gather_active_ios(scm, &sld, &ln->cmpl_q);
- /* No I/Os pending on this lnode */
- if (list_empty(&ln->cmpl_q))
- return 0;
- /* Wait until all active I/Os on this lnode are completed */
- while (!list_empty(&ln->cmpl_q) && count--) {
- spin_unlock_irq(&hw->lock);
- msleep(CSIO_SCSI_ABORT_Q_POLL_MS);
- spin_lock_irq(&hw->lock);
- }
- /* all I/Os completed */
- if (list_empty(&ln->cmpl_q))
- return 0;
- csio_dbg(hw, "Some I/Os pending on ln:%p, aborting them..\n", ln);
- /* I/Os are pending, abort them */
- rv = csio_scsi_abort_io_q(scm, &ln->cmpl_q, 30000);
- if (rv != 0) {
- csio_dbg(hw, "Some I/O aborts timed out, cleaning up..\n");
- csio_scsi_cleanup_io_q(scm, &ln->cmpl_q);
- }
- CSIO_DB_ASSERT(list_empty(&ln->cmpl_q));
- return rv;
- }
- static ssize_t
- csio_show_hw_state(struct device *dev,
- struct device_attribute *attr, char *buf)
- {
- struct csio_lnode *ln = shost_priv(class_to_shost(dev));
- struct csio_hw *hw = csio_lnode_to_hw(ln);
- if (csio_is_hw_ready(hw))
- return snprintf(buf, PAGE_SIZE, "ready\n");
- else
- return snprintf(buf, PAGE_SIZE, "not ready\n");
- }
- /* Device reset */
- static ssize_t
- csio_device_reset(struct device *dev,
- struct device_attribute *attr, const char *buf, size_t count)
- {
- struct csio_lnode *ln = shost_priv(class_to_shost(dev));
- struct csio_hw *hw = csio_lnode_to_hw(ln);
- if (*buf != '1')
- return -EINVAL;
- /* Delete NPIV lnodes */
- csio_lnodes_exit(hw, 1);
- /* Block upper IOs */
- csio_lnodes_block_request(hw);
- spin_lock_irq(&hw->lock);
- csio_hw_reset(hw);
- spin_unlock_irq(&hw->lock);
- /* Unblock upper IOs */
- csio_lnodes_unblock_request(hw);
- return count;
- }
- /* disable port */
- static ssize_t
- csio_disable_port(struct device *dev,
- struct device_attribute *attr, const char *buf, size_t count)
- {
- struct csio_lnode *ln = shost_priv(class_to_shost(dev));
- struct csio_hw *hw = csio_lnode_to_hw(ln);
- bool disable;
- if (*buf == '1' || *buf == '0')
- disable = (*buf == '1') ? true : false;
- else
- return -EINVAL;
- /* Block upper IOs */
- csio_lnodes_block_by_port(hw, ln->portid);
- spin_lock_irq(&hw->lock);
- csio_disable_lnodes(hw, ln->portid, disable);
- spin_unlock_irq(&hw->lock);
- /* Unblock upper IOs */
- csio_lnodes_unblock_by_port(hw, ln->portid);
- return count;
- }
- /* Show debug level */
- static ssize_t
- csio_show_dbg_level(struct device *dev,
- struct device_attribute *attr, char *buf)
- {
- struct csio_lnode *ln = shost_priv(class_to_shost(dev));
- return snprintf(buf, PAGE_SIZE, "%x\n", ln->params.log_level);
- }
- /* Store debug level */
- static ssize_t
- csio_store_dbg_level(struct device *dev,
- struct device_attribute *attr, const char *buf, size_t count)
- {
- struct csio_lnode *ln = shost_priv(class_to_shost(dev));
- struct csio_hw *hw = csio_lnode_to_hw(ln);
- uint32_t dbg_level = 0;
- if (!isdigit(buf[0]))
- return -EINVAL;
- if (sscanf(buf, "%i", &dbg_level))
- return -EINVAL;
- ln->params.log_level = dbg_level;
- hw->params.log_level = dbg_level;
- return 0;
- }
- static DEVICE_ATTR(hw_state, S_IRUGO, csio_show_hw_state, NULL);
- static DEVICE_ATTR(device_reset, S_IWUSR, NULL, csio_device_reset);
- static DEVICE_ATTR(disable_port, S_IWUSR, NULL, csio_disable_port);
- static DEVICE_ATTR(dbg_level, S_IRUGO | S_IWUSR, csio_show_dbg_level,
- csio_store_dbg_level);
- static struct device_attribute *csio_fcoe_lport_attrs[] = {
- &dev_attr_hw_state,
- &dev_attr_device_reset,
- &dev_attr_disable_port,
- &dev_attr_dbg_level,
- NULL,
- };
- static ssize_t
- csio_show_num_reg_rnodes(struct device *dev,
- struct device_attribute *attr, char *buf)
- {
- struct csio_lnode *ln = shost_priv(class_to_shost(dev));
- return snprintf(buf, PAGE_SIZE, "%d\n", ln->num_reg_rnodes);
- }
- static DEVICE_ATTR(num_reg_rnodes, S_IRUGO, csio_show_num_reg_rnodes, NULL);
- static struct device_attribute *csio_fcoe_vport_attrs[] = {
- &dev_attr_num_reg_rnodes,
- &dev_attr_dbg_level,
- NULL,
- };
- static inline uint32_t
- csio_scsi_copy_to_sgl(struct csio_hw *hw, struct csio_ioreq *req)
- {
- struct scsi_cmnd *scmnd = (struct scsi_cmnd *)csio_scsi_cmnd(req);
- struct scatterlist *sg;
- uint32_t bytes_left;
- uint32_t bytes_copy;
- uint32_t buf_off = 0;
- uint32_t start_off = 0;
- uint32_t sg_off = 0;
- void *sg_addr;
- void *buf_addr;
- struct csio_dma_buf *dma_buf;
- bytes_left = scsi_bufflen(scmnd);
- sg = scsi_sglist(scmnd);
- dma_buf = (struct csio_dma_buf *)csio_list_next(&req->gen_list);
- /* Copy data from driver buffer to SGs of SCSI CMD */
- while (bytes_left > 0 && sg && dma_buf) {
- if (buf_off >= dma_buf->len) {
- buf_off = 0;
- dma_buf = (struct csio_dma_buf *)
- csio_list_next(dma_buf);
- continue;
- }
- if (start_off >= sg->length) {
- start_off -= sg->length;
- sg = sg_next(sg);
- continue;
- }
- buf_addr = dma_buf->vaddr + buf_off;
- sg_off = sg->offset + start_off;
- bytes_copy = min((dma_buf->len - buf_off),
- sg->length - start_off);
- bytes_copy = min((uint32_t)(PAGE_SIZE - (sg_off & ~PAGE_MASK)),
- bytes_copy);
- sg_addr = kmap_atomic(sg_page(sg) + (sg_off >> PAGE_SHIFT));
- if (!sg_addr) {
- csio_err(hw, "failed to kmap sg:%p of ioreq:%p\n",
- sg, req);
- break;
- }
- csio_dbg(hw, "copy_to_sgl:sg_addr %p sg_off %d buf %p len %d\n",
- sg_addr, sg_off, buf_addr, bytes_copy);
- memcpy(sg_addr + (sg_off & ~PAGE_MASK), buf_addr, bytes_copy);
- kunmap_atomic(sg_addr);
- start_off += bytes_copy;
- buf_off += bytes_copy;
- bytes_left -= bytes_copy;
- }
- if (bytes_left > 0)
- return DID_ERROR;
- else
- return DID_OK;
- }
- /*
- * csio_scsi_err_handler - SCSI error handler.
- * @hw: HW module.
- * @req: IO request.
- *
- */
- static inline void
- csio_scsi_err_handler(struct csio_hw *hw, struct csio_ioreq *req)
- {
- struct scsi_cmnd *cmnd = (struct scsi_cmnd *)csio_scsi_cmnd(req);
- struct csio_scsim *scm = csio_hw_to_scsim(hw);
- struct fcp_resp_with_ext *fcp_resp;
- struct fcp_resp_rsp_info *rsp_info;
- struct csio_dma_buf *dma_buf;
- uint8_t flags, scsi_status = 0;
- uint32_t host_status = DID_OK;
- uint32_t rsp_len = 0, sns_len = 0;
- struct csio_rnode *rn = (struct csio_rnode *)(cmnd->device->hostdata);
- switch (req->wr_status) {
- case FW_HOSTERROR:
- if (unlikely(!csio_is_hw_ready(hw)))
- return;
- host_status = DID_ERROR;
- CSIO_INC_STATS(scm, n_hosterror);
- break;
- case FW_SCSI_RSP_ERR:
- dma_buf = &req->dma_buf;
- fcp_resp = (struct fcp_resp_with_ext *)dma_buf->vaddr;
- rsp_info = (struct fcp_resp_rsp_info *)(fcp_resp + 1);
- flags = fcp_resp->resp.fr_flags;
- scsi_status = fcp_resp->resp.fr_status;
- if (flags & FCP_RSP_LEN_VAL) {
- rsp_len = be32_to_cpu(fcp_resp->ext.fr_rsp_len);
- if ((rsp_len != 0 && rsp_len != 4 && rsp_len != 8) ||
- (rsp_info->rsp_code != FCP_TMF_CMPL)) {
- host_status = DID_ERROR;
- goto out;
- }
- }
- if ((flags & FCP_SNS_LEN_VAL) && fcp_resp->ext.fr_sns_len) {
- sns_len = be32_to_cpu(fcp_resp->ext.fr_sns_len);
- if (sns_len > SCSI_SENSE_BUFFERSIZE)
- sns_len = SCSI_SENSE_BUFFERSIZE;
- memcpy(cmnd->sense_buffer,
- &rsp_info->_fr_resvd[0] + rsp_len, sns_len);
- CSIO_INC_STATS(scm, n_autosense);
- }
- scsi_set_resid(cmnd, 0);
- /* Under run */
- if (flags & FCP_RESID_UNDER) {
- scsi_set_resid(cmnd,
- be32_to_cpu(fcp_resp->ext.fr_resid));
- if (!(flags & FCP_SNS_LEN_VAL) &&
- (scsi_status == SAM_STAT_GOOD) &&
- ((scsi_bufflen(cmnd) - scsi_get_resid(cmnd))
- < cmnd->underflow))
- host_status = DID_ERROR;
- } else if (flags & FCP_RESID_OVER)
- host_status = DID_ERROR;
- CSIO_INC_STATS(scm, n_rsperror);
- break;
- case FW_SCSI_OVER_FLOW_ERR:
- csio_warn(hw,
- "Over-flow error,cmnd:0x%x expected len:0x%x"
- " resid:0x%x\n", cmnd->cmnd[0],
- scsi_bufflen(cmnd), scsi_get_resid(cmnd));
- host_status = DID_ERROR;
- CSIO_INC_STATS(scm, n_ovflerror);
- break;
- case FW_SCSI_UNDER_FLOW_ERR:
- csio_warn(hw,
- "Under-flow error,cmnd:0x%x expected"
- " len:0x%x resid:0x%x lun:0x%llx ssn:0x%x\n",
- cmnd->cmnd[0], scsi_bufflen(cmnd),
- scsi_get_resid(cmnd), cmnd->device->lun,
- rn->flowid);
- host_status = DID_ERROR;
- CSIO_INC_STATS(scm, n_unflerror);
- break;
- case FW_SCSI_ABORT_REQUESTED:
- case FW_SCSI_ABORTED:
- case FW_SCSI_CLOSE_REQUESTED:
- csio_dbg(hw, "Req %p cmd:%p op:%x %s\n", req, cmnd,
- cmnd->cmnd[0],
- (req->wr_status == FW_SCSI_CLOSE_REQUESTED) ?
- "closed" : "aborted");
- /*
- * csio_eh_abort_handler checks this value to
- * succeed or fail the abort request.
- */
- host_status = DID_REQUEUE;
- if (req->wr_status == FW_SCSI_CLOSE_REQUESTED)
- CSIO_INC_STATS(scm, n_closed);
- else
- CSIO_INC_STATS(scm, n_aborted);
- break;
- case FW_SCSI_ABORT_TIMEDOUT:
- /* FW timed out the abort itself */
- csio_dbg(hw, "FW timed out abort req:%p cmnd:%p status:%x\n",
- req, cmnd, req->wr_status);
- host_status = DID_ERROR;
- CSIO_INC_STATS(scm, n_abrt_timedout);
- break;
- case FW_RDEV_NOT_READY:
- /*
- * In firmware, a RDEV can get into this state
- * temporarily, before moving into dissapeared/lost
- * state. So, the driver should complete the request equivalent
- * to device-disappeared!
- */
- CSIO_INC_STATS(scm, n_rdev_nr_error);
- host_status = DID_ERROR;
- break;
- case FW_ERR_RDEV_LOST:
- CSIO_INC_STATS(scm, n_rdev_lost_error);
- host_status = DID_ERROR;
- break;
- case FW_ERR_RDEV_LOGO:
- CSIO_INC_STATS(scm, n_rdev_logo_error);
- host_status = DID_ERROR;
- break;
- case FW_ERR_RDEV_IMPL_LOGO:
- host_status = DID_ERROR;
- break;
- case FW_ERR_LINK_DOWN:
- CSIO_INC_STATS(scm, n_link_down_error);
- host_status = DID_ERROR;
- break;
- case FW_FCOE_NO_XCHG:
- CSIO_INC_STATS(scm, n_no_xchg_error);
- host_status = DID_ERROR;
- break;
- default:
- csio_err(hw, "Unknown SCSI FW WR status:%d req:%p cmnd:%p\n",
- req->wr_status, req, cmnd);
- CSIO_DB_ASSERT(0);
- CSIO_INC_STATS(scm, n_unknown_error);
- host_status = DID_ERROR;
- break;
- }
- out:
- if (req->nsge > 0)
- scsi_dma_unmap(cmnd);
- cmnd->result = (((host_status) << 16) | scsi_status);
- cmnd->scsi_done(cmnd);
- /* Wake up waiting threads */
- csio_scsi_cmnd(req) = NULL;
- complete(&req->cmplobj);
- }
- /*
- * csio_scsi_cbfn - SCSI callback function.
- * @hw: HW module.
- * @req: IO request.
- *
- */
- static void
- csio_scsi_cbfn(struct csio_hw *hw, struct csio_ioreq *req)
- {
- struct scsi_cmnd *cmnd = (struct scsi_cmnd *)csio_scsi_cmnd(req);
- uint8_t scsi_status = SAM_STAT_GOOD;
- uint32_t host_status = DID_OK;
- if (likely(req->wr_status == FW_SUCCESS)) {
- if (req->nsge > 0) {
- scsi_dma_unmap(cmnd);
- if (req->dcopy)
- host_status = csio_scsi_copy_to_sgl(hw, req);
- }
- cmnd->result = (((host_status) << 16) | scsi_status);
- cmnd->scsi_done(cmnd);
- csio_scsi_cmnd(req) = NULL;
- CSIO_INC_STATS(csio_hw_to_scsim(hw), n_tot_success);
- } else {
- /* Error handling */
- csio_scsi_err_handler(hw, req);
- }
- }
- /**
- * csio_queuecommand - Entry point to kickstart an I/O request.
- * @host: The scsi_host pointer.
- * @cmnd: The I/O request from ML.
- *
- * This routine does the following:
- * - Checks for HW and Rnode module readiness.
- * - Gets a free ioreq structure (which is already initialized
- * to uninit during its allocation).
- * - Maps SG elements.
- * - Initializes ioreq members.
- * - Kicks off the SCSI state machine for this IO.
- * - Returns busy status on error.
- */
- static int
- csio_queuecommand(struct Scsi_Host *host, struct scsi_cmnd *cmnd)
- {
- struct csio_lnode *ln = shost_priv(host);
- struct csio_hw *hw = csio_lnode_to_hw(ln);
- struct csio_scsim *scsim = csio_hw_to_scsim(hw);
- struct csio_rnode *rn = (struct csio_rnode *)(cmnd->device->hostdata);
- struct csio_ioreq *ioreq = NULL;
- unsigned long flags;
- int nsge = 0;
- int rv = SCSI_MLQUEUE_HOST_BUSY, nr;
- int retval;
- int cpu;
- struct csio_scsi_qset *sqset;
- struct fc_rport *rport = starget_to_rport(scsi_target(cmnd->device));
- if (!blk_rq_cpu_valid(cmnd->request))
- cpu = smp_processor_id();
- else
- cpu = cmnd->request->cpu;
- sqset = &hw->sqset[ln->portid][cpu];
- nr = fc_remote_port_chkready(rport);
- if (nr) {
- cmnd->result = nr;
- CSIO_INC_STATS(scsim, n_rn_nr_error);
- goto err_done;
- }
- if (unlikely(!csio_is_hw_ready(hw))) {
- cmnd->result = (DID_REQUEUE << 16);
- CSIO_INC_STATS(scsim, n_hw_nr_error);
- goto err_done;
- }
- /* Get req->nsge, if there are SG elements to be mapped */
- nsge = scsi_dma_map(cmnd);
- if (unlikely(nsge < 0)) {
- CSIO_INC_STATS(scsim, n_dmamap_error);
- goto err;
- }
- /* Do we support so many mappings? */
- if (unlikely(nsge > scsim->max_sge)) {
- csio_warn(hw,
- "More SGEs than can be supported."
- " SGEs: %d, Max SGEs: %d\n", nsge, scsim->max_sge);
- CSIO_INC_STATS(scsim, n_unsupp_sge_error);
- goto err_dma_unmap;
- }
- /* Get a free ioreq structure - SM is already set to uninit */
- ioreq = csio_get_scsi_ioreq_lock(hw, scsim);
- if (!ioreq) {
- csio_err(hw, "Out of I/O request elements. Active #:%d\n",
- scsim->stats.n_active);
- CSIO_INC_STATS(scsim, n_no_req_error);
- goto err_dma_unmap;
- }
- ioreq->nsge = nsge;
- ioreq->lnode = ln;
- ioreq->rnode = rn;
- ioreq->iq_idx = sqset->iq_idx;
- ioreq->eq_idx = sqset->eq_idx;
- ioreq->wr_status = 0;
- ioreq->drv_status = 0;
- csio_scsi_cmnd(ioreq) = (void *)cmnd;
- ioreq->tmo = 0;
- ioreq->datadir = cmnd->sc_data_direction;
- if (cmnd->sc_data_direction == DMA_TO_DEVICE) {
- CSIO_INC_STATS(ln, n_output_requests);
- ln->stats.n_output_bytes += scsi_bufflen(cmnd);
- } else if (cmnd->sc_data_direction == DMA_FROM_DEVICE) {
- CSIO_INC_STATS(ln, n_input_requests);
- ln->stats.n_input_bytes += scsi_bufflen(cmnd);
- } else
- CSIO_INC_STATS(ln, n_control_requests);
- /* Set cbfn */
- ioreq->io_cbfn = csio_scsi_cbfn;
- /* Needed during abort */
- cmnd->host_scribble = (unsigned char *)ioreq;
- cmnd->SCp.Message = 0;
- /* Kick off SCSI IO SM on the ioreq */
- spin_lock_irqsave(&hw->lock, flags);
- retval = csio_scsi_start_io(ioreq);
- spin_unlock_irqrestore(&hw->lock, flags);
- if (retval != 0) {
- csio_err(hw, "ioreq: %p couldnt be started, status:%d\n",
- ioreq, retval);
- CSIO_INC_STATS(scsim, n_busy_error);
- goto err_put_req;
- }
- return 0;
- err_put_req:
- csio_put_scsi_ioreq_lock(hw, scsim, ioreq);
- err_dma_unmap:
- if (nsge > 0)
- scsi_dma_unmap(cmnd);
- err:
- return rv;
- err_done:
- cmnd->scsi_done(cmnd);
- return 0;
- }
- static int
- csio_do_abrt_cls(struct csio_hw *hw, struct csio_ioreq *ioreq, bool abort)
- {
- int rv;
- int cpu = smp_processor_id();
- struct csio_lnode *ln = ioreq->lnode;
- struct csio_scsi_qset *sqset = &hw->sqset[ln->portid][cpu];
- ioreq->tmo = CSIO_SCSI_ABRT_TMO_MS;
- /*
- * Use current processor queue for posting the abort/close, but retain
- * the ingress queue ID of the original I/O being aborted/closed - we
- * need the abort/close completion to be received on the same queue
- * as the original I/O.
- */
- ioreq->eq_idx = sqset->eq_idx;
- if (abort == SCSI_ABORT)
- rv = csio_scsi_abort(ioreq);
- else
- rv = csio_scsi_close(ioreq);
- return rv;
- }
- static int
- csio_eh_abort_handler(struct scsi_cmnd *cmnd)
- {
- struct csio_ioreq *ioreq;
- struct csio_lnode *ln = shost_priv(cmnd->device->host);
- struct csio_hw *hw = csio_lnode_to_hw(ln);
- struct csio_scsim *scsim = csio_hw_to_scsim(hw);
- int ready = 0, ret;
- unsigned long tmo = 0;
- int rv;
- struct csio_rnode *rn = (struct csio_rnode *)(cmnd->device->hostdata);
- ret = fc_block_scsi_eh(cmnd);
- if (ret)
- return ret;
- ioreq = (struct csio_ioreq *)cmnd->host_scribble;
- if (!ioreq)
- return SUCCESS;
- if (!rn)
- return FAILED;
- csio_dbg(hw,
- "Request to abort ioreq:%p cmd:%p cdb:%08llx"
- " ssni:0x%x lun:%llu iq:0x%x\n",
- ioreq, cmnd, *((uint64_t *)cmnd->cmnd), rn->flowid,
- cmnd->device->lun, csio_q_physiqid(hw, ioreq->iq_idx));
- if (((struct scsi_cmnd *)csio_scsi_cmnd(ioreq)) != cmnd) {
- CSIO_INC_STATS(scsim, n_abrt_race_comp);
- return SUCCESS;
- }
- ready = csio_is_lnode_ready(ln);
- tmo = CSIO_SCSI_ABRT_TMO_MS;
- reinit_completion(&ioreq->cmplobj);
- spin_lock_irq(&hw->lock);
- rv = csio_do_abrt_cls(hw, ioreq, (ready ? SCSI_ABORT : SCSI_CLOSE));
- spin_unlock_irq(&hw->lock);
- if (rv != 0) {
- if (rv == -EINVAL) {
- /* Return success, if abort/close request issued on
- * already completed IO
- */
- return SUCCESS;
- }
- if (ready)
- CSIO_INC_STATS(scsim, n_abrt_busy_error);
- else
- CSIO_INC_STATS(scsim, n_cls_busy_error);
- goto inval_scmnd;
- }
- wait_for_completion_timeout(&ioreq->cmplobj, msecs_to_jiffies(tmo));
- /* FW didnt respond to abort within our timeout */
- if (((struct scsi_cmnd *)csio_scsi_cmnd(ioreq)) == cmnd) {
- csio_err(hw, "Abort timed out -- req: %p\n", ioreq);
- CSIO_INC_STATS(scsim, n_abrt_timedout);
- inval_scmnd:
- if (ioreq->nsge > 0)
- scsi_dma_unmap(cmnd);
- spin_lock_irq(&hw->lock);
- csio_scsi_cmnd(ioreq) = NULL;
- spin_unlock_irq(&hw->lock);
- cmnd->result = (DID_ERROR << 16);
- cmnd->scsi_done(cmnd);
- return FAILED;
- }
- /* FW successfully aborted the request */
- if (host_byte(cmnd->result) == DID_REQUEUE) {
- csio_info(hw,
- "Aborted SCSI command to (%d:%llu) serial#:0x%lx\n",
- cmnd->device->id, cmnd->device->lun,
- cmnd->serial_number);
- return SUCCESS;
- } else {
- csio_info(hw,
- "Failed to abort SCSI command, (%d:%llu) serial#:0x%lx\n",
- cmnd->device->id, cmnd->device->lun,
- cmnd->serial_number);
- return FAILED;
- }
- }
- /*
- * csio_tm_cbfn - TM callback function.
- * @hw: HW module.
- * @req: IO request.
- *
- * Cache the result in 'cmnd', since ioreq will be freed soon
- * after we return from here, and the waiting thread shouldnt trust
- * the ioreq contents.
- */
- static void
- csio_tm_cbfn(struct csio_hw *hw, struct csio_ioreq *req)
- {
- struct scsi_cmnd *cmnd = (struct scsi_cmnd *)csio_scsi_cmnd(req);
- struct csio_dma_buf *dma_buf;
- uint8_t flags = 0;
- struct fcp_resp_with_ext *fcp_resp;
- struct fcp_resp_rsp_info *rsp_info;
- csio_dbg(hw, "req: %p in csio_tm_cbfn status: %d\n",
- req, req->wr_status);
- /* Cache FW return status */
- cmnd->SCp.Status = req->wr_status;
- /* Special handling based on FCP response */
- /*
- * FW returns us this error, if flags were set. FCP4 says
- * FCP_RSP_LEN_VAL in flags shall be set for TM completions.
- * So if a target were to set this bit, we expect that the
- * rsp_code is set to FCP_TMF_CMPL for a successful TM
- * completion. Any other rsp_code means TM operation failed.
- * If a target were to just ignore setting flags, we treat
- * the TM operation as success, and FW returns FW_SUCCESS.
- */
- if (req->wr_status == FW_SCSI_RSP_ERR) {
- dma_buf = &req->dma_buf;
- fcp_resp = (struct fcp_resp_with_ext *)dma_buf->vaddr;
- rsp_info = (struct fcp_resp_rsp_info *)(fcp_resp + 1);
- flags = fcp_resp->resp.fr_flags;
- /* Modify return status if flags indicate success */
- if (flags & FCP_RSP_LEN_VAL)
- if (rsp_info->rsp_code == FCP_TMF_CMPL)
- cmnd->SCp.Status = FW_SUCCESS;
- csio_dbg(hw, "TM FCP rsp code: %d\n", rsp_info->rsp_code);
- }
- /* Wake up the TM handler thread */
- csio_scsi_cmnd(req) = NULL;
- }
- static int
- csio_eh_lun_reset_handler(struct scsi_cmnd *cmnd)
- {
- struct csio_lnode *ln = shost_priv(cmnd->device->host);
- struct csio_hw *hw = csio_lnode_to_hw(ln);
- struct csio_scsim *scsim = csio_hw_to_scsim(hw);
- struct csio_rnode *rn = (struct csio_rnode *)(cmnd->device->hostdata);
- struct csio_ioreq *ioreq = NULL;
- struct csio_scsi_qset *sqset;
- unsigned long flags;
- int retval;
- int count, ret;
- LIST_HEAD(local_q);
- struct csio_scsi_level_data sld;
- if (!rn)
- goto fail;
- csio_dbg(hw, "Request to reset LUN:%llu (ssni:0x%x tgtid:%d)\n",
- cmnd->device->lun, rn->flowid, rn->scsi_id);
- if (!csio_is_lnode_ready(ln)) {
- csio_err(hw,
- "LUN reset cannot be issued on non-ready"
- " local node vnpi:0x%x (LUN:%llu)\n",
- ln->vnp_flowid, cmnd->device->lun);
- goto fail;
- }
- /* Lnode is ready, now wait on rport node readiness */
- ret = fc_block_scsi_eh(cmnd);
- if (ret)
- return ret;
- /*
- * If we have blocked in the previous call, at this point, either the
- * remote node has come back online, or device loss timer has fired
- * and the remote node is destroyed. Allow the LUN reset only for
- * the former case, since LUN reset is a TMF I/O on the wire, and we
- * need a valid session to issue it.
- */
- if (fc_remote_port_chkready(rn->rport)) {
- csio_err(hw,
- "LUN reset cannot be issued on non-ready"
- " remote node ssni:0x%x (LUN:%llu)\n",
- rn->flowid, cmnd->device->lun);
- goto fail;
- }
- /* Get a free ioreq structure - SM is already set to uninit */
- ioreq = csio_get_scsi_ioreq_lock(hw, scsim);
- if (!ioreq) {
- csio_err(hw, "Out of IO request elements. Active # :%d\n",
- scsim->stats.n_active);
- goto fail;
- }
- sqset = &hw->sqset[ln->portid][smp_processor_id()];
- ioreq->nsge = 0;
- ioreq->lnode = ln;
- ioreq->rnode = rn;
- ioreq->iq_idx = sqset->iq_idx;
- ioreq->eq_idx = sqset->eq_idx;
- csio_scsi_cmnd(ioreq) = cmnd;
- cmnd->host_scribble = (unsigned char *)ioreq;
- cmnd->SCp.Status = 0;
- cmnd->SCp.Message = FCP_TMF_LUN_RESET;
- ioreq->tmo = CSIO_SCSI_LUNRST_TMO_MS / 1000;
- /*
- * FW times the LUN reset for ioreq->tmo, so we got to wait a little
- * longer (10s for now) than that to allow FW to return the timed
- * out command.
- */
- count = DIV_ROUND_UP((ioreq->tmo + 10) * 1000, CSIO_SCSI_TM_POLL_MS);
- /* Set cbfn */
- ioreq->io_cbfn = csio_tm_cbfn;
- /* Save of the ioreq info for later use */
- sld.level = CSIO_LEV_LUN;
- sld.lnode = ioreq->lnode;
- sld.rnode = ioreq->rnode;
- sld.oslun = cmnd->device->lun;
- spin_lock_irqsave(&hw->lock, flags);
- /* Kick off TM SM on the ioreq */
- retval = csio_scsi_start_tm(ioreq);
- spin_unlock_irqrestore(&hw->lock, flags);
- if (retval != 0) {
- csio_err(hw, "Failed to issue LUN reset, req:%p, status:%d\n",
- ioreq, retval);
- goto fail_ret_ioreq;
- }
- csio_dbg(hw, "Waiting max %d secs for LUN reset completion\n",
- count * (CSIO_SCSI_TM_POLL_MS / 1000));
- /* Wait for completion */
- while ((((struct scsi_cmnd *)csio_scsi_cmnd(ioreq)) == cmnd)
- && count--)
- msleep(CSIO_SCSI_TM_POLL_MS);
- /* LUN reset timed-out */
- if (((struct scsi_cmnd *)csio_scsi_cmnd(ioreq)) == cmnd) {
- csio_err(hw, "LUN reset (%d:%llu) timed out\n",
- cmnd->device->id, cmnd->device->lun);
- spin_lock_irq(&hw->lock);
- csio_scsi_drvcleanup(ioreq);
- list_del_init(&ioreq->sm.sm_list);
- spin_unlock_irq(&hw->lock);
- goto fail_ret_ioreq;
- }
- /* LUN reset returned, check cached status */
- if (cmnd->SCp.Status != FW_SUCCESS) {
- csio_err(hw, "LUN reset failed (%d:%llu), status: %d\n",
- cmnd->device->id, cmnd->device->lun, cmnd->SCp.Status);
- goto fail;
- }
- /* LUN reset succeeded, Start aborting affected I/Os */
- /*
- * Since the host guarantees during LUN reset that there
- * will not be any more I/Os to that LUN, until the LUN reset
- * completes, we gather pending I/Os after the LUN reset.
- */
- spin_lock_irq(&hw->lock);
- csio_scsi_gather_active_ios(scsim, &sld, &local_q);
- retval = csio_scsi_abort_io_q(scsim, &local_q, 30000);
- spin_unlock_irq(&hw->lock);
- /* Aborts may have timed out */
- if (retval != 0) {
- csio_err(hw,
- "Attempt to abort I/Os during LUN reset of %llu"
- " returned %d\n", cmnd->device->lun, retval);
- /* Return I/Os back to active_q */
- spin_lock_irq(&hw->lock);
- list_splice_tail_init(&local_q, &scsim->active_q);
- spin_unlock_irq(&hw->lock);
- goto fail;
- }
- CSIO_INC_STATS(rn, n_lun_rst);
- csio_info(hw, "LUN reset occurred (%d:%llu)\n",
- cmnd->device->id, cmnd->device->lun);
- return SUCCESS;
- fail_ret_ioreq:
- csio_put_scsi_ioreq_lock(hw, scsim, ioreq);
- fail:
- CSIO_INC_STATS(rn, n_lun_rst_fail);
- return FAILED;
- }
- static int
- csio_slave_alloc(struct scsi_device *sdev)
- {
- struct fc_rport *rport = starget_to_rport(scsi_target(sdev));
- if (!rport || fc_remote_port_chkready(rport))
- return -ENXIO;
- sdev->hostdata = *((struct csio_lnode **)(rport->dd_data));
- return 0;
- }
- static int
- csio_slave_configure(struct scsi_device *sdev)
- {
- scsi_change_queue_depth(sdev, csio_lun_qdepth);
- return 0;
- }
- static void
- csio_slave_destroy(struct scsi_device *sdev)
- {
- sdev->hostdata = NULL;
- }
- static int
- csio_scan_finished(struct Scsi_Host *shost, unsigned long time)
- {
- struct csio_lnode *ln = shost_priv(shost);
- int rv = 1;
- spin_lock_irq(shost->host_lock);
- if (!ln->hwp || csio_list_deleted(&ln->sm.sm_list))
- goto out;
- rv = csio_scan_done(ln, jiffies, time, csio_max_scan_tmo * HZ,
- csio_delta_scan_tmo * HZ);
- out:
- spin_unlock_irq(shost->host_lock);
- return rv;
- }
- struct scsi_host_template csio_fcoe_shost_template = {
- .module = THIS_MODULE,
- .name = CSIO_DRV_DESC,
- .proc_name = KBUILD_MODNAME,
- .queuecommand = csio_queuecommand,
- .eh_abort_handler = csio_eh_abort_handler,
- .eh_device_reset_handler = csio_eh_lun_reset_handler,
- .slave_alloc = csio_slave_alloc,
- .slave_configure = csio_slave_configure,
- .slave_destroy = csio_slave_destroy,
- .scan_finished = csio_scan_finished,
- .this_id = -1,
- .sg_tablesize = CSIO_SCSI_MAX_SGE,
- .cmd_per_lun = CSIO_MAX_CMD_PER_LUN,
- .use_clustering = ENABLE_CLUSTERING,
- .shost_attrs = csio_fcoe_lport_attrs,
- .max_sectors = CSIO_MAX_SECTOR_SIZE,
- };
- struct scsi_host_template csio_fcoe_shost_vport_template = {
- .module = THIS_MODULE,
- .name = CSIO_DRV_DESC,
- .proc_name = KBUILD_MODNAME,
- .queuecommand = csio_queuecommand,
- .eh_abort_handler = csio_eh_abort_handler,
- .eh_device_reset_handler = csio_eh_lun_reset_handler,
- .slave_alloc = csio_slave_alloc,
- .slave_configure = csio_slave_configure,
- .slave_destroy = csio_slave_destroy,
- .scan_finished = csio_scan_finished,
- .this_id = -1,
- .sg_tablesize = CSIO_SCSI_MAX_SGE,
- .cmd_per_lun = CSIO_MAX_CMD_PER_LUN,
- .use_clustering = ENABLE_CLUSTERING,
- .shost_attrs = csio_fcoe_vport_attrs,
- .max_sectors = CSIO_MAX_SECTOR_SIZE,
- };
- /*
- * csio_scsi_alloc_ddp_bufs - Allocate buffers for DDP of unaligned SGLs.
- * @scm: SCSI Module
- * @hw: HW device.
- * @buf_size: buffer size
- * @num_buf : Number of buffers.
- *
- * This routine allocates DMA buffers required for SCSI Data xfer, if
- * each SGL buffer for a SCSI Read request posted by SCSI midlayer are
- * not virtually contiguous.
- */
- static int
- csio_scsi_alloc_ddp_bufs(struct csio_scsim *scm, struct csio_hw *hw,
- int buf_size, int num_buf)
- {
- int n = 0;
- struct list_head *tmp;
- struct csio_dma_buf *ddp_desc = NULL;
- uint32_t unit_size = 0;
- if (!num_buf)
- return 0;
- if (!buf_size)
- return -EINVAL;
- INIT_LIST_HEAD(&scm->ddp_freelist);
- /* Align buf size to page size */
- buf_size = (buf_size + PAGE_SIZE - 1) & PAGE_MASK;
- /* Initialize dma descriptors */
- for (n = 0; n < num_buf; n++) {
- /* Set unit size to request size */
- unit_size = buf_size;
- ddp_desc = kzalloc(sizeof(struct csio_dma_buf), GFP_KERNEL);
- if (!ddp_desc) {
- csio_err(hw,
- "Failed to allocate ddp descriptors,"
- " Num allocated = %d.\n",
- scm->stats.n_free_ddp);
- goto no_mem;
- }
- /* Allocate Dma buffers for DDP */
- ddp_desc->vaddr = pci_alloc_consistent(hw->pdev, unit_size,
- &ddp_desc->paddr);
- if (!ddp_desc->vaddr) {
- csio_err(hw,
- "SCSI response DMA buffer (ddp) allocation"
- " failed!\n");
- kfree(ddp_desc);
- goto no_mem;
- }
- ddp_desc->len = unit_size;
- /* Added it to scsi ddp freelist */
- list_add_tail(&ddp_desc->list, &scm->ddp_freelist);
- CSIO_INC_STATS(scm, n_free_ddp);
- }
- return 0;
- no_mem:
- /* release dma descs back to freelist and free dma memory */
- list_for_each(tmp, &scm->ddp_freelist) {
- ddp_desc = (struct csio_dma_buf *) tmp;
- tmp = csio_list_prev(tmp);
- pci_free_consistent(hw->pdev, ddp_desc->len, ddp_desc->vaddr,
- ddp_desc->paddr);
- list_del_init(&ddp_desc->list);
- kfree(ddp_desc);
- }
- scm->stats.n_free_ddp = 0;
- return -ENOMEM;
- }
- /*
- * csio_scsi_free_ddp_bufs - free DDP buffers of unaligned SGLs.
- * @scm: SCSI Module
- * @hw: HW device.
- *
- * This routine frees ddp buffers.
- */
- static void
- csio_scsi_free_ddp_bufs(struct csio_scsim *scm, struct csio_hw *hw)
- {
- struct list_head *tmp;
- struct csio_dma_buf *ddp_desc;
- /* release dma descs back to freelist and free dma memory */
- list_for_each(tmp, &scm->ddp_freelist) {
- ddp_desc = (struct csio_dma_buf *) tmp;
- tmp = csio_list_prev(tmp);
- pci_free_consistent(hw->pdev, ddp_desc->len, ddp_desc->vaddr,
- ddp_desc->paddr);
- list_del_init(&ddp_desc->list);
- kfree(ddp_desc);
- }
- scm->stats.n_free_ddp = 0;
- }
- /**
- * csio_scsim_init - Initialize SCSI Module
- * @scm: SCSI Module
- * @hw: HW module
- *
- */
- int
- csio_scsim_init(struct csio_scsim *scm, struct csio_hw *hw)
- {
- int i;
- struct csio_ioreq *ioreq;
- struct csio_dma_buf *dma_buf;
- INIT_LIST_HEAD(&scm->active_q);
- scm->hw = hw;
- scm->proto_cmd_len = sizeof(struct fcp_cmnd);
- scm->proto_rsp_len = CSIO_SCSI_RSP_LEN;
- scm->max_sge = CSIO_SCSI_MAX_SGE;
- spin_lock_init(&scm->freelist_lock);
- /* Pre-allocate ioreqs and initialize them */
- INIT_LIST_HEAD(&scm->ioreq_freelist);
- for (i = 0; i < csio_scsi_ioreqs; i++) {
- ioreq = kzalloc(sizeof(struct csio_ioreq), GFP_KERNEL);
- if (!ioreq) {
- csio_err(hw,
- "I/O request element allocation failed, "
- " Num allocated = %d.\n",
- scm->stats.n_free_ioreq);
- goto free_ioreq;
- }
- /* Allocate Dma buffers for Response Payload */
- dma_buf = &ioreq->dma_buf;
- dma_buf->vaddr = pci_pool_alloc(hw->scsi_pci_pool, GFP_KERNEL,
- &dma_buf->paddr);
- if (!dma_buf->vaddr) {
- csio_err(hw,
- "SCSI response DMA buffer allocation"
- " failed!\n");
- kfree(ioreq);
- goto free_ioreq;
- }
- dma_buf->len = scm->proto_rsp_len;
- /* Set state to uninit */
- csio_init_state(&ioreq->sm, csio_scsis_uninit);
- INIT_LIST_HEAD(&ioreq->gen_list);
- init_completion(&ioreq->cmplobj);
- list_add_tail(&ioreq->sm.sm_list, &scm->ioreq_freelist);
- CSIO_INC_STATS(scm, n_free_ioreq);
- }
- if (csio_scsi_alloc_ddp_bufs(scm, hw, PAGE_SIZE, csio_ddp_descs))
- goto free_ioreq;
- return 0;
- free_ioreq:
- /*
- * Free up existing allocations, since an error
- * from here means we are returning for good
- */
- while (!list_empty(&scm->ioreq_freelist)) {
- struct csio_sm *tmp;
- tmp = list_first_entry(&scm->ioreq_freelist,
- struct csio_sm, sm_list);
- list_del_init(&tmp->sm_list);
- ioreq = (struct csio_ioreq *)tmp;
- dma_buf = &ioreq->dma_buf;
- pci_pool_free(hw->scsi_pci_pool, dma_buf->vaddr,
- dma_buf->paddr);
- kfree(ioreq);
- }
- scm->stats.n_free_ioreq = 0;
- return -ENOMEM;
- }
- /**
- * csio_scsim_exit: Uninitialize SCSI Module
- * @scm: SCSI Module
- *
- */
- void
- csio_scsim_exit(struct csio_scsim *scm)
- {
- struct csio_ioreq *ioreq;
- struct csio_dma_buf *dma_buf;
- while (!list_empty(&scm->ioreq_freelist)) {
- struct csio_sm *tmp;
- tmp = list_first_entry(&scm->ioreq_freelist,
- struct csio_sm, sm_list);
- list_del_init(&tmp->sm_list);
- ioreq = (struct csio_ioreq *)tmp;
- dma_buf = &ioreq->dma_buf;
- pci_pool_free(scm->hw->scsi_pci_pool, dma_buf->vaddr,
- dma_buf->paddr);
- kfree(ioreq);
- }
- scm->stats.n_free_ioreq = 0;
- csio_scsi_free_ddp_bufs(scm, scm->hw);
- }
|