mpi3mr_app.c 97 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283848586878889909192939495969798991001011021031041051061071081091101111121131141151161171181191201211221231241251261271281291301311321331341351361371381391401411421431441451461471481491501511521531541551561571581591601611621631641651661671681691701711721731741751761771781791801811821831841851861871881891901911921931941951961971981992002012022032042052062072082092102112122132142152162172182192202212222232242252262272282292302312322332342352362372382392402412422432442452462472482492502512522532542552562572582592602612622632642652662672682692702712722732742752762772782792802812822832842852862872882892902912922932942952962972982993003013023033043053063073083093103113123133143153163173183193203213223233243253263273283293303313323333343353363373383393403413423433443453463473483493503513523533543553563573583593603613623633643653663673683693703713723733743753763773783793803813823833843853863873883893903913923933943953963973983994004014024034044054064074084094104114124134144154164174184194204214224234244254264274284294304314324334344354364374384394404414424434444454464474484494504514524534544554564574584594604614624634644654664674684694704714724734744754764774784794804814824834844854864874884894904914924934944954964974984995005015025035045055065075085095105115125135145155165175185195205215225235245255265275285295305315325335345355365375385395405415425435445455465475485495505515525535545555565575585595605615625635645655665675685695705715725735745755765775785795805815825835845855865875885895905915925935945955965975985996006016026036046056066076086096106116126136146156166176186196206216226236246256266276286296306316326336346356366376386396406416426436446456466476486496506516526536546556566576586596606616626636646656666676686696706716726736746756766776786796806816826836846856866876886896906916926936946956966976986997007017027037047057067077087097107117127137147157167177187197207217227237247257267277287297307317327337347357367377387397407417427437447457467477487497507517527537547557567577587597607617627637647657667677687697707717727737747757767777787797807817827837847857867877887897907917927937947957967977987998008018028038048058068078088098108118128138148158168178188198208218228238248258268278288298308318328338348358368378388398408418428438448458468478488498508518528538548558568578588598608618628638648658668678688698708718728738748758768778788798808818828838848858868878888898908918928938948958968978988999009019029039049059069079089099109119129139149159169179189199209219229239249259269279289299309319329339349359369379389399409419429439449459469479489499509519529539549559569579589599609619629639649659669679689699709719729739749759769779789799809819829839849859869879889899909919929939949959969979989991000100110021003100410051006100710081009101010111012101310141015101610171018101910201021102210231024102510261027102810291030103110321033103410351036103710381039104010411042104310441045104610471048104910501051105210531054105510561057105810591060106110621063106410651066106710681069107010711072107310741075107610771078107910801081108210831084108510861087108810891090109110921093109410951096109710981099110011011102110311041105110611071108110911101111111211131114111511161117111811191120112111221123112411251126112711281129113011311132113311341135113611371138113911401141114211431144114511461147114811491150115111521153115411551156115711581159116011611162116311641165116611671168116911701171117211731174117511761177117811791180118111821183118411851186118711881189119011911192119311941195119611971198119912001201120212031204120512061207120812091210121112121213121412151216121712181219122012211222122312241225122612271228122912301231123212331234123512361237123812391240124112421243124412451246124712481249125012511252125312541255125612571258125912601261126212631264126512661267126812691270127112721273127412751276127712781279128012811282128312841285128612871288128912901291129212931294129512961297129812991300130113021303130413051306130713081309131013111312131313141315131613171318131913201321132213231324132513261327132813291330133113321333133413351336133713381339134013411342134313441345134613471348134913501351135213531354135513561357135813591360136113621363136413651366136713681369137013711372137313741375137613771378137913801381138213831384138513861387138813891390139113921393139413951396139713981399140014011402140314041405140614071408140914101411141214131414141514161417141814191420142114221423142414251426142714281429143014311432143314341435143614371438143914401441144214431444144514461447144814491450145114521453145414551456145714581459146014611462146314641465146614671468146914701471147214731474147514761477147814791480148114821483148414851486148714881489149014911492149314941495149614971498149915001501150215031504150515061507150815091510151115121513151415151516151715181519152015211522152315241525152615271528152915301531153215331534153515361537153815391540154115421543154415451546154715481549155015511552155315541555155615571558155915601561156215631564156515661567156815691570157115721573157415751576157715781579158015811582158315841585158615871588158915901591159215931594159515961597159815991600160116021603160416051606160716081609161016111612161316141615161616171618161916201621162216231624162516261627162816291630163116321633163416351636163716381639164016411642164316441645164616471648164916501651165216531654165516561657165816591660166116621663166416651666166716681669167016711672167316741675167616771678167916801681168216831684168516861687168816891690169116921693169416951696169716981699170017011702170317041705170617071708170917101711171217131714171517161717171817191720172117221723172417251726172717281729173017311732173317341735173617371738173917401741174217431744174517461747174817491750175117521753175417551756175717581759176017611762176317641765176617671768176917701771177217731774177517761777177817791780178117821783178417851786178717881789179017911792179317941795179617971798179918001801180218031804180518061807180818091810181118121813181418151816181718181819182018211822182318241825182618271828182918301831183218331834183518361837183818391840184118421843184418451846184718481849185018511852185318541855185618571858185918601861186218631864186518661867186818691870187118721873187418751876187718781879188018811882188318841885188618871888188918901891189218931894189518961897189818991900190119021903190419051906190719081909191019111912191319141915191619171918191919201921192219231924192519261927192819291930193119321933193419351936193719381939194019411942194319441945194619471948194919501951195219531954195519561957195819591960196119621963196419651966196719681969197019711972197319741975197619771978197919801981198219831984198519861987198819891990199119921993199419951996199719981999200020012002200320042005200620072008200920102011201220132014201520162017201820192020202120222023202420252026202720282029203020312032203320342035203620372038203920402041204220432044204520462047204820492050205120522053205420552056205720582059206020612062206320642065206620672068206920702071207220732074207520762077207820792080208120822083208420852086208720882089209020912092209320942095209620972098209921002101210221032104210521062107210821092110211121122113211421152116211721182119212021212122212321242125212621272128212921302131213221332134213521362137213821392140214121422143214421452146214721482149215021512152215321542155215621572158215921602161216221632164216521662167216821692170217121722173217421752176217721782179218021812182218321842185218621872188218921902191219221932194219521962197219821992200220122022203220422052206220722082209221022112212221322142215221622172218221922202221222222232224222522262227222822292230223122322233223422352236223722382239224022412242224322442245224622472248224922502251225222532254225522562257225822592260226122622263226422652266226722682269227022712272227322742275227622772278227922802281228222832284228522862287228822892290229122922293229422952296229722982299230023012302230323042305230623072308230923102311231223132314231523162317231823192320232123222323232423252326232723282329233023312332233323342335233623372338233923402341234223432344234523462347234823492350235123522353235423552356235723582359236023612362236323642365236623672368236923702371237223732374237523762377237823792380238123822383238423852386238723882389239023912392239323942395239623972398239924002401240224032404240524062407240824092410241124122413241424152416241724182419242024212422242324242425242624272428242924302431243224332434243524362437243824392440244124422443244424452446244724482449245024512452245324542455245624572458245924602461246224632464246524662467246824692470247124722473247424752476247724782479248024812482248324842485248624872488248924902491249224932494249524962497249824992500250125022503250425052506250725082509251025112512251325142515251625172518251925202521252225232524252525262527252825292530253125322533253425352536253725382539254025412542254325442545254625472548254925502551255225532554255525562557255825592560256125622563256425652566256725682569257025712572257325742575257625772578257925802581258225832584258525862587258825892590259125922593259425952596259725982599260026012602260326042605260626072608260926102611261226132614261526162617261826192620262126222623262426252626262726282629263026312632263326342635263626372638263926402641264226432644264526462647264826492650265126522653265426552656265726582659266026612662266326642665266626672668266926702671267226732674267526762677267826792680268126822683268426852686268726882689269026912692269326942695269626972698269927002701270227032704270527062707270827092710271127122713271427152716271727182719272027212722272327242725272627272728272927302731273227332734273527362737273827392740274127422743274427452746274727482749275027512752275327542755275627572758275927602761276227632764276527662767276827692770277127722773277427752776277727782779278027812782278327842785278627872788278927902791279227932794279527962797279827992800280128022803280428052806280728082809281028112812281328142815281628172818281928202821282228232824282528262827282828292830283128322833283428352836283728382839284028412842284328442845284628472848284928502851285228532854285528562857285828592860286128622863286428652866286728682869287028712872287328742875287628772878287928802881288228832884288528862887288828892890289128922893289428952896289728982899290029012902290329042905290629072908290929102911291229132914291529162917291829192920292129222923292429252926292729282929293029312932293329342935293629372938293929402941294229432944294529462947294829492950295129522953295429552956295729582959296029612962296329642965296629672968296929702971297229732974297529762977297829792980298129822983298429852986298729882989299029912992299329942995299629972998299930003001300230033004300530063007300830093010301130123013301430153016301730183019302030213022302330243025302630273028302930303031303230333034303530363037303830393040304130423043304430453046304730483049305030513052305330543055305630573058305930603061306230633064306530663067306830693070307130723073307430753076307730783079308030813082308330843085308630873088308930903091309230933094309530963097309830993100310131023103310431053106310731083109311031113112311331143115311631173118311931203121312231233124312531263127312831293130313131323133313431353136313731383139314031413142314331443145314631473148314931503151315231533154315531563157315831593160316131623163316431653166316731683169317031713172317331743175317631773178317931803181318231833184318531863187318831893190319131923193319431953196319731983199320032013202320332043205320632073208320932103211321232133214321532163217321832193220322132223223322432253226322732283229323032313232323332343235323632373238323932403241324232433244324532463247324832493250325132523253325432553256325732583259326032613262326332643265326632673268326932703271327232733274327532763277327832793280328132823283328432853286328732883289329032913292329332943295329632973298329933003301330233033304330533063307330833093310331133123313331433153316331733183319332033213322332333243325332633273328332933303331333233333334333533363337333833393340334133423343334433453346334733483349
  1. // SPDX-License-Identifier: GPL-2.0-or-later
  2. /*
  3. * Driver for Broadcom MPI3 Storage Controllers
  4. *
  5. * Copyright (C) 2017-2023 Broadcom Inc.
  6. * (mailto: mpi3mr-linuxdrv.pdl@broadcom.com)
  7. *
  8. */
  9. #include "mpi3mr.h"
  10. #include <linux/bsg-lib.h>
  11. #include <uapi/scsi/scsi_bsg_mpi3mr.h>
  12. /**
  13. * mpi3mr_alloc_trace_buffer: Allocate trace buffer
  14. * @mrioc: Adapter instance reference
  15. * @trace_size: Trace buffer size
  16. *
  17. * Allocate trace buffer
  18. * Return: 0 on success, non-zero on failure.
  19. */
  20. static int mpi3mr_alloc_trace_buffer(struct mpi3mr_ioc *mrioc, u32 trace_size)
  21. {
  22. struct diag_buffer_desc *diag_buffer = &mrioc->diag_buffers[0];
  23. diag_buffer->addr = dma_alloc_coherent(&mrioc->pdev->dev,
  24. trace_size, &diag_buffer->dma_addr, GFP_KERNEL);
  25. if (diag_buffer->addr) {
  26. dprint_init(mrioc, "trace diag buffer is allocated successfully\n");
  27. return 0;
  28. }
  29. return -1;
  30. }
  31. /**
  32. * mpi3mr_alloc_diag_bufs - Allocate memory for diag buffers
  33. * @mrioc: Adapter instance reference
  34. *
  35. * This functions checks whether the driver defined buffer sizes
  36. * are greater than IOCFacts provided controller local buffer
  37. * sizes and if the driver defined sizes are more then the
  38. * driver allocates the specific buffer by reading driver page1
  39. *
  40. * Return: Nothing.
  41. */
  42. void mpi3mr_alloc_diag_bufs(struct mpi3mr_ioc *mrioc)
  43. {
  44. struct diag_buffer_desc *diag_buffer;
  45. struct mpi3_driver_page1 driver_pg1;
  46. u32 trace_dec_size, trace_min_size, fw_dec_size, fw_min_size,
  47. trace_size, fw_size;
  48. u16 pg_sz = sizeof(driver_pg1);
  49. int retval = 0;
  50. bool retry = false;
  51. if (mrioc->diag_buffers[0].addr || mrioc->diag_buffers[1].addr)
  52. return;
  53. retval = mpi3mr_cfg_get_driver_pg1(mrioc, &driver_pg1, pg_sz);
  54. if (retval) {
  55. ioc_warn(mrioc,
  56. "%s: driver page 1 read failed, allocating trace\n"
  57. "and firmware diag buffers of default size\n", __func__);
  58. trace_size = fw_size = MPI3MR_DEFAULT_HDB_MAX_SZ;
  59. trace_dec_size = fw_dec_size = MPI3MR_DEFAULT_HDB_DEC_SZ;
  60. trace_min_size = fw_min_size = MPI3MR_DEFAULT_HDB_MIN_SZ;
  61. } else {
  62. trace_size = driver_pg1.host_diag_trace_max_size * 1024;
  63. trace_dec_size = driver_pg1.host_diag_trace_decrement_size
  64. * 1024;
  65. trace_min_size = driver_pg1.host_diag_trace_min_size * 1024;
  66. fw_size = driver_pg1.host_diag_fw_max_size * 1024;
  67. fw_dec_size = driver_pg1.host_diag_fw_decrement_size * 1024;
  68. fw_min_size = driver_pg1.host_diag_fw_min_size * 1024;
  69. dprint_init(mrioc,
  70. "%s:trace diag buffer sizes read from driver\n"
  71. "page1: maximum size = %dKB, decrement size = %dKB\n"
  72. ", minimum size = %dKB\n", __func__, driver_pg1.host_diag_trace_max_size,
  73. driver_pg1.host_diag_trace_decrement_size,
  74. driver_pg1.host_diag_trace_min_size);
  75. dprint_init(mrioc,
  76. "%s:firmware diag buffer sizes read from driver\n"
  77. "page1: maximum size = %dKB, decrement size = %dKB\n"
  78. ", minimum size = %dKB\n", __func__, driver_pg1.host_diag_fw_max_size,
  79. driver_pg1.host_diag_fw_decrement_size,
  80. driver_pg1.host_diag_fw_min_size);
  81. if ((trace_size == 0) && (fw_size == 0))
  82. return;
  83. }
  84. retry_trace:
  85. diag_buffer = &mrioc->diag_buffers[0];
  86. diag_buffer->type = MPI3_DIAG_BUFFER_TYPE_TRACE;
  87. diag_buffer->status = MPI3MR_HDB_BUFSTATUS_NOT_ALLOCATED;
  88. if ((mrioc->facts.diag_trace_sz < trace_size) && (trace_size >=
  89. trace_min_size)) {
  90. if (!retry)
  91. dprint_init(mrioc,
  92. "trying to allocate trace diag buffer of size = %dKB\n",
  93. trace_size / 1024);
  94. if (get_order(trace_size) > MAX_PAGE_ORDER ||
  95. mpi3mr_alloc_trace_buffer(mrioc, trace_size)) {
  96. retry = true;
  97. trace_size -= trace_dec_size;
  98. dprint_init(mrioc, "trace diag buffer allocation failed\n"
  99. "retrying smaller size %dKB\n", trace_size / 1024);
  100. goto retry_trace;
  101. } else
  102. diag_buffer->size = trace_size;
  103. }
  104. retry = false;
  105. retry_fw:
  106. diag_buffer = &mrioc->diag_buffers[1];
  107. diag_buffer->type = MPI3_DIAG_BUFFER_TYPE_FW;
  108. diag_buffer->status = MPI3MR_HDB_BUFSTATUS_NOT_ALLOCATED;
  109. if ((mrioc->facts.diag_fw_sz < fw_size) && (fw_size >= fw_min_size)) {
  110. if (get_order(fw_size) <= MAX_PAGE_ORDER) {
  111. diag_buffer->addr
  112. = dma_alloc_coherent(&mrioc->pdev->dev, fw_size,
  113. &diag_buffer->dma_addr,
  114. GFP_KERNEL);
  115. }
  116. if (!retry)
  117. dprint_init(mrioc,
  118. "%s:trying to allocate firmware diag buffer of size = %dKB\n",
  119. __func__, fw_size / 1024);
  120. if (diag_buffer->addr) {
  121. dprint_init(mrioc, "%s:firmware diag buffer allocated successfully\n",
  122. __func__);
  123. diag_buffer->size = fw_size;
  124. } else {
  125. retry = true;
  126. fw_size -= fw_dec_size;
  127. dprint_init(mrioc, "%s:trace diag buffer allocation failed,\n"
  128. "retrying smaller size %dKB\n",
  129. __func__, fw_size / 1024);
  130. goto retry_fw;
  131. }
  132. }
  133. }
  134. /**
  135. * mpi3mr_issue_diag_buf_post - Send diag buffer post req
  136. * @mrioc: Adapter instance reference
  137. * @diag_buffer: Diagnostic buffer descriptor
  138. *
  139. * Issue diagnostic buffer post MPI request through admin queue
  140. * and wait for the completion of it or time out.
  141. *
  142. * Return: 0 on success, non-zero on failures.
  143. */
  144. int mpi3mr_issue_diag_buf_post(struct mpi3mr_ioc *mrioc,
  145. struct diag_buffer_desc *diag_buffer)
  146. {
  147. struct mpi3_diag_buffer_post_request diag_buf_post_req;
  148. u8 prev_status;
  149. int retval = 0;
  150. memset(&diag_buf_post_req, 0, sizeof(diag_buf_post_req));
  151. mutex_lock(&mrioc->init_cmds.mutex);
  152. if (mrioc->init_cmds.state & MPI3MR_CMD_PENDING) {
  153. dprint_bsg_err(mrioc, "%s: command is in use\n", __func__);
  154. mutex_unlock(&mrioc->init_cmds.mutex);
  155. return -1;
  156. }
  157. mrioc->init_cmds.state = MPI3MR_CMD_PENDING;
  158. mrioc->init_cmds.is_waiting = 1;
  159. mrioc->init_cmds.callback = NULL;
  160. diag_buf_post_req.host_tag = cpu_to_le16(MPI3MR_HOSTTAG_INITCMDS);
  161. diag_buf_post_req.function = MPI3_FUNCTION_DIAG_BUFFER_POST;
  162. diag_buf_post_req.type = diag_buffer->type;
  163. diag_buf_post_req.address = le64_to_cpu(diag_buffer->dma_addr);
  164. diag_buf_post_req.length = le32_to_cpu(diag_buffer->size);
  165. dprint_bsg_info(mrioc, "%s: posting diag buffer type %d\n", __func__,
  166. diag_buffer->type);
  167. prev_status = diag_buffer->status;
  168. diag_buffer->status = MPI3MR_HDB_BUFSTATUS_POSTED_UNPAUSED;
  169. init_completion(&mrioc->init_cmds.done);
  170. retval = mpi3mr_admin_request_post(mrioc, &diag_buf_post_req,
  171. sizeof(diag_buf_post_req), 1);
  172. if (retval) {
  173. dprint_bsg_err(mrioc, "%s: admin request post failed\n",
  174. __func__);
  175. goto out_unlock;
  176. }
  177. wait_for_completion_timeout(&mrioc->init_cmds.done,
  178. (MPI3MR_INTADMCMD_TIMEOUT * HZ));
  179. if (!(mrioc->init_cmds.state & MPI3MR_CMD_COMPLETE)) {
  180. mrioc->init_cmds.is_waiting = 0;
  181. dprint_bsg_err(mrioc, "%s: command timedout\n", __func__);
  182. mpi3mr_check_rh_fault_ioc(mrioc,
  183. MPI3MR_RESET_FROM_DIAG_BUFFER_POST_TIMEOUT);
  184. retval = -1;
  185. goto out_unlock;
  186. }
  187. if ((mrioc->init_cmds.ioc_status & MPI3_IOCSTATUS_STATUS_MASK)
  188. != MPI3_IOCSTATUS_SUCCESS) {
  189. dprint_bsg_err(mrioc,
  190. "%s: command failed, buffer_type (%d) ioc_status(0x%04x) log_info(0x%08x)\n",
  191. __func__, diag_buffer->type,
  192. (mrioc->init_cmds.ioc_status & MPI3_IOCSTATUS_STATUS_MASK),
  193. mrioc->init_cmds.ioc_loginfo);
  194. retval = -1;
  195. goto out_unlock;
  196. }
  197. dprint_bsg_info(mrioc, "%s: diag buffer type %d posted successfully\n",
  198. __func__, diag_buffer->type);
  199. out_unlock:
  200. if (retval)
  201. diag_buffer->status = prev_status;
  202. mrioc->init_cmds.state = MPI3MR_CMD_NOTUSED;
  203. mutex_unlock(&mrioc->init_cmds.mutex);
  204. return retval;
  205. }
  206. /**
  207. * mpi3mr_post_diag_bufs - Post diag buffers to the controller
  208. * @mrioc: Adapter instance reference
  209. *
  210. * This function calls helper function to post both trace and
  211. * firmware buffers to the controller.
  212. *
  213. * Return: None
  214. */
  215. int mpi3mr_post_diag_bufs(struct mpi3mr_ioc *mrioc)
  216. {
  217. u8 i;
  218. struct diag_buffer_desc *diag_buffer;
  219. for (i = 0; i < MPI3MR_MAX_NUM_HDB; i++) {
  220. diag_buffer = &mrioc->diag_buffers[i];
  221. if (!(diag_buffer->addr))
  222. continue;
  223. if (mpi3mr_issue_diag_buf_post(mrioc, diag_buffer))
  224. return -1;
  225. }
  226. return 0;
  227. }
  228. /**
  229. * mpi3mr_issue_diag_buf_release - Send diag buffer release req
  230. * @mrioc: Adapter instance reference
  231. * @diag_buffer: Diagnostic buffer descriptor
  232. *
  233. * Issue diagnostic buffer manage MPI request with release
  234. * action request through admin queue and wait for the
  235. * completion of it or time out.
  236. *
  237. * Return: 0 on success, non-zero on failures.
  238. */
  239. int mpi3mr_issue_diag_buf_release(struct mpi3mr_ioc *mrioc,
  240. struct diag_buffer_desc *diag_buffer)
  241. {
  242. struct mpi3_diag_buffer_manage_request diag_buf_manage_req;
  243. int retval = 0;
  244. if ((diag_buffer->status != MPI3MR_HDB_BUFSTATUS_POSTED_UNPAUSED) &&
  245. (diag_buffer->status != MPI3MR_HDB_BUFSTATUS_POSTED_PAUSED))
  246. return retval;
  247. memset(&diag_buf_manage_req, 0, sizeof(diag_buf_manage_req));
  248. mutex_lock(&mrioc->init_cmds.mutex);
  249. if (mrioc->init_cmds.state & MPI3MR_CMD_PENDING) {
  250. dprint_reset(mrioc, "%s: command is in use\n", __func__);
  251. mutex_unlock(&mrioc->init_cmds.mutex);
  252. return -1;
  253. }
  254. mrioc->init_cmds.state = MPI3MR_CMD_PENDING;
  255. mrioc->init_cmds.is_waiting = 1;
  256. mrioc->init_cmds.callback = NULL;
  257. diag_buf_manage_req.host_tag = cpu_to_le16(MPI3MR_HOSTTAG_INITCMDS);
  258. diag_buf_manage_req.function = MPI3_FUNCTION_DIAG_BUFFER_MANAGE;
  259. diag_buf_manage_req.type = diag_buffer->type;
  260. diag_buf_manage_req.action = MPI3_DIAG_BUFFER_ACTION_RELEASE;
  261. dprint_reset(mrioc, "%s: releasing diag buffer type %d\n", __func__,
  262. diag_buffer->type);
  263. init_completion(&mrioc->init_cmds.done);
  264. retval = mpi3mr_admin_request_post(mrioc, &diag_buf_manage_req,
  265. sizeof(diag_buf_manage_req), 1);
  266. if (retval) {
  267. dprint_reset(mrioc, "%s: admin request post failed\n", __func__);
  268. mpi3mr_set_trigger_data_in_hdb(diag_buffer,
  269. MPI3MR_HDB_TRIGGER_TYPE_UNKNOWN, NULL, 1);
  270. goto out_unlock;
  271. }
  272. wait_for_completion_timeout(&mrioc->init_cmds.done,
  273. (MPI3MR_INTADMCMD_TIMEOUT * HZ));
  274. if (!(mrioc->init_cmds.state & MPI3MR_CMD_COMPLETE)) {
  275. mrioc->init_cmds.is_waiting = 0;
  276. dprint_reset(mrioc, "%s: command timedout\n", __func__);
  277. mpi3mr_check_rh_fault_ioc(mrioc,
  278. MPI3MR_RESET_FROM_DIAG_BUFFER_RELEASE_TIMEOUT);
  279. retval = -1;
  280. goto out_unlock;
  281. }
  282. if ((mrioc->init_cmds.ioc_status & MPI3_IOCSTATUS_STATUS_MASK)
  283. != MPI3_IOCSTATUS_SUCCESS) {
  284. dprint_reset(mrioc,
  285. "%s: command failed, buffer_type (%d) ioc_status(0x%04x) log_info(0x%08x)\n",
  286. __func__, diag_buffer->type,
  287. (mrioc->init_cmds.ioc_status & MPI3_IOCSTATUS_STATUS_MASK),
  288. mrioc->init_cmds.ioc_loginfo);
  289. retval = -1;
  290. goto out_unlock;
  291. }
  292. dprint_reset(mrioc, "%s: diag buffer type %d released successfully\n",
  293. __func__, diag_buffer->type);
  294. out_unlock:
  295. mrioc->init_cmds.state = MPI3MR_CMD_NOTUSED;
  296. mutex_unlock(&mrioc->init_cmds.mutex);
  297. return retval;
  298. }
  299. /**
  300. * mpi3mr_process_trigger - Generic HDB Trigger handler
  301. * @mrioc: Adapter instance reference
  302. * @trigger_type: Trigger type
  303. * @trigger_data: Trigger data
  304. * @trigger_flags: Trigger flags
  305. *
  306. * This function checks validity of HDB, triggers and based on
  307. * trigger information, creates an event to be processed in the
  308. * firmware event worker thread .
  309. *
  310. * This function should be called with trigger spinlock held
  311. *
  312. * Return: Nothing
  313. */
  314. static void mpi3mr_process_trigger(struct mpi3mr_ioc *mrioc, u8 trigger_type,
  315. union mpi3mr_trigger_data *trigger_data, u8 trigger_flags)
  316. {
  317. struct trigger_event_data event_data;
  318. struct diag_buffer_desc *trace_hdb = NULL;
  319. struct diag_buffer_desc *fw_hdb = NULL;
  320. u64 global_trigger;
  321. trace_hdb = mpi3mr_diag_buffer_for_type(mrioc,
  322. MPI3_DIAG_BUFFER_TYPE_TRACE);
  323. if (trace_hdb &&
  324. (trace_hdb->status != MPI3MR_HDB_BUFSTATUS_POSTED_UNPAUSED) &&
  325. (trace_hdb->status != MPI3MR_HDB_BUFSTATUS_POSTED_PAUSED))
  326. trace_hdb = NULL;
  327. fw_hdb = mpi3mr_diag_buffer_for_type(mrioc, MPI3_DIAG_BUFFER_TYPE_FW);
  328. if (fw_hdb &&
  329. (fw_hdb->status != MPI3MR_HDB_BUFSTATUS_POSTED_UNPAUSED) &&
  330. (fw_hdb->status != MPI3MR_HDB_BUFSTATUS_POSTED_PAUSED))
  331. fw_hdb = NULL;
  332. if (mrioc->snapdump_trigger_active || (mrioc->fw_release_trigger_active
  333. && mrioc->trace_release_trigger_active) ||
  334. (!trace_hdb && !fw_hdb) || (!mrioc->driver_pg2) ||
  335. ((trigger_type == MPI3MR_HDB_TRIGGER_TYPE_ELEMENT)
  336. && (!mrioc->driver_pg2->num_triggers)))
  337. return;
  338. memset(&event_data, 0, sizeof(event_data));
  339. event_data.trigger_type = trigger_type;
  340. memcpy(&event_data.trigger_specific_data, trigger_data,
  341. sizeof(*trigger_data));
  342. global_trigger = le64_to_cpu(mrioc->driver_pg2->global_trigger);
  343. if (global_trigger & MPI3_DRIVER2_GLOBALTRIGGER_SNAPDUMP_ENABLED) {
  344. event_data.snapdump = true;
  345. event_data.trace_hdb = trace_hdb;
  346. event_data.fw_hdb = fw_hdb;
  347. mrioc->snapdump_trigger_active = true;
  348. } else if (trigger_type == MPI3MR_HDB_TRIGGER_TYPE_GLOBAL) {
  349. if ((trace_hdb) && (global_trigger &
  350. MPI3_DRIVER2_GLOBALTRIGGER_DIAG_TRACE_RELEASE) &&
  351. (!mrioc->trace_release_trigger_active)) {
  352. event_data.trace_hdb = trace_hdb;
  353. mrioc->trace_release_trigger_active = true;
  354. }
  355. if ((fw_hdb) && (global_trigger &
  356. MPI3_DRIVER2_GLOBALTRIGGER_DIAG_FW_RELEASE) &&
  357. (!mrioc->fw_release_trigger_active)) {
  358. event_data.fw_hdb = fw_hdb;
  359. mrioc->fw_release_trigger_active = true;
  360. }
  361. } else if (trigger_type == MPI3MR_HDB_TRIGGER_TYPE_ELEMENT) {
  362. if ((trace_hdb) && (trigger_flags &
  363. MPI3_DRIVER2_TRIGGER_FLAGS_DIAG_TRACE_RELEASE) &&
  364. (!mrioc->trace_release_trigger_active)) {
  365. event_data.trace_hdb = trace_hdb;
  366. mrioc->trace_release_trigger_active = true;
  367. }
  368. if ((fw_hdb) && (trigger_flags &
  369. MPI3_DRIVER2_TRIGGER_FLAGS_DIAG_FW_RELEASE) &&
  370. (!mrioc->fw_release_trigger_active)) {
  371. event_data.fw_hdb = fw_hdb;
  372. mrioc->fw_release_trigger_active = true;
  373. }
  374. }
  375. if (event_data.trace_hdb || event_data.fw_hdb)
  376. mpi3mr_hdb_trigger_data_event(mrioc, &event_data);
  377. }
  378. /**
  379. * mpi3mr_global_trigger - Global HDB trigger handler
  380. * @mrioc: Adapter instance reference
  381. * @trigger_data: Trigger data
  382. *
  383. * This function checks whether the given global trigger is
  384. * enabled in the driver page 2 and if so calls generic trigger
  385. * handler to queue event for HDB release.
  386. *
  387. * Return: Nothing
  388. */
  389. void mpi3mr_global_trigger(struct mpi3mr_ioc *mrioc, u64 trigger_data)
  390. {
  391. unsigned long flags;
  392. union mpi3mr_trigger_data trigger_specific_data;
  393. spin_lock_irqsave(&mrioc->trigger_lock, flags);
  394. if (le64_to_cpu(mrioc->driver_pg2->global_trigger) & trigger_data) {
  395. memset(&trigger_specific_data, 0,
  396. sizeof(trigger_specific_data));
  397. trigger_specific_data.global = trigger_data;
  398. mpi3mr_process_trigger(mrioc, MPI3MR_HDB_TRIGGER_TYPE_GLOBAL,
  399. &trigger_specific_data, 0);
  400. }
  401. spin_unlock_irqrestore(&mrioc->trigger_lock, flags);
  402. }
  403. /**
  404. * mpi3mr_scsisense_trigger - SCSI sense HDB trigger handler
  405. * @mrioc: Adapter instance reference
  406. * @sensekey: Sense Key
  407. * @asc: Additional Sense Code
  408. * @ascq: Additional Sense Code Qualifier
  409. *
  410. * This function compares SCSI sense trigger values with driver
  411. * page 2 values and calls generic trigger handler to release
  412. * HDBs if match found
  413. *
  414. * Return: Nothing
  415. */
  416. void mpi3mr_scsisense_trigger(struct mpi3mr_ioc *mrioc, u8 sensekey, u8 asc,
  417. u8 ascq)
  418. {
  419. struct mpi3_driver2_trigger_scsi_sense *scsi_sense_trigger = NULL;
  420. u64 i = 0;
  421. unsigned long flags;
  422. u8 num_triggers, trigger_flags;
  423. if (mrioc->scsisense_trigger_present) {
  424. spin_lock_irqsave(&mrioc->trigger_lock, flags);
  425. scsi_sense_trigger = (struct mpi3_driver2_trigger_scsi_sense *)
  426. mrioc->driver_pg2->trigger;
  427. num_triggers = mrioc->driver_pg2->num_triggers;
  428. for (i = 0; i < num_triggers; i++, scsi_sense_trigger++) {
  429. if (scsi_sense_trigger->type !=
  430. MPI3_DRIVER2_TRIGGER_TYPE_SCSI_SENSE)
  431. continue;
  432. if (!(scsi_sense_trigger->sense_key ==
  433. MPI3_DRIVER2_TRIGGER_SCSI_SENSE_SENSE_KEY_MATCH_ALL
  434. || scsi_sense_trigger->sense_key == sensekey))
  435. continue;
  436. if (!(scsi_sense_trigger->asc ==
  437. MPI3_DRIVER2_TRIGGER_SCSI_SENSE_ASC_MATCH_ALL ||
  438. scsi_sense_trigger->asc == asc))
  439. continue;
  440. if (!(scsi_sense_trigger->ascq ==
  441. MPI3_DRIVER2_TRIGGER_SCSI_SENSE_ASCQ_MATCH_ALL ||
  442. scsi_sense_trigger->ascq == ascq))
  443. continue;
  444. trigger_flags = scsi_sense_trigger->flags;
  445. mpi3mr_process_trigger(mrioc,
  446. MPI3MR_HDB_TRIGGER_TYPE_ELEMENT,
  447. (union mpi3mr_trigger_data *)scsi_sense_trigger,
  448. trigger_flags);
  449. break;
  450. }
  451. spin_unlock_irqrestore(&mrioc->trigger_lock, flags);
  452. }
  453. }
  454. /**
  455. * mpi3mr_event_trigger - MPI event HDB trigger handler
  456. * @mrioc: Adapter instance reference
  457. * @event: MPI Event
  458. *
  459. * This function compares event trigger values with driver page
  460. * 2 values and calls generic trigger handler to release
  461. * HDBs if match found.
  462. *
  463. * Return: Nothing
  464. */
  465. void mpi3mr_event_trigger(struct mpi3mr_ioc *mrioc, u8 event)
  466. {
  467. struct mpi3_driver2_trigger_event *event_trigger = NULL;
  468. u64 i = 0;
  469. unsigned long flags;
  470. u8 num_triggers, trigger_flags;
  471. if (mrioc->event_trigger_present) {
  472. spin_lock_irqsave(&mrioc->trigger_lock, flags);
  473. event_trigger = (struct mpi3_driver2_trigger_event *)
  474. mrioc->driver_pg2->trigger;
  475. num_triggers = mrioc->driver_pg2->num_triggers;
  476. for (i = 0; i < num_triggers; i++, event_trigger++) {
  477. if (event_trigger->type !=
  478. MPI3_DRIVER2_TRIGGER_TYPE_EVENT)
  479. continue;
  480. if (event_trigger->event != event)
  481. continue;
  482. trigger_flags = event_trigger->flags;
  483. mpi3mr_process_trigger(mrioc,
  484. MPI3MR_HDB_TRIGGER_TYPE_ELEMENT,
  485. (union mpi3mr_trigger_data *)event_trigger,
  486. trigger_flags);
  487. break;
  488. }
  489. spin_unlock_irqrestore(&mrioc->trigger_lock, flags);
  490. }
  491. }
  492. /**
  493. * mpi3mr_reply_trigger - MPI Reply HDB trigger handler
  494. * @mrioc: Adapter instance reference
  495. * @ioc_status: Masked value of IOC Status from MPI Reply
  496. * @ioc_loginfo: IOC Log Info from MPI Reply
  497. *
  498. * This function compares IOC status and IOC log info trigger
  499. * values with driver page 2 values and calls generic trigger
  500. * handler to release HDBs if match found.
  501. *
  502. * Return: Nothing
  503. */
  504. void mpi3mr_reply_trigger(struct mpi3mr_ioc *mrioc, u16 ioc_status,
  505. u32 ioc_loginfo)
  506. {
  507. struct mpi3_driver2_trigger_reply *reply_trigger = NULL;
  508. u64 i = 0;
  509. unsigned long flags;
  510. u8 num_triggers, trigger_flags;
  511. if (mrioc->reply_trigger_present) {
  512. spin_lock_irqsave(&mrioc->trigger_lock, flags);
  513. reply_trigger = (struct mpi3_driver2_trigger_reply *)
  514. mrioc->driver_pg2->trigger;
  515. num_triggers = mrioc->driver_pg2->num_triggers;
  516. for (i = 0; i < num_triggers; i++, reply_trigger++) {
  517. if (reply_trigger->type !=
  518. MPI3_DRIVER2_TRIGGER_TYPE_REPLY)
  519. continue;
  520. if ((le16_to_cpu(reply_trigger->ioc_status) !=
  521. ioc_status)
  522. && (le16_to_cpu(reply_trigger->ioc_status) !=
  523. MPI3_DRIVER2_TRIGGER_REPLY_IOCSTATUS_MATCH_ALL))
  524. continue;
  525. if ((le32_to_cpu(reply_trigger->ioc_log_info) !=
  526. (le32_to_cpu(reply_trigger->ioc_log_info_mask) &
  527. ioc_loginfo)))
  528. continue;
  529. trigger_flags = reply_trigger->flags;
  530. mpi3mr_process_trigger(mrioc,
  531. MPI3MR_HDB_TRIGGER_TYPE_ELEMENT,
  532. (union mpi3mr_trigger_data *)reply_trigger,
  533. trigger_flags);
  534. break;
  535. }
  536. spin_unlock_irqrestore(&mrioc->trigger_lock, flags);
  537. }
  538. }
  539. /**
  540. * mpi3mr_get_num_trigger - Gets number of HDB triggers
  541. * @mrioc: Adapter instance reference
  542. * @num_triggers: Number of triggers
  543. * @page_action: Page action
  544. *
  545. * This function reads number of triggers by reading driver page
  546. * 2
  547. *
  548. * Return: 0 on success and proper error codes on failure
  549. */
  550. static int mpi3mr_get_num_trigger(struct mpi3mr_ioc *mrioc, u8 *num_triggers,
  551. u8 page_action)
  552. {
  553. struct mpi3_driver_page2 drvr_page2;
  554. int retval = 0;
  555. *num_triggers = 0;
  556. retval = mpi3mr_cfg_get_driver_pg2(mrioc, &drvr_page2,
  557. sizeof(struct mpi3_driver_page2), page_action);
  558. if (retval) {
  559. dprint_init(mrioc, "%s: driver page 2 read failed\n", __func__);
  560. return retval;
  561. }
  562. *num_triggers = drvr_page2.num_triggers;
  563. return retval;
  564. }
  565. /**
  566. * mpi3mr_refresh_trigger - Handler for Refresh trigger BSG
  567. * @mrioc: Adapter instance reference
  568. * @page_action: Page action
  569. *
  570. * This function caches the driver page 2 in the driver's memory
  571. * by reading driver page 2 from the controller for a given page
  572. * type and updates the HDB trigger values
  573. *
  574. * Return: 0 on success and proper error codes on failure
  575. */
  576. int mpi3mr_refresh_trigger(struct mpi3mr_ioc *mrioc, u8 page_action)
  577. {
  578. u16 pg_sz = sizeof(struct mpi3_driver_page2);
  579. struct mpi3_driver_page2 *drvr_page2 = NULL;
  580. u8 trigger_type, num_triggers;
  581. int retval;
  582. int i = 0;
  583. unsigned long flags;
  584. retval = mpi3mr_get_num_trigger(mrioc, &num_triggers, page_action);
  585. if (retval)
  586. goto out;
  587. pg_sz = offsetof(struct mpi3_driver_page2, trigger) +
  588. (num_triggers * sizeof(union mpi3_driver2_trigger_element));
  589. drvr_page2 = kzalloc(pg_sz, GFP_KERNEL);
  590. if (!drvr_page2) {
  591. retval = -ENOMEM;
  592. goto out;
  593. }
  594. retval = mpi3mr_cfg_get_driver_pg2(mrioc, drvr_page2, pg_sz, page_action);
  595. if (retval) {
  596. dprint_init(mrioc, "%s: driver page 2 read failed\n", __func__);
  597. kfree(drvr_page2);
  598. goto out;
  599. }
  600. spin_lock_irqsave(&mrioc->trigger_lock, flags);
  601. kfree(mrioc->driver_pg2);
  602. mrioc->driver_pg2 = drvr_page2;
  603. mrioc->reply_trigger_present = false;
  604. mrioc->event_trigger_present = false;
  605. mrioc->scsisense_trigger_present = false;
  606. for (i = 0; (i < mrioc->driver_pg2->num_triggers); i++) {
  607. trigger_type = mrioc->driver_pg2->trigger[i].event.type;
  608. switch (trigger_type) {
  609. case MPI3_DRIVER2_TRIGGER_TYPE_REPLY:
  610. mrioc->reply_trigger_present = true;
  611. break;
  612. case MPI3_DRIVER2_TRIGGER_TYPE_EVENT:
  613. mrioc->event_trigger_present = true;
  614. break;
  615. case MPI3_DRIVER2_TRIGGER_TYPE_SCSI_SENSE:
  616. mrioc->scsisense_trigger_present = true;
  617. break;
  618. default:
  619. break;
  620. }
  621. }
  622. spin_unlock_irqrestore(&mrioc->trigger_lock, flags);
  623. out:
  624. return retval;
  625. }
  626. /**
  627. * mpi3mr_release_diag_bufs - Release diag buffers
  628. * @mrioc: Adapter instance reference
  629. * @skip_rel_action: Skip release action and set buffer state
  630. *
  631. * This function calls helper function to release both trace and
  632. * firmware buffers from the controller.
  633. *
  634. * Return: None
  635. */
  636. void mpi3mr_release_diag_bufs(struct mpi3mr_ioc *mrioc, u8 skip_rel_action)
  637. {
  638. u8 i;
  639. struct diag_buffer_desc *diag_buffer;
  640. for (i = 0; i < MPI3MR_MAX_NUM_HDB; i++) {
  641. diag_buffer = &mrioc->diag_buffers[i];
  642. if (!(diag_buffer->addr))
  643. continue;
  644. if (diag_buffer->status == MPI3MR_HDB_BUFSTATUS_RELEASED)
  645. continue;
  646. if (!skip_rel_action)
  647. mpi3mr_issue_diag_buf_release(mrioc, diag_buffer);
  648. diag_buffer->status = MPI3MR_HDB_BUFSTATUS_RELEASED;
  649. atomic64_inc(&event_counter);
  650. }
  651. }
  652. /**
  653. * mpi3mr_set_trigger_data_in_hdb - Updates HDB trigger type and
  654. * trigger data
  655. *
  656. * @hdb: HDB pointer
  657. * @type: Trigger type
  658. * @data: Trigger data
  659. * @force: Trigger overwrite flag
  660. * @trigger_data: Pointer to trigger data information
  661. *
  662. * Updates trigger type and trigger data based on parameter
  663. * passed to this function
  664. *
  665. * Return: Nothing
  666. */
  667. void mpi3mr_set_trigger_data_in_hdb(struct diag_buffer_desc *hdb,
  668. u8 type, union mpi3mr_trigger_data *trigger_data, bool force)
  669. {
  670. if ((!force) && (hdb->trigger_type != MPI3MR_HDB_TRIGGER_TYPE_UNKNOWN))
  671. return;
  672. hdb->trigger_type = type;
  673. if (!trigger_data)
  674. memset(&hdb->trigger_data, 0, sizeof(*trigger_data));
  675. else
  676. memcpy(&hdb->trigger_data, trigger_data, sizeof(*trigger_data));
  677. }
  678. /**
  679. * mpi3mr_set_trigger_data_in_all_hdb - Updates HDB trigger type
  680. * and trigger data for all HDB
  681. *
  682. * @mrioc: Adapter instance reference
  683. * @type: Trigger type
  684. * @data: Trigger data
  685. * @force: Trigger overwrite flag
  686. * @trigger_data: Pointer to trigger data information
  687. *
  688. * Updates trigger type and trigger data based on parameter
  689. * passed to this function
  690. *
  691. * Return: Nothing
  692. */
  693. void mpi3mr_set_trigger_data_in_all_hdb(struct mpi3mr_ioc *mrioc,
  694. u8 type, union mpi3mr_trigger_data *trigger_data, bool force)
  695. {
  696. struct diag_buffer_desc *hdb = NULL;
  697. hdb = mpi3mr_diag_buffer_for_type(mrioc, MPI3_DIAG_BUFFER_TYPE_TRACE);
  698. if (hdb)
  699. mpi3mr_set_trigger_data_in_hdb(hdb, type, trigger_data, force);
  700. hdb = mpi3mr_diag_buffer_for_type(mrioc, MPI3_DIAG_BUFFER_TYPE_FW);
  701. if (hdb)
  702. mpi3mr_set_trigger_data_in_hdb(hdb, type, trigger_data, force);
  703. }
  704. /**
  705. * mpi3mr_hdbstatuschg_evt_th - HDB status change evt tophalf
  706. * @mrioc: Adapter instance reference
  707. * @event_reply: event data
  708. *
  709. * Modifies the status of the applicable diag buffer descriptors
  710. *
  711. * Return: Nothing
  712. */
  713. void mpi3mr_hdbstatuschg_evt_th(struct mpi3mr_ioc *mrioc,
  714. struct mpi3_event_notification_reply *event_reply)
  715. {
  716. struct mpi3_event_data_diag_buffer_status_change *evtdata;
  717. struct diag_buffer_desc *diag_buffer;
  718. evtdata = (struct mpi3_event_data_diag_buffer_status_change *)
  719. event_reply->event_data;
  720. diag_buffer = mpi3mr_diag_buffer_for_type(mrioc, evtdata->type);
  721. if (!diag_buffer)
  722. return;
  723. if ((diag_buffer->status != MPI3MR_HDB_BUFSTATUS_POSTED_UNPAUSED) &&
  724. (diag_buffer->status != MPI3MR_HDB_BUFSTATUS_POSTED_PAUSED))
  725. return;
  726. switch (evtdata->reason_code) {
  727. case MPI3_EVENT_DIAG_BUFFER_STATUS_CHANGE_RC_RELEASED:
  728. {
  729. diag_buffer->status = MPI3MR_HDB_BUFSTATUS_RELEASED;
  730. mpi3mr_set_trigger_data_in_hdb(diag_buffer,
  731. MPI3MR_HDB_TRIGGER_TYPE_FW_RELEASED, NULL, 0);
  732. atomic64_inc(&event_counter);
  733. break;
  734. }
  735. case MPI3_EVENT_DIAG_BUFFER_STATUS_CHANGE_RC_RESUMED:
  736. {
  737. diag_buffer->status = MPI3MR_HDB_BUFSTATUS_POSTED_UNPAUSED;
  738. break;
  739. }
  740. case MPI3_EVENT_DIAG_BUFFER_STATUS_CHANGE_RC_PAUSED:
  741. {
  742. diag_buffer->status = MPI3MR_HDB_BUFSTATUS_POSTED_PAUSED;
  743. break;
  744. }
  745. default:
  746. dprint_event_th(mrioc, "%s: unknown reason_code(%d)\n",
  747. __func__, evtdata->reason_code);
  748. break;
  749. }
  750. }
  751. /**
  752. * mpi3mr_diag_buffer_for_type - returns buffer desc for type
  753. * @mrioc: Adapter instance reference
  754. * @buf_type: Diagnostic buffer type
  755. *
  756. * Identifies matching diag descriptor from mrioc for given diag
  757. * buffer type.
  758. *
  759. * Return: diag buffer descriptor on success, NULL on failures.
  760. */
  761. struct diag_buffer_desc *
  762. mpi3mr_diag_buffer_for_type(struct mpi3mr_ioc *mrioc, u8 buf_type)
  763. {
  764. u8 i;
  765. for (i = 0; i < MPI3MR_MAX_NUM_HDB; i++) {
  766. if (mrioc->diag_buffers[i].type == buf_type)
  767. return &mrioc->diag_buffers[i];
  768. }
  769. return NULL;
  770. }
  771. /**
  772. * mpi3mr_bsg_pel_abort - sends PEL abort request
  773. * @mrioc: Adapter instance reference
  774. *
  775. * This function sends PEL abort request to the firmware through
  776. * admin request queue.
  777. *
  778. * Return: 0 on success, -1 on failure
  779. */
  780. static int mpi3mr_bsg_pel_abort(struct mpi3mr_ioc *mrioc)
  781. {
  782. struct mpi3_pel_req_action_abort pel_abort_req;
  783. struct mpi3_pel_reply *pel_reply;
  784. int retval = 0;
  785. u16 pe_log_status;
  786. if (mrioc->reset_in_progress) {
  787. dprint_bsg_err(mrioc, "%s: reset in progress\n", __func__);
  788. return -1;
  789. }
  790. if (mrioc->stop_bsgs || mrioc->block_on_pci_err) {
  791. dprint_bsg_err(mrioc, "%s: bsgs are blocked\n", __func__);
  792. return -1;
  793. }
  794. memset(&pel_abort_req, 0, sizeof(pel_abort_req));
  795. mutex_lock(&mrioc->pel_abort_cmd.mutex);
  796. if (mrioc->pel_abort_cmd.state & MPI3MR_CMD_PENDING) {
  797. dprint_bsg_err(mrioc, "%s: command is in use\n", __func__);
  798. mutex_unlock(&mrioc->pel_abort_cmd.mutex);
  799. return -1;
  800. }
  801. mrioc->pel_abort_cmd.state = MPI3MR_CMD_PENDING;
  802. mrioc->pel_abort_cmd.is_waiting = 1;
  803. mrioc->pel_abort_cmd.callback = NULL;
  804. pel_abort_req.host_tag = cpu_to_le16(MPI3MR_HOSTTAG_PEL_ABORT);
  805. pel_abort_req.function = MPI3_FUNCTION_PERSISTENT_EVENT_LOG;
  806. pel_abort_req.action = MPI3_PEL_ACTION_ABORT;
  807. pel_abort_req.abort_host_tag = cpu_to_le16(MPI3MR_HOSTTAG_PEL_WAIT);
  808. mrioc->pel_abort_requested = 1;
  809. init_completion(&mrioc->pel_abort_cmd.done);
  810. retval = mpi3mr_admin_request_post(mrioc, &pel_abort_req,
  811. sizeof(pel_abort_req), 0);
  812. if (retval) {
  813. retval = -1;
  814. dprint_bsg_err(mrioc, "%s: admin request post failed\n",
  815. __func__);
  816. mrioc->pel_abort_requested = 0;
  817. goto out_unlock;
  818. }
  819. wait_for_completion_timeout(&mrioc->pel_abort_cmd.done,
  820. (MPI3MR_INTADMCMD_TIMEOUT * HZ));
  821. if (!(mrioc->pel_abort_cmd.state & MPI3MR_CMD_COMPLETE)) {
  822. mrioc->pel_abort_cmd.is_waiting = 0;
  823. dprint_bsg_err(mrioc, "%s: command timedout\n", __func__);
  824. if (!(mrioc->pel_abort_cmd.state & MPI3MR_CMD_RESET))
  825. mpi3mr_soft_reset_handler(mrioc,
  826. MPI3MR_RESET_FROM_PELABORT_TIMEOUT, 1);
  827. retval = -1;
  828. goto out_unlock;
  829. }
  830. if ((mrioc->pel_abort_cmd.ioc_status & MPI3_IOCSTATUS_STATUS_MASK)
  831. != MPI3_IOCSTATUS_SUCCESS) {
  832. dprint_bsg_err(mrioc,
  833. "%s: command failed, ioc_status(0x%04x) log_info(0x%08x)\n",
  834. __func__, (mrioc->pel_abort_cmd.ioc_status &
  835. MPI3_IOCSTATUS_STATUS_MASK),
  836. mrioc->pel_abort_cmd.ioc_loginfo);
  837. retval = -1;
  838. goto out_unlock;
  839. }
  840. if (mrioc->pel_abort_cmd.state & MPI3MR_CMD_REPLY_VALID) {
  841. pel_reply = (struct mpi3_pel_reply *)mrioc->pel_abort_cmd.reply;
  842. pe_log_status = le16_to_cpu(pel_reply->pe_log_status);
  843. if (pe_log_status != MPI3_PEL_STATUS_SUCCESS) {
  844. dprint_bsg_err(mrioc,
  845. "%s: command failed, pel_status(0x%04x)\n",
  846. __func__, pe_log_status);
  847. retval = -1;
  848. }
  849. }
  850. out_unlock:
  851. mrioc->pel_abort_cmd.state = MPI3MR_CMD_NOTUSED;
  852. mutex_unlock(&mrioc->pel_abort_cmd.mutex);
  853. return retval;
  854. }
  855. /**
  856. * mpi3mr_bsg_verify_adapter - verify adapter number is valid
  857. * @ioc_number: Adapter number
  858. *
  859. * This function returns the adapter instance pointer of given
  860. * adapter number. If adapter number does not match with the
  861. * driver's adapter list, driver returns NULL.
  862. *
  863. * Return: adapter instance reference
  864. */
  865. static struct mpi3mr_ioc *mpi3mr_bsg_verify_adapter(int ioc_number)
  866. {
  867. struct mpi3mr_ioc *mrioc = NULL;
  868. spin_lock(&mrioc_list_lock);
  869. list_for_each_entry(mrioc, &mrioc_list, list) {
  870. if (mrioc->id == ioc_number) {
  871. spin_unlock(&mrioc_list_lock);
  872. return mrioc;
  873. }
  874. }
  875. spin_unlock(&mrioc_list_lock);
  876. return NULL;
  877. }
  878. /**
  879. * mpi3mr_bsg_refresh_hdb_triggers - Refresh HDB trigger data
  880. * @mrioc: Adapter instance reference
  881. * @job: BSG Job pointer
  882. *
  883. * This function reads the controller trigger config page as
  884. * defined by the input page type and refreshes the driver's
  885. * local trigger information structures with the controller's
  886. * config page data.
  887. *
  888. * Return: 0 on success and proper error codes on failure
  889. */
  890. static long
  891. mpi3mr_bsg_refresh_hdb_triggers(struct mpi3mr_ioc *mrioc,
  892. struct bsg_job *job)
  893. {
  894. struct mpi3mr_bsg_out_refresh_hdb_triggers refresh_triggers;
  895. uint32_t data_out_sz;
  896. u8 page_action;
  897. long rval = -EINVAL;
  898. data_out_sz = job->request_payload.payload_len;
  899. if (data_out_sz != sizeof(refresh_triggers)) {
  900. dprint_bsg_err(mrioc, "%s: invalid size argument\n",
  901. __func__);
  902. return rval;
  903. }
  904. if (mrioc->unrecoverable) {
  905. dprint_bsg_err(mrioc, "%s: unrecoverable controller\n",
  906. __func__);
  907. return -EFAULT;
  908. }
  909. if (mrioc->reset_in_progress) {
  910. dprint_bsg_err(mrioc, "%s: reset in progress\n", __func__);
  911. return -EAGAIN;
  912. }
  913. sg_copy_to_buffer(job->request_payload.sg_list,
  914. job->request_payload.sg_cnt,
  915. &refresh_triggers, sizeof(refresh_triggers));
  916. switch (refresh_triggers.page_type) {
  917. case MPI3MR_HDB_REFRESH_TYPE_CURRENT:
  918. page_action = MPI3_CONFIG_ACTION_READ_CURRENT;
  919. break;
  920. case MPI3MR_HDB_REFRESH_TYPE_DEFAULT:
  921. page_action = MPI3_CONFIG_ACTION_READ_DEFAULT;
  922. break;
  923. case MPI3MR_HDB_HDB_REFRESH_TYPE_PERSISTENT:
  924. page_action = MPI3_CONFIG_ACTION_READ_PERSISTENT;
  925. break;
  926. default:
  927. dprint_bsg_err(mrioc,
  928. "%s: unsupported refresh trigger, page_type %d\n",
  929. __func__, refresh_triggers.page_type);
  930. return rval;
  931. }
  932. rval = mpi3mr_refresh_trigger(mrioc, page_action);
  933. return rval;
  934. }
  935. /**
  936. * mpi3mr_bsg_upload_hdb - Upload a specific HDB to user space
  937. * @mrioc: Adapter instance reference
  938. * @job: BSG Job pointer
  939. *
  940. * Return: 0 on success and proper error codes on failure
  941. */
  942. static long mpi3mr_bsg_upload_hdb(struct mpi3mr_ioc *mrioc,
  943. struct bsg_job *job)
  944. {
  945. struct mpi3mr_bsg_out_upload_hdb upload_hdb;
  946. struct diag_buffer_desc *diag_buffer;
  947. uint32_t data_out_size;
  948. uint32_t data_in_size;
  949. data_out_size = job->request_payload.payload_len;
  950. data_in_size = job->reply_payload.payload_len;
  951. if (data_out_size != sizeof(upload_hdb)) {
  952. dprint_bsg_err(mrioc, "%s: invalid size argument\n",
  953. __func__);
  954. return -EINVAL;
  955. }
  956. sg_copy_to_buffer(job->request_payload.sg_list,
  957. job->request_payload.sg_cnt,
  958. &upload_hdb, sizeof(upload_hdb));
  959. if ((!upload_hdb.length) || (data_in_size != upload_hdb.length)) {
  960. dprint_bsg_err(mrioc, "%s: invalid length argument\n",
  961. __func__);
  962. return -EINVAL;
  963. }
  964. diag_buffer = mpi3mr_diag_buffer_for_type(mrioc, upload_hdb.buf_type);
  965. if ((!diag_buffer) || (!diag_buffer->addr)) {
  966. dprint_bsg_err(mrioc, "%s: invalid buffer type %d\n",
  967. __func__, upload_hdb.buf_type);
  968. return -EINVAL;
  969. }
  970. if ((diag_buffer->status != MPI3MR_HDB_BUFSTATUS_RELEASED) &&
  971. (diag_buffer->status != MPI3MR_HDB_BUFSTATUS_POSTED_PAUSED)) {
  972. dprint_bsg_err(mrioc,
  973. "%s: invalid buffer status %d for type %d\n",
  974. __func__, diag_buffer->status, upload_hdb.buf_type);
  975. return -EINVAL;
  976. }
  977. if ((upload_hdb.start_offset + upload_hdb.length) > diag_buffer->size) {
  978. dprint_bsg_err(mrioc,
  979. "%s: invalid start offset %d, length %d for type %d\n",
  980. __func__, upload_hdb.start_offset, upload_hdb.length,
  981. upload_hdb.buf_type);
  982. return -EINVAL;
  983. }
  984. sg_copy_from_buffer(job->reply_payload.sg_list,
  985. job->reply_payload.sg_cnt,
  986. (diag_buffer->addr + upload_hdb.start_offset),
  987. data_in_size);
  988. return 0;
  989. }
  990. /**
  991. * mpi3mr_bsg_repost_hdb - Re-post HDB
  992. * @mrioc: Adapter instance reference
  993. * @job: BSG job pointer
  994. *
  995. * This function retrieves the HDB descriptor corresponding to a
  996. * given buffer type and if the HDB is in released status then
  997. * posts the HDB with the firmware.
  998. *
  999. * Return: 0 on success and proper error codes on failure
  1000. */
  1001. static long mpi3mr_bsg_repost_hdb(struct mpi3mr_ioc *mrioc,
  1002. struct bsg_job *job)
  1003. {
  1004. struct mpi3mr_bsg_out_repost_hdb repost_hdb;
  1005. struct diag_buffer_desc *diag_buffer;
  1006. uint32_t data_out_sz;
  1007. data_out_sz = job->request_payload.payload_len;
  1008. if (data_out_sz != sizeof(repost_hdb)) {
  1009. dprint_bsg_err(mrioc, "%s: invalid size argument\n",
  1010. __func__);
  1011. return -EINVAL;
  1012. }
  1013. if (mrioc->unrecoverable) {
  1014. dprint_bsg_err(mrioc, "%s: unrecoverable controller\n",
  1015. __func__);
  1016. return -EFAULT;
  1017. }
  1018. if (mrioc->reset_in_progress) {
  1019. dprint_bsg_err(mrioc, "%s: reset in progress\n", __func__);
  1020. return -EAGAIN;
  1021. }
  1022. sg_copy_to_buffer(job->request_payload.sg_list,
  1023. job->request_payload.sg_cnt,
  1024. &repost_hdb, sizeof(repost_hdb));
  1025. diag_buffer = mpi3mr_diag_buffer_for_type(mrioc, repost_hdb.buf_type);
  1026. if ((!diag_buffer) || (!diag_buffer->addr)) {
  1027. dprint_bsg_err(mrioc, "%s: invalid buffer type %d\n",
  1028. __func__, repost_hdb.buf_type);
  1029. return -EINVAL;
  1030. }
  1031. if (diag_buffer->status != MPI3MR_HDB_BUFSTATUS_RELEASED) {
  1032. dprint_bsg_err(mrioc,
  1033. "%s: invalid buffer status %d for type %d\n",
  1034. __func__, diag_buffer->status, repost_hdb.buf_type);
  1035. return -EINVAL;
  1036. }
  1037. if (mpi3mr_issue_diag_buf_post(mrioc, diag_buffer)) {
  1038. dprint_bsg_err(mrioc, "%s: post failed for type %d\n",
  1039. __func__, repost_hdb.buf_type);
  1040. return -EFAULT;
  1041. }
  1042. mpi3mr_set_trigger_data_in_hdb(diag_buffer,
  1043. MPI3MR_HDB_TRIGGER_TYPE_UNKNOWN, NULL, 1);
  1044. return 0;
  1045. }
  1046. /**
  1047. * mpi3mr_bsg_query_hdb - Handler for query HDB command
  1048. * @mrioc: Adapter instance reference
  1049. * @job: BSG job pointer
  1050. *
  1051. * This function prepares and copies the host diagnostic buffer
  1052. * entries to the user buffer.
  1053. *
  1054. * Return: 0 on success and proper error codes on failure
  1055. */
  1056. static long mpi3mr_bsg_query_hdb(struct mpi3mr_ioc *mrioc,
  1057. struct bsg_job *job)
  1058. {
  1059. long rval = 0;
  1060. struct mpi3mr_bsg_in_hdb_status *hbd_status;
  1061. struct mpi3mr_hdb_entry *hbd_status_entry;
  1062. u32 length, min_length;
  1063. u8 i;
  1064. struct diag_buffer_desc *diag_buffer;
  1065. uint32_t data_in_sz = 0;
  1066. data_in_sz = job->request_payload.payload_len;
  1067. length = (sizeof(*hbd_status) + ((MPI3MR_MAX_NUM_HDB - 1) *
  1068. sizeof(*hbd_status_entry)));
  1069. hbd_status = kmalloc(length, GFP_KERNEL);
  1070. if (!hbd_status)
  1071. return -ENOMEM;
  1072. hbd_status_entry = &hbd_status->entry[0];
  1073. hbd_status->num_hdb_types = MPI3MR_MAX_NUM_HDB;
  1074. for (i = 0; i < MPI3MR_MAX_NUM_HDB; i++) {
  1075. diag_buffer = &mrioc->diag_buffers[i];
  1076. hbd_status_entry->buf_type = diag_buffer->type;
  1077. hbd_status_entry->status = diag_buffer->status;
  1078. hbd_status_entry->trigger_type = diag_buffer->trigger_type;
  1079. memcpy(&hbd_status_entry->trigger_data,
  1080. &diag_buffer->trigger_data,
  1081. sizeof(hbd_status_entry->trigger_data));
  1082. hbd_status_entry->size = (diag_buffer->size / 1024);
  1083. hbd_status_entry++;
  1084. }
  1085. hbd_status->element_trigger_format =
  1086. MPI3MR_HDB_QUERY_ELEMENT_TRIGGER_FORMAT_DATA;
  1087. if (data_in_sz < 4) {
  1088. dprint_bsg_err(mrioc, "%s: invalid size passed\n", __func__);
  1089. rval = -EINVAL;
  1090. goto out;
  1091. }
  1092. min_length = min(data_in_sz, length);
  1093. if (job->request_payload.payload_len >= min_length) {
  1094. sg_copy_from_buffer(job->request_payload.sg_list,
  1095. job->request_payload.sg_cnt,
  1096. hbd_status, min_length);
  1097. rval = 0;
  1098. }
  1099. out:
  1100. kfree(hbd_status);
  1101. return rval;
  1102. }
  1103. /**
  1104. * mpi3mr_enable_logdata - Handler for log data enable
  1105. * @mrioc: Adapter instance reference
  1106. * @job: BSG job reference
  1107. *
  1108. * This function enables log data caching in the driver if not
  1109. * already enabled and return the maximum number of log data
  1110. * entries that can be cached in the driver.
  1111. *
  1112. * Return: 0 on success and proper error codes on failure
  1113. */
  1114. static long mpi3mr_enable_logdata(struct mpi3mr_ioc *mrioc,
  1115. struct bsg_job *job)
  1116. {
  1117. struct mpi3mr_logdata_enable logdata_enable;
  1118. if (!mrioc->logdata_buf) {
  1119. mrioc->logdata_entry_sz =
  1120. (mrioc->reply_sz - (sizeof(struct mpi3_event_notification_reply) - 4))
  1121. + MPI3MR_BSG_LOGDATA_ENTRY_HEADER_SZ;
  1122. mrioc->logdata_buf_idx = 0;
  1123. mrioc->logdata_buf = kcalloc(MPI3MR_BSG_LOGDATA_MAX_ENTRIES,
  1124. mrioc->logdata_entry_sz, GFP_KERNEL);
  1125. if (!mrioc->logdata_buf)
  1126. return -ENOMEM;
  1127. }
  1128. memset(&logdata_enable, 0, sizeof(logdata_enable));
  1129. logdata_enable.max_entries =
  1130. MPI3MR_BSG_LOGDATA_MAX_ENTRIES;
  1131. if (job->request_payload.payload_len >= sizeof(logdata_enable)) {
  1132. sg_copy_from_buffer(job->request_payload.sg_list,
  1133. job->request_payload.sg_cnt,
  1134. &logdata_enable, sizeof(logdata_enable));
  1135. return 0;
  1136. }
  1137. return -EINVAL;
  1138. }
  1139. /**
  1140. * mpi3mr_get_logdata - Handler for get log data
  1141. * @mrioc: Adapter instance reference
  1142. * @job: BSG job pointer
  1143. * This function copies the log data entries to the user buffer
  1144. * when log caching is enabled in the driver.
  1145. *
  1146. * Return: 0 on success and proper error codes on failure
  1147. */
  1148. static long mpi3mr_get_logdata(struct mpi3mr_ioc *mrioc,
  1149. struct bsg_job *job)
  1150. {
  1151. u16 num_entries, sz, entry_sz = mrioc->logdata_entry_sz;
  1152. if ((!mrioc->logdata_buf) || (job->request_payload.payload_len < entry_sz))
  1153. return -EINVAL;
  1154. num_entries = job->request_payload.payload_len / entry_sz;
  1155. if (num_entries > MPI3MR_BSG_LOGDATA_MAX_ENTRIES)
  1156. num_entries = MPI3MR_BSG_LOGDATA_MAX_ENTRIES;
  1157. sz = num_entries * entry_sz;
  1158. if (job->request_payload.payload_len >= sz) {
  1159. sg_copy_from_buffer(job->request_payload.sg_list,
  1160. job->request_payload.sg_cnt,
  1161. mrioc->logdata_buf, sz);
  1162. return 0;
  1163. }
  1164. return -EINVAL;
  1165. }
  1166. /**
  1167. * mpi3mr_bsg_pel_enable - Handler for PEL enable driver
  1168. * @mrioc: Adapter instance reference
  1169. * @job: BSG job pointer
  1170. *
  1171. * This function is the handler for PEL enable driver.
  1172. * Validates the application given class and locale and if
  1173. * requires aborts the existing PEL wait request and/or issues
  1174. * new PEL wait request to the firmware and returns.
  1175. *
  1176. * Return: 0 on success and proper error codes on failure.
  1177. */
  1178. static long mpi3mr_bsg_pel_enable(struct mpi3mr_ioc *mrioc,
  1179. struct bsg_job *job)
  1180. {
  1181. long rval = -EINVAL;
  1182. struct mpi3mr_bsg_out_pel_enable pel_enable;
  1183. u8 issue_pel_wait;
  1184. u8 tmp_class;
  1185. u16 tmp_locale;
  1186. if (job->request_payload.payload_len != sizeof(pel_enable)) {
  1187. dprint_bsg_err(mrioc, "%s: invalid size argument\n",
  1188. __func__);
  1189. return rval;
  1190. }
  1191. if (mrioc->unrecoverable) {
  1192. dprint_bsg_err(mrioc, "%s: unrecoverable controller\n",
  1193. __func__);
  1194. return -EFAULT;
  1195. }
  1196. if (mrioc->reset_in_progress) {
  1197. dprint_bsg_err(mrioc, "%s: reset in progress\n", __func__);
  1198. return -EAGAIN;
  1199. }
  1200. if (mrioc->stop_bsgs) {
  1201. dprint_bsg_err(mrioc, "%s: bsgs are blocked\n", __func__);
  1202. return -EAGAIN;
  1203. }
  1204. sg_copy_to_buffer(job->request_payload.sg_list,
  1205. job->request_payload.sg_cnt,
  1206. &pel_enable, sizeof(pel_enable));
  1207. if (pel_enable.pel_class > MPI3_PEL_CLASS_FAULT) {
  1208. dprint_bsg_err(mrioc, "%s: out of range class %d sent\n",
  1209. __func__, pel_enable.pel_class);
  1210. rval = 0;
  1211. goto out;
  1212. }
  1213. if (!mrioc->pel_enabled)
  1214. issue_pel_wait = 1;
  1215. else {
  1216. if ((mrioc->pel_class <= pel_enable.pel_class) &&
  1217. !((mrioc->pel_locale & pel_enable.pel_locale) ^
  1218. pel_enable.pel_locale)) {
  1219. issue_pel_wait = 0;
  1220. rval = 0;
  1221. } else {
  1222. pel_enable.pel_locale |= mrioc->pel_locale;
  1223. if (mrioc->pel_class < pel_enable.pel_class)
  1224. pel_enable.pel_class = mrioc->pel_class;
  1225. rval = mpi3mr_bsg_pel_abort(mrioc);
  1226. if (rval) {
  1227. dprint_bsg_err(mrioc,
  1228. "%s: pel_abort failed, status(%ld)\n",
  1229. __func__, rval);
  1230. goto out;
  1231. }
  1232. issue_pel_wait = 1;
  1233. }
  1234. }
  1235. if (issue_pel_wait) {
  1236. tmp_class = mrioc->pel_class;
  1237. tmp_locale = mrioc->pel_locale;
  1238. mrioc->pel_class = pel_enable.pel_class;
  1239. mrioc->pel_locale = pel_enable.pel_locale;
  1240. mrioc->pel_enabled = 1;
  1241. rval = mpi3mr_pel_get_seqnum_post(mrioc, NULL);
  1242. if (rval) {
  1243. mrioc->pel_class = tmp_class;
  1244. mrioc->pel_locale = tmp_locale;
  1245. mrioc->pel_enabled = 0;
  1246. dprint_bsg_err(mrioc,
  1247. "%s: pel get sequence number failed, status(%ld)\n",
  1248. __func__, rval);
  1249. }
  1250. }
  1251. out:
  1252. return rval;
  1253. }
  1254. /**
  1255. * mpi3mr_get_all_tgt_info - Get all target information
  1256. * @mrioc: Adapter instance reference
  1257. * @job: BSG job reference
  1258. *
  1259. * This function copies the driver managed target devices device
  1260. * handle, persistent ID, bus ID and taret ID to the user
  1261. * provided buffer for the specific controller. This function
  1262. * also provides the number of devices managed by the driver for
  1263. * the specific controller.
  1264. *
  1265. * Return: 0 on success and proper error codes on failure
  1266. */
  1267. static long mpi3mr_get_all_tgt_info(struct mpi3mr_ioc *mrioc,
  1268. struct bsg_job *job)
  1269. {
  1270. u16 num_devices = 0, i = 0, size;
  1271. unsigned long flags;
  1272. struct mpi3mr_tgt_dev *tgtdev;
  1273. struct mpi3mr_device_map_info *devmap_info = NULL;
  1274. struct mpi3mr_all_tgt_info *alltgt_info = NULL;
  1275. uint32_t min_entrylen = 0, kern_entrylen = 0, usr_entrylen = 0;
  1276. if (job->request_payload.payload_len < sizeof(u32)) {
  1277. dprint_bsg_err(mrioc, "%s: invalid size argument\n",
  1278. __func__);
  1279. return -EINVAL;
  1280. }
  1281. spin_lock_irqsave(&mrioc->tgtdev_lock, flags);
  1282. list_for_each_entry(tgtdev, &mrioc->tgtdev_list, list)
  1283. num_devices++;
  1284. spin_unlock_irqrestore(&mrioc->tgtdev_lock, flags);
  1285. if ((job->request_payload.payload_len <= sizeof(u64)) ||
  1286. list_empty(&mrioc->tgtdev_list)) {
  1287. sg_copy_from_buffer(job->request_payload.sg_list,
  1288. job->request_payload.sg_cnt,
  1289. &num_devices, sizeof(num_devices));
  1290. return 0;
  1291. }
  1292. kern_entrylen = num_devices * sizeof(*devmap_info);
  1293. size = sizeof(u64) + kern_entrylen;
  1294. alltgt_info = kzalloc(size, GFP_KERNEL);
  1295. if (!alltgt_info)
  1296. return -ENOMEM;
  1297. devmap_info = alltgt_info->dmi;
  1298. memset((u8 *)devmap_info, 0xFF, kern_entrylen);
  1299. spin_lock_irqsave(&mrioc->tgtdev_lock, flags);
  1300. list_for_each_entry(tgtdev, &mrioc->tgtdev_list, list) {
  1301. if (i < num_devices) {
  1302. devmap_info[i].handle = tgtdev->dev_handle;
  1303. devmap_info[i].perst_id = tgtdev->perst_id;
  1304. if (tgtdev->host_exposed && tgtdev->starget) {
  1305. devmap_info[i].target_id = tgtdev->starget->id;
  1306. devmap_info[i].bus_id =
  1307. tgtdev->starget->channel;
  1308. }
  1309. i++;
  1310. }
  1311. }
  1312. num_devices = i;
  1313. spin_unlock_irqrestore(&mrioc->tgtdev_lock, flags);
  1314. alltgt_info->num_devices = num_devices;
  1315. usr_entrylen = (job->request_payload.payload_len - sizeof(u64)) /
  1316. sizeof(*devmap_info);
  1317. usr_entrylen *= sizeof(*devmap_info);
  1318. min_entrylen = min(usr_entrylen, kern_entrylen);
  1319. sg_copy_from_buffer(job->request_payload.sg_list,
  1320. job->request_payload.sg_cnt,
  1321. alltgt_info, (min_entrylen + sizeof(u64)));
  1322. kfree(alltgt_info);
  1323. return 0;
  1324. }
  1325. /**
  1326. * mpi3mr_get_change_count - Get topology change count
  1327. * @mrioc: Adapter instance reference
  1328. * @job: BSG job reference
  1329. *
  1330. * This function copies the toplogy change count provided by the
  1331. * driver in events and cached in the driver to the user
  1332. * provided buffer for the specific controller.
  1333. *
  1334. * Return: 0 on success and proper error codes on failure
  1335. */
  1336. static long mpi3mr_get_change_count(struct mpi3mr_ioc *mrioc,
  1337. struct bsg_job *job)
  1338. {
  1339. struct mpi3mr_change_count chgcnt;
  1340. memset(&chgcnt, 0, sizeof(chgcnt));
  1341. chgcnt.change_count = mrioc->change_count;
  1342. if (job->request_payload.payload_len >= sizeof(chgcnt)) {
  1343. sg_copy_from_buffer(job->request_payload.sg_list,
  1344. job->request_payload.sg_cnt,
  1345. &chgcnt, sizeof(chgcnt));
  1346. return 0;
  1347. }
  1348. return -EINVAL;
  1349. }
  1350. /**
  1351. * mpi3mr_bsg_adp_reset - Issue controller reset
  1352. * @mrioc: Adapter instance reference
  1353. * @job: BSG job reference
  1354. *
  1355. * This function identifies the user provided reset type and
  1356. * issues approporiate reset to the controller and wait for that
  1357. * to complete and reinitialize the controller and then returns
  1358. *
  1359. * Return: 0 on success and proper error codes on failure
  1360. */
  1361. static long mpi3mr_bsg_adp_reset(struct mpi3mr_ioc *mrioc,
  1362. struct bsg_job *job)
  1363. {
  1364. long rval = -EINVAL;
  1365. u8 save_snapdump;
  1366. struct mpi3mr_bsg_adp_reset adpreset;
  1367. if (job->request_payload.payload_len !=
  1368. sizeof(adpreset)) {
  1369. dprint_bsg_err(mrioc, "%s: invalid size argument\n",
  1370. __func__);
  1371. goto out;
  1372. }
  1373. if (mrioc->unrecoverable || mrioc->block_on_pci_err)
  1374. return -EINVAL;
  1375. sg_copy_to_buffer(job->request_payload.sg_list,
  1376. job->request_payload.sg_cnt,
  1377. &adpreset, sizeof(adpreset));
  1378. switch (adpreset.reset_type) {
  1379. case MPI3MR_BSG_ADPRESET_SOFT:
  1380. save_snapdump = 0;
  1381. break;
  1382. case MPI3MR_BSG_ADPRESET_DIAG_FAULT:
  1383. save_snapdump = 1;
  1384. break;
  1385. default:
  1386. dprint_bsg_err(mrioc, "%s: unknown reset_type(%d)\n",
  1387. __func__, adpreset.reset_type);
  1388. goto out;
  1389. }
  1390. rval = mpi3mr_soft_reset_handler(mrioc, MPI3MR_RESET_FROM_APP,
  1391. save_snapdump);
  1392. if (rval)
  1393. dprint_bsg_err(mrioc,
  1394. "%s: reset handler returned error(%ld) for reset type %d\n",
  1395. __func__, rval, adpreset.reset_type);
  1396. out:
  1397. return rval;
  1398. }
  1399. /**
  1400. * mpi3mr_bsg_populate_adpinfo - Get adapter info command handler
  1401. * @mrioc: Adapter instance reference
  1402. * @job: BSG job reference
  1403. *
  1404. * This function provides adapter information for the given
  1405. * controller
  1406. *
  1407. * Return: 0 on success and proper error codes on failure
  1408. */
  1409. static long mpi3mr_bsg_populate_adpinfo(struct mpi3mr_ioc *mrioc,
  1410. struct bsg_job *job)
  1411. {
  1412. enum mpi3mr_iocstate ioc_state;
  1413. struct mpi3mr_bsg_in_adpinfo adpinfo;
  1414. memset(&adpinfo, 0, sizeof(adpinfo));
  1415. adpinfo.adp_type = MPI3MR_BSG_ADPTYPE_AVGFAMILY;
  1416. adpinfo.pci_dev_id = mrioc->pdev->device;
  1417. adpinfo.pci_dev_hw_rev = mrioc->pdev->revision;
  1418. adpinfo.pci_subsys_dev_id = mrioc->pdev->subsystem_device;
  1419. adpinfo.pci_subsys_ven_id = mrioc->pdev->subsystem_vendor;
  1420. adpinfo.pci_bus = mrioc->pdev->bus->number;
  1421. adpinfo.pci_dev = PCI_SLOT(mrioc->pdev->devfn);
  1422. adpinfo.pci_func = PCI_FUNC(mrioc->pdev->devfn);
  1423. adpinfo.pci_seg_id = pci_domain_nr(mrioc->pdev->bus);
  1424. adpinfo.app_intfc_ver = MPI3MR_IOCTL_VERSION;
  1425. ioc_state = mpi3mr_get_iocstate(mrioc);
  1426. if (ioc_state == MRIOC_STATE_UNRECOVERABLE)
  1427. adpinfo.adp_state = MPI3MR_BSG_ADPSTATE_UNRECOVERABLE;
  1428. else if ((mrioc->reset_in_progress) || (mrioc->stop_bsgs))
  1429. adpinfo.adp_state = MPI3MR_BSG_ADPSTATE_IN_RESET;
  1430. else if (ioc_state == MRIOC_STATE_FAULT)
  1431. adpinfo.adp_state = MPI3MR_BSG_ADPSTATE_FAULT;
  1432. else
  1433. adpinfo.adp_state = MPI3MR_BSG_ADPSTATE_OPERATIONAL;
  1434. memcpy((u8 *)&adpinfo.driver_info, (u8 *)&mrioc->driver_info,
  1435. sizeof(adpinfo.driver_info));
  1436. if (job->request_payload.payload_len >= sizeof(adpinfo)) {
  1437. sg_copy_from_buffer(job->request_payload.sg_list,
  1438. job->request_payload.sg_cnt,
  1439. &adpinfo, sizeof(adpinfo));
  1440. return 0;
  1441. }
  1442. return -EINVAL;
  1443. }
  1444. /**
  1445. * mpi3mr_bsg_process_drv_cmds - Driver Command handler
  1446. * @job: BSG job reference
  1447. *
  1448. * This function is the top level handler for driver commands,
  1449. * this does basic validation of the buffer and identifies the
  1450. * opcode and switches to correct sub handler.
  1451. *
  1452. * Return: 0 on success and proper error codes on failure
  1453. */
  1454. static long mpi3mr_bsg_process_drv_cmds(struct bsg_job *job)
  1455. {
  1456. long rval = -EINVAL;
  1457. struct mpi3mr_ioc *mrioc = NULL;
  1458. struct mpi3mr_bsg_packet *bsg_req = NULL;
  1459. struct mpi3mr_bsg_drv_cmd *drvrcmd = NULL;
  1460. bsg_req = job->request;
  1461. drvrcmd = &bsg_req->cmd.drvrcmd;
  1462. mrioc = mpi3mr_bsg_verify_adapter(drvrcmd->mrioc_id);
  1463. if (!mrioc)
  1464. return -ENODEV;
  1465. if (drvrcmd->opcode == MPI3MR_DRVBSG_OPCODE_ADPINFO) {
  1466. rval = mpi3mr_bsg_populate_adpinfo(mrioc, job);
  1467. return rval;
  1468. }
  1469. if (mutex_lock_interruptible(&mrioc->bsg_cmds.mutex))
  1470. return -ERESTARTSYS;
  1471. switch (drvrcmd->opcode) {
  1472. case MPI3MR_DRVBSG_OPCODE_ADPRESET:
  1473. rval = mpi3mr_bsg_adp_reset(mrioc, job);
  1474. break;
  1475. case MPI3MR_DRVBSG_OPCODE_ALLTGTDEVINFO:
  1476. rval = mpi3mr_get_all_tgt_info(mrioc, job);
  1477. break;
  1478. case MPI3MR_DRVBSG_OPCODE_GETCHGCNT:
  1479. rval = mpi3mr_get_change_count(mrioc, job);
  1480. break;
  1481. case MPI3MR_DRVBSG_OPCODE_LOGDATAENABLE:
  1482. rval = mpi3mr_enable_logdata(mrioc, job);
  1483. break;
  1484. case MPI3MR_DRVBSG_OPCODE_GETLOGDATA:
  1485. rval = mpi3mr_get_logdata(mrioc, job);
  1486. break;
  1487. case MPI3MR_DRVBSG_OPCODE_PELENABLE:
  1488. rval = mpi3mr_bsg_pel_enable(mrioc, job);
  1489. break;
  1490. case MPI3MR_DRVBSG_OPCODE_QUERY_HDB:
  1491. rval = mpi3mr_bsg_query_hdb(mrioc, job);
  1492. break;
  1493. case MPI3MR_DRVBSG_OPCODE_REPOST_HDB:
  1494. rval = mpi3mr_bsg_repost_hdb(mrioc, job);
  1495. break;
  1496. case MPI3MR_DRVBSG_OPCODE_UPLOAD_HDB:
  1497. rval = mpi3mr_bsg_upload_hdb(mrioc, job);
  1498. break;
  1499. case MPI3MR_DRVBSG_OPCODE_REFRESH_HDB_TRIGGERS:
  1500. rval = mpi3mr_bsg_refresh_hdb_triggers(mrioc, job);
  1501. break;
  1502. case MPI3MR_DRVBSG_OPCODE_UNKNOWN:
  1503. default:
  1504. pr_err("%s: unsupported driver command opcode %d\n",
  1505. MPI3MR_DRIVER_NAME, drvrcmd->opcode);
  1506. break;
  1507. }
  1508. mutex_unlock(&mrioc->bsg_cmds.mutex);
  1509. return rval;
  1510. }
  1511. /**
  1512. * mpi3mr_total_num_ioctl_sges - Count number of SGEs required
  1513. * @drv_bufs: DMA address of the buffers to be placed in sgl
  1514. * @bufcnt: Number of DMA buffers
  1515. *
  1516. * This function returns total number of data SGEs required
  1517. * including zero length SGEs and excluding management request
  1518. * and response buffer for the given list of data buffer
  1519. * descriptors
  1520. *
  1521. * Return: Number of SGE elements needed
  1522. */
  1523. static inline u16 mpi3mr_total_num_ioctl_sges(struct mpi3mr_buf_map *drv_bufs,
  1524. u8 bufcnt)
  1525. {
  1526. u16 i, sge_count = 0;
  1527. for (i = 0; i < bufcnt; i++, drv_bufs++) {
  1528. if (drv_bufs->data_dir == DMA_NONE ||
  1529. drv_bufs->kern_buf)
  1530. continue;
  1531. sge_count += drv_bufs->num_dma_desc;
  1532. if (!drv_bufs->num_dma_desc)
  1533. sge_count++;
  1534. }
  1535. return sge_count;
  1536. }
  1537. /**
  1538. * mpi3mr_bsg_build_sgl - SGL construction for MPI commands
  1539. * @mrioc: Adapter instance reference
  1540. * @mpi_req: MPI request
  1541. * @sgl_offset: offset to start sgl in the MPI request
  1542. * @drv_bufs: DMA address of the buffers to be placed in sgl
  1543. * @bufcnt: Number of DMA buffers
  1544. * @is_rmc: Does the buffer list has management command buffer
  1545. * @is_rmr: Does the buffer list has management response buffer
  1546. * @num_datasges: Number of data buffers in the list
  1547. *
  1548. * This function places the DMA address of the given buffers in
  1549. * proper format as SGEs in the given MPI request.
  1550. *
  1551. * Return: 0 on success,-1 on failure
  1552. */
  1553. static int mpi3mr_bsg_build_sgl(struct mpi3mr_ioc *mrioc, u8 *mpi_req,
  1554. u32 sgl_offset, struct mpi3mr_buf_map *drv_bufs,
  1555. u8 bufcnt, u8 is_rmc, u8 is_rmr, u8 num_datasges)
  1556. {
  1557. struct mpi3_request_header *mpi_header =
  1558. (struct mpi3_request_header *)mpi_req;
  1559. u8 *sgl = (mpi_req + sgl_offset), count = 0;
  1560. struct mpi3_mgmt_passthrough_request *rmgmt_req =
  1561. (struct mpi3_mgmt_passthrough_request *)mpi_req;
  1562. struct mpi3mr_buf_map *drv_buf_iter = drv_bufs;
  1563. u8 flag, sgl_flags, sgl_flag_eob, sgl_flags_last, last_chain_sgl_flag;
  1564. u16 available_sges, i, sges_needed;
  1565. u32 sge_element_size = sizeof(struct mpi3_sge_common);
  1566. bool chain_used = false;
  1567. sgl_flags = MPI3_SGE_FLAGS_ELEMENT_TYPE_SIMPLE |
  1568. MPI3_SGE_FLAGS_DLAS_SYSTEM;
  1569. sgl_flag_eob = sgl_flags | MPI3_SGE_FLAGS_END_OF_BUFFER;
  1570. sgl_flags_last = sgl_flag_eob | MPI3_SGE_FLAGS_END_OF_LIST;
  1571. last_chain_sgl_flag = MPI3_SGE_FLAGS_ELEMENT_TYPE_LAST_CHAIN |
  1572. MPI3_SGE_FLAGS_DLAS_SYSTEM;
  1573. sges_needed = mpi3mr_total_num_ioctl_sges(drv_bufs, bufcnt);
  1574. if (is_rmc) {
  1575. mpi3mr_add_sg_single(&rmgmt_req->command_sgl,
  1576. sgl_flags_last, drv_buf_iter->kern_buf_len,
  1577. drv_buf_iter->kern_buf_dma);
  1578. sgl = (u8 *)drv_buf_iter->kern_buf +
  1579. drv_buf_iter->bsg_buf_len;
  1580. available_sges = (drv_buf_iter->kern_buf_len -
  1581. drv_buf_iter->bsg_buf_len) / sge_element_size;
  1582. if (sges_needed > available_sges)
  1583. return -1;
  1584. chain_used = true;
  1585. drv_buf_iter++;
  1586. count++;
  1587. if (is_rmr) {
  1588. mpi3mr_add_sg_single(&rmgmt_req->response_sgl,
  1589. sgl_flags_last, drv_buf_iter->kern_buf_len,
  1590. drv_buf_iter->kern_buf_dma);
  1591. drv_buf_iter++;
  1592. count++;
  1593. } else
  1594. mpi3mr_build_zero_len_sge(
  1595. &rmgmt_req->response_sgl);
  1596. if (num_datasges) {
  1597. i = 0;
  1598. goto build_sges;
  1599. }
  1600. } else {
  1601. if (sgl_offset >= MPI3MR_ADMIN_REQ_FRAME_SZ)
  1602. return -1;
  1603. available_sges = (MPI3MR_ADMIN_REQ_FRAME_SZ - sgl_offset) /
  1604. sge_element_size;
  1605. if (!available_sges)
  1606. return -1;
  1607. }
  1608. if (!num_datasges) {
  1609. mpi3mr_build_zero_len_sge(sgl);
  1610. return 0;
  1611. }
  1612. if (mpi_header->function == MPI3_BSG_FUNCTION_SMP_PASSTHROUGH) {
  1613. if ((sges_needed > 2) || (sges_needed > available_sges))
  1614. return -1;
  1615. for (; count < bufcnt; count++, drv_buf_iter++) {
  1616. if (drv_buf_iter->data_dir == DMA_NONE ||
  1617. !drv_buf_iter->num_dma_desc)
  1618. continue;
  1619. mpi3mr_add_sg_single(sgl, sgl_flags_last,
  1620. drv_buf_iter->dma_desc[0].size,
  1621. drv_buf_iter->dma_desc[0].dma_addr);
  1622. sgl += sge_element_size;
  1623. }
  1624. return 0;
  1625. }
  1626. i = 0;
  1627. build_sges:
  1628. for (; count < bufcnt; count++, drv_buf_iter++) {
  1629. if (drv_buf_iter->data_dir == DMA_NONE)
  1630. continue;
  1631. if (!drv_buf_iter->num_dma_desc) {
  1632. if (chain_used && !available_sges)
  1633. return -1;
  1634. if (!chain_used && (available_sges == 1) &&
  1635. (sges_needed > 1))
  1636. goto setup_chain;
  1637. flag = sgl_flag_eob;
  1638. if (num_datasges == 1)
  1639. flag = sgl_flags_last;
  1640. mpi3mr_add_sg_single(sgl, flag, 0, 0);
  1641. sgl += sge_element_size;
  1642. sges_needed--;
  1643. available_sges--;
  1644. num_datasges--;
  1645. continue;
  1646. }
  1647. for (; i < drv_buf_iter->num_dma_desc; i++) {
  1648. if (chain_used && !available_sges)
  1649. return -1;
  1650. if (!chain_used && (available_sges == 1) &&
  1651. (sges_needed > 1))
  1652. goto setup_chain;
  1653. flag = sgl_flags;
  1654. if (i == (drv_buf_iter->num_dma_desc - 1)) {
  1655. if (num_datasges == 1)
  1656. flag = sgl_flags_last;
  1657. else
  1658. flag = sgl_flag_eob;
  1659. }
  1660. mpi3mr_add_sg_single(sgl, flag,
  1661. drv_buf_iter->dma_desc[i].size,
  1662. drv_buf_iter->dma_desc[i].dma_addr);
  1663. sgl += sge_element_size;
  1664. available_sges--;
  1665. sges_needed--;
  1666. }
  1667. num_datasges--;
  1668. i = 0;
  1669. }
  1670. return 0;
  1671. setup_chain:
  1672. available_sges = mrioc->ioctl_chain_sge.size / sge_element_size;
  1673. if (sges_needed > available_sges)
  1674. return -1;
  1675. mpi3mr_add_sg_single(sgl, last_chain_sgl_flag,
  1676. (sges_needed * sge_element_size),
  1677. mrioc->ioctl_chain_sge.dma_addr);
  1678. memset(mrioc->ioctl_chain_sge.addr, 0, mrioc->ioctl_chain_sge.size);
  1679. sgl = (u8 *)mrioc->ioctl_chain_sge.addr;
  1680. chain_used = true;
  1681. goto build_sges;
  1682. }
  1683. /**
  1684. * mpi3mr_get_nvme_data_fmt - returns the NVMe data format
  1685. * @nvme_encap_request: NVMe encapsulated MPI request
  1686. *
  1687. * This function returns the type of the data format specified
  1688. * in user provided NVMe command in NVMe encapsulated request.
  1689. *
  1690. * Return: Data format of the NVMe command (PRP/SGL etc)
  1691. */
  1692. static unsigned int mpi3mr_get_nvme_data_fmt(
  1693. struct mpi3_nvme_encapsulated_request *nvme_encap_request)
  1694. {
  1695. u8 format = 0;
  1696. format = ((nvme_encap_request->command[0] & 0xc000) >> 14);
  1697. return format;
  1698. }
  1699. /**
  1700. * mpi3mr_build_nvme_sgl - SGL constructor for NVME
  1701. * encapsulated request
  1702. * @mrioc: Adapter instance reference
  1703. * @nvme_encap_request: NVMe encapsulated MPI request
  1704. * @drv_bufs: DMA address of the buffers to be placed in sgl
  1705. * @bufcnt: Number of DMA buffers
  1706. *
  1707. * This function places the DMA address of the given buffers in
  1708. * proper format as SGEs in the given NVMe encapsulated request.
  1709. *
  1710. * Return: 0 on success, -1 on failure
  1711. */
  1712. static int mpi3mr_build_nvme_sgl(struct mpi3mr_ioc *mrioc,
  1713. struct mpi3_nvme_encapsulated_request *nvme_encap_request,
  1714. struct mpi3mr_buf_map *drv_bufs, u8 bufcnt)
  1715. {
  1716. struct mpi3mr_nvme_pt_sge *nvme_sgl;
  1717. __le64 sgl_dma;
  1718. u8 count;
  1719. size_t length = 0;
  1720. u16 available_sges = 0, i;
  1721. u32 sge_element_size = sizeof(struct mpi3mr_nvme_pt_sge);
  1722. struct mpi3mr_buf_map *drv_buf_iter = drv_bufs;
  1723. u64 sgemod_mask = ((u64)((mrioc->facts.sge_mod_mask) <<
  1724. mrioc->facts.sge_mod_shift) << 32);
  1725. u64 sgemod_val = ((u64)(mrioc->facts.sge_mod_value) <<
  1726. mrioc->facts.sge_mod_shift) << 32;
  1727. u32 size;
  1728. nvme_sgl = (struct mpi3mr_nvme_pt_sge *)
  1729. ((u8 *)(nvme_encap_request->command) + MPI3MR_NVME_CMD_SGL_OFFSET);
  1730. /*
  1731. * Not all commands require a data transfer. If no data, just return
  1732. * without constructing any sgl.
  1733. */
  1734. for (count = 0; count < bufcnt; count++, drv_buf_iter++) {
  1735. if (drv_buf_iter->data_dir == DMA_NONE)
  1736. continue;
  1737. length = drv_buf_iter->kern_buf_len;
  1738. break;
  1739. }
  1740. if (!length || !drv_buf_iter->num_dma_desc)
  1741. return 0;
  1742. if (drv_buf_iter->num_dma_desc == 1) {
  1743. available_sges = 1;
  1744. goto build_sges;
  1745. }
  1746. sgl_dma = cpu_to_le64(mrioc->ioctl_chain_sge.dma_addr);
  1747. if (sgl_dma & sgemod_mask) {
  1748. dprint_bsg_err(mrioc,
  1749. "%s: SGL chain address collides with SGE modifier\n",
  1750. __func__);
  1751. return -1;
  1752. }
  1753. sgl_dma &= ~sgemod_mask;
  1754. sgl_dma |= sgemod_val;
  1755. memset(mrioc->ioctl_chain_sge.addr, 0, mrioc->ioctl_chain_sge.size);
  1756. available_sges = mrioc->ioctl_chain_sge.size / sge_element_size;
  1757. if (available_sges < drv_buf_iter->num_dma_desc)
  1758. return -1;
  1759. memset(nvme_sgl, 0, sizeof(struct mpi3mr_nvme_pt_sge));
  1760. nvme_sgl->base_addr = sgl_dma;
  1761. size = drv_buf_iter->num_dma_desc * sizeof(struct mpi3mr_nvme_pt_sge);
  1762. nvme_sgl->length = cpu_to_le32(size);
  1763. nvme_sgl->type = MPI3MR_NVMESGL_LAST_SEGMENT;
  1764. nvme_sgl = (struct mpi3mr_nvme_pt_sge *)mrioc->ioctl_chain_sge.addr;
  1765. build_sges:
  1766. for (i = 0; i < drv_buf_iter->num_dma_desc; i++) {
  1767. sgl_dma = cpu_to_le64(drv_buf_iter->dma_desc[i].dma_addr);
  1768. if (sgl_dma & sgemod_mask) {
  1769. dprint_bsg_err(mrioc,
  1770. "%s: SGL address collides with SGE modifier\n",
  1771. __func__);
  1772. return -1;
  1773. }
  1774. sgl_dma &= ~sgemod_mask;
  1775. sgl_dma |= sgemod_val;
  1776. nvme_sgl->base_addr = sgl_dma;
  1777. nvme_sgl->length = cpu_to_le32(drv_buf_iter->dma_desc[i].size);
  1778. nvme_sgl->type = MPI3MR_NVMESGL_DATA_SEGMENT;
  1779. nvme_sgl++;
  1780. available_sges--;
  1781. }
  1782. return 0;
  1783. }
  1784. /**
  1785. * mpi3mr_build_nvme_prp - PRP constructor for NVME
  1786. * encapsulated request
  1787. * @mrioc: Adapter instance reference
  1788. * @nvme_encap_request: NVMe encapsulated MPI request
  1789. * @drv_bufs: DMA address of the buffers to be placed in SGL
  1790. * @bufcnt: Number of DMA buffers
  1791. *
  1792. * This function places the DMA address of the given buffers in
  1793. * proper format as PRP entries in the given NVMe encapsulated
  1794. * request.
  1795. *
  1796. * Return: 0 on success, -1 on failure
  1797. */
  1798. static int mpi3mr_build_nvme_prp(struct mpi3mr_ioc *mrioc,
  1799. struct mpi3_nvme_encapsulated_request *nvme_encap_request,
  1800. struct mpi3mr_buf_map *drv_bufs, u8 bufcnt)
  1801. {
  1802. int prp_size = MPI3MR_NVME_PRP_SIZE;
  1803. __le64 *prp_entry, *prp1_entry, *prp2_entry;
  1804. __le64 *prp_page;
  1805. dma_addr_t prp_entry_dma, prp_page_dma, dma_addr;
  1806. u32 offset, entry_len, dev_pgsz;
  1807. u32 page_mask_result, page_mask;
  1808. size_t length = 0, desc_len;
  1809. u8 count;
  1810. struct mpi3mr_buf_map *drv_buf_iter = drv_bufs;
  1811. u64 sgemod_mask = ((u64)((mrioc->facts.sge_mod_mask) <<
  1812. mrioc->facts.sge_mod_shift) << 32);
  1813. u64 sgemod_val = ((u64)(mrioc->facts.sge_mod_value) <<
  1814. mrioc->facts.sge_mod_shift) << 32;
  1815. u16 dev_handle = nvme_encap_request->dev_handle;
  1816. struct mpi3mr_tgt_dev *tgtdev;
  1817. u16 desc_count = 0;
  1818. tgtdev = mpi3mr_get_tgtdev_by_handle(mrioc, dev_handle);
  1819. if (!tgtdev) {
  1820. dprint_bsg_err(mrioc, "%s: invalid device handle 0x%04x\n",
  1821. __func__, dev_handle);
  1822. return -1;
  1823. }
  1824. if (tgtdev->dev_spec.pcie_inf.pgsz == 0) {
  1825. dprint_bsg_err(mrioc,
  1826. "%s: NVMe device page size is zero for handle 0x%04x\n",
  1827. __func__, dev_handle);
  1828. mpi3mr_tgtdev_put(tgtdev);
  1829. return -1;
  1830. }
  1831. dev_pgsz = 1 << (tgtdev->dev_spec.pcie_inf.pgsz);
  1832. mpi3mr_tgtdev_put(tgtdev);
  1833. page_mask = dev_pgsz - 1;
  1834. if (dev_pgsz > MPI3MR_IOCTL_SGE_SIZE) {
  1835. dprint_bsg_err(mrioc,
  1836. "%s: NVMe device page size(%d) is greater than ioctl data sge size(%d) for handle 0x%04x\n",
  1837. __func__, dev_pgsz, MPI3MR_IOCTL_SGE_SIZE, dev_handle);
  1838. return -1;
  1839. }
  1840. if (MPI3MR_IOCTL_SGE_SIZE % dev_pgsz) {
  1841. dprint_bsg_err(mrioc,
  1842. "%s: ioctl data sge size(%d) is not a multiple of NVMe device page size(%d) for handle 0x%04x\n",
  1843. __func__, MPI3MR_IOCTL_SGE_SIZE, dev_pgsz, dev_handle);
  1844. return -1;
  1845. }
  1846. /*
  1847. * Not all commands require a data transfer. If no data, just return
  1848. * without constructing any PRP.
  1849. */
  1850. for (count = 0; count < bufcnt; count++, drv_buf_iter++) {
  1851. if (drv_buf_iter->data_dir == DMA_NONE)
  1852. continue;
  1853. length = drv_buf_iter->kern_buf_len;
  1854. break;
  1855. }
  1856. if (!length || !drv_buf_iter->num_dma_desc)
  1857. return 0;
  1858. for (count = 0; count < drv_buf_iter->num_dma_desc; count++) {
  1859. dma_addr = drv_buf_iter->dma_desc[count].dma_addr;
  1860. if (dma_addr & page_mask) {
  1861. dprint_bsg_err(mrioc,
  1862. "%s:dma_addr %pad is not aligned with page size 0x%x\n",
  1863. __func__, &dma_addr, dev_pgsz);
  1864. return -1;
  1865. }
  1866. }
  1867. dma_addr = drv_buf_iter->dma_desc[0].dma_addr;
  1868. desc_len = drv_buf_iter->dma_desc[0].size;
  1869. mrioc->prp_sz = 0;
  1870. mrioc->prp_list_virt = dma_alloc_coherent(&mrioc->pdev->dev,
  1871. dev_pgsz, &mrioc->prp_list_dma, GFP_KERNEL);
  1872. if (!mrioc->prp_list_virt)
  1873. return -1;
  1874. mrioc->prp_sz = dev_pgsz;
  1875. /*
  1876. * Set pointers to PRP1 and PRP2, which are in the NVMe command.
  1877. * PRP1 is located at a 24 byte offset from the start of the NVMe
  1878. * command. Then set the current PRP entry pointer to PRP1.
  1879. */
  1880. prp1_entry = (__le64 *)((u8 *)(nvme_encap_request->command) +
  1881. MPI3MR_NVME_CMD_PRP1_OFFSET);
  1882. prp2_entry = (__le64 *)((u8 *)(nvme_encap_request->command) +
  1883. MPI3MR_NVME_CMD_PRP2_OFFSET);
  1884. prp_entry = prp1_entry;
  1885. /*
  1886. * For the PRP entries, use the specially allocated buffer of
  1887. * contiguous memory.
  1888. */
  1889. prp_page = (__le64 *)mrioc->prp_list_virt;
  1890. prp_page_dma = mrioc->prp_list_dma;
  1891. /*
  1892. * Check if we are within 1 entry of a page boundary we don't
  1893. * want our first entry to be a PRP List entry.
  1894. */
  1895. page_mask_result = (uintptr_t)((u8 *)prp_page + prp_size) & page_mask;
  1896. if (!page_mask_result) {
  1897. dprint_bsg_err(mrioc, "%s: PRP page is not page aligned\n",
  1898. __func__);
  1899. goto err_out;
  1900. }
  1901. /*
  1902. * Set PRP physical pointer, which initially points to the current PRP
  1903. * DMA memory page.
  1904. */
  1905. prp_entry_dma = prp_page_dma;
  1906. /* Loop while the length is not zero. */
  1907. while (length) {
  1908. page_mask_result = (prp_entry_dma + prp_size) & page_mask;
  1909. if (!page_mask_result && (length > dev_pgsz)) {
  1910. dprint_bsg_err(mrioc,
  1911. "%s: single PRP page is not sufficient\n",
  1912. __func__);
  1913. goto err_out;
  1914. }
  1915. /* Need to handle if entry will be part of a page. */
  1916. offset = dma_addr & page_mask;
  1917. entry_len = dev_pgsz - offset;
  1918. if (prp_entry == prp1_entry) {
  1919. /*
  1920. * Must fill in the first PRP pointer (PRP1) before
  1921. * moving on.
  1922. */
  1923. *prp1_entry = cpu_to_le64(dma_addr);
  1924. if (*prp1_entry & sgemod_mask) {
  1925. dprint_bsg_err(mrioc,
  1926. "%s: PRP1 address collides with SGE modifier\n",
  1927. __func__);
  1928. goto err_out;
  1929. }
  1930. *prp1_entry &= ~sgemod_mask;
  1931. *prp1_entry |= sgemod_val;
  1932. /*
  1933. * Now point to the second PRP entry within the
  1934. * command (PRP2).
  1935. */
  1936. prp_entry = prp2_entry;
  1937. } else if (prp_entry == prp2_entry) {
  1938. /*
  1939. * Should the PRP2 entry be a PRP List pointer or just
  1940. * a regular PRP pointer? If there is more than one
  1941. * more page of data, must use a PRP List pointer.
  1942. */
  1943. if (length > dev_pgsz) {
  1944. /*
  1945. * PRP2 will contain a PRP List pointer because
  1946. * more PRP's are needed with this command. The
  1947. * list will start at the beginning of the
  1948. * contiguous buffer.
  1949. */
  1950. *prp2_entry = cpu_to_le64(prp_entry_dma);
  1951. if (*prp2_entry & sgemod_mask) {
  1952. dprint_bsg_err(mrioc,
  1953. "%s: PRP list address collides with SGE modifier\n",
  1954. __func__);
  1955. goto err_out;
  1956. }
  1957. *prp2_entry &= ~sgemod_mask;
  1958. *prp2_entry |= sgemod_val;
  1959. /*
  1960. * The next PRP Entry will be the start of the
  1961. * first PRP List.
  1962. */
  1963. prp_entry = prp_page;
  1964. continue;
  1965. } else {
  1966. /*
  1967. * After this, the PRP Entries are complete.
  1968. * This command uses 2 PRP's and no PRP list.
  1969. */
  1970. *prp2_entry = cpu_to_le64(dma_addr);
  1971. if (*prp2_entry & sgemod_mask) {
  1972. dprint_bsg_err(mrioc,
  1973. "%s: PRP2 collides with SGE modifier\n",
  1974. __func__);
  1975. goto err_out;
  1976. }
  1977. *prp2_entry &= ~sgemod_mask;
  1978. *prp2_entry |= sgemod_val;
  1979. }
  1980. } else {
  1981. /*
  1982. * Put entry in list and bump the addresses.
  1983. *
  1984. * After PRP1 and PRP2 are filled in, this will fill in
  1985. * all remaining PRP entries in a PRP List, one per
  1986. * each time through the loop.
  1987. */
  1988. *prp_entry = cpu_to_le64(dma_addr);
  1989. if (*prp_entry & sgemod_mask) {
  1990. dprint_bsg_err(mrioc,
  1991. "%s: PRP address collides with SGE modifier\n",
  1992. __func__);
  1993. goto err_out;
  1994. }
  1995. *prp_entry &= ~sgemod_mask;
  1996. *prp_entry |= sgemod_val;
  1997. prp_entry++;
  1998. prp_entry_dma += prp_size;
  1999. }
  2000. /* decrement length accounting for last partial page. */
  2001. if (entry_len >= length) {
  2002. length = 0;
  2003. } else {
  2004. if (entry_len <= desc_len) {
  2005. dma_addr += entry_len;
  2006. desc_len -= entry_len;
  2007. }
  2008. if (!desc_len) {
  2009. if ((++desc_count) >=
  2010. drv_buf_iter->num_dma_desc) {
  2011. dprint_bsg_err(mrioc,
  2012. "%s: Invalid len %zd while building PRP\n",
  2013. __func__, length);
  2014. goto err_out;
  2015. }
  2016. dma_addr =
  2017. drv_buf_iter->dma_desc[desc_count].dma_addr;
  2018. desc_len =
  2019. drv_buf_iter->dma_desc[desc_count].size;
  2020. }
  2021. length -= entry_len;
  2022. }
  2023. }
  2024. return 0;
  2025. err_out:
  2026. if (mrioc->prp_list_virt) {
  2027. dma_free_coherent(&mrioc->pdev->dev, mrioc->prp_sz,
  2028. mrioc->prp_list_virt, mrioc->prp_list_dma);
  2029. mrioc->prp_list_virt = NULL;
  2030. }
  2031. return -1;
  2032. }
  2033. /**
  2034. * mpi3mr_map_data_buffer_dma - build dma descriptors for data
  2035. * buffers
  2036. * @mrioc: Adapter instance reference
  2037. * @drv_buf: buffer map descriptor
  2038. * @desc_count: Number of already consumed dma descriptors
  2039. *
  2040. * This function computes how many pre-allocated DMA descriptors
  2041. * are required for the given data buffer and if those number of
  2042. * descriptors are free, then setup the mapping of the scattered
  2043. * DMA address to the given data buffer, if the data direction
  2044. * of the buffer is DMA_TO_DEVICE then the actual data is copied to
  2045. * the DMA buffers
  2046. *
  2047. * Return: 0 on success, -1 on failure
  2048. */
  2049. static int mpi3mr_map_data_buffer_dma(struct mpi3mr_ioc *mrioc,
  2050. struct mpi3mr_buf_map *drv_buf,
  2051. u16 desc_count)
  2052. {
  2053. u16 i, needed_desc = drv_buf->kern_buf_len / MPI3MR_IOCTL_SGE_SIZE;
  2054. u32 buf_len = drv_buf->kern_buf_len, copied_len = 0;
  2055. if (drv_buf->kern_buf_len % MPI3MR_IOCTL_SGE_SIZE)
  2056. needed_desc++;
  2057. if ((needed_desc + desc_count) > MPI3MR_NUM_IOCTL_SGE) {
  2058. dprint_bsg_err(mrioc, "%s: DMA descriptor mapping error %d:%d:%d\n",
  2059. __func__, needed_desc, desc_count, MPI3MR_NUM_IOCTL_SGE);
  2060. return -1;
  2061. }
  2062. drv_buf->dma_desc = kzalloc(sizeof(*drv_buf->dma_desc) * needed_desc,
  2063. GFP_KERNEL);
  2064. if (!drv_buf->dma_desc)
  2065. return -1;
  2066. for (i = 0; i < needed_desc; i++, desc_count++) {
  2067. drv_buf->dma_desc[i].addr = mrioc->ioctl_sge[desc_count].addr;
  2068. drv_buf->dma_desc[i].dma_addr =
  2069. mrioc->ioctl_sge[desc_count].dma_addr;
  2070. if (buf_len < mrioc->ioctl_sge[desc_count].size)
  2071. drv_buf->dma_desc[i].size = buf_len;
  2072. else
  2073. drv_buf->dma_desc[i].size =
  2074. mrioc->ioctl_sge[desc_count].size;
  2075. buf_len -= drv_buf->dma_desc[i].size;
  2076. memset(drv_buf->dma_desc[i].addr, 0,
  2077. mrioc->ioctl_sge[desc_count].size);
  2078. if (drv_buf->data_dir == DMA_TO_DEVICE) {
  2079. memcpy(drv_buf->dma_desc[i].addr,
  2080. drv_buf->bsg_buf + copied_len,
  2081. drv_buf->dma_desc[i].size);
  2082. copied_len += drv_buf->dma_desc[i].size;
  2083. }
  2084. }
  2085. drv_buf->num_dma_desc = needed_desc;
  2086. return 0;
  2087. }
  2088. /**
  2089. * mpi3mr_bsg_process_mpt_cmds - MPI Pass through BSG handler
  2090. * @job: BSG job reference
  2091. *
  2092. * This function is the top level handler for MPI Pass through
  2093. * command, this does basic validation of the input data buffers,
  2094. * identifies the given buffer types and MPI command, allocates
  2095. * DMAable memory for user given buffers, construstcs SGL
  2096. * properly and passes the command to the firmware.
  2097. *
  2098. * Once the MPI command is completed the driver copies the data
  2099. * if any and reply, sense information to user provided buffers.
  2100. * If the command is timed out then issues controller reset
  2101. * prior to returning.
  2102. *
  2103. * Return: 0 on success and proper error codes on failure
  2104. */
  2105. static long mpi3mr_bsg_process_mpt_cmds(struct bsg_job *job)
  2106. {
  2107. long rval = -EINVAL;
  2108. struct mpi3mr_ioc *mrioc = NULL;
  2109. u8 *mpi_req = NULL, *sense_buff_k = NULL;
  2110. u8 mpi_msg_size = 0;
  2111. struct mpi3mr_bsg_packet *bsg_req = NULL;
  2112. struct mpi3mr_bsg_mptcmd *karg;
  2113. struct mpi3mr_buf_entry *buf_entries = NULL;
  2114. struct mpi3mr_buf_map *drv_bufs = NULL, *drv_buf_iter = NULL;
  2115. u8 count, bufcnt = 0, is_rmcb = 0, is_rmrb = 0;
  2116. u8 din_cnt = 0, dout_cnt = 0;
  2117. u8 invalid_be = 0, erb_offset = 0xFF, mpirep_offset = 0xFF;
  2118. u8 block_io = 0, nvme_fmt = 0, resp_code = 0;
  2119. struct mpi3_request_header *mpi_header = NULL;
  2120. struct mpi3_status_reply_descriptor *status_desc;
  2121. struct mpi3_scsi_task_mgmt_request *tm_req;
  2122. u32 erbsz = MPI3MR_SENSE_BUF_SZ, tmplen;
  2123. u16 dev_handle;
  2124. struct mpi3mr_tgt_dev *tgtdev;
  2125. struct mpi3mr_stgt_priv_data *stgt_priv = NULL;
  2126. struct mpi3mr_bsg_in_reply_buf *bsg_reply_buf = NULL;
  2127. u32 din_size = 0, dout_size = 0;
  2128. u8 *din_buf = NULL, *dout_buf = NULL;
  2129. u8 *sgl_iter = NULL, *sgl_din_iter = NULL, *sgl_dout_iter = NULL;
  2130. u16 rmc_size = 0, desc_count = 0;
  2131. bsg_req = job->request;
  2132. karg = (struct mpi3mr_bsg_mptcmd *)&bsg_req->cmd.mptcmd;
  2133. mrioc = mpi3mr_bsg_verify_adapter(karg->mrioc_id);
  2134. if (!mrioc)
  2135. return -ENODEV;
  2136. if (mutex_lock_interruptible(&mrioc->bsg_cmds.mutex))
  2137. return -ERESTARTSYS;
  2138. if (mrioc->bsg_cmds.state & MPI3MR_CMD_PENDING) {
  2139. dprint_bsg_err(mrioc, "%s: command is in use\n", __func__);
  2140. mutex_unlock(&mrioc->bsg_cmds.mutex);
  2141. return -EAGAIN;
  2142. }
  2143. if (!mrioc->ioctl_sges_allocated) {
  2144. dprint_bsg_err(mrioc, "%s: DMA memory was not allocated\n",
  2145. __func__);
  2146. return -ENOMEM;
  2147. }
  2148. if (karg->timeout < MPI3MR_APP_DEFAULT_TIMEOUT)
  2149. karg->timeout = MPI3MR_APP_DEFAULT_TIMEOUT;
  2150. mpi_req = kzalloc(MPI3MR_ADMIN_REQ_FRAME_SZ, GFP_KERNEL);
  2151. if (!mpi_req) {
  2152. mutex_unlock(&mrioc->bsg_cmds.mutex);
  2153. return -ENOMEM;
  2154. }
  2155. mpi_header = (struct mpi3_request_header *)mpi_req;
  2156. bufcnt = karg->buf_entry_list.num_of_entries;
  2157. drv_bufs = kzalloc((sizeof(*drv_bufs) * bufcnt), GFP_KERNEL);
  2158. if (!drv_bufs) {
  2159. mutex_unlock(&mrioc->bsg_cmds.mutex);
  2160. rval = -ENOMEM;
  2161. goto out;
  2162. }
  2163. dout_buf = kzalloc(job->request_payload.payload_len,
  2164. GFP_KERNEL);
  2165. if (!dout_buf) {
  2166. mutex_unlock(&mrioc->bsg_cmds.mutex);
  2167. rval = -ENOMEM;
  2168. goto out;
  2169. }
  2170. din_buf = kzalloc(job->reply_payload.payload_len,
  2171. GFP_KERNEL);
  2172. if (!din_buf) {
  2173. mutex_unlock(&mrioc->bsg_cmds.mutex);
  2174. rval = -ENOMEM;
  2175. goto out;
  2176. }
  2177. sg_copy_to_buffer(job->request_payload.sg_list,
  2178. job->request_payload.sg_cnt,
  2179. dout_buf, job->request_payload.payload_len);
  2180. buf_entries = karg->buf_entry_list.buf_entry;
  2181. sgl_din_iter = din_buf;
  2182. sgl_dout_iter = dout_buf;
  2183. drv_buf_iter = drv_bufs;
  2184. for (count = 0; count < bufcnt; count++, buf_entries++, drv_buf_iter++) {
  2185. switch (buf_entries->buf_type) {
  2186. case MPI3MR_BSG_BUFTYPE_RAIDMGMT_CMD:
  2187. sgl_iter = sgl_dout_iter;
  2188. sgl_dout_iter += buf_entries->buf_len;
  2189. drv_buf_iter->data_dir = DMA_TO_DEVICE;
  2190. is_rmcb = 1;
  2191. if ((count != 0) || !buf_entries->buf_len)
  2192. invalid_be = 1;
  2193. break;
  2194. case MPI3MR_BSG_BUFTYPE_RAIDMGMT_RESP:
  2195. sgl_iter = sgl_din_iter;
  2196. sgl_din_iter += buf_entries->buf_len;
  2197. drv_buf_iter->data_dir = DMA_FROM_DEVICE;
  2198. is_rmrb = 1;
  2199. if (count != 1 || !is_rmcb || !buf_entries->buf_len)
  2200. invalid_be = 1;
  2201. break;
  2202. case MPI3MR_BSG_BUFTYPE_DATA_IN:
  2203. sgl_iter = sgl_din_iter;
  2204. sgl_din_iter += buf_entries->buf_len;
  2205. drv_buf_iter->data_dir = DMA_FROM_DEVICE;
  2206. din_cnt++;
  2207. din_size += buf_entries->buf_len;
  2208. if ((din_cnt > 1) && !is_rmcb)
  2209. invalid_be = 1;
  2210. break;
  2211. case MPI3MR_BSG_BUFTYPE_DATA_OUT:
  2212. sgl_iter = sgl_dout_iter;
  2213. sgl_dout_iter += buf_entries->buf_len;
  2214. drv_buf_iter->data_dir = DMA_TO_DEVICE;
  2215. dout_cnt++;
  2216. dout_size += buf_entries->buf_len;
  2217. if ((dout_cnt > 1) && !is_rmcb)
  2218. invalid_be = 1;
  2219. break;
  2220. case MPI3MR_BSG_BUFTYPE_MPI_REPLY:
  2221. sgl_iter = sgl_din_iter;
  2222. sgl_din_iter += buf_entries->buf_len;
  2223. drv_buf_iter->data_dir = DMA_NONE;
  2224. mpirep_offset = count;
  2225. if (!buf_entries->buf_len)
  2226. invalid_be = 1;
  2227. break;
  2228. case MPI3MR_BSG_BUFTYPE_ERR_RESPONSE:
  2229. sgl_iter = sgl_din_iter;
  2230. sgl_din_iter += buf_entries->buf_len;
  2231. drv_buf_iter->data_dir = DMA_NONE;
  2232. erb_offset = count;
  2233. if (!buf_entries->buf_len)
  2234. invalid_be = 1;
  2235. break;
  2236. case MPI3MR_BSG_BUFTYPE_MPI_REQUEST:
  2237. sgl_iter = sgl_dout_iter;
  2238. sgl_dout_iter += buf_entries->buf_len;
  2239. drv_buf_iter->data_dir = DMA_NONE;
  2240. mpi_msg_size = buf_entries->buf_len;
  2241. if ((!mpi_msg_size || (mpi_msg_size % 4)) ||
  2242. (mpi_msg_size > MPI3MR_ADMIN_REQ_FRAME_SZ)) {
  2243. dprint_bsg_err(mrioc, "%s: invalid MPI message size\n",
  2244. __func__);
  2245. mutex_unlock(&mrioc->bsg_cmds.mutex);
  2246. rval = -EINVAL;
  2247. goto out;
  2248. }
  2249. memcpy(mpi_req, sgl_iter, buf_entries->buf_len);
  2250. break;
  2251. default:
  2252. invalid_be = 1;
  2253. break;
  2254. }
  2255. if (invalid_be) {
  2256. dprint_bsg_err(mrioc, "%s: invalid buffer entries passed\n",
  2257. __func__);
  2258. mutex_unlock(&mrioc->bsg_cmds.mutex);
  2259. rval = -EINVAL;
  2260. goto out;
  2261. }
  2262. if (sgl_dout_iter > (dout_buf + job->request_payload.payload_len)) {
  2263. dprint_bsg_err(mrioc, "%s: data_out buffer length mismatch\n",
  2264. __func__);
  2265. mutex_unlock(&mrioc->bsg_cmds.mutex);
  2266. rval = -EINVAL;
  2267. goto out;
  2268. }
  2269. if (sgl_din_iter > (din_buf + job->reply_payload.payload_len)) {
  2270. dprint_bsg_err(mrioc, "%s: data_in buffer length mismatch\n",
  2271. __func__);
  2272. mutex_unlock(&mrioc->bsg_cmds.mutex);
  2273. rval = -EINVAL;
  2274. goto out;
  2275. }
  2276. drv_buf_iter->bsg_buf = sgl_iter;
  2277. drv_buf_iter->bsg_buf_len = buf_entries->buf_len;
  2278. }
  2279. if (is_rmcb && ((din_size + dout_size) > MPI3MR_MAX_APP_XFER_SIZE)) {
  2280. dprint_bsg_err(mrioc, "%s:%d: invalid data transfer size passed for function 0x%x din_size = %d, dout_size = %d\n",
  2281. __func__, __LINE__, mpi_header->function, din_size,
  2282. dout_size);
  2283. mutex_unlock(&mrioc->bsg_cmds.mutex);
  2284. rval = -EINVAL;
  2285. goto out;
  2286. }
  2287. if (din_size > MPI3MR_MAX_APP_XFER_SIZE) {
  2288. dprint_bsg_err(mrioc,
  2289. "%s:%d: invalid data transfer size passed for function 0x%x din_size=%d\n",
  2290. __func__, __LINE__, mpi_header->function, din_size);
  2291. mutex_unlock(&mrioc->bsg_cmds.mutex);
  2292. rval = -EINVAL;
  2293. goto out;
  2294. }
  2295. if (dout_size > MPI3MR_MAX_APP_XFER_SIZE) {
  2296. dprint_bsg_err(mrioc,
  2297. "%s:%d: invalid data transfer size passed for function 0x%x dout_size = %d\n",
  2298. __func__, __LINE__, mpi_header->function, dout_size);
  2299. mutex_unlock(&mrioc->bsg_cmds.mutex);
  2300. rval = -EINVAL;
  2301. goto out;
  2302. }
  2303. if (mpi_header->function == MPI3_BSG_FUNCTION_SMP_PASSTHROUGH) {
  2304. if (din_size > MPI3MR_IOCTL_SGE_SIZE ||
  2305. dout_size > MPI3MR_IOCTL_SGE_SIZE) {
  2306. dprint_bsg_err(mrioc, "%s:%d: invalid message size passed:%d:%d:%d:%d\n",
  2307. __func__, __LINE__, din_cnt, dout_cnt, din_size,
  2308. dout_size);
  2309. mutex_unlock(&mrioc->bsg_cmds.mutex);
  2310. rval = -EINVAL;
  2311. goto out;
  2312. }
  2313. }
  2314. drv_buf_iter = drv_bufs;
  2315. for (count = 0; count < bufcnt; count++, drv_buf_iter++) {
  2316. if (drv_buf_iter->data_dir == DMA_NONE)
  2317. continue;
  2318. drv_buf_iter->kern_buf_len = drv_buf_iter->bsg_buf_len;
  2319. if (is_rmcb && !count) {
  2320. drv_buf_iter->kern_buf_len =
  2321. mrioc->ioctl_chain_sge.size;
  2322. drv_buf_iter->kern_buf =
  2323. mrioc->ioctl_chain_sge.addr;
  2324. drv_buf_iter->kern_buf_dma =
  2325. mrioc->ioctl_chain_sge.dma_addr;
  2326. drv_buf_iter->dma_desc = NULL;
  2327. drv_buf_iter->num_dma_desc = 0;
  2328. memset(drv_buf_iter->kern_buf, 0,
  2329. drv_buf_iter->kern_buf_len);
  2330. tmplen = min(drv_buf_iter->kern_buf_len,
  2331. drv_buf_iter->bsg_buf_len);
  2332. rmc_size = tmplen;
  2333. memcpy(drv_buf_iter->kern_buf, drv_buf_iter->bsg_buf, tmplen);
  2334. } else if (is_rmrb && (count == 1)) {
  2335. drv_buf_iter->kern_buf_len =
  2336. mrioc->ioctl_resp_sge.size;
  2337. drv_buf_iter->kern_buf =
  2338. mrioc->ioctl_resp_sge.addr;
  2339. drv_buf_iter->kern_buf_dma =
  2340. mrioc->ioctl_resp_sge.dma_addr;
  2341. drv_buf_iter->dma_desc = NULL;
  2342. drv_buf_iter->num_dma_desc = 0;
  2343. memset(drv_buf_iter->kern_buf, 0,
  2344. drv_buf_iter->kern_buf_len);
  2345. tmplen = min(drv_buf_iter->kern_buf_len,
  2346. drv_buf_iter->bsg_buf_len);
  2347. drv_buf_iter->kern_buf_len = tmplen;
  2348. memset(drv_buf_iter->bsg_buf, 0,
  2349. drv_buf_iter->bsg_buf_len);
  2350. } else {
  2351. if (!drv_buf_iter->kern_buf_len)
  2352. continue;
  2353. if (mpi3mr_map_data_buffer_dma(mrioc, drv_buf_iter, desc_count)) {
  2354. rval = -ENOMEM;
  2355. mutex_unlock(&mrioc->bsg_cmds.mutex);
  2356. dprint_bsg_err(mrioc, "%s:%d: mapping data buffers failed\n",
  2357. __func__, __LINE__);
  2358. goto out;
  2359. }
  2360. desc_count += drv_buf_iter->num_dma_desc;
  2361. }
  2362. }
  2363. if (erb_offset != 0xFF) {
  2364. sense_buff_k = kzalloc(erbsz, GFP_KERNEL);
  2365. if (!sense_buff_k) {
  2366. rval = -ENOMEM;
  2367. mutex_unlock(&mrioc->bsg_cmds.mutex);
  2368. goto out;
  2369. }
  2370. }
  2371. if (mrioc->unrecoverable) {
  2372. dprint_bsg_err(mrioc, "%s: unrecoverable controller\n",
  2373. __func__);
  2374. rval = -EFAULT;
  2375. mutex_unlock(&mrioc->bsg_cmds.mutex);
  2376. goto out;
  2377. }
  2378. if (mrioc->reset_in_progress) {
  2379. dprint_bsg_err(mrioc, "%s: reset in progress\n", __func__);
  2380. rval = -EAGAIN;
  2381. mutex_unlock(&mrioc->bsg_cmds.mutex);
  2382. goto out;
  2383. }
  2384. if (mrioc->stop_bsgs || mrioc->block_on_pci_err) {
  2385. dprint_bsg_err(mrioc, "%s: bsgs are blocked\n", __func__);
  2386. rval = -EAGAIN;
  2387. mutex_unlock(&mrioc->bsg_cmds.mutex);
  2388. goto out;
  2389. }
  2390. if (mpi_header->function == MPI3_BSG_FUNCTION_NVME_ENCAPSULATED) {
  2391. nvme_fmt = mpi3mr_get_nvme_data_fmt(
  2392. (struct mpi3_nvme_encapsulated_request *)mpi_req);
  2393. if (nvme_fmt == MPI3MR_NVME_DATA_FORMAT_PRP) {
  2394. if (mpi3mr_build_nvme_prp(mrioc,
  2395. (struct mpi3_nvme_encapsulated_request *)mpi_req,
  2396. drv_bufs, bufcnt)) {
  2397. rval = -ENOMEM;
  2398. mutex_unlock(&mrioc->bsg_cmds.mutex);
  2399. goto out;
  2400. }
  2401. } else if (nvme_fmt == MPI3MR_NVME_DATA_FORMAT_SGL1 ||
  2402. nvme_fmt == MPI3MR_NVME_DATA_FORMAT_SGL2) {
  2403. if (mpi3mr_build_nvme_sgl(mrioc,
  2404. (struct mpi3_nvme_encapsulated_request *)mpi_req,
  2405. drv_bufs, bufcnt)) {
  2406. rval = -EINVAL;
  2407. mutex_unlock(&mrioc->bsg_cmds.mutex);
  2408. goto out;
  2409. }
  2410. } else {
  2411. dprint_bsg_err(mrioc,
  2412. "%s:invalid NVMe command format\n", __func__);
  2413. rval = -EINVAL;
  2414. mutex_unlock(&mrioc->bsg_cmds.mutex);
  2415. goto out;
  2416. }
  2417. } else {
  2418. if (mpi3mr_bsg_build_sgl(mrioc, mpi_req, mpi_msg_size,
  2419. drv_bufs, bufcnt, is_rmcb, is_rmrb,
  2420. (dout_cnt + din_cnt))) {
  2421. dprint_bsg_err(mrioc, "%s: sgl build failed\n", __func__);
  2422. rval = -EAGAIN;
  2423. mutex_unlock(&mrioc->bsg_cmds.mutex);
  2424. goto out;
  2425. }
  2426. }
  2427. if (mpi_header->function == MPI3_BSG_FUNCTION_SCSI_TASK_MGMT) {
  2428. tm_req = (struct mpi3_scsi_task_mgmt_request *)mpi_req;
  2429. if (tm_req->task_type !=
  2430. MPI3_SCSITASKMGMT_TASKTYPE_ABORT_TASK) {
  2431. dev_handle = tm_req->dev_handle;
  2432. block_io = 1;
  2433. }
  2434. }
  2435. if (block_io) {
  2436. tgtdev = mpi3mr_get_tgtdev_by_handle(mrioc, dev_handle);
  2437. if (tgtdev && tgtdev->starget && tgtdev->starget->hostdata) {
  2438. stgt_priv = (struct mpi3mr_stgt_priv_data *)
  2439. tgtdev->starget->hostdata;
  2440. atomic_inc(&stgt_priv->block_io);
  2441. mpi3mr_tgtdev_put(tgtdev);
  2442. }
  2443. }
  2444. mrioc->bsg_cmds.state = MPI3MR_CMD_PENDING;
  2445. mrioc->bsg_cmds.is_waiting = 1;
  2446. mrioc->bsg_cmds.callback = NULL;
  2447. mrioc->bsg_cmds.is_sense = 0;
  2448. mrioc->bsg_cmds.sensebuf = sense_buff_k;
  2449. memset(mrioc->bsg_cmds.reply, 0, mrioc->reply_sz);
  2450. mpi_header->host_tag = cpu_to_le16(MPI3MR_HOSTTAG_BSG_CMDS);
  2451. if (mrioc->logging_level & MPI3_DEBUG_BSG_INFO) {
  2452. dprint_bsg_info(mrioc,
  2453. "%s: posting bsg request to the controller\n", __func__);
  2454. dprint_dump(mpi_req, MPI3MR_ADMIN_REQ_FRAME_SZ,
  2455. "bsg_mpi3_req");
  2456. if (mpi_header->function == MPI3_BSG_FUNCTION_MGMT_PASSTHROUGH) {
  2457. drv_buf_iter = &drv_bufs[0];
  2458. dprint_dump(drv_buf_iter->kern_buf,
  2459. rmc_size, "mpi3_mgmt_req");
  2460. }
  2461. }
  2462. init_completion(&mrioc->bsg_cmds.done);
  2463. rval = mpi3mr_admin_request_post(mrioc, mpi_req,
  2464. MPI3MR_ADMIN_REQ_FRAME_SZ, 0);
  2465. if (rval) {
  2466. mrioc->bsg_cmds.is_waiting = 0;
  2467. dprint_bsg_err(mrioc,
  2468. "%s: posting bsg request is failed\n", __func__);
  2469. rval = -EAGAIN;
  2470. goto out_unlock;
  2471. }
  2472. wait_for_completion_timeout(&mrioc->bsg_cmds.done,
  2473. (karg->timeout * HZ));
  2474. if (block_io && stgt_priv)
  2475. atomic_dec(&stgt_priv->block_io);
  2476. if (!(mrioc->bsg_cmds.state & MPI3MR_CMD_COMPLETE)) {
  2477. mrioc->bsg_cmds.is_waiting = 0;
  2478. rval = -EAGAIN;
  2479. if (mrioc->bsg_cmds.state & MPI3MR_CMD_RESET)
  2480. goto out_unlock;
  2481. if (((mpi_header->function != MPI3_FUNCTION_SCSI_IO) &&
  2482. (mpi_header->function != MPI3_FUNCTION_NVME_ENCAPSULATED))
  2483. || (mrioc->logging_level & MPI3_DEBUG_BSG_ERROR)) {
  2484. ioc_info(mrioc, "%s: bsg request timedout after %d seconds\n",
  2485. __func__, karg->timeout);
  2486. if (!(mrioc->logging_level & MPI3_DEBUG_BSG_INFO)) {
  2487. dprint_dump(mpi_req, MPI3MR_ADMIN_REQ_FRAME_SZ,
  2488. "bsg_mpi3_req");
  2489. if (mpi_header->function ==
  2490. MPI3_FUNCTION_MGMT_PASSTHROUGH) {
  2491. drv_buf_iter = &drv_bufs[0];
  2492. dprint_dump(drv_buf_iter->kern_buf,
  2493. rmc_size, "mpi3_mgmt_req");
  2494. }
  2495. }
  2496. }
  2497. if ((mpi_header->function == MPI3_BSG_FUNCTION_NVME_ENCAPSULATED) ||
  2498. (mpi_header->function == MPI3_BSG_FUNCTION_SCSI_IO)) {
  2499. dprint_bsg_err(mrioc, "%s: bsg request timedout after %d seconds,\n"
  2500. "issuing target reset to (0x%04x)\n", __func__,
  2501. karg->timeout, mpi_header->function_dependent);
  2502. mpi3mr_issue_tm(mrioc,
  2503. MPI3_SCSITASKMGMT_TASKTYPE_TARGET_RESET,
  2504. mpi_header->function_dependent, 0,
  2505. MPI3MR_HOSTTAG_BLK_TMS, MPI3MR_RESETTM_TIMEOUT,
  2506. &mrioc->host_tm_cmds, &resp_code, NULL);
  2507. }
  2508. if (!(mrioc->bsg_cmds.state & MPI3MR_CMD_COMPLETE) &&
  2509. !(mrioc->bsg_cmds.state & MPI3MR_CMD_RESET))
  2510. mpi3mr_soft_reset_handler(mrioc,
  2511. MPI3MR_RESET_FROM_APP_TIMEOUT, 1);
  2512. goto out_unlock;
  2513. }
  2514. dprint_bsg_info(mrioc, "%s: bsg request is completed\n", __func__);
  2515. if (mrioc->prp_list_virt) {
  2516. dma_free_coherent(&mrioc->pdev->dev, mrioc->prp_sz,
  2517. mrioc->prp_list_virt, mrioc->prp_list_dma);
  2518. mrioc->prp_list_virt = NULL;
  2519. }
  2520. if ((mrioc->bsg_cmds.ioc_status & MPI3_IOCSTATUS_STATUS_MASK)
  2521. != MPI3_IOCSTATUS_SUCCESS) {
  2522. dprint_bsg_info(mrioc,
  2523. "%s: command failed, ioc_status(0x%04x) log_info(0x%08x)\n",
  2524. __func__,
  2525. (mrioc->bsg_cmds.ioc_status & MPI3_IOCSTATUS_STATUS_MASK),
  2526. mrioc->bsg_cmds.ioc_loginfo);
  2527. }
  2528. if ((mpirep_offset != 0xFF) &&
  2529. drv_bufs[mpirep_offset].bsg_buf_len) {
  2530. drv_buf_iter = &drv_bufs[mpirep_offset];
  2531. drv_buf_iter->kern_buf_len = (sizeof(*bsg_reply_buf) +
  2532. mrioc->reply_sz);
  2533. bsg_reply_buf = kzalloc(drv_buf_iter->kern_buf_len, GFP_KERNEL);
  2534. if (!bsg_reply_buf) {
  2535. rval = -ENOMEM;
  2536. goto out_unlock;
  2537. }
  2538. if (mrioc->bsg_cmds.state & MPI3MR_CMD_REPLY_VALID) {
  2539. bsg_reply_buf->mpi_reply_type =
  2540. MPI3MR_BSG_MPI_REPLY_BUFTYPE_ADDRESS;
  2541. memcpy(bsg_reply_buf->reply_buf,
  2542. mrioc->bsg_cmds.reply, mrioc->reply_sz);
  2543. } else {
  2544. bsg_reply_buf->mpi_reply_type =
  2545. MPI3MR_BSG_MPI_REPLY_BUFTYPE_STATUS;
  2546. status_desc = (struct mpi3_status_reply_descriptor *)
  2547. bsg_reply_buf->reply_buf;
  2548. status_desc->ioc_status = mrioc->bsg_cmds.ioc_status;
  2549. status_desc->ioc_log_info = mrioc->bsg_cmds.ioc_loginfo;
  2550. }
  2551. tmplen = min(drv_buf_iter->kern_buf_len,
  2552. drv_buf_iter->bsg_buf_len);
  2553. memcpy(drv_buf_iter->bsg_buf, bsg_reply_buf, tmplen);
  2554. }
  2555. if (erb_offset != 0xFF && mrioc->bsg_cmds.sensebuf &&
  2556. mrioc->bsg_cmds.is_sense) {
  2557. drv_buf_iter = &drv_bufs[erb_offset];
  2558. tmplen = min(erbsz, drv_buf_iter->bsg_buf_len);
  2559. memcpy(drv_buf_iter->bsg_buf, sense_buff_k, tmplen);
  2560. }
  2561. drv_buf_iter = drv_bufs;
  2562. for (count = 0; count < bufcnt; count++, drv_buf_iter++) {
  2563. if (drv_buf_iter->data_dir == DMA_NONE)
  2564. continue;
  2565. if ((count == 1) && is_rmrb) {
  2566. memcpy(drv_buf_iter->bsg_buf,
  2567. drv_buf_iter->kern_buf,
  2568. drv_buf_iter->kern_buf_len);
  2569. } else if (drv_buf_iter->data_dir == DMA_FROM_DEVICE) {
  2570. tmplen = 0;
  2571. for (desc_count = 0;
  2572. desc_count < drv_buf_iter->num_dma_desc;
  2573. desc_count++) {
  2574. memcpy(((u8 *)drv_buf_iter->bsg_buf + tmplen),
  2575. drv_buf_iter->dma_desc[desc_count].addr,
  2576. drv_buf_iter->dma_desc[desc_count].size);
  2577. tmplen +=
  2578. drv_buf_iter->dma_desc[desc_count].size;
  2579. }
  2580. }
  2581. }
  2582. out_unlock:
  2583. if (din_buf) {
  2584. job->reply_payload_rcv_len =
  2585. sg_copy_from_buffer(job->reply_payload.sg_list,
  2586. job->reply_payload.sg_cnt,
  2587. din_buf, job->reply_payload.payload_len);
  2588. }
  2589. mrioc->bsg_cmds.is_sense = 0;
  2590. mrioc->bsg_cmds.sensebuf = NULL;
  2591. mrioc->bsg_cmds.state = MPI3MR_CMD_NOTUSED;
  2592. mutex_unlock(&mrioc->bsg_cmds.mutex);
  2593. out:
  2594. kfree(sense_buff_k);
  2595. kfree(dout_buf);
  2596. kfree(din_buf);
  2597. kfree(mpi_req);
  2598. if (drv_bufs) {
  2599. drv_buf_iter = drv_bufs;
  2600. for (count = 0; count < bufcnt; count++, drv_buf_iter++)
  2601. kfree(drv_buf_iter->dma_desc);
  2602. kfree(drv_bufs);
  2603. }
  2604. kfree(bsg_reply_buf);
  2605. return rval;
  2606. }
  2607. /**
  2608. * mpi3mr_app_save_logdata - Save Log Data events
  2609. * @mrioc: Adapter instance reference
  2610. * @event_data: event data associated with log data event
  2611. * @event_data_size: event data size to copy
  2612. *
  2613. * If log data event caching is enabled by the applicatiobns,
  2614. * then this function saves the log data in the circular queue
  2615. * and Sends async signal SIGIO to indicate there is an async
  2616. * event from the firmware to the event monitoring applications.
  2617. *
  2618. * Return:Nothing
  2619. */
  2620. void mpi3mr_app_save_logdata(struct mpi3mr_ioc *mrioc, char *event_data,
  2621. u16 event_data_size)
  2622. {
  2623. u32 index = mrioc->logdata_buf_idx, sz;
  2624. struct mpi3mr_logdata_entry *entry;
  2625. if (!(mrioc->logdata_buf))
  2626. return;
  2627. entry = (struct mpi3mr_logdata_entry *)
  2628. (mrioc->logdata_buf + (index * mrioc->logdata_entry_sz));
  2629. entry->valid_entry = 1;
  2630. sz = min(mrioc->logdata_entry_sz, event_data_size);
  2631. memcpy(entry->data, event_data, sz);
  2632. mrioc->logdata_buf_idx =
  2633. ((++index) % MPI3MR_BSG_LOGDATA_MAX_ENTRIES);
  2634. atomic64_inc(&event_counter);
  2635. }
  2636. /**
  2637. * mpi3mr_bsg_request - bsg request entry point
  2638. * @job: BSG job reference
  2639. *
  2640. * This is driver's entry point for bsg requests
  2641. *
  2642. * Return: 0 on success and proper error codes on failure
  2643. */
  2644. static int mpi3mr_bsg_request(struct bsg_job *job)
  2645. {
  2646. long rval = -EINVAL;
  2647. unsigned int reply_payload_rcv_len = 0;
  2648. struct mpi3mr_bsg_packet *bsg_req = job->request;
  2649. switch (bsg_req->cmd_type) {
  2650. case MPI3MR_DRV_CMD:
  2651. rval = mpi3mr_bsg_process_drv_cmds(job);
  2652. break;
  2653. case MPI3MR_MPT_CMD:
  2654. rval = mpi3mr_bsg_process_mpt_cmds(job);
  2655. break;
  2656. default:
  2657. pr_err("%s: unsupported BSG command(0x%08x)\n",
  2658. MPI3MR_DRIVER_NAME, bsg_req->cmd_type);
  2659. break;
  2660. }
  2661. bsg_job_done(job, rval, reply_payload_rcv_len);
  2662. return 0;
  2663. }
  2664. /**
  2665. * mpi3mr_bsg_exit - de-registration from bsg layer
  2666. * @mrioc: Adapter instance reference
  2667. *
  2668. * This will be called during driver unload and all
  2669. * bsg resources allocated during load will be freed.
  2670. *
  2671. * Return:Nothing
  2672. */
  2673. void mpi3mr_bsg_exit(struct mpi3mr_ioc *mrioc)
  2674. {
  2675. struct device *bsg_dev = &mrioc->bsg_dev;
  2676. if (!mrioc->bsg_queue)
  2677. return;
  2678. bsg_remove_queue(mrioc->bsg_queue);
  2679. mrioc->bsg_queue = NULL;
  2680. device_del(bsg_dev);
  2681. put_device(bsg_dev);
  2682. }
  2683. /**
  2684. * mpi3mr_bsg_node_release -release bsg device node
  2685. * @dev: bsg device node
  2686. *
  2687. * decrements bsg dev parent reference count
  2688. *
  2689. * Return:Nothing
  2690. */
  2691. static void mpi3mr_bsg_node_release(struct device *dev)
  2692. {
  2693. put_device(dev->parent);
  2694. }
  2695. /**
  2696. * mpi3mr_bsg_init - registration with bsg layer
  2697. * @mrioc: Adapter instance reference
  2698. *
  2699. * This will be called during driver load and it will
  2700. * register driver with bsg layer
  2701. *
  2702. * Return:Nothing
  2703. */
  2704. void mpi3mr_bsg_init(struct mpi3mr_ioc *mrioc)
  2705. {
  2706. struct device *bsg_dev = &mrioc->bsg_dev;
  2707. struct device *parent = &mrioc->shost->shost_gendev;
  2708. struct queue_limits lim = {
  2709. .max_hw_sectors = MPI3MR_MAX_APP_XFER_SECTORS,
  2710. .max_segments = MPI3MR_MAX_APP_XFER_SEGMENTS,
  2711. };
  2712. struct request_queue *q;
  2713. device_initialize(bsg_dev);
  2714. bsg_dev->parent = get_device(parent);
  2715. bsg_dev->release = mpi3mr_bsg_node_release;
  2716. dev_set_name(bsg_dev, "mpi3mrctl%u", mrioc->id);
  2717. if (device_add(bsg_dev)) {
  2718. ioc_err(mrioc, "%s: bsg device add failed\n",
  2719. dev_name(bsg_dev));
  2720. put_device(bsg_dev);
  2721. return;
  2722. }
  2723. q = bsg_setup_queue(bsg_dev, dev_name(bsg_dev), &lim,
  2724. mpi3mr_bsg_request, NULL, 0);
  2725. if (IS_ERR(q)) {
  2726. ioc_err(mrioc, "%s: bsg registration failed\n",
  2727. dev_name(bsg_dev));
  2728. device_del(bsg_dev);
  2729. put_device(bsg_dev);
  2730. return;
  2731. }
  2732. mrioc->bsg_queue = q;
  2733. }
  2734. /**
  2735. * version_fw_show - SysFS callback for firmware version read
  2736. * @dev: class device
  2737. * @attr: Device attributes
  2738. * @buf: Buffer to copy
  2739. *
  2740. * Return: sysfs_emit() return after copying firmware version
  2741. */
  2742. static ssize_t
  2743. version_fw_show(struct device *dev, struct device_attribute *attr,
  2744. char *buf)
  2745. {
  2746. struct Scsi_Host *shost = class_to_shost(dev);
  2747. struct mpi3mr_ioc *mrioc = shost_priv(shost);
  2748. struct mpi3mr_compimg_ver *fwver = &mrioc->facts.fw_ver;
  2749. return sysfs_emit(buf, "%d.%d.%d.%d.%05d-%05d\n",
  2750. fwver->gen_major, fwver->gen_minor, fwver->ph_major,
  2751. fwver->ph_minor, fwver->cust_id, fwver->build_num);
  2752. }
  2753. static DEVICE_ATTR_RO(version_fw);
  2754. /**
  2755. * fw_queue_depth_show - SysFS callback for firmware max cmds
  2756. * @dev: class device
  2757. * @attr: Device attributes
  2758. * @buf: Buffer to copy
  2759. *
  2760. * Return: sysfs_emit() return after copying firmware max commands
  2761. */
  2762. static ssize_t
  2763. fw_queue_depth_show(struct device *dev, struct device_attribute *attr,
  2764. char *buf)
  2765. {
  2766. struct Scsi_Host *shost = class_to_shost(dev);
  2767. struct mpi3mr_ioc *mrioc = shost_priv(shost);
  2768. return sysfs_emit(buf, "%d\n", mrioc->facts.max_reqs);
  2769. }
  2770. static DEVICE_ATTR_RO(fw_queue_depth);
  2771. /**
  2772. * op_req_q_count_show - SysFS callback for request queue count
  2773. * @dev: class device
  2774. * @attr: Device attributes
  2775. * @buf: Buffer to copy
  2776. *
  2777. * Return: sysfs_emit() return after copying request queue count
  2778. */
  2779. static ssize_t
  2780. op_req_q_count_show(struct device *dev, struct device_attribute *attr,
  2781. char *buf)
  2782. {
  2783. struct Scsi_Host *shost = class_to_shost(dev);
  2784. struct mpi3mr_ioc *mrioc = shost_priv(shost);
  2785. return sysfs_emit(buf, "%d\n", mrioc->num_op_req_q);
  2786. }
  2787. static DEVICE_ATTR_RO(op_req_q_count);
  2788. /**
  2789. * reply_queue_count_show - SysFS callback for reply queue count
  2790. * @dev: class device
  2791. * @attr: Device attributes
  2792. * @buf: Buffer to copy
  2793. *
  2794. * Return: sysfs_emit() return after copying reply queue count
  2795. */
  2796. static ssize_t
  2797. reply_queue_count_show(struct device *dev, struct device_attribute *attr,
  2798. char *buf)
  2799. {
  2800. struct Scsi_Host *shost = class_to_shost(dev);
  2801. struct mpi3mr_ioc *mrioc = shost_priv(shost);
  2802. return sysfs_emit(buf, "%d\n", mrioc->num_op_reply_q);
  2803. }
  2804. static DEVICE_ATTR_RO(reply_queue_count);
  2805. /**
  2806. * logging_level_show - Show controller debug level
  2807. * @dev: class device
  2808. * @attr: Device attributes
  2809. * @buf: Buffer to copy
  2810. *
  2811. * A sysfs 'read/write' shost attribute, to show the current
  2812. * debug log level used by the driver for the specific
  2813. * controller.
  2814. *
  2815. * Return: sysfs_emit() return
  2816. */
  2817. static ssize_t
  2818. logging_level_show(struct device *dev,
  2819. struct device_attribute *attr, char *buf)
  2820. {
  2821. struct Scsi_Host *shost = class_to_shost(dev);
  2822. struct mpi3mr_ioc *mrioc = shost_priv(shost);
  2823. return sysfs_emit(buf, "%08xh\n", mrioc->logging_level);
  2824. }
  2825. /**
  2826. * logging_level_store- Change controller debug level
  2827. * @dev: class device
  2828. * @attr: Device attributes
  2829. * @buf: Buffer to copy
  2830. * @count: size of the buffer
  2831. *
  2832. * A sysfs 'read/write' shost attribute, to change the current
  2833. * debug log level used by the driver for the specific
  2834. * controller.
  2835. *
  2836. * Return: strlen() return
  2837. */
  2838. static ssize_t
  2839. logging_level_store(struct device *dev,
  2840. struct device_attribute *attr,
  2841. const char *buf, size_t count)
  2842. {
  2843. struct Scsi_Host *shost = class_to_shost(dev);
  2844. struct mpi3mr_ioc *mrioc = shost_priv(shost);
  2845. int val = 0;
  2846. if (kstrtoint(buf, 0, &val) != 0)
  2847. return -EINVAL;
  2848. mrioc->logging_level = val;
  2849. ioc_info(mrioc, "logging_level=%08xh\n", mrioc->logging_level);
  2850. return strlen(buf);
  2851. }
  2852. static DEVICE_ATTR_RW(logging_level);
  2853. /**
  2854. * adp_state_show() - SysFS callback for adapter state show
  2855. * @dev: class device
  2856. * @attr: Device attributes
  2857. * @buf: Buffer to copy
  2858. *
  2859. * Return: sysfs_emit() return after copying adapter state
  2860. */
  2861. static ssize_t
  2862. adp_state_show(struct device *dev, struct device_attribute *attr,
  2863. char *buf)
  2864. {
  2865. struct Scsi_Host *shost = class_to_shost(dev);
  2866. struct mpi3mr_ioc *mrioc = shost_priv(shost);
  2867. enum mpi3mr_iocstate ioc_state;
  2868. uint8_t adp_state;
  2869. ioc_state = mpi3mr_get_iocstate(mrioc);
  2870. if (ioc_state == MRIOC_STATE_UNRECOVERABLE)
  2871. adp_state = MPI3MR_BSG_ADPSTATE_UNRECOVERABLE;
  2872. else if (mrioc->reset_in_progress || mrioc->stop_bsgs ||
  2873. mrioc->block_on_pci_err)
  2874. adp_state = MPI3MR_BSG_ADPSTATE_IN_RESET;
  2875. else if (ioc_state == MRIOC_STATE_FAULT)
  2876. adp_state = MPI3MR_BSG_ADPSTATE_FAULT;
  2877. else
  2878. adp_state = MPI3MR_BSG_ADPSTATE_OPERATIONAL;
  2879. return sysfs_emit(buf, "%u\n", adp_state);
  2880. }
  2881. static DEVICE_ATTR_RO(adp_state);
  2882. static struct attribute *mpi3mr_host_attrs[] = {
  2883. &dev_attr_version_fw.attr,
  2884. &dev_attr_fw_queue_depth.attr,
  2885. &dev_attr_op_req_q_count.attr,
  2886. &dev_attr_reply_queue_count.attr,
  2887. &dev_attr_logging_level.attr,
  2888. &dev_attr_adp_state.attr,
  2889. NULL,
  2890. };
  2891. static const struct attribute_group mpi3mr_host_attr_group = {
  2892. .attrs = mpi3mr_host_attrs
  2893. };
  2894. const struct attribute_group *mpi3mr_host_groups[] = {
  2895. &mpi3mr_host_attr_group,
  2896. NULL,
  2897. };
  2898. /*
  2899. * SCSI Device attributes under sysfs
  2900. */
  2901. /**
  2902. * sas_address_show - SysFS callback for dev SASaddress display
  2903. * @dev: class device
  2904. * @attr: Device attributes
  2905. * @buf: Buffer to copy
  2906. *
  2907. * Return: sysfs_emit() return after copying SAS address of the
  2908. * specific SAS/SATA end device.
  2909. */
  2910. static ssize_t
  2911. sas_address_show(struct device *dev, struct device_attribute *attr,
  2912. char *buf)
  2913. {
  2914. struct scsi_device *sdev = to_scsi_device(dev);
  2915. struct mpi3mr_sdev_priv_data *sdev_priv_data;
  2916. struct mpi3mr_stgt_priv_data *tgt_priv_data;
  2917. struct mpi3mr_tgt_dev *tgtdev;
  2918. sdev_priv_data = sdev->hostdata;
  2919. if (!sdev_priv_data)
  2920. return 0;
  2921. tgt_priv_data = sdev_priv_data->tgt_priv_data;
  2922. if (!tgt_priv_data)
  2923. return 0;
  2924. tgtdev = tgt_priv_data->tgt_dev;
  2925. if (!tgtdev || tgtdev->dev_type != MPI3_DEVICE_DEVFORM_SAS_SATA)
  2926. return 0;
  2927. return sysfs_emit(buf, "0x%016llx\n",
  2928. (unsigned long long)tgtdev->dev_spec.sas_sata_inf.sas_address);
  2929. }
  2930. static DEVICE_ATTR_RO(sas_address);
  2931. /**
  2932. * device_handle_show - SysFS callback for device handle display
  2933. * @dev: class device
  2934. * @attr: Device attributes
  2935. * @buf: Buffer to copy
  2936. *
  2937. * Return: sysfs_emit() return after copying firmware internal
  2938. * device handle of the specific device.
  2939. */
  2940. static ssize_t
  2941. device_handle_show(struct device *dev, struct device_attribute *attr,
  2942. char *buf)
  2943. {
  2944. struct scsi_device *sdev = to_scsi_device(dev);
  2945. struct mpi3mr_sdev_priv_data *sdev_priv_data;
  2946. struct mpi3mr_stgt_priv_data *tgt_priv_data;
  2947. struct mpi3mr_tgt_dev *tgtdev;
  2948. sdev_priv_data = sdev->hostdata;
  2949. if (!sdev_priv_data)
  2950. return 0;
  2951. tgt_priv_data = sdev_priv_data->tgt_priv_data;
  2952. if (!tgt_priv_data)
  2953. return 0;
  2954. tgtdev = tgt_priv_data->tgt_dev;
  2955. if (!tgtdev)
  2956. return 0;
  2957. return sysfs_emit(buf, "0x%04x\n", tgtdev->dev_handle);
  2958. }
  2959. static DEVICE_ATTR_RO(device_handle);
  2960. /**
  2961. * persistent_id_show - SysFS callback for persisten ID display
  2962. * @dev: class device
  2963. * @attr: Device attributes
  2964. * @buf: Buffer to copy
  2965. *
  2966. * Return: sysfs_emit() return after copying persistent ID of the
  2967. * of the specific device.
  2968. */
  2969. static ssize_t
  2970. persistent_id_show(struct device *dev, struct device_attribute *attr,
  2971. char *buf)
  2972. {
  2973. struct scsi_device *sdev = to_scsi_device(dev);
  2974. struct mpi3mr_sdev_priv_data *sdev_priv_data;
  2975. struct mpi3mr_stgt_priv_data *tgt_priv_data;
  2976. struct mpi3mr_tgt_dev *tgtdev;
  2977. sdev_priv_data = sdev->hostdata;
  2978. if (!sdev_priv_data)
  2979. return 0;
  2980. tgt_priv_data = sdev_priv_data->tgt_priv_data;
  2981. if (!tgt_priv_data)
  2982. return 0;
  2983. tgtdev = tgt_priv_data->tgt_dev;
  2984. if (!tgtdev)
  2985. return 0;
  2986. return sysfs_emit(buf, "%d\n", tgtdev->perst_id);
  2987. }
  2988. static DEVICE_ATTR_RO(persistent_id);
  2989. /**
  2990. * sas_ncq_prio_supported_show - Indicate if device supports NCQ priority
  2991. * @dev: pointer to embedded device
  2992. * @attr: sas_ncq_prio_supported attribute descriptor
  2993. * @buf: the buffer returned
  2994. *
  2995. * A sysfs 'read-only' sdev attribute, only works with SATA devices
  2996. */
  2997. static ssize_t
  2998. sas_ncq_prio_supported_show(struct device *dev,
  2999. struct device_attribute *attr, char *buf)
  3000. {
  3001. struct scsi_device *sdev = to_scsi_device(dev);
  3002. return sysfs_emit(buf, "%d\n", sas_ata_ncq_prio_supported(sdev));
  3003. }
  3004. static DEVICE_ATTR_RO(sas_ncq_prio_supported);
  3005. /**
  3006. * sas_ncq_prio_enable_show - send prioritized io commands to device
  3007. * @dev: pointer to embedded device
  3008. * @attr: sas_ncq_prio_enable attribute descriptor
  3009. * @buf: the buffer returned
  3010. *
  3011. * A sysfs 'read/write' sdev attribute, only works with SATA devices
  3012. */
  3013. static ssize_t
  3014. sas_ncq_prio_enable_show(struct device *dev,
  3015. struct device_attribute *attr, char *buf)
  3016. {
  3017. struct scsi_device *sdev = to_scsi_device(dev);
  3018. struct mpi3mr_sdev_priv_data *sdev_priv_data = sdev->hostdata;
  3019. if (!sdev_priv_data)
  3020. return 0;
  3021. return sysfs_emit(buf, "%d\n", sdev_priv_data->ncq_prio_enable);
  3022. }
  3023. static ssize_t
  3024. sas_ncq_prio_enable_store(struct device *dev,
  3025. struct device_attribute *attr,
  3026. const char *buf, size_t count)
  3027. {
  3028. struct scsi_device *sdev = to_scsi_device(dev);
  3029. struct mpi3mr_sdev_priv_data *sdev_priv_data = sdev->hostdata;
  3030. bool ncq_prio_enable = 0;
  3031. if (kstrtobool(buf, &ncq_prio_enable))
  3032. return -EINVAL;
  3033. if (!sas_ata_ncq_prio_supported(sdev))
  3034. return -EINVAL;
  3035. sdev_priv_data->ncq_prio_enable = ncq_prio_enable;
  3036. return strlen(buf);
  3037. }
  3038. static DEVICE_ATTR_RW(sas_ncq_prio_enable);
  3039. static struct attribute *mpi3mr_dev_attrs[] = {
  3040. &dev_attr_sas_address.attr,
  3041. &dev_attr_device_handle.attr,
  3042. &dev_attr_persistent_id.attr,
  3043. &dev_attr_sas_ncq_prio_supported.attr,
  3044. &dev_attr_sas_ncq_prio_enable.attr,
  3045. NULL,
  3046. };
  3047. static const struct attribute_group mpi3mr_dev_attr_group = {
  3048. .attrs = mpi3mr_dev_attrs
  3049. };
  3050. const struct attribute_group *mpi3mr_dev_groups[] = {
  3051. &mpi3mr_dev_attr_group,
  3052. NULL,
  3053. };