dasd_fba.c 23 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835
  1. // SPDX-License-Identifier: GPL-2.0
  2. /*
  3. * Author(s)......: Holger Smolinski <Holger.Smolinski@de.ibm.com>
  4. * Bugreports.to..: <Linux390@de.ibm.com>
  5. * Copyright IBM Corp. 1999, 2009
  6. */
  7. #define KMSG_COMPONENT "dasd-fba"
  8. #include <linux/stddef.h>
  9. #include <linux/kernel.h>
  10. #include <asm/debug.h>
  11. #include <linux/slab.h>
  12. #include <linux/hdreg.h> /* HDIO_GETGEO */
  13. #include <linux/bio.h>
  14. #include <linux/module.h>
  15. #include <linux/init.h>
  16. #include <asm/idals.h>
  17. #include <asm/ebcdic.h>
  18. #include <asm/io.h>
  19. #include <asm/ccwdev.h>
  20. #include "dasd_int.h"
  21. #include "dasd_fba.h"
  22. #ifdef PRINTK_HEADER
  23. #undef PRINTK_HEADER
  24. #endif /* PRINTK_HEADER */
  25. #define PRINTK_HEADER "dasd(fba):"
  26. #define FBA_DEFAULT_RETRIES 32
  27. #define DASD_FBA_CCW_WRITE 0x41
  28. #define DASD_FBA_CCW_READ 0x42
  29. #define DASD_FBA_CCW_LOCATE 0x43
  30. #define DASD_FBA_CCW_DEFINE_EXTENT 0x63
  31. MODULE_LICENSE("GPL");
  32. static struct dasd_discipline dasd_fba_discipline;
  33. static void *dasd_fba_zero_page;
  34. struct dasd_fba_private {
  35. struct dasd_fba_characteristics rdc_data;
  36. };
  37. static struct ccw_device_id dasd_fba_ids[] = {
  38. { CCW_DEVICE_DEVTYPE (0x6310, 0, 0x9336, 0), .driver_info = 0x1},
  39. { CCW_DEVICE_DEVTYPE (0x3880, 0, 0x3370, 0), .driver_info = 0x2},
  40. { /* end of list */ },
  41. };
  42. MODULE_DEVICE_TABLE(ccw, dasd_fba_ids);
  43. static struct ccw_driver dasd_fba_driver; /* see below */
  44. static int
  45. dasd_fba_probe(struct ccw_device *cdev)
  46. {
  47. return dasd_generic_probe(cdev, &dasd_fba_discipline);
  48. }
  49. static int
  50. dasd_fba_set_online(struct ccw_device *cdev)
  51. {
  52. return dasd_generic_set_online(cdev, &dasd_fba_discipline);
  53. }
  54. static struct ccw_driver dasd_fba_driver = {
  55. .driver = {
  56. .name = "dasd-fba",
  57. .owner = THIS_MODULE,
  58. },
  59. .ids = dasd_fba_ids,
  60. .probe = dasd_fba_probe,
  61. .remove = dasd_generic_remove,
  62. .set_offline = dasd_generic_set_offline,
  63. .set_online = dasd_fba_set_online,
  64. .notify = dasd_generic_notify,
  65. .path_event = dasd_generic_path_event,
  66. .freeze = dasd_generic_pm_freeze,
  67. .thaw = dasd_generic_restore_device,
  68. .restore = dasd_generic_restore_device,
  69. .int_class = IRQIO_DAS,
  70. };
  71. static void
  72. define_extent(struct ccw1 * ccw, struct DE_fba_data *data, int rw,
  73. int blksize, int beg, int nr)
  74. {
  75. ccw->cmd_code = DASD_FBA_CCW_DEFINE_EXTENT;
  76. ccw->flags = 0;
  77. ccw->count = 16;
  78. ccw->cda = (__u32) __pa(data);
  79. memset(data, 0, sizeof (struct DE_fba_data));
  80. if (rw == WRITE)
  81. (data->mask).perm = 0x0;
  82. else if (rw == READ)
  83. (data->mask).perm = 0x1;
  84. else
  85. data->mask.perm = 0x2;
  86. data->blk_size = blksize;
  87. data->ext_loc = beg;
  88. data->ext_end = nr - 1;
  89. }
  90. static void
  91. locate_record(struct ccw1 * ccw, struct LO_fba_data *data, int rw,
  92. int block_nr, int block_ct)
  93. {
  94. ccw->cmd_code = DASD_FBA_CCW_LOCATE;
  95. ccw->flags = 0;
  96. ccw->count = 8;
  97. ccw->cda = (__u32) __pa(data);
  98. memset(data, 0, sizeof (struct LO_fba_data));
  99. if (rw == WRITE)
  100. data->operation.cmd = 0x5;
  101. else if (rw == READ)
  102. data->operation.cmd = 0x6;
  103. else
  104. data->operation.cmd = 0x8;
  105. data->blk_nr = block_nr;
  106. data->blk_ct = block_ct;
  107. }
  108. static int
  109. dasd_fba_check_characteristics(struct dasd_device *device)
  110. {
  111. struct dasd_fba_private *private = device->private;
  112. struct ccw_device *cdev = device->cdev;
  113. struct dasd_block *block;
  114. int readonly, rc;
  115. if (!private) {
  116. private = kzalloc(sizeof(*private), GFP_KERNEL | GFP_DMA);
  117. if (!private) {
  118. dev_warn(&device->cdev->dev,
  119. "Allocating memory for private DASD "
  120. "data failed\n");
  121. return -ENOMEM;
  122. }
  123. device->private = private;
  124. } else {
  125. memset(private, 0, sizeof(*private));
  126. }
  127. block = dasd_alloc_block();
  128. if (IS_ERR(block)) {
  129. DBF_EVENT_DEVID(DBF_WARNING, cdev, "%s", "could not allocate "
  130. "dasd block structure");
  131. device->private = NULL;
  132. kfree(private);
  133. return PTR_ERR(block);
  134. }
  135. device->block = block;
  136. block->base = device;
  137. /* Read Device Characteristics */
  138. rc = dasd_generic_read_dev_chars(device, DASD_FBA_MAGIC,
  139. &private->rdc_data, 32);
  140. if (rc) {
  141. DBF_EVENT_DEVID(DBF_WARNING, cdev, "Read device "
  142. "characteristics returned error %d", rc);
  143. device->block = NULL;
  144. dasd_free_block(block);
  145. device->private = NULL;
  146. kfree(private);
  147. return rc;
  148. }
  149. device->default_expires = DASD_EXPIRES;
  150. device->default_retries = FBA_DEFAULT_RETRIES;
  151. dasd_path_set_opm(device, LPM_ANYPATH);
  152. readonly = dasd_device_is_ro(device);
  153. if (readonly)
  154. set_bit(DASD_FLAG_DEVICE_RO, &device->flags);
  155. /* FBA supports discard, set the according feature bit */
  156. dasd_set_feature(cdev, DASD_FEATURE_DISCARD, 1);
  157. dev_info(&device->cdev->dev,
  158. "New FBA DASD %04X/%02X (CU %04X/%02X) with %d MB "
  159. "and %d B/blk%s\n",
  160. cdev->id.dev_type,
  161. cdev->id.dev_model,
  162. cdev->id.cu_type,
  163. cdev->id.cu_model,
  164. ((private->rdc_data.blk_bdsa *
  165. (private->rdc_data.blk_size >> 9)) >> 11),
  166. private->rdc_data.blk_size,
  167. readonly ? ", read-only device" : "");
  168. return 0;
  169. }
  170. static int dasd_fba_do_analysis(struct dasd_block *block)
  171. {
  172. struct dasd_fba_private *private = block->base->private;
  173. int sb, rc;
  174. rc = dasd_check_blocksize(private->rdc_data.blk_size);
  175. if (rc) {
  176. DBF_DEV_EVENT(DBF_WARNING, block->base, "unknown blocksize %d",
  177. private->rdc_data.blk_size);
  178. return rc;
  179. }
  180. block->blocks = private->rdc_data.blk_bdsa;
  181. block->bp_block = private->rdc_data.blk_size;
  182. block->s2b_shift = 0; /* bits to shift 512 to get a block */
  183. for (sb = 512; sb < private->rdc_data.blk_size; sb = sb << 1)
  184. block->s2b_shift++;
  185. return 0;
  186. }
  187. static int dasd_fba_fill_geometry(struct dasd_block *block,
  188. struct hd_geometry *geo)
  189. {
  190. if (dasd_check_blocksize(block->bp_block) != 0)
  191. return -EINVAL;
  192. geo->cylinders = (block->blocks << block->s2b_shift) >> 10;
  193. geo->heads = 16;
  194. geo->sectors = 128 >> block->s2b_shift;
  195. return 0;
  196. }
  197. static dasd_erp_fn_t
  198. dasd_fba_erp_action(struct dasd_ccw_req * cqr)
  199. {
  200. return dasd_default_erp_action;
  201. }
  202. static dasd_erp_fn_t
  203. dasd_fba_erp_postaction(struct dasd_ccw_req * cqr)
  204. {
  205. if (cqr->function == dasd_default_erp_action)
  206. return dasd_default_erp_postaction;
  207. DBF_DEV_EVENT(DBF_WARNING, cqr->startdev, "unknown ERP action %p",
  208. cqr->function);
  209. return NULL;
  210. }
  211. static void dasd_fba_check_for_device_change(struct dasd_device *device,
  212. struct dasd_ccw_req *cqr,
  213. struct irb *irb)
  214. {
  215. char mask;
  216. /* first of all check for state change pending interrupt */
  217. mask = DEV_STAT_ATTENTION | DEV_STAT_DEV_END | DEV_STAT_UNIT_EXCEP;
  218. if ((irb->scsw.cmd.dstat & mask) == mask)
  219. dasd_generic_handle_state_change(device);
  220. };
  221. /*
  222. * Builds a CCW with no data payload
  223. */
  224. static void ccw_write_no_data(struct ccw1 *ccw)
  225. {
  226. ccw->cmd_code = DASD_FBA_CCW_WRITE;
  227. ccw->flags |= CCW_FLAG_SLI;
  228. ccw->count = 0;
  229. }
  230. /*
  231. * Builds a CCW that writes only zeroes.
  232. */
  233. static void ccw_write_zero(struct ccw1 *ccw, int count)
  234. {
  235. ccw->cmd_code = DASD_FBA_CCW_WRITE;
  236. ccw->flags |= CCW_FLAG_SLI;
  237. ccw->count = count;
  238. ccw->cda = (__u32) (addr_t) dasd_fba_zero_page;
  239. }
  240. /*
  241. * Helper function to count the amount of necessary CCWs within a given range
  242. * with 4k alignment and command chaining in mind.
  243. */
  244. static int count_ccws(sector_t first_rec, sector_t last_rec,
  245. unsigned int blocks_per_page)
  246. {
  247. sector_t wz_stop = 0, d_stop = 0;
  248. int cur_pos = 0;
  249. int count = 0;
  250. if (first_rec % blocks_per_page != 0) {
  251. wz_stop = first_rec + blocks_per_page -
  252. (first_rec % blocks_per_page) - 1;
  253. if (wz_stop > last_rec)
  254. wz_stop = last_rec;
  255. cur_pos = wz_stop - first_rec + 1;
  256. count++;
  257. }
  258. if (last_rec - (first_rec + cur_pos) + 1 >= blocks_per_page) {
  259. if ((last_rec - blocks_per_page + 1) % blocks_per_page != 0)
  260. d_stop = last_rec - ((last_rec - blocks_per_page + 1) %
  261. blocks_per_page);
  262. else
  263. d_stop = last_rec;
  264. cur_pos += d_stop - (first_rec + cur_pos) + 1;
  265. count++;
  266. }
  267. if (cur_pos == 0 || first_rec + cur_pos - 1 < last_rec)
  268. count++;
  269. return count;
  270. }
  271. /*
  272. * This function builds a CCW request for block layer discard requests.
  273. * Each page in the z/VM hypervisor that represents certain records of an FBA
  274. * device will be padded with zeros. This is a special behaviour of the WRITE
  275. * command which is triggered when no data payload is added to the CCW.
  276. *
  277. * Note: Due to issues in some z/VM versions, we can't fully utilise this
  278. * special behaviour. We have to keep a 4k (or 8 block) alignment in mind to
  279. * work around those issues and write actual zeroes to the unaligned parts in
  280. * the request. This workaround might be removed in the future.
  281. */
  282. static struct dasd_ccw_req *dasd_fba_build_cp_discard(
  283. struct dasd_device *memdev,
  284. struct dasd_block *block,
  285. struct request *req)
  286. {
  287. struct LO_fba_data *LO_data;
  288. struct dasd_ccw_req *cqr;
  289. struct ccw1 *ccw;
  290. sector_t wz_stop = 0, d_stop = 0;
  291. sector_t first_rec, last_rec;
  292. unsigned int blksize = block->bp_block;
  293. unsigned int blocks_per_page;
  294. int wz_count = 0;
  295. int d_count = 0;
  296. int cur_pos = 0; /* Current position within the extent */
  297. int count = 0;
  298. int cplength;
  299. int datasize;
  300. int nr_ccws;
  301. first_rec = blk_rq_pos(req) >> block->s2b_shift;
  302. last_rec =
  303. (blk_rq_pos(req) + blk_rq_sectors(req) - 1) >> block->s2b_shift;
  304. count = last_rec - first_rec + 1;
  305. blocks_per_page = BLOCKS_PER_PAGE(blksize);
  306. nr_ccws = count_ccws(first_rec, last_rec, blocks_per_page);
  307. /* define extent + nr_ccws * locate record + nr_ccws * single CCW */
  308. cplength = 1 + 2 * nr_ccws;
  309. datasize = sizeof(struct DE_fba_data) +
  310. nr_ccws * (sizeof(struct LO_fba_data) + sizeof(struct ccw1));
  311. cqr = dasd_smalloc_request(DASD_FBA_MAGIC, cplength, datasize, memdev,
  312. blk_mq_rq_to_pdu(req));
  313. if (IS_ERR(cqr))
  314. return cqr;
  315. ccw = cqr->cpaddr;
  316. define_extent(ccw++, cqr->data, WRITE, blksize, first_rec, count);
  317. LO_data = cqr->data + sizeof(struct DE_fba_data);
  318. /* First part is not aligned. Calculate range to write zeroes. */
  319. if (first_rec % blocks_per_page != 0) {
  320. wz_stop = first_rec + blocks_per_page -
  321. (first_rec % blocks_per_page) - 1;
  322. if (wz_stop > last_rec)
  323. wz_stop = last_rec;
  324. wz_count = wz_stop - first_rec + 1;
  325. ccw[-1].flags |= CCW_FLAG_CC;
  326. locate_record(ccw++, LO_data++, WRITE, cur_pos, wz_count);
  327. ccw[-1].flags |= CCW_FLAG_CC;
  328. ccw_write_zero(ccw++, wz_count * blksize);
  329. cur_pos = wz_count;
  330. }
  331. /* We can do proper discard when we've got at least blocks_per_page blocks. */
  332. if (last_rec - (first_rec + cur_pos) + 1 >= blocks_per_page) {
  333. /* is last record at page boundary? */
  334. if ((last_rec - blocks_per_page + 1) % blocks_per_page != 0)
  335. d_stop = last_rec - ((last_rec - blocks_per_page + 1) %
  336. blocks_per_page);
  337. else
  338. d_stop = last_rec;
  339. d_count = d_stop - (first_rec + cur_pos) + 1;
  340. ccw[-1].flags |= CCW_FLAG_CC;
  341. locate_record(ccw++, LO_data++, WRITE, cur_pos, d_count);
  342. ccw[-1].flags |= CCW_FLAG_CC;
  343. ccw_write_no_data(ccw++);
  344. cur_pos += d_count;
  345. }
  346. /* We might still have some bits left which need to be zeroed. */
  347. if (cur_pos == 0 || first_rec + cur_pos - 1 < last_rec) {
  348. if (d_stop != 0)
  349. wz_count = last_rec - d_stop;
  350. else if (wz_stop != 0)
  351. wz_count = last_rec - wz_stop;
  352. else
  353. wz_count = count;
  354. ccw[-1].flags |= CCW_FLAG_CC;
  355. locate_record(ccw++, LO_data++, WRITE, cur_pos, wz_count);
  356. ccw[-1].flags |= CCW_FLAG_CC;
  357. ccw_write_zero(ccw++, wz_count * blksize);
  358. }
  359. if (blk_noretry_request(req) ||
  360. block->base->features & DASD_FEATURE_FAILFAST)
  361. set_bit(DASD_CQR_FLAGS_FAILFAST, &cqr->flags);
  362. cqr->startdev = memdev;
  363. cqr->memdev = memdev;
  364. cqr->block = block;
  365. cqr->expires = memdev->default_expires * HZ; /* default 5 minutes */
  366. cqr->retries = memdev->default_retries;
  367. cqr->buildclk = get_tod_clock();
  368. cqr->status = DASD_CQR_FILLED;
  369. return cqr;
  370. }
  371. static struct dasd_ccw_req *dasd_fba_build_cp_regular(
  372. struct dasd_device *memdev,
  373. struct dasd_block *block,
  374. struct request *req)
  375. {
  376. struct dasd_fba_private *private = block->base->private;
  377. unsigned long *idaws;
  378. struct LO_fba_data *LO_data;
  379. struct dasd_ccw_req *cqr;
  380. struct ccw1 *ccw;
  381. struct req_iterator iter;
  382. struct bio_vec bv;
  383. char *dst;
  384. int count, cidaw, cplength, datasize;
  385. sector_t recid, first_rec, last_rec;
  386. unsigned int blksize, off;
  387. unsigned char cmd;
  388. if (rq_data_dir(req) == READ) {
  389. cmd = DASD_FBA_CCW_READ;
  390. } else if (rq_data_dir(req) == WRITE) {
  391. cmd = DASD_FBA_CCW_WRITE;
  392. } else
  393. return ERR_PTR(-EINVAL);
  394. blksize = block->bp_block;
  395. /* Calculate record id of first and last block. */
  396. first_rec = blk_rq_pos(req) >> block->s2b_shift;
  397. last_rec =
  398. (blk_rq_pos(req) + blk_rq_sectors(req) - 1) >> block->s2b_shift;
  399. /* Check struct bio and count the number of blocks for the request. */
  400. count = 0;
  401. cidaw = 0;
  402. rq_for_each_segment(bv, req, iter) {
  403. if (bv.bv_len & (blksize - 1))
  404. /* Fba can only do full blocks. */
  405. return ERR_PTR(-EINVAL);
  406. count += bv.bv_len >> (block->s2b_shift + 9);
  407. if (idal_is_needed (page_address(bv.bv_page), bv.bv_len))
  408. cidaw += bv.bv_len / blksize;
  409. }
  410. /* Paranoia. */
  411. if (count != last_rec - first_rec + 1)
  412. return ERR_PTR(-EINVAL);
  413. /* 1x define extent + 1x locate record + number of blocks */
  414. cplength = 2 + count;
  415. /* 1x define extent + 1x locate record */
  416. datasize = sizeof(struct DE_fba_data) + sizeof(struct LO_fba_data) +
  417. cidaw * sizeof(unsigned long);
  418. /*
  419. * Find out number of additional locate record ccws if the device
  420. * can't do data chaining.
  421. */
  422. if (private->rdc_data.mode.bits.data_chain == 0) {
  423. cplength += count - 1;
  424. datasize += (count - 1)*sizeof(struct LO_fba_data);
  425. }
  426. /* Allocate the ccw request. */
  427. cqr = dasd_smalloc_request(DASD_FBA_MAGIC, cplength, datasize, memdev,
  428. blk_mq_rq_to_pdu(req));
  429. if (IS_ERR(cqr))
  430. return cqr;
  431. ccw = cqr->cpaddr;
  432. /* First ccw is define extent. */
  433. define_extent(ccw++, cqr->data, rq_data_dir(req),
  434. block->bp_block, blk_rq_pos(req), blk_rq_sectors(req));
  435. /* Build locate_record + read/write ccws. */
  436. idaws = (unsigned long *) (cqr->data + sizeof(struct DE_fba_data));
  437. LO_data = (struct LO_fba_data *) (idaws + cidaw);
  438. /* Locate record for all blocks for smart devices. */
  439. if (private->rdc_data.mode.bits.data_chain != 0) {
  440. ccw[-1].flags |= CCW_FLAG_CC;
  441. locate_record(ccw++, LO_data++, rq_data_dir(req), 0, count);
  442. }
  443. recid = first_rec;
  444. rq_for_each_segment(bv, req, iter) {
  445. dst = page_address(bv.bv_page) + bv.bv_offset;
  446. if (dasd_page_cache) {
  447. char *copy = kmem_cache_alloc(dasd_page_cache,
  448. GFP_DMA | __GFP_NOWARN);
  449. if (copy && rq_data_dir(req) == WRITE)
  450. memcpy(copy + bv.bv_offset, dst, bv.bv_len);
  451. if (copy)
  452. dst = copy + bv.bv_offset;
  453. }
  454. for (off = 0; off < bv.bv_len; off += blksize) {
  455. /* Locate record for stupid devices. */
  456. if (private->rdc_data.mode.bits.data_chain == 0) {
  457. ccw[-1].flags |= CCW_FLAG_CC;
  458. locate_record(ccw, LO_data++,
  459. rq_data_dir(req),
  460. recid - first_rec, 1);
  461. ccw->flags = CCW_FLAG_CC;
  462. ccw++;
  463. } else {
  464. if (recid > first_rec)
  465. ccw[-1].flags |= CCW_FLAG_DC;
  466. else
  467. ccw[-1].flags |= CCW_FLAG_CC;
  468. }
  469. ccw->cmd_code = cmd;
  470. ccw->count = block->bp_block;
  471. if (idal_is_needed(dst, blksize)) {
  472. ccw->cda = (__u32)(addr_t) idaws;
  473. ccw->flags = CCW_FLAG_IDA;
  474. idaws = idal_create_words(idaws, dst, blksize);
  475. } else {
  476. ccw->cda = (__u32)(addr_t) dst;
  477. ccw->flags = 0;
  478. }
  479. ccw++;
  480. dst += blksize;
  481. recid++;
  482. }
  483. }
  484. if (blk_noretry_request(req) ||
  485. block->base->features & DASD_FEATURE_FAILFAST)
  486. set_bit(DASD_CQR_FLAGS_FAILFAST, &cqr->flags);
  487. cqr->startdev = memdev;
  488. cqr->memdev = memdev;
  489. cqr->block = block;
  490. cqr->expires = memdev->default_expires * HZ; /* default 5 minutes */
  491. cqr->retries = memdev->default_retries;
  492. cqr->buildclk = get_tod_clock();
  493. cqr->status = DASD_CQR_FILLED;
  494. return cqr;
  495. }
  496. static struct dasd_ccw_req *dasd_fba_build_cp(struct dasd_device *memdev,
  497. struct dasd_block *block,
  498. struct request *req)
  499. {
  500. if (req_op(req) == REQ_OP_DISCARD || req_op(req) == REQ_OP_WRITE_ZEROES)
  501. return dasd_fba_build_cp_discard(memdev, block, req);
  502. else
  503. return dasd_fba_build_cp_regular(memdev, block, req);
  504. }
  505. static int
  506. dasd_fba_free_cp(struct dasd_ccw_req *cqr, struct request *req)
  507. {
  508. struct dasd_fba_private *private = cqr->block->base->private;
  509. struct ccw1 *ccw;
  510. struct req_iterator iter;
  511. struct bio_vec bv;
  512. char *dst, *cda;
  513. unsigned int blksize, off;
  514. int status;
  515. if (!dasd_page_cache)
  516. goto out;
  517. blksize = cqr->block->bp_block;
  518. ccw = cqr->cpaddr;
  519. /* Skip over define extent & locate record. */
  520. ccw++;
  521. if (private->rdc_data.mode.bits.data_chain != 0)
  522. ccw++;
  523. rq_for_each_segment(bv, req, iter) {
  524. dst = page_address(bv.bv_page) + bv.bv_offset;
  525. for (off = 0; off < bv.bv_len; off += blksize) {
  526. /* Skip locate record. */
  527. if (private->rdc_data.mode.bits.data_chain == 0)
  528. ccw++;
  529. if (dst) {
  530. if (ccw->flags & CCW_FLAG_IDA)
  531. cda = *((char **)((addr_t) ccw->cda));
  532. else
  533. cda = (char *)((addr_t) ccw->cda);
  534. if (dst != cda) {
  535. if (rq_data_dir(req) == READ)
  536. memcpy(dst, cda, bv.bv_len);
  537. kmem_cache_free(dasd_page_cache,
  538. (void *)((addr_t)cda & PAGE_MASK));
  539. }
  540. dst = NULL;
  541. }
  542. ccw++;
  543. }
  544. }
  545. out:
  546. status = cqr->status == DASD_CQR_DONE;
  547. dasd_sfree_request(cqr, cqr->memdev);
  548. return status;
  549. }
  550. static void dasd_fba_handle_terminated_request(struct dasd_ccw_req *cqr)
  551. {
  552. if (cqr->retries < 0)
  553. cqr->status = DASD_CQR_FAILED;
  554. else
  555. cqr->status = DASD_CQR_FILLED;
  556. };
  557. static int
  558. dasd_fba_fill_info(struct dasd_device * device,
  559. struct dasd_information2_t * info)
  560. {
  561. struct dasd_fba_private *private = device->private;
  562. info->label_block = 1;
  563. info->FBA_layout = 1;
  564. info->format = DASD_FORMAT_LDL;
  565. info->characteristics_size = sizeof(private->rdc_data);
  566. memcpy(info->characteristics, &private->rdc_data,
  567. sizeof(private->rdc_data));
  568. info->confdata_size = 0;
  569. return 0;
  570. }
  571. static void
  572. dasd_fba_dump_sense_dbf(struct dasd_device *device, struct irb *irb,
  573. char *reason)
  574. {
  575. u64 *sense;
  576. sense = (u64 *) dasd_get_sense(irb);
  577. if (sense) {
  578. DBF_DEV_EVENT(DBF_EMERG, device,
  579. "%s: %s %02x%02x%02x %016llx %016llx %016llx "
  580. "%016llx", reason,
  581. scsw_is_tm(&irb->scsw) ? "t" : "c",
  582. scsw_cc(&irb->scsw), scsw_cstat(&irb->scsw),
  583. scsw_dstat(&irb->scsw), sense[0], sense[1],
  584. sense[2], sense[3]);
  585. } else {
  586. DBF_DEV_EVENT(DBF_EMERG, device, "%s",
  587. "SORRY - NO VALID SENSE AVAILABLE\n");
  588. }
  589. }
  590. static void
  591. dasd_fba_dump_sense(struct dasd_device *device, struct dasd_ccw_req * req,
  592. struct irb *irb)
  593. {
  594. char *page;
  595. struct ccw1 *act, *end, *last;
  596. int len, sl, sct, count;
  597. page = (char *) get_zeroed_page(GFP_ATOMIC);
  598. if (page == NULL) {
  599. DBF_DEV_EVENT(DBF_WARNING, device, "%s",
  600. "No memory to dump sense data");
  601. return;
  602. }
  603. len = sprintf(page, PRINTK_HEADER
  604. " I/O status report for device %s:\n",
  605. dev_name(&device->cdev->dev));
  606. len += sprintf(page + len, PRINTK_HEADER
  607. " in req: %p CS: 0x%02X DS: 0x%02X\n", req,
  608. irb->scsw.cmd.cstat, irb->scsw.cmd.dstat);
  609. len += sprintf(page + len, PRINTK_HEADER
  610. " device %s: Failing CCW: %p\n",
  611. dev_name(&device->cdev->dev),
  612. (void *) (addr_t) irb->scsw.cmd.cpa);
  613. if (irb->esw.esw0.erw.cons) {
  614. for (sl = 0; sl < 4; sl++) {
  615. len += sprintf(page + len, PRINTK_HEADER
  616. " Sense(hex) %2d-%2d:",
  617. (8 * sl), ((8 * sl) + 7));
  618. for (sct = 0; sct < 8; sct++) {
  619. len += sprintf(page + len, " %02x",
  620. irb->ecw[8 * sl + sct]);
  621. }
  622. len += sprintf(page + len, "\n");
  623. }
  624. } else {
  625. len += sprintf(page + len, PRINTK_HEADER
  626. " SORRY - NO VALID SENSE AVAILABLE\n");
  627. }
  628. printk(KERN_ERR "%s", page);
  629. /* dump the Channel Program */
  630. /* print first CCWs (maximum 8) */
  631. act = req->cpaddr;
  632. for (last = act; last->flags & (CCW_FLAG_CC | CCW_FLAG_DC); last++);
  633. end = min(act + 8, last);
  634. len = sprintf(page, PRINTK_HEADER " Related CP in req: %p\n", req);
  635. while (act <= end) {
  636. len += sprintf(page + len, PRINTK_HEADER
  637. " CCW %p: %08X %08X DAT:",
  638. act, ((int *) act)[0], ((int *) act)[1]);
  639. for (count = 0; count < 32 && count < act->count;
  640. count += sizeof(int))
  641. len += sprintf(page + len, " %08X",
  642. ((int *) (addr_t) act->cda)
  643. [(count>>2)]);
  644. len += sprintf(page + len, "\n");
  645. act++;
  646. }
  647. printk(KERN_ERR "%s", page);
  648. /* print failing CCW area */
  649. len = 0;
  650. if (act < ((struct ccw1 *)(addr_t) irb->scsw.cmd.cpa) - 2) {
  651. act = ((struct ccw1 *)(addr_t) irb->scsw.cmd.cpa) - 2;
  652. len += sprintf(page + len, PRINTK_HEADER "......\n");
  653. }
  654. end = min((struct ccw1 *)(addr_t) irb->scsw.cmd.cpa + 2, last);
  655. while (act <= end) {
  656. len += sprintf(page + len, PRINTK_HEADER
  657. " CCW %p: %08X %08X DAT:",
  658. act, ((int *) act)[0], ((int *) act)[1]);
  659. for (count = 0; count < 32 && count < act->count;
  660. count += sizeof(int))
  661. len += sprintf(page + len, " %08X",
  662. ((int *) (addr_t) act->cda)
  663. [(count>>2)]);
  664. len += sprintf(page + len, "\n");
  665. act++;
  666. }
  667. /* print last CCWs */
  668. if (act < last - 2) {
  669. act = last - 2;
  670. len += sprintf(page + len, PRINTK_HEADER "......\n");
  671. }
  672. while (act <= last) {
  673. len += sprintf(page + len, PRINTK_HEADER
  674. " CCW %p: %08X %08X DAT:",
  675. act, ((int *) act)[0], ((int *) act)[1]);
  676. for (count = 0; count < 32 && count < act->count;
  677. count += sizeof(int))
  678. len += sprintf(page + len, " %08X",
  679. ((int *) (addr_t) act->cda)
  680. [(count>>2)]);
  681. len += sprintf(page + len, "\n");
  682. act++;
  683. }
  684. if (len > 0)
  685. printk(KERN_ERR "%s", page);
  686. free_page((unsigned long) page);
  687. }
  688. /*
  689. * max_blocks is dependent on the amount of storage that is available
  690. * in the static io buffer for each device. Currently each device has
  691. * 8192 bytes (=2 pages). For 64 bit one dasd_mchunkt_t structure has
  692. * 24 bytes, the struct dasd_ccw_req has 136 bytes and each block can use
  693. * up to 16 bytes (8 for the ccw and 8 for the idal pointer). In
  694. * addition we have one define extent ccw + 16 bytes of data and a
  695. * locate record ccw for each block (stupid devices!) + 16 bytes of data.
  696. * That makes:
  697. * (8192 - 24 - 136 - 8 - 16) / 40 = 200.2 blocks at maximum.
  698. * We want to fit two into the available memory so that we can immediately
  699. * start the next request if one finishes off. That makes 100.1 blocks
  700. * for one request. Give a little safety and the result is 96.
  701. */
  702. static struct dasd_discipline dasd_fba_discipline = {
  703. .owner = THIS_MODULE,
  704. .name = "FBA ",
  705. .ebcname = "FBA ",
  706. .max_blocks = 96,
  707. .check_device = dasd_fba_check_characteristics,
  708. .do_analysis = dasd_fba_do_analysis,
  709. .verify_path = dasd_generic_verify_path,
  710. .fill_geometry = dasd_fba_fill_geometry,
  711. .start_IO = dasd_start_IO,
  712. .term_IO = dasd_term_IO,
  713. .handle_terminated_request = dasd_fba_handle_terminated_request,
  714. .erp_action = dasd_fba_erp_action,
  715. .erp_postaction = dasd_fba_erp_postaction,
  716. .check_for_device_change = dasd_fba_check_for_device_change,
  717. .build_cp = dasd_fba_build_cp,
  718. .free_cp = dasd_fba_free_cp,
  719. .dump_sense = dasd_fba_dump_sense,
  720. .dump_sense_dbf = dasd_fba_dump_sense_dbf,
  721. .fill_info = dasd_fba_fill_info,
  722. };
  723. static int __init
  724. dasd_fba_init(void)
  725. {
  726. int ret;
  727. ASCEBC(dasd_fba_discipline.ebcname, 4);
  728. dasd_fba_zero_page = (void *)get_zeroed_page(GFP_KERNEL | GFP_DMA);
  729. if (!dasd_fba_zero_page)
  730. return -ENOMEM;
  731. ret = ccw_driver_register(&dasd_fba_driver);
  732. if (!ret)
  733. wait_for_device_probe();
  734. return ret;
  735. }
  736. static void __exit
  737. dasd_fba_cleanup(void)
  738. {
  739. ccw_driver_unregister(&dasd_fba_driver);
  740. free_page((unsigned long)dasd_fba_zero_page);
  741. }
  742. module_init(dasd_fba_init);
  743. module_exit(dasd_fba_cleanup);