scatterlist.c 37 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352135313541355135613571358135913601361136213631364136513661367136813691370137113721373137413751376137713781379138013811382138313841385138613871388138913901391139213931394139513961397139813991400140114021403140414051406140714081409141014111412141314141415141614171418141914201421142214231424142514261427142814291430143114321433
  1. // SPDX-License-Identifier: GPL-2.0-only
  2. /*
  3. * Copyright (C) 2007 Jens Axboe <jens.axboe@oracle.com>
  4. *
  5. * Scatterlist handling helpers.
  6. */
  7. #include <linux/export.h>
  8. #include <linux/slab.h>
  9. #include <linux/scatterlist.h>
  10. #include <linux/highmem.h>
  11. #include <linux/kmemleak.h>
  12. #include <linux/bvec.h>
  13. #include <linux/uio.h>
  14. #include <linux/folio_queue.h>
  15. /**
  16. * sg_next - return the next scatterlist entry in a list
  17. * @sg: The current sg entry
  18. *
  19. * Description:
  20. * Usually the next entry will be @sg@ + 1, but if this sg element is part
  21. * of a chained scatterlist, it could jump to the start of a new
  22. * scatterlist array.
  23. *
  24. **/
  25. struct scatterlist *sg_next(struct scatterlist *sg)
  26. {
  27. if (sg_is_last(sg))
  28. return NULL;
  29. sg++;
  30. if (unlikely(sg_is_chain(sg)))
  31. sg = sg_chain_ptr(sg);
  32. return sg;
  33. }
  34. EXPORT_SYMBOL(sg_next);
  35. /**
  36. * sg_nents - return total count of entries in scatterlist
  37. * @sg: The scatterlist
  38. *
  39. * Description:
  40. * Allows to know how many entries are in sg, taking into account
  41. * chaining as well
  42. *
  43. **/
  44. int sg_nents(struct scatterlist *sg)
  45. {
  46. int nents;
  47. for (nents = 0; sg; sg = sg_next(sg))
  48. nents++;
  49. return nents;
  50. }
  51. EXPORT_SYMBOL(sg_nents);
  52. /**
  53. * sg_nents_for_len - return total count of entries in scatterlist
  54. * needed to satisfy the supplied length
  55. * @sg: The scatterlist
  56. * @len: The total required length
  57. *
  58. * Description:
  59. * Determines the number of entries in sg that are required to meet
  60. * the supplied length, taking into account chaining as well
  61. *
  62. * Returns:
  63. * the number of sg entries needed, negative error on failure
  64. *
  65. **/
  66. int sg_nents_for_len(struct scatterlist *sg, u64 len)
  67. {
  68. int nents;
  69. u64 total;
  70. if (!len)
  71. return 0;
  72. for (nents = 0, total = 0; sg; sg = sg_next(sg)) {
  73. nents++;
  74. total += sg->length;
  75. if (total >= len)
  76. return nents;
  77. }
  78. return -EINVAL;
  79. }
  80. EXPORT_SYMBOL(sg_nents_for_len);
  81. /**
  82. * sg_last - return the last scatterlist entry in a list
  83. * @sgl: First entry in the scatterlist
  84. * @nents: Number of entries in the scatterlist
  85. *
  86. * Description:
  87. * Should only be used casually, it (currently) scans the entire list
  88. * to get the last entry.
  89. *
  90. * Note that the @sgl@ pointer passed in need not be the first one,
  91. * the important bit is that @nents@ denotes the number of entries that
  92. * exist from @sgl@.
  93. *
  94. **/
  95. struct scatterlist *sg_last(struct scatterlist *sgl, unsigned int nents)
  96. {
  97. struct scatterlist *sg, *ret = NULL;
  98. unsigned int i;
  99. for_each_sg(sgl, sg, nents, i)
  100. ret = sg;
  101. BUG_ON(!sg_is_last(ret));
  102. return ret;
  103. }
  104. EXPORT_SYMBOL(sg_last);
  105. /**
  106. * sg_init_table - Initialize SG table
  107. * @sgl: The SG table
  108. * @nents: Number of entries in table
  109. *
  110. * Notes:
  111. * If this is part of a chained sg table, sg_mark_end() should be
  112. * used only on the last table part.
  113. *
  114. **/
  115. void sg_init_table(struct scatterlist *sgl, unsigned int nents)
  116. {
  117. memset(sgl, 0, sizeof(*sgl) * nents);
  118. sg_init_marker(sgl, nents);
  119. }
  120. EXPORT_SYMBOL(sg_init_table);
  121. /**
  122. * sg_init_one - Initialize a single entry sg list
  123. * @sg: SG entry
  124. * @buf: Virtual address for IO
  125. * @buflen: IO length
  126. *
  127. **/
  128. void sg_init_one(struct scatterlist *sg, const void *buf, unsigned int buflen)
  129. {
  130. sg_init_table(sg, 1);
  131. sg_set_buf(sg, buf, buflen);
  132. }
  133. EXPORT_SYMBOL(sg_init_one);
  134. /*
  135. * The default behaviour of sg_alloc_table() is to use these kmalloc/kfree
  136. * helpers.
  137. */
  138. static struct scatterlist *sg_kmalloc(unsigned int nents, gfp_t gfp_mask)
  139. {
  140. if (nents == SG_MAX_SINGLE_ALLOC) {
  141. /*
  142. * Kmemleak doesn't track page allocations as they are not
  143. * commonly used (in a raw form) for kernel data structures.
  144. * As we chain together a list of pages and then a normal
  145. * kmalloc (tracked by kmemleak), in order to for that last
  146. * allocation not to become decoupled (and thus a
  147. * false-positive) we need to inform kmemleak of all the
  148. * intermediate allocations.
  149. */
  150. void *ptr = (void *) __get_free_page(gfp_mask);
  151. kmemleak_alloc(ptr, PAGE_SIZE, 1, gfp_mask);
  152. return ptr;
  153. } else
  154. return kmalloc_array(nents, sizeof(struct scatterlist),
  155. gfp_mask);
  156. }
  157. static void sg_kfree(struct scatterlist *sg, unsigned int nents)
  158. {
  159. if (nents == SG_MAX_SINGLE_ALLOC) {
  160. kmemleak_free(sg);
  161. free_page((unsigned long) sg);
  162. } else
  163. kfree(sg);
  164. }
  165. /**
  166. * __sg_free_table - Free a previously mapped sg table
  167. * @table: The sg table header to use
  168. * @max_ents: The maximum number of entries per single scatterlist
  169. * @nents_first_chunk: Number of entries int the (preallocated) first
  170. * scatterlist chunk, 0 means no such preallocated first chunk
  171. * @free_fn: Free function
  172. * @num_ents: Number of entries in the table
  173. *
  174. * Description:
  175. * Free an sg table previously allocated and setup with
  176. * __sg_alloc_table(). The @max_ents value must be identical to
  177. * that previously used with __sg_alloc_table().
  178. *
  179. **/
  180. void __sg_free_table(struct sg_table *table, unsigned int max_ents,
  181. unsigned int nents_first_chunk, sg_free_fn *free_fn,
  182. unsigned int num_ents)
  183. {
  184. struct scatterlist *sgl, *next;
  185. unsigned curr_max_ents = nents_first_chunk ?: max_ents;
  186. if (unlikely(!table->sgl))
  187. return;
  188. sgl = table->sgl;
  189. while (num_ents) {
  190. unsigned int alloc_size = num_ents;
  191. unsigned int sg_size;
  192. /*
  193. * If we have more than max_ents segments left,
  194. * then assign 'next' to the sg table after the current one.
  195. * sg_size is then one less than alloc size, since the last
  196. * element is the chain pointer.
  197. */
  198. if (alloc_size > curr_max_ents) {
  199. next = sg_chain_ptr(&sgl[curr_max_ents - 1]);
  200. alloc_size = curr_max_ents;
  201. sg_size = alloc_size - 1;
  202. } else {
  203. sg_size = alloc_size;
  204. next = NULL;
  205. }
  206. num_ents -= sg_size;
  207. if (nents_first_chunk)
  208. nents_first_chunk = 0;
  209. else
  210. free_fn(sgl, alloc_size);
  211. sgl = next;
  212. curr_max_ents = max_ents;
  213. }
  214. table->sgl = NULL;
  215. }
  216. EXPORT_SYMBOL(__sg_free_table);
  217. /**
  218. * sg_free_append_table - Free a previously allocated append sg table.
  219. * @table: The mapped sg append table header
  220. *
  221. **/
  222. void sg_free_append_table(struct sg_append_table *table)
  223. {
  224. __sg_free_table(&table->sgt, SG_MAX_SINGLE_ALLOC, 0, sg_kfree,
  225. table->total_nents);
  226. }
  227. EXPORT_SYMBOL(sg_free_append_table);
  228. /**
  229. * sg_free_table - Free a previously allocated sg table
  230. * @table: The mapped sg table header
  231. *
  232. **/
  233. void sg_free_table(struct sg_table *table)
  234. {
  235. __sg_free_table(table, SG_MAX_SINGLE_ALLOC, 0, sg_kfree,
  236. table->orig_nents);
  237. }
  238. EXPORT_SYMBOL(sg_free_table);
  239. /**
  240. * __sg_alloc_table - Allocate and initialize an sg table with given allocator
  241. * @table: The sg table header to use
  242. * @nents: Number of entries in sg list
  243. * @max_ents: The maximum number of entries the allocator returns per call
  244. * @first_chunk: first SGL if preallocated (may be %NULL)
  245. * @nents_first_chunk: Number of entries in the (preallocated) first
  246. * scatterlist chunk, 0 means no such preallocated chunk provided by user
  247. * @gfp_mask: GFP allocation mask
  248. * @alloc_fn: Allocator to use
  249. *
  250. * Description:
  251. * This function returns a @table @nents long. The allocator is
  252. * defined to return scatterlist chunks of maximum size @max_ents.
  253. * Thus if @nents is bigger than @max_ents, the scatterlists will be
  254. * chained in units of @max_ents.
  255. *
  256. * Notes:
  257. * If this function returns non-0 (eg failure), the caller must call
  258. * __sg_free_table() to cleanup any leftover allocations.
  259. *
  260. **/
  261. int __sg_alloc_table(struct sg_table *table, unsigned int nents,
  262. unsigned int max_ents, struct scatterlist *first_chunk,
  263. unsigned int nents_first_chunk, gfp_t gfp_mask,
  264. sg_alloc_fn *alloc_fn)
  265. {
  266. struct scatterlist *sg, *prv;
  267. unsigned int left;
  268. unsigned curr_max_ents = nents_first_chunk ?: max_ents;
  269. unsigned prv_max_ents;
  270. memset(table, 0, sizeof(*table));
  271. if (nents == 0)
  272. return -EINVAL;
  273. #ifdef CONFIG_ARCH_NO_SG_CHAIN
  274. if (WARN_ON_ONCE(nents > max_ents))
  275. return -EINVAL;
  276. #endif
  277. left = nents;
  278. prv = NULL;
  279. do {
  280. unsigned int sg_size, alloc_size = left;
  281. if (alloc_size > curr_max_ents) {
  282. alloc_size = curr_max_ents;
  283. sg_size = alloc_size - 1;
  284. } else
  285. sg_size = alloc_size;
  286. left -= sg_size;
  287. if (first_chunk) {
  288. sg = first_chunk;
  289. first_chunk = NULL;
  290. } else {
  291. sg = alloc_fn(alloc_size, gfp_mask);
  292. }
  293. if (unlikely(!sg)) {
  294. /*
  295. * Adjust entry count to reflect that the last
  296. * entry of the previous table won't be used for
  297. * linkage. Without this, sg_kfree() may get
  298. * confused.
  299. */
  300. if (prv)
  301. table->nents = ++table->orig_nents;
  302. return -ENOMEM;
  303. }
  304. sg_init_table(sg, alloc_size);
  305. table->nents = table->orig_nents += sg_size;
  306. /*
  307. * If this is the first mapping, assign the sg table header.
  308. * If this is not the first mapping, chain previous part.
  309. */
  310. if (prv)
  311. sg_chain(prv, prv_max_ents, sg);
  312. else
  313. table->sgl = sg;
  314. /*
  315. * If no more entries after this one, mark the end
  316. */
  317. if (!left)
  318. sg_mark_end(&sg[sg_size - 1]);
  319. prv = sg;
  320. prv_max_ents = curr_max_ents;
  321. curr_max_ents = max_ents;
  322. } while (left);
  323. return 0;
  324. }
  325. EXPORT_SYMBOL(__sg_alloc_table);
  326. /**
  327. * sg_alloc_table - Allocate and initialize an sg table
  328. * @table: The sg table header to use
  329. * @nents: Number of entries in sg list
  330. * @gfp_mask: GFP allocation mask
  331. *
  332. * Description:
  333. * Allocate and initialize an sg table. If @nents@ is larger than
  334. * SG_MAX_SINGLE_ALLOC a chained sg table will be setup.
  335. *
  336. **/
  337. int sg_alloc_table(struct sg_table *table, unsigned int nents, gfp_t gfp_mask)
  338. {
  339. int ret;
  340. ret = __sg_alloc_table(table, nents, SG_MAX_SINGLE_ALLOC,
  341. NULL, 0, gfp_mask, sg_kmalloc);
  342. if (unlikely(ret))
  343. sg_free_table(table);
  344. return ret;
  345. }
  346. EXPORT_SYMBOL(sg_alloc_table);
  347. static struct scatterlist *get_next_sg(struct sg_append_table *table,
  348. struct scatterlist *cur,
  349. unsigned long needed_sges,
  350. gfp_t gfp_mask)
  351. {
  352. struct scatterlist *new_sg, *next_sg;
  353. unsigned int alloc_size;
  354. if (cur) {
  355. next_sg = sg_next(cur);
  356. /* Check if last entry should be keeped for chainning */
  357. if (!sg_is_last(next_sg) || needed_sges == 1)
  358. return next_sg;
  359. }
  360. alloc_size = min_t(unsigned long, needed_sges, SG_MAX_SINGLE_ALLOC);
  361. new_sg = sg_kmalloc(alloc_size, gfp_mask);
  362. if (!new_sg)
  363. return ERR_PTR(-ENOMEM);
  364. sg_init_table(new_sg, alloc_size);
  365. if (cur) {
  366. table->total_nents += alloc_size - 1;
  367. __sg_chain(next_sg, new_sg);
  368. } else {
  369. table->sgt.sgl = new_sg;
  370. table->total_nents = alloc_size;
  371. }
  372. return new_sg;
  373. }
  374. static bool pages_are_mergeable(struct page *a, struct page *b)
  375. {
  376. if (page_to_pfn(a) != page_to_pfn(b) + 1)
  377. return false;
  378. if (!zone_device_pages_have_same_pgmap(a, b))
  379. return false;
  380. return true;
  381. }
  382. /**
  383. * sg_alloc_append_table_from_pages - Allocate and initialize an append sg
  384. * table from an array of pages
  385. * @sgt_append: The sg append table to use
  386. * @pages: Pointer to an array of page pointers
  387. * @n_pages: Number of pages in the pages array
  388. * @offset: Offset from start of the first page to the start of a buffer
  389. * @size: Number of valid bytes in the buffer (after offset)
  390. * @max_segment: Maximum size of a scatterlist element in bytes
  391. * @left_pages: Left pages caller have to set after this call
  392. * @gfp_mask: GFP allocation mask
  393. *
  394. * Description:
  395. * In the first call it allocate and initialize an sg table from a list of
  396. * pages, else reuse the scatterlist from sgt_append. Contiguous ranges of
  397. * the pages are squashed into a single scatterlist entry up to the maximum
  398. * size specified in @max_segment. A user may provide an offset at a start
  399. * and a size of valid data in a buffer specified by the page array. The
  400. * returned sg table is released by sg_free_append_table
  401. *
  402. * Returns:
  403. * 0 on success, negative error on failure
  404. *
  405. * Notes:
  406. * If this function returns non-0 (eg failure), the caller must call
  407. * sg_free_append_table() to cleanup any leftover allocations.
  408. *
  409. * In the fist call, sgt_append must by initialized.
  410. */
  411. int sg_alloc_append_table_from_pages(struct sg_append_table *sgt_append,
  412. struct page **pages, unsigned int n_pages, unsigned int offset,
  413. unsigned long size, unsigned int max_segment,
  414. unsigned int left_pages, gfp_t gfp_mask)
  415. {
  416. unsigned int chunks, cur_page, seg_len, i, prv_len = 0;
  417. unsigned int added_nents = 0;
  418. struct scatterlist *s = sgt_append->prv;
  419. struct page *last_pg;
  420. /*
  421. * The algorithm below requires max_segment to be aligned to PAGE_SIZE
  422. * otherwise it can overshoot.
  423. */
  424. max_segment = ALIGN_DOWN(max_segment, PAGE_SIZE);
  425. if (WARN_ON(max_segment < PAGE_SIZE))
  426. return -EINVAL;
  427. if (IS_ENABLED(CONFIG_ARCH_NO_SG_CHAIN) && sgt_append->prv)
  428. return -EOPNOTSUPP;
  429. if (sgt_append->prv) {
  430. unsigned long next_pfn = (page_to_phys(sg_page(sgt_append->prv)) +
  431. sgt_append->prv->offset + sgt_append->prv->length) / PAGE_SIZE;
  432. if (WARN_ON(offset))
  433. return -EINVAL;
  434. /* Merge contiguous pages into the last SG */
  435. prv_len = sgt_append->prv->length;
  436. if (page_to_pfn(pages[0]) == next_pfn) {
  437. last_pg = pfn_to_page(next_pfn - 1);
  438. while (n_pages && pages_are_mergeable(pages[0], last_pg)) {
  439. if (sgt_append->prv->length + PAGE_SIZE > max_segment)
  440. break;
  441. sgt_append->prv->length += PAGE_SIZE;
  442. last_pg = pages[0];
  443. pages++;
  444. n_pages--;
  445. }
  446. if (!n_pages)
  447. goto out;
  448. }
  449. }
  450. /* compute number of contiguous chunks */
  451. chunks = 1;
  452. seg_len = 0;
  453. for (i = 1; i < n_pages; i++) {
  454. seg_len += PAGE_SIZE;
  455. if (seg_len >= max_segment ||
  456. !pages_are_mergeable(pages[i], pages[i - 1])) {
  457. chunks++;
  458. seg_len = 0;
  459. }
  460. }
  461. /* merging chunks and putting them into the scatterlist */
  462. cur_page = 0;
  463. for (i = 0; i < chunks; i++) {
  464. unsigned int j, chunk_size;
  465. /* look for the end of the current chunk */
  466. seg_len = 0;
  467. for (j = cur_page + 1; j < n_pages; j++) {
  468. seg_len += PAGE_SIZE;
  469. if (seg_len >= max_segment ||
  470. !pages_are_mergeable(pages[j], pages[j - 1]))
  471. break;
  472. }
  473. /* Pass how many chunks might be left */
  474. s = get_next_sg(sgt_append, s, chunks - i + left_pages,
  475. gfp_mask);
  476. if (IS_ERR(s)) {
  477. /*
  478. * Adjust entry length to be as before function was
  479. * called.
  480. */
  481. if (sgt_append->prv)
  482. sgt_append->prv->length = prv_len;
  483. return PTR_ERR(s);
  484. }
  485. chunk_size = ((j - cur_page) << PAGE_SHIFT) - offset;
  486. sg_set_page(s, pages[cur_page],
  487. min_t(unsigned long, size, chunk_size), offset);
  488. added_nents++;
  489. size -= chunk_size;
  490. offset = 0;
  491. cur_page = j;
  492. }
  493. sgt_append->sgt.nents += added_nents;
  494. sgt_append->sgt.orig_nents = sgt_append->sgt.nents;
  495. sgt_append->prv = s;
  496. out:
  497. if (!left_pages)
  498. sg_mark_end(s);
  499. return 0;
  500. }
  501. EXPORT_SYMBOL(sg_alloc_append_table_from_pages);
  502. /**
  503. * sg_alloc_table_from_pages_segment - Allocate and initialize an sg table from
  504. * an array of pages and given maximum
  505. * segment.
  506. * @sgt: The sg table header to use
  507. * @pages: Pointer to an array of page pointers
  508. * @n_pages: Number of pages in the pages array
  509. * @offset: Offset from start of the first page to the start of a buffer
  510. * @size: Number of valid bytes in the buffer (after offset)
  511. * @max_segment: Maximum size of a scatterlist element in bytes
  512. * @gfp_mask: GFP allocation mask
  513. *
  514. * Description:
  515. * Allocate and initialize an sg table from a list of pages. Contiguous
  516. * ranges of the pages are squashed into a single scatterlist node up to the
  517. * maximum size specified in @max_segment. A user may provide an offset at a
  518. * start and a size of valid data in a buffer specified by the page array.
  519. *
  520. * The returned sg table is released by sg_free_table.
  521. *
  522. * Returns:
  523. * 0 on success, negative error on failure
  524. */
  525. int sg_alloc_table_from_pages_segment(struct sg_table *sgt, struct page **pages,
  526. unsigned int n_pages, unsigned int offset,
  527. unsigned long size, unsigned int max_segment,
  528. gfp_t gfp_mask)
  529. {
  530. struct sg_append_table append = {};
  531. int err;
  532. err = sg_alloc_append_table_from_pages(&append, pages, n_pages, offset,
  533. size, max_segment, 0, gfp_mask);
  534. if (err) {
  535. sg_free_append_table(&append);
  536. return err;
  537. }
  538. memcpy(sgt, &append.sgt, sizeof(*sgt));
  539. WARN_ON(append.total_nents != sgt->orig_nents);
  540. return 0;
  541. }
  542. EXPORT_SYMBOL(sg_alloc_table_from_pages_segment);
  543. #ifdef CONFIG_SGL_ALLOC
  544. /**
  545. * sgl_alloc_order - allocate a scatterlist and its pages
  546. * @length: Length in bytes of the scatterlist. Must be at least one
  547. * @order: Second argument for alloc_pages()
  548. * @chainable: Whether or not to allocate an extra element in the scatterlist
  549. * for scatterlist chaining purposes
  550. * @gfp: Memory allocation flags
  551. * @nent_p: [out] Number of entries in the scatterlist that have pages
  552. *
  553. * Returns: A pointer to an initialized scatterlist or %NULL upon failure.
  554. */
  555. struct scatterlist *sgl_alloc_order(unsigned long long length,
  556. unsigned int order, bool chainable,
  557. gfp_t gfp, unsigned int *nent_p)
  558. {
  559. struct scatterlist *sgl, *sg;
  560. struct page *page;
  561. unsigned int nent, nalloc;
  562. u32 elem_len;
  563. nent = round_up(length, PAGE_SIZE << order) >> (PAGE_SHIFT + order);
  564. /* Check for integer overflow */
  565. if (length > (nent << (PAGE_SHIFT + order)))
  566. return NULL;
  567. nalloc = nent;
  568. if (chainable) {
  569. /* Check for integer overflow */
  570. if (nalloc + 1 < nalloc)
  571. return NULL;
  572. nalloc++;
  573. }
  574. sgl = kmalloc_array(nalloc, sizeof(struct scatterlist),
  575. gfp & ~GFP_DMA);
  576. if (!sgl)
  577. return NULL;
  578. sg_init_table(sgl, nalloc);
  579. sg = sgl;
  580. while (length) {
  581. elem_len = min_t(u64, length, PAGE_SIZE << order);
  582. page = alloc_pages(gfp, order);
  583. if (!page) {
  584. sgl_free_order(sgl, order);
  585. return NULL;
  586. }
  587. sg_set_page(sg, page, elem_len, 0);
  588. length -= elem_len;
  589. sg = sg_next(sg);
  590. }
  591. WARN_ONCE(length, "length = %lld\n", length);
  592. if (nent_p)
  593. *nent_p = nent;
  594. return sgl;
  595. }
  596. EXPORT_SYMBOL(sgl_alloc_order);
  597. /**
  598. * sgl_alloc - allocate a scatterlist and its pages
  599. * @length: Length in bytes of the scatterlist
  600. * @gfp: Memory allocation flags
  601. * @nent_p: [out] Number of entries in the scatterlist
  602. *
  603. * Returns: A pointer to an initialized scatterlist or %NULL upon failure.
  604. */
  605. struct scatterlist *sgl_alloc(unsigned long long length, gfp_t gfp,
  606. unsigned int *nent_p)
  607. {
  608. return sgl_alloc_order(length, 0, false, gfp, nent_p);
  609. }
  610. EXPORT_SYMBOL(sgl_alloc);
  611. /**
  612. * sgl_free_n_order - free a scatterlist and its pages
  613. * @sgl: Scatterlist with one or more elements
  614. * @nents: Maximum number of elements to free
  615. * @order: Second argument for __free_pages()
  616. *
  617. * Notes:
  618. * - If several scatterlists have been chained and each chain element is
  619. * freed separately then it's essential to set nents correctly to avoid that a
  620. * page would get freed twice.
  621. * - All pages in a chained scatterlist can be freed at once by setting @nents
  622. * to a high number.
  623. */
  624. void sgl_free_n_order(struct scatterlist *sgl, int nents, int order)
  625. {
  626. struct scatterlist *sg;
  627. struct page *page;
  628. int i;
  629. for_each_sg(sgl, sg, nents, i) {
  630. if (!sg)
  631. break;
  632. page = sg_page(sg);
  633. if (page)
  634. __free_pages(page, order);
  635. }
  636. kfree(sgl);
  637. }
  638. EXPORT_SYMBOL(sgl_free_n_order);
  639. /**
  640. * sgl_free_order - free a scatterlist and its pages
  641. * @sgl: Scatterlist with one or more elements
  642. * @order: Second argument for __free_pages()
  643. */
  644. void sgl_free_order(struct scatterlist *sgl, int order)
  645. {
  646. sgl_free_n_order(sgl, INT_MAX, order);
  647. }
  648. EXPORT_SYMBOL(sgl_free_order);
  649. /**
  650. * sgl_free - free a scatterlist and its pages
  651. * @sgl: Scatterlist with one or more elements
  652. */
  653. void sgl_free(struct scatterlist *sgl)
  654. {
  655. sgl_free_order(sgl, 0);
  656. }
  657. EXPORT_SYMBOL(sgl_free);
  658. #endif /* CONFIG_SGL_ALLOC */
  659. void __sg_page_iter_start(struct sg_page_iter *piter,
  660. struct scatterlist *sglist, unsigned int nents,
  661. unsigned long pgoffset)
  662. {
  663. piter->__pg_advance = 0;
  664. piter->__nents = nents;
  665. piter->sg = sglist;
  666. piter->sg_pgoffset = pgoffset;
  667. }
  668. EXPORT_SYMBOL(__sg_page_iter_start);
  669. static int sg_page_count(struct scatterlist *sg)
  670. {
  671. return PAGE_ALIGN(sg->offset + sg->length) >> PAGE_SHIFT;
  672. }
  673. bool __sg_page_iter_next(struct sg_page_iter *piter)
  674. {
  675. if (!piter->__nents || !piter->sg)
  676. return false;
  677. piter->sg_pgoffset += piter->__pg_advance;
  678. piter->__pg_advance = 1;
  679. while (piter->sg_pgoffset >= sg_page_count(piter->sg)) {
  680. piter->sg_pgoffset -= sg_page_count(piter->sg);
  681. piter->sg = sg_next(piter->sg);
  682. if (!--piter->__nents || !piter->sg)
  683. return false;
  684. }
  685. return true;
  686. }
  687. EXPORT_SYMBOL(__sg_page_iter_next);
  688. static int sg_dma_page_count(struct scatterlist *sg)
  689. {
  690. return PAGE_ALIGN(sg->offset + sg_dma_len(sg)) >> PAGE_SHIFT;
  691. }
  692. bool __sg_page_iter_dma_next(struct sg_dma_page_iter *dma_iter)
  693. {
  694. struct sg_page_iter *piter = &dma_iter->base;
  695. if (!piter->__nents || !piter->sg)
  696. return false;
  697. piter->sg_pgoffset += piter->__pg_advance;
  698. piter->__pg_advance = 1;
  699. while (piter->sg_pgoffset >= sg_dma_page_count(piter->sg)) {
  700. piter->sg_pgoffset -= sg_dma_page_count(piter->sg);
  701. piter->sg = sg_next(piter->sg);
  702. if (!--piter->__nents || !piter->sg)
  703. return false;
  704. }
  705. return true;
  706. }
  707. EXPORT_SYMBOL(__sg_page_iter_dma_next);
  708. /**
  709. * sg_miter_start - start mapping iteration over a sg list
  710. * @miter: sg mapping iter to be started
  711. * @sgl: sg list to iterate over
  712. * @nents: number of sg entries
  713. * @flags: sg iterator flags
  714. *
  715. * Description:
  716. * Starts mapping iterator @miter.
  717. *
  718. * Context:
  719. * Don't care.
  720. */
  721. void sg_miter_start(struct sg_mapping_iter *miter, struct scatterlist *sgl,
  722. unsigned int nents, unsigned int flags)
  723. {
  724. memset(miter, 0, sizeof(struct sg_mapping_iter));
  725. __sg_page_iter_start(&miter->piter, sgl, nents, 0);
  726. WARN_ON(!(flags & (SG_MITER_TO_SG | SG_MITER_FROM_SG)));
  727. miter->__flags = flags;
  728. }
  729. EXPORT_SYMBOL(sg_miter_start);
  730. static bool sg_miter_get_next_page(struct sg_mapping_iter *miter)
  731. {
  732. if (!miter->__remaining) {
  733. struct scatterlist *sg;
  734. if (!__sg_page_iter_next(&miter->piter))
  735. return false;
  736. sg = miter->piter.sg;
  737. miter->__offset = miter->piter.sg_pgoffset ? 0 : sg->offset;
  738. miter->piter.sg_pgoffset += miter->__offset >> PAGE_SHIFT;
  739. miter->__offset &= PAGE_SIZE - 1;
  740. miter->__remaining = sg->offset + sg->length -
  741. (miter->piter.sg_pgoffset << PAGE_SHIFT) -
  742. miter->__offset;
  743. miter->__remaining = min_t(unsigned long, miter->__remaining,
  744. PAGE_SIZE - miter->__offset);
  745. }
  746. return true;
  747. }
  748. /**
  749. * sg_miter_skip - reposition mapping iterator
  750. * @miter: sg mapping iter to be skipped
  751. * @offset: number of bytes to plus the current location
  752. *
  753. * Description:
  754. * Sets the offset of @miter to its current location plus @offset bytes.
  755. * If mapping iterator @miter has been proceeded by sg_miter_next(), this
  756. * stops @miter.
  757. *
  758. * Context:
  759. * Don't care.
  760. *
  761. * Returns:
  762. * true if @miter contains the valid mapping. false if end of sg
  763. * list is reached.
  764. */
  765. bool sg_miter_skip(struct sg_mapping_iter *miter, off_t offset)
  766. {
  767. sg_miter_stop(miter);
  768. while (offset) {
  769. off_t consumed;
  770. if (!sg_miter_get_next_page(miter))
  771. return false;
  772. consumed = min_t(off_t, offset, miter->__remaining);
  773. miter->__offset += consumed;
  774. miter->__remaining -= consumed;
  775. offset -= consumed;
  776. }
  777. return true;
  778. }
  779. EXPORT_SYMBOL(sg_miter_skip);
  780. /**
  781. * sg_miter_next - proceed mapping iterator to the next mapping
  782. * @miter: sg mapping iter to proceed
  783. *
  784. * Description:
  785. * Proceeds @miter to the next mapping. @miter should have been started
  786. * using sg_miter_start(). On successful return, @miter->page,
  787. * @miter->addr and @miter->length point to the current mapping.
  788. *
  789. * Context:
  790. * May sleep if !SG_MITER_ATOMIC.
  791. *
  792. * Returns:
  793. * true if @miter contains the next mapping. false if end of sg
  794. * list is reached.
  795. */
  796. bool sg_miter_next(struct sg_mapping_iter *miter)
  797. {
  798. sg_miter_stop(miter);
  799. /*
  800. * Get to the next page if necessary.
  801. * __remaining, __offset is adjusted by sg_miter_stop
  802. */
  803. if (!sg_miter_get_next_page(miter))
  804. return false;
  805. miter->page = sg_page_iter_page(&miter->piter);
  806. miter->consumed = miter->length = miter->__remaining;
  807. if (miter->__flags & SG_MITER_ATOMIC)
  808. miter->addr = kmap_atomic(miter->page) + miter->__offset;
  809. else
  810. miter->addr = kmap(miter->page) + miter->__offset;
  811. return true;
  812. }
  813. EXPORT_SYMBOL(sg_miter_next);
  814. /**
  815. * sg_miter_stop - stop mapping iteration
  816. * @miter: sg mapping iter to be stopped
  817. *
  818. * Description:
  819. * Stops mapping iterator @miter. @miter should have been started
  820. * using sg_miter_start(). A stopped iteration can be resumed by
  821. * calling sg_miter_next() on it. This is useful when resources (kmap)
  822. * need to be released during iteration.
  823. *
  824. * Context:
  825. * Don't care otherwise.
  826. */
  827. void sg_miter_stop(struct sg_mapping_iter *miter)
  828. {
  829. WARN_ON(miter->consumed > miter->length);
  830. /* drop resources from the last iteration */
  831. if (miter->addr) {
  832. miter->__offset += miter->consumed;
  833. miter->__remaining -= miter->consumed;
  834. if (miter->__flags & SG_MITER_TO_SG)
  835. flush_dcache_page(miter->page);
  836. if (miter->__flags & SG_MITER_ATOMIC) {
  837. WARN_ON_ONCE(!pagefault_disabled());
  838. kunmap_atomic(miter->addr);
  839. } else
  840. kunmap(miter->page);
  841. miter->page = NULL;
  842. miter->addr = NULL;
  843. miter->length = 0;
  844. miter->consumed = 0;
  845. }
  846. }
  847. EXPORT_SYMBOL(sg_miter_stop);
  848. /**
  849. * sg_copy_buffer - Copy data between a linear buffer and an SG list
  850. * @sgl: The SG list
  851. * @nents: Number of SG entries
  852. * @buf: Where to copy from
  853. * @buflen: The number of bytes to copy
  854. * @skip: Number of bytes to skip before copying
  855. * @to_buffer: transfer direction (true == from an sg list to a
  856. * buffer, false == from a buffer to an sg list)
  857. *
  858. * Returns the number of copied bytes.
  859. *
  860. **/
  861. size_t sg_copy_buffer(struct scatterlist *sgl, unsigned int nents, void *buf,
  862. size_t buflen, off_t skip, bool to_buffer)
  863. {
  864. unsigned int offset = 0;
  865. struct sg_mapping_iter miter;
  866. unsigned int sg_flags = SG_MITER_ATOMIC;
  867. if (to_buffer)
  868. sg_flags |= SG_MITER_FROM_SG;
  869. else
  870. sg_flags |= SG_MITER_TO_SG;
  871. sg_miter_start(&miter, sgl, nents, sg_flags);
  872. if (!sg_miter_skip(&miter, skip))
  873. return 0;
  874. while ((offset < buflen) && sg_miter_next(&miter)) {
  875. unsigned int len;
  876. len = min(miter.length, buflen - offset);
  877. if (to_buffer)
  878. memcpy(buf + offset, miter.addr, len);
  879. else
  880. memcpy(miter.addr, buf + offset, len);
  881. offset += len;
  882. }
  883. sg_miter_stop(&miter);
  884. return offset;
  885. }
  886. EXPORT_SYMBOL(sg_copy_buffer);
  887. /**
  888. * sg_copy_from_buffer - Copy from a linear buffer to an SG list
  889. * @sgl: The SG list
  890. * @nents: Number of SG entries
  891. * @buf: Where to copy from
  892. * @buflen: The number of bytes to copy
  893. *
  894. * Returns the number of copied bytes.
  895. *
  896. **/
  897. size_t sg_copy_from_buffer(struct scatterlist *sgl, unsigned int nents,
  898. const void *buf, size_t buflen)
  899. {
  900. return sg_copy_buffer(sgl, nents, (void *)buf, buflen, 0, false);
  901. }
  902. EXPORT_SYMBOL(sg_copy_from_buffer);
  903. /**
  904. * sg_copy_to_buffer - Copy from an SG list to a linear buffer
  905. * @sgl: The SG list
  906. * @nents: Number of SG entries
  907. * @buf: Where to copy to
  908. * @buflen: The number of bytes to copy
  909. *
  910. * Returns the number of copied bytes.
  911. *
  912. **/
  913. size_t sg_copy_to_buffer(struct scatterlist *sgl, unsigned int nents,
  914. void *buf, size_t buflen)
  915. {
  916. return sg_copy_buffer(sgl, nents, buf, buflen, 0, true);
  917. }
  918. EXPORT_SYMBOL(sg_copy_to_buffer);
  919. /**
  920. * sg_pcopy_from_buffer - Copy from a linear buffer to an SG list
  921. * @sgl: The SG list
  922. * @nents: Number of SG entries
  923. * @buf: Where to copy from
  924. * @buflen: The number of bytes to copy
  925. * @skip: Number of bytes to skip before copying
  926. *
  927. * Returns the number of copied bytes.
  928. *
  929. **/
  930. size_t sg_pcopy_from_buffer(struct scatterlist *sgl, unsigned int nents,
  931. const void *buf, size_t buflen, off_t skip)
  932. {
  933. return sg_copy_buffer(sgl, nents, (void *)buf, buflen, skip, false);
  934. }
  935. EXPORT_SYMBOL(sg_pcopy_from_buffer);
  936. /**
  937. * sg_pcopy_to_buffer - Copy from an SG list to a linear buffer
  938. * @sgl: The SG list
  939. * @nents: Number of SG entries
  940. * @buf: Where to copy to
  941. * @buflen: The number of bytes to copy
  942. * @skip: Number of bytes to skip before copying
  943. *
  944. * Returns the number of copied bytes.
  945. *
  946. **/
  947. size_t sg_pcopy_to_buffer(struct scatterlist *sgl, unsigned int nents,
  948. void *buf, size_t buflen, off_t skip)
  949. {
  950. return sg_copy_buffer(sgl, nents, buf, buflen, skip, true);
  951. }
  952. EXPORT_SYMBOL(sg_pcopy_to_buffer);
  953. /**
  954. * sg_zero_buffer - Zero-out a part of a SG list
  955. * @sgl: The SG list
  956. * @nents: Number of SG entries
  957. * @buflen: The number of bytes to zero out
  958. * @skip: Number of bytes to skip before zeroing
  959. *
  960. * Returns the number of bytes zeroed.
  961. **/
  962. size_t sg_zero_buffer(struct scatterlist *sgl, unsigned int nents,
  963. size_t buflen, off_t skip)
  964. {
  965. unsigned int offset = 0;
  966. struct sg_mapping_iter miter;
  967. unsigned int sg_flags = SG_MITER_ATOMIC | SG_MITER_TO_SG;
  968. sg_miter_start(&miter, sgl, nents, sg_flags);
  969. if (!sg_miter_skip(&miter, skip))
  970. return false;
  971. while (offset < buflen && sg_miter_next(&miter)) {
  972. unsigned int len;
  973. len = min(miter.length, buflen - offset);
  974. memset(miter.addr, 0, len);
  975. offset += len;
  976. }
  977. sg_miter_stop(&miter);
  978. return offset;
  979. }
  980. EXPORT_SYMBOL(sg_zero_buffer);
  981. /*
  982. * Extract and pin a list of up to sg_max pages from UBUF- or IOVEC-class
  983. * iterators, and add them to the scatterlist.
  984. */
  985. static ssize_t extract_user_to_sg(struct iov_iter *iter,
  986. ssize_t maxsize,
  987. struct sg_table *sgtable,
  988. unsigned int sg_max,
  989. iov_iter_extraction_t extraction_flags)
  990. {
  991. struct scatterlist *sg = sgtable->sgl + sgtable->nents;
  992. struct page **pages;
  993. unsigned int npages;
  994. ssize_t ret = 0, res;
  995. size_t len, off;
  996. /* We decant the page list into the tail of the scatterlist */
  997. pages = (void *)sgtable->sgl +
  998. array_size(sg_max, sizeof(struct scatterlist));
  999. pages -= sg_max;
  1000. do {
  1001. res = iov_iter_extract_pages(iter, &pages, maxsize, sg_max,
  1002. extraction_flags, &off);
  1003. if (res <= 0)
  1004. goto failed;
  1005. len = res;
  1006. maxsize -= len;
  1007. ret += len;
  1008. npages = DIV_ROUND_UP(off + len, PAGE_SIZE);
  1009. sg_max -= npages;
  1010. for (; npages > 0; npages--) {
  1011. struct page *page = *pages;
  1012. size_t seg = min_t(size_t, PAGE_SIZE - off, len);
  1013. *pages++ = NULL;
  1014. sg_set_page(sg, page, seg, off);
  1015. sgtable->nents++;
  1016. sg++;
  1017. len -= seg;
  1018. off = 0;
  1019. }
  1020. } while (maxsize > 0 && sg_max > 0);
  1021. return ret;
  1022. failed:
  1023. while (sgtable->nents > sgtable->orig_nents)
  1024. unpin_user_page(sg_page(&sgtable->sgl[--sgtable->nents]));
  1025. return res;
  1026. }
  1027. /*
  1028. * Extract up to sg_max pages from a BVEC-type iterator and add them to the
  1029. * scatterlist. The pages are not pinned.
  1030. */
  1031. static ssize_t extract_bvec_to_sg(struct iov_iter *iter,
  1032. ssize_t maxsize,
  1033. struct sg_table *sgtable,
  1034. unsigned int sg_max,
  1035. iov_iter_extraction_t extraction_flags)
  1036. {
  1037. const struct bio_vec *bv = iter->bvec;
  1038. struct scatterlist *sg = sgtable->sgl + sgtable->nents;
  1039. unsigned long start = iter->iov_offset;
  1040. unsigned int i;
  1041. ssize_t ret = 0;
  1042. for (i = 0; i < iter->nr_segs; i++) {
  1043. size_t off, len;
  1044. len = bv[i].bv_len;
  1045. if (start >= len) {
  1046. start -= len;
  1047. continue;
  1048. }
  1049. len = min_t(size_t, maxsize, len - start);
  1050. off = bv[i].bv_offset + start;
  1051. sg_set_page(sg, bv[i].bv_page, len, off);
  1052. sgtable->nents++;
  1053. sg++;
  1054. sg_max--;
  1055. ret += len;
  1056. maxsize -= len;
  1057. if (maxsize <= 0 || sg_max == 0)
  1058. break;
  1059. start = 0;
  1060. }
  1061. if (ret > 0)
  1062. iov_iter_advance(iter, ret);
  1063. return ret;
  1064. }
  1065. /*
  1066. * Extract up to sg_max pages from a KVEC-type iterator and add them to the
  1067. * scatterlist. This can deal with vmalloc'd buffers as well as kmalloc'd or
  1068. * static buffers. The pages are not pinned.
  1069. */
  1070. static ssize_t extract_kvec_to_sg(struct iov_iter *iter,
  1071. ssize_t maxsize,
  1072. struct sg_table *sgtable,
  1073. unsigned int sg_max,
  1074. iov_iter_extraction_t extraction_flags)
  1075. {
  1076. const struct kvec *kv = iter->kvec;
  1077. struct scatterlist *sg = sgtable->sgl + sgtable->nents;
  1078. unsigned long start = iter->iov_offset;
  1079. unsigned int i;
  1080. ssize_t ret = 0;
  1081. for (i = 0; i < iter->nr_segs; i++) {
  1082. struct page *page;
  1083. unsigned long kaddr;
  1084. size_t off, len, seg;
  1085. len = kv[i].iov_len;
  1086. if (start >= len) {
  1087. start -= len;
  1088. continue;
  1089. }
  1090. kaddr = (unsigned long)kv[i].iov_base + start;
  1091. off = kaddr & ~PAGE_MASK;
  1092. len = min_t(size_t, maxsize, len - start);
  1093. kaddr &= PAGE_MASK;
  1094. maxsize -= len;
  1095. ret += len;
  1096. do {
  1097. seg = min_t(size_t, len, PAGE_SIZE - off);
  1098. if (is_vmalloc_or_module_addr((void *)kaddr))
  1099. page = vmalloc_to_page((void *)kaddr);
  1100. else
  1101. page = virt_to_page((void *)kaddr);
  1102. sg_set_page(sg, page, len, off);
  1103. sgtable->nents++;
  1104. sg++;
  1105. sg_max--;
  1106. len -= seg;
  1107. kaddr += PAGE_SIZE;
  1108. off = 0;
  1109. } while (len > 0 && sg_max > 0);
  1110. if (maxsize <= 0 || sg_max == 0)
  1111. break;
  1112. start = 0;
  1113. }
  1114. if (ret > 0)
  1115. iov_iter_advance(iter, ret);
  1116. return ret;
  1117. }
  1118. /*
  1119. * Extract up to sg_max folios from an FOLIOQ-type iterator and add them to
  1120. * the scatterlist. The pages are not pinned.
  1121. */
  1122. static ssize_t extract_folioq_to_sg(struct iov_iter *iter,
  1123. ssize_t maxsize,
  1124. struct sg_table *sgtable,
  1125. unsigned int sg_max,
  1126. iov_iter_extraction_t extraction_flags)
  1127. {
  1128. const struct folio_queue *folioq = iter->folioq;
  1129. struct scatterlist *sg = sgtable->sgl + sgtable->nents;
  1130. unsigned int slot = iter->folioq_slot;
  1131. ssize_t ret = 0;
  1132. size_t offset = iter->iov_offset;
  1133. BUG_ON(!folioq);
  1134. if (slot >= folioq_nr_slots(folioq)) {
  1135. folioq = folioq->next;
  1136. if (WARN_ON_ONCE(!folioq))
  1137. return 0;
  1138. slot = 0;
  1139. }
  1140. do {
  1141. struct folio *folio = folioq_folio(folioq, slot);
  1142. size_t fsize = folioq_folio_size(folioq, slot);
  1143. if (offset < fsize) {
  1144. size_t part = umin(maxsize - ret, fsize - offset);
  1145. sg_set_page(sg, folio_page(folio, 0), part, offset);
  1146. sgtable->nents++;
  1147. sg++;
  1148. sg_max--;
  1149. offset += part;
  1150. ret += part;
  1151. }
  1152. if (offset >= fsize) {
  1153. offset = 0;
  1154. slot++;
  1155. if (slot >= folioq_nr_slots(folioq)) {
  1156. if (!folioq->next) {
  1157. WARN_ON_ONCE(ret < iter->count);
  1158. break;
  1159. }
  1160. folioq = folioq->next;
  1161. slot = 0;
  1162. }
  1163. }
  1164. } while (sg_max > 0 && ret < maxsize);
  1165. iter->folioq = folioq;
  1166. iter->folioq_slot = slot;
  1167. iter->iov_offset = offset;
  1168. iter->count -= ret;
  1169. return ret;
  1170. }
  1171. /*
  1172. * Extract up to sg_max folios from an XARRAY-type iterator and add them to
  1173. * the scatterlist. The pages are not pinned.
  1174. */
  1175. static ssize_t extract_xarray_to_sg(struct iov_iter *iter,
  1176. ssize_t maxsize,
  1177. struct sg_table *sgtable,
  1178. unsigned int sg_max,
  1179. iov_iter_extraction_t extraction_flags)
  1180. {
  1181. struct scatterlist *sg = sgtable->sgl + sgtable->nents;
  1182. struct xarray *xa = iter->xarray;
  1183. struct folio *folio;
  1184. loff_t start = iter->xarray_start + iter->iov_offset;
  1185. pgoff_t index = start / PAGE_SIZE;
  1186. ssize_t ret = 0;
  1187. size_t offset, len;
  1188. XA_STATE(xas, xa, index);
  1189. rcu_read_lock();
  1190. xas_for_each(&xas, folio, ULONG_MAX) {
  1191. if (xas_retry(&xas, folio))
  1192. continue;
  1193. if (WARN_ON(xa_is_value(folio)))
  1194. break;
  1195. if (WARN_ON(folio_test_hugetlb(folio)))
  1196. break;
  1197. offset = offset_in_folio(folio, start);
  1198. len = min_t(size_t, maxsize, folio_size(folio) - offset);
  1199. sg_set_page(sg, folio_page(folio, 0), len, offset);
  1200. sgtable->nents++;
  1201. sg++;
  1202. sg_max--;
  1203. maxsize -= len;
  1204. ret += len;
  1205. if (maxsize <= 0 || sg_max == 0)
  1206. break;
  1207. }
  1208. rcu_read_unlock();
  1209. if (ret > 0)
  1210. iov_iter_advance(iter, ret);
  1211. return ret;
  1212. }
  1213. /**
  1214. * extract_iter_to_sg - Extract pages from an iterator and add to an sglist
  1215. * @iter: The iterator to extract from
  1216. * @maxsize: The amount of iterator to copy
  1217. * @sgtable: The scatterlist table to fill in
  1218. * @sg_max: Maximum number of elements in @sgtable that may be filled
  1219. * @extraction_flags: Flags to qualify the request
  1220. *
  1221. * Extract the page fragments from the given amount of the source iterator and
  1222. * add them to a scatterlist that refers to all of those bits, to a maximum
  1223. * addition of @sg_max elements.
  1224. *
  1225. * The pages referred to by UBUF- and IOVEC-type iterators are extracted and
  1226. * pinned; BVEC-, KVEC-, FOLIOQ- and XARRAY-type are extracted but aren't
  1227. * pinned; DISCARD-type is not supported.
  1228. *
  1229. * No end mark is placed on the scatterlist; that's left to the caller.
  1230. *
  1231. * @extraction_flags can have ITER_ALLOW_P2PDMA set to request peer-to-peer DMA
  1232. * be allowed on the pages extracted.
  1233. *
  1234. * If successful, @sgtable->nents is updated to include the number of elements
  1235. * added and the number of bytes added is returned. @sgtable->orig_nents is
  1236. * left unaltered.
  1237. *
  1238. * The iov_iter_extract_mode() function should be used to query how cleanup
  1239. * should be performed.
  1240. */
  1241. ssize_t extract_iter_to_sg(struct iov_iter *iter, size_t maxsize,
  1242. struct sg_table *sgtable, unsigned int sg_max,
  1243. iov_iter_extraction_t extraction_flags)
  1244. {
  1245. if (maxsize == 0)
  1246. return 0;
  1247. switch (iov_iter_type(iter)) {
  1248. case ITER_UBUF:
  1249. case ITER_IOVEC:
  1250. return extract_user_to_sg(iter, maxsize, sgtable, sg_max,
  1251. extraction_flags);
  1252. case ITER_BVEC:
  1253. return extract_bvec_to_sg(iter, maxsize, sgtable, sg_max,
  1254. extraction_flags);
  1255. case ITER_KVEC:
  1256. return extract_kvec_to_sg(iter, maxsize, sgtable, sg_max,
  1257. extraction_flags);
  1258. case ITER_FOLIOQ:
  1259. return extract_folioq_to_sg(iter, maxsize, sgtable, sg_max,
  1260. extraction_flags);
  1261. case ITER_XARRAY:
  1262. return extract_xarray_to_sg(iter, maxsize, sgtable, sg_max,
  1263. extraction_flags);
  1264. default:
  1265. pr_err("%s(%u) unsupported\n", __func__, iov_iter_type(iter));
  1266. WARN_ON_ONCE(1);
  1267. return -EIO;
  1268. }
  1269. }
  1270. EXPORT_SYMBOL_GPL(extract_iter_to_sg);