iov_iter.c 17 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746
  1. #include <linux/export.h>
  2. #include <linux/uio.h>
  3. #include <linux/pagemap.h>
  4. #include <linux/slab.h>
  5. #include <linux/vmalloc.h>
  6. static size_t copy_page_to_iter_iovec(struct page *page, size_t offset, size_t bytes,
  7. struct iov_iter *i)
  8. {
  9. size_t skip, copy, left, wanted;
  10. const struct iovec *iov;
  11. char __user *buf;
  12. void *kaddr, *from;
  13. if (unlikely(bytes > i->count))
  14. bytes = i->count;
  15. if (unlikely(!bytes))
  16. return 0;
  17. wanted = bytes;
  18. iov = i->iov;
  19. skip = i->iov_offset;
  20. buf = iov->iov_base + skip;
  21. copy = min(bytes, iov->iov_len - skip);
  22. if (!fault_in_pages_writeable(buf, copy)) {
  23. kaddr = kmap_atomic(page);
  24. from = kaddr + offset;
  25. /* first chunk, usually the only one */
  26. left = __copy_to_user_inatomic(buf, from, copy);
  27. copy -= left;
  28. skip += copy;
  29. from += copy;
  30. bytes -= copy;
  31. while (unlikely(!left && bytes)) {
  32. iov++;
  33. buf = iov->iov_base;
  34. copy = min(bytes, iov->iov_len);
  35. left = __copy_to_user_inatomic(buf, from, copy);
  36. copy -= left;
  37. skip = copy;
  38. from += copy;
  39. bytes -= copy;
  40. }
  41. if (likely(!bytes)) {
  42. kunmap_atomic(kaddr);
  43. goto done;
  44. }
  45. offset = from - kaddr;
  46. buf += copy;
  47. kunmap_atomic(kaddr);
  48. copy = min(bytes, iov->iov_len - skip);
  49. }
  50. /* Too bad - revert to non-atomic kmap */
  51. kaddr = kmap(page);
  52. from = kaddr + offset;
  53. left = __copy_to_user(buf, from, copy);
  54. copy -= left;
  55. skip += copy;
  56. from += copy;
  57. bytes -= copy;
  58. while (unlikely(!left && bytes)) {
  59. iov++;
  60. buf = iov->iov_base;
  61. copy = min(bytes, iov->iov_len);
  62. left = __copy_to_user(buf, from, copy);
  63. copy -= left;
  64. skip = copy;
  65. from += copy;
  66. bytes -= copy;
  67. }
  68. kunmap(page);
  69. done:
  70. if (skip == iov->iov_len) {
  71. iov++;
  72. skip = 0;
  73. }
  74. i->count -= wanted - bytes;
  75. i->nr_segs -= iov - i->iov;
  76. i->iov = iov;
  77. i->iov_offset = skip;
  78. return wanted - bytes;
  79. }
  80. static size_t copy_page_from_iter_iovec(struct page *page, size_t offset, size_t bytes,
  81. struct iov_iter *i)
  82. {
  83. size_t skip, copy, left, wanted;
  84. const struct iovec *iov;
  85. char __user *buf;
  86. void *kaddr, *to;
  87. if (unlikely(bytes > i->count))
  88. bytes = i->count;
  89. if (unlikely(!bytes))
  90. return 0;
  91. wanted = bytes;
  92. iov = i->iov;
  93. skip = i->iov_offset;
  94. buf = iov->iov_base + skip;
  95. copy = min(bytes, iov->iov_len - skip);
  96. if (!fault_in_pages_readable(buf, copy)) {
  97. kaddr = kmap_atomic(page);
  98. to = kaddr + offset;
  99. /* first chunk, usually the only one */
  100. left = __copy_from_user_inatomic(to, buf, copy);
  101. copy -= left;
  102. skip += copy;
  103. to += copy;
  104. bytes -= copy;
  105. while (unlikely(!left && bytes)) {
  106. iov++;
  107. buf = iov->iov_base;
  108. copy = min(bytes, iov->iov_len);
  109. left = __copy_from_user_inatomic(to, buf, copy);
  110. copy -= left;
  111. skip = copy;
  112. to += copy;
  113. bytes -= copy;
  114. }
  115. if (likely(!bytes)) {
  116. kunmap_atomic(kaddr);
  117. goto done;
  118. }
  119. offset = to - kaddr;
  120. buf += copy;
  121. kunmap_atomic(kaddr);
  122. copy = min(bytes, iov->iov_len - skip);
  123. }
  124. /* Too bad - revert to non-atomic kmap */
  125. kaddr = kmap(page);
  126. to = kaddr + offset;
  127. left = __copy_from_user(to, buf, copy);
  128. copy -= left;
  129. skip += copy;
  130. to += copy;
  131. bytes -= copy;
  132. while (unlikely(!left && bytes)) {
  133. iov++;
  134. buf = iov->iov_base;
  135. copy = min(bytes, iov->iov_len);
  136. left = __copy_from_user(to, buf, copy);
  137. copy -= left;
  138. skip = copy;
  139. to += copy;
  140. bytes -= copy;
  141. }
  142. kunmap(page);
  143. done:
  144. if (skip == iov->iov_len) {
  145. iov++;
  146. skip = 0;
  147. }
  148. i->count -= wanted - bytes;
  149. i->nr_segs -= iov - i->iov;
  150. i->iov = iov;
  151. i->iov_offset = skip;
  152. return wanted - bytes;
  153. }
  154. static size_t __iovec_copy_from_user_inatomic(char *vaddr,
  155. const struct iovec *iov, size_t base, size_t bytes)
  156. {
  157. size_t copied = 0, left = 0;
  158. while (bytes) {
  159. char __user *buf = iov->iov_base + base;
  160. int copy = min(bytes, iov->iov_len - base);
  161. base = 0;
  162. left = __copy_from_user_inatomic(vaddr, buf, copy);
  163. copied += copy;
  164. bytes -= copy;
  165. vaddr += copy;
  166. iov++;
  167. if (unlikely(left))
  168. break;
  169. }
  170. return copied - left;
  171. }
  172. /*
  173. * Copy as much as we can into the page and return the number of bytes which
  174. * were successfully copied. If a fault is encountered then return the number of
  175. * bytes which were copied.
  176. */
  177. static size_t copy_from_user_atomic_iovec(struct page *page,
  178. struct iov_iter *i, unsigned long offset, size_t bytes)
  179. {
  180. char *kaddr;
  181. size_t copied;
  182. kaddr = kmap_atomic(page);
  183. if (likely(i->nr_segs == 1)) {
  184. int left;
  185. char __user *buf = i->iov->iov_base + i->iov_offset;
  186. left = __copy_from_user_inatomic(kaddr + offset, buf, bytes);
  187. copied = bytes - left;
  188. } else {
  189. copied = __iovec_copy_from_user_inatomic(kaddr + offset,
  190. i->iov, i->iov_offset, bytes);
  191. }
  192. kunmap_atomic(kaddr);
  193. return copied;
  194. }
  195. static void advance_iovec(struct iov_iter *i, size_t bytes)
  196. {
  197. BUG_ON(i->count < bytes);
  198. if (likely(i->nr_segs == 1)) {
  199. i->iov_offset += bytes;
  200. i->count -= bytes;
  201. } else {
  202. const struct iovec *iov = i->iov;
  203. size_t base = i->iov_offset;
  204. unsigned long nr_segs = i->nr_segs;
  205. /*
  206. * The !iov->iov_len check ensures we skip over unlikely
  207. * zero-length segments (without overruning the iovec).
  208. */
  209. while (bytes || unlikely(i->count && !iov->iov_len)) {
  210. int copy;
  211. copy = min(bytes, iov->iov_len - base);
  212. BUG_ON(!i->count || i->count < copy);
  213. i->count -= copy;
  214. bytes -= copy;
  215. base += copy;
  216. if (iov->iov_len == base) {
  217. iov++;
  218. nr_segs--;
  219. base = 0;
  220. }
  221. }
  222. i->iov = iov;
  223. i->iov_offset = base;
  224. i->nr_segs = nr_segs;
  225. }
  226. }
  227. /*
  228. * Fault in the first iovec of the given iov_iter, to a maximum length
  229. * of bytes. Returns 0 on success, or non-zero if the memory could not be
  230. * accessed (ie. because it is an invalid address).
  231. *
  232. * writev-intensive code may want this to prefault several iovecs -- that
  233. * would be possible (callers must not rely on the fact that _only_ the
  234. * first iovec will be faulted with the current implementation).
  235. */
  236. int iov_iter_fault_in_readable(struct iov_iter *i, size_t bytes)
  237. {
  238. if (!(i->type & ITER_BVEC)) {
  239. char __user *buf = i->iov->iov_base + i->iov_offset;
  240. bytes = min(bytes, i->iov->iov_len - i->iov_offset);
  241. return fault_in_pages_readable(buf, bytes);
  242. }
  243. return 0;
  244. }
  245. EXPORT_SYMBOL(iov_iter_fault_in_readable);
  246. static unsigned long alignment_iovec(const struct iov_iter *i)
  247. {
  248. const struct iovec *iov = i->iov;
  249. unsigned long res;
  250. size_t size = i->count;
  251. size_t n;
  252. if (!size)
  253. return 0;
  254. res = (unsigned long)iov->iov_base + i->iov_offset;
  255. n = iov->iov_len - i->iov_offset;
  256. if (n >= size)
  257. return res | size;
  258. size -= n;
  259. res |= n;
  260. while (size > (++iov)->iov_len) {
  261. res |= (unsigned long)iov->iov_base | iov->iov_len;
  262. size -= iov->iov_len;
  263. }
  264. res |= (unsigned long)iov->iov_base | size;
  265. return res;
  266. }
  267. void iov_iter_init(struct iov_iter *i, int direction,
  268. const struct iovec *iov, unsigned long nr_segs,
  269. size_t count)
  270. {
  271. /* It will get better. Eventually... */
  272. if (segment_eq(get_fs(), KERNEL_DS))
  273. direction |= ITER_KVEC;
  274. i->type = direction;
  275. i->iov = iov;
  276. i->nr_segs = nr_segs;
  277. i->iov_offset = 0;
  278. i->count = count;
  279. }
  280. EXPORT_SYMBOL(iov_iter_init);
  281. static ssize_t get_pages_iovec(struct iov_iter *i,
  282. struct page **pages, size_t maxsize, unsigned maxpages,
  283. size_t *start)
  284. {
  285. size_t offset = i->iov_offset;
  286. const struct iovec *iov = i->iov;
  287. size_t len;
  288. unsigned long addr;
  289. int n;
  290. int res;
  291. len = iov->iov_len - offset;
  292. if (len > i->count)
  293. len = i->count;
  294. if (len > maxsize)
  295. len = maxsize;
  296. addr = (unsigned long)iov->iov_base + offset;
  297. len += *start = addr & (PAGE_SIZE - 1);
  298. if (len > maxpages * PAGE_SIZE)
  299. len = maxpages * PAGE_SIZE;
  300. addr &= ~(PAGE_SIZE - 1);
  301. n = (len + PAGE_SIZE - 1) / PAGE_SIZE;
  302. res = get_user_pages_fast(addr, n, (i->type & WRITE) != WRITE, pages);
  303. if (unlikely(res < 0))
  304. return res;
  305. return (res == n ? len : res * PAGE_SIZE) - *start;
  306. }
  307. static ssize_t get_pages_alloc_iovec(struct iov_iter *i,
  308. struct page ***pages, size_t maxsize,
  309. size_t *start)
  310. {
  311. size_t offset = i->iov_offset;
  312. const struct iovec *iov = i->iov;
  313. size_t len;
  314. unsigned long addr;
  315. void *p;
  316. int n;
  317. int res;
  318. len = iov->iov_len - offset;
  319. if (len > i->count)
  320. len = i->count;
  321. if (len > maxsize)
  322. len = maxsize;
  323. addr = (unsigned long)iov->iov_base + offset;
  324. len += *start = addr & (PAGE_SIZE - 1);
  325. addr &= ~(PAGE_SIZE - 1);
  326. n = (len + PAGE_SIZE - 1) / PAGE_SIZE;
  327. p = kmalloc(n * sizeof(struct page *), GFP_KERNEL);
  328. if (!p)
  329. p = vmalloc(n * sizeof(struct page *));
  330. if (!p)
  331. return -ENOMEM;
  332. res = get_user_pages_fast(addr, n, (i->type & WRITE) != WRITE, p);
  333. if (unlikely(res < 0)) {
  334. kvfree(p);
  335. return res;
  336. }
  337. *pages = p;
  338. return (res == n ? len : res * PAGE_SIZE) - *start;
  339. }
  340. static int iov_iter_npages_iovec(const struct iov_iter *i, int maxpages)
  341. {
  342. size_t offset = i->iov_offset;
  343. size_t size = i->count;
  344. const struct iovec *iov = i->iov;
  345. int npages = 0;
  346. int n;
  347. for (n = 0; size && n < i->nr_segs; n++, iov++) {
  348. unsigned long addr = (unsigned long)iov->iov_base + offset;
  349. size_t len = iov->iov_len - offset;
  350. offset = 0;
  351. if (unlikely(!len)) /* empty segment */
  352. continue;
  353. if (len > size)
  354. len = size;
  355. npages += (addr + len + PAGE_SIZE - 1) / PAGE_SIZE
  356. - addr / PAGE_SIZE;
  357. if (npages >= maxpages) /* don't bother going further */
  358. return maxpages;
  359. size -= len;
  360. offset = 0;
  361. }
  362. return min(npages, maxpages);
  363. }
  364. static void memcpy_from_page(char *to, struct page *page, size_t offset, size_t len)
  365. {
  366. char *from = kmap_atomic(page);
  367. memcpy(to, from + offset, len);
  368. kunmap_atomic(from);
  369. }
  370. static void memcpy_to_page(struct page *page, size_t offset, char *from, size_t len)
  371. {
  372. char *to = kmap_atomic(page);
  373. memcpy(to + offset, from, len);
  374. kunmap_atomic(to);
  375. }
  376. static size_t copy_page_to_iter_bvec(struct page *page, size_t offset, size_t bytes,
  377. struct iov_iter *i)
  378. {
  379. size_t skip, copy, wanted;
  380. const struct bio_vec *bvec;
  381. void *kaddr, *from;
  382. if (unlikely(bytes > i->count))
  383. bytes = i->count;
  384. if (unlikely(!bytes))
  385. return 0;
  386. wanted = bytes;
  387. bvec = i->bvec;
  388. skip = i->iov_offset;
  389. copy = min_t(size_t, bytes, bvec->bv_len - skip);
  390. kaddr = kmap_atomic(page);
  391. from = kaddr + offset;
  392. memcpy_to_page(bvec->bv_page, skip + bvec->bv_offset, from, copy);
  393. skip += copy;
  394. from += copy;
  395. bytes -= copy;
  396. while (bytes) {
  397. bvec++;
  398. copy = min(bytes, (size_t)bvec->bv_len);
  399. memcpy_to_page(bvec->bv_page, bvec->bv_offset, from, copy);
  400. skip = copy;
  401. from += copy;
  402. bytes -= copy;
  403. }
  404. kunmap_atomic(kaddr);
  405. if (skip == bvec->bv_len) {
  406. bvec++;
  407. skip = 0;
  408. }
  409. i->count -= wanted - bytes;
  410. i->nr_segs -= bvec - i->bvec;
  411. i->bvec = bvec;
  412. i->iov_offset = skip;
  413. return wanted - bytes;
  414. }
  415. static size_t copy_page_from_iter_bvec(struct page *page, size_t offset, size_t bytes,
  416. struct iov_iter *i)
  417. {
  418. size_t skip, copy, wanted;
  419. const struct bio_vec *bvec;
  420. void *kaddr, *to;
  421. if (unlikely(bytes > i->count))
  422. bytes = i->count;
  423. if (unlikely(!bytes))
  424. return 0;
  425. wanted = bytes;
  426. bvec = i->bvec;
  427. skip = i->iov_offset;
  428. kaddr = kmap_atomic(page);
  429. to = kaddr + offset;
  430. copy = min(bytes, bvec->bv_len - skip);
  431. memcpy_from_page(to, bvec->bv_page, bvec->bv_offset + skip, copy);
  432. to += copy;
  433. skip += copy;
  434. bytes -= copy;
  435. while (bytes) {
  436. bvec++;
  437. copy = min(bytes, (size_t)bvec->bv_len);
  438. memcpy_from_page(to, bvec->bv_page, bvec->bv_offset, copy);
  439. skip = copy;
  440. to += copy;
  441. bytes -= copy;
  442. }
  443. kunmap_atomic(kaddr);
  444. if (skip == bvec->bv_len) {
  445. bvec++;
  446. skip = 0;
  447. }
  448. i->count -= wanted;
  449. i->nr_segs -= bvec - i->bvec;
  450. i->bvec = bvec;
  451. i->iov_offset = skip;
  452. return wanted;
  453. }
  454. static size_t copy_from_user_bvec(struct page *page,
  455. struct iov_iter *i, unsigned long offset, size_t bytes)
  456. {
  457. char *kaddr;
  458. size_t left;
  459. const struct bio_vec *bvec;
  460. size_t base = i->iov_offset;
  461. kaddr = kmap_atomic(page);
  462. for (left = bytes, bvec = i->bvec; left; bvec++, base = 0) {
  463. size_t copy = min(left, bvec->bv_len - base);
  464. if (!bvec->bv_len)
  465. continue;
  466. memcpy_from_page(kaddr + offset, bvec->bv_page,
  467. bvec->bv_offset + base, copy);
  468. offset += copy;
  469. left -= copy;
  470. }
  471. kunmap_atomic(kaddr);
  472. return bytes;
  473. }
  474. static void advance_bvec(struct iov_iter *i, size_t bytes)
  475. {
  476. BUG_ON(i->count < bytes);
  477. if (likely(i->nr_segs == 1)) {
  478. i->iov_offset += bytes;
  479. i->count -= bytes;
  480. } else {
  481. const struct bio_vec *bvec = i->bvec;
  482. size_t base = i->iov_offset;
  483. unsigned long nr_segs = i->nr_segs;
  484. /*
  485. * The !iov->iov_len check ensures we skip over unlikely
  486. * zero-length segments (without overruning the iovec).
  487. */
  488. while (bytes || unlikely(i->count && !bvec->bv_len)) {
  489. int copy;
  490. copy = min(bytes, bvec->bv_len - base);
  491. BUG_ON(!i->count || i->count < copy);
  492. i->count -= copy;
  493. bytes -= copy;
  494. base += copy;
  495. if (bvec->bv_len == base) {
  496. bvec++;
  497. nr_segs--;
  498. base = 0;
  499. }
  500. }
  501. i->bvec = bvec;
  502. i->iov_offset = base;
  503. i->nr_segs = nr_segs;
  504. }
  505. }
  506. static unsigned long alignment_bvec(const struct iov_iter *i)
  507. {
  508. const struct bio_vec *bvec = i->bvec;
  509. unsigned long res;
  510. size_t size = i->count;
  511. size_t n;
  512. if (!size)
  513. return 0;
  514. res = bvec->bv_offset + i->iov_offset;
  515. n = bvec->bv_len - i->iov_offset;
  516. if (n >= size)
  517. return res | size;
  518. size -= n;
  519. res |= n;
  520. while (size > (++bvec)->bv_len) {
  521. res |= bvec->bv_offset | bvec->bv_len;
  522. size -= bvec->bv_len;
  523. }
  524. res |= bvec->bv_offset | size;
  525. return res;
  526. }
  527. static ssize_t get_pages_bvec(struct iov_iter *i,
  528. struct page **pages, size_t maxsize, unsigned maxpages,
  529. size_t *start)
  530. {
  531. const struct bio_vec *bvec = i->bvec;
  532. size_t len = bvec->bv_len - i->iov_offset;
  533. if (len > i->count)
  534. len = i->count;
  535. if (len > maxsize)
  536. len = maxsize;
  537. /* can't be more than PAGE_SIZE */
  538. *start = bvec->bv_offset + i->iov_offset;
  539. get_page(*pages = bvec->bv_page);
  540. return len;
  541. }
  542. static ssize_t get_pages_alloc_bvec(struct iov_iter *i,
  543. struct page ***pages, size_t maxsize,
  544. size_t *start)
  545. {
  546. const struct bio_vec *bvec = i->bvec;
  547. size_t len = bvec->bv_len - i->iov_offset;
  548. if (len > i->count)
  549. len = i->count;
  550. if (len > maxsize)
  551. len = maxsize;
  552. *start = bvec->bv_offset + i->iov_offset;
  553. *pages = kmalloc(sizeof(struct page *), GFP_KERNEL);
  554. if (!*pages)
  555. return -ENOMEM;
  556. get_page(**pages = bvec->bv_page);
  557. return len;
  558. }
  559. static int iov_iter_npages_bvec(const struct iov_iter *i, int maxpages)
  560. {
  561. size_t offset = i->iov_offset;
  562. size_t size = i->count;
  563. const struct bio_vec *bvec = i->bvec;
  564. int npages = 0;
  565. int n;
  566. for (n = 0; size && n < i->nr_segs; n++, bvec++) {
  567. size_t len = bvec->bv_len - offset;
  568. offset = 0;
  569. if (unlikely(!len)) /* empty segment */
  570. continue;
  571. if (len > size)
  572. len = size;
  573. npages++;
  574. if (npages >= maxpages) /* don't bother going further */
  575. return maxpages;
  576. size -= len;
  577. offset = 0;
  578. }
  579. return min(npages, maxpages);
  580. }
  581. size_t copy_page_to_iter(struct page *page, size_t offset, size_t bytes,
  582. struct iov_iter *i)
  583. {
  584. if (i->type & ITER_BVEC)
  585. return copy_page_to_iter_bvec(page, offset, bytes, i);
  586. else
  587. return copy_page_to_iter_iovec(page, offset, bytes, i);
  588. }
  589. EXPORT_SYMBOL(copy_page_to_iter);
  590. size_t copy_page_from_iter(struct page *page, size_t offset, size_t bytes,
  591. struct iov_iter *i)
  592. {
  593. if (i->type & ITER_BVEC)
  594. return copy_page_from_iter_bvec(page, offset, bytes, i);
  595. else
  596. return copy_page_from_iter_iovec(page, offset, bytes, i);
  597. }
  598. EXPORT_SYMBOL(copy_page_from_iter);
  599. size_t iov_iter_copy_from_user_atomic(struct page *page,
  600. struct iov_iter *i, unsigned long offset, size_t bytes)
  601. {
  602. if (i->type & ITER_BVEC)
  603. return copy_from_user_bvec(page, i, offset, bytes);
  604. else
  605. return copy_from_user_atomic_iovec(page, i, offset, bytes);
  606. }
  607. EXPORT_SYMBOL(iov_iter_copy_from_user_atomic);
  608. void iov_iter_advance(struct iov_iter *i, size_t size)
  609. {
  610. if (i->type & ITER_BVEC)
  611. advance_bvec(i, size);
  612. else
  613. advance_iovec(i, size);
  614. }
  615. EXPORT_SYMBOL(iov_iter_advance);
  616. /*
  617. * Return the count of just the current iov_iter segment.
  618. */
  619. size_t iov_iter_single_seg_count(const struct iov_iter *i)
  620. {
  621. if (i->nr_segs == 1)
  622. return i->count;
  623. else if (i->type & ITER_BVEC)
  624. return min(i->count, i->iov->iov_len - i->iov_offset);
  625. else
  626. return min(i->count, i->bvec->bv_len - i->iov_offset);
  627. }
  628. EXPORT_SYMBOL(iov_iter_single_seg_count);
  629. unsigned long iov_iter_alignment(const struct iov_iter *i)
  630. {
  631. if (i->type & ITER_BVEC)
  632. return alignment_bvec(i);
  633. else
  634. return alignment_iovec(i);
  635. }
  636. EXPORT_SYMBOL(iov_iter_alignment);
  637. ssize_t iov_iter_get_pages(struct iov_iter *i,
  638. struct page **pages, size_t maxsize, unsigned maxpages,
  639. size_t *start)
  640. {
  641. if (i->type & ITER_BVEC)
  642. return get_pages_bvec(i, pages, maxsize, maxpages, start);
  643. else
  644. return get_pages_iovec(i, pages, maxsize, maxpages, start);
  645. }
  646. EXPORT_SYMBOL(iov_iter_get_pages);
  647. ssize_t iov_iter_get_pages_alloc(struct iov_iter *i,
  648. struct page ***pages, size_t maxsize,
  649. size_t *start)
  650. {
  651. if (i->type & ITER_BVEC)
  652. return get_pages_alloc_bvec(i, pages, maxsize, start);
  653. else
  654. return get_pages_alloc_iovec(i, pages, maxsize, start);
  655. }
  656. EXPORT_SYMBOL(iov_iter_get_pages_alloc);
  657. int iov_iter_npages(const struct iov_iter *i, int maxpages)
  658. {
  659. if (i->type & ITER_BVEC)
  660. return iov_iter_npages_bvec(i, maxpages);
  661. else
  662. return iov_iter_npages_iovec(i, maxpages);
  663. }
  664. EXPORT_SYMBOL(iov_iter_npages);