iov_iter.c 17 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743
  1. #include <linux/export.h>
  2. #include <linux/uio.h>
  3. #include <linux/pagemap.h>
  4. #include <linux/slab.h>
  5. #include <linux/vmalloc.h>
  6. static size_t copy_page_to_iter_iovec(struct page *page, size_t offset, size_t bytes,
  7. struct iov_iter *i)
  8. {
  9. size_t skip, copy, left, wanted;
  10. const struct iovec *iov;
  11. char __user *buf;
  12. void *kaddr, *from;
  13. if (unlikely(bytes > i->count))
  14. bytes = i->count;
  15. if (unlikely(!bytes))
  16. return 0;
  17. wanted = bytes;
  18. iov = i->iov;
  19. skip = i->iov_offset;
  20. buf = iov->iov_base + skip;
  21. copy = min(bytes, iov->iov_len - skip);
  22. if (!fault_in_pages_writeable(buf, copy)) {
  23. kaddr = kmap_atomic(page);
  24. from = kaddr + offset;
  25. /* first chunk, usually the only one */
  26. left = __copy_to_user_inatomic(buf, from, copy);
  27. copy -= left;
  28. skip += copy;
  29. from += copy;
  30. bytes -= copy;
  31. while (unlikely(!left && bytes)) {
  32. iov++;
  33. buf = iov->iov_base;
  34. copy = min(bytes, iov->iov_len);
  35. left = __copy_to_user_inatomic(buf, from, copy);
  36. copy -= left;
  37. skip = copy;
  38. from += copy;
  39. bytes -= copy;
  40. }
  41. if (likely(!bytes)) {
  42. kunmap_atomic(kaddr);
  43. goto done;
  44. }
  45. offset = from - kaddr;
  46. buf += copy;
  47. kunmap_atomic(kaddr);
  48. copy = min(bytes, iov->iov_len - skip);
  49. }
  50. /* Too bad - revert to non-atomic kmap */
  51. kaddr = kmap(page);
  52. from = kaddr + offset;
  53. left = __copy_to_user(buf, from, copy);
  54. copy -= left;
  55. skip += copy;
  56. from += copy;
  57. bytes -= copy;
  58. while (unlikely(!left && bytes)) {
  59. iov++;
  60. buf = iov->iov_base;
  61. copy = min(bytes, iov->iov_len);
  62. left = __copy_to_user(buf, from, copy);
  63. copy -= left;
  64. skip = copy;
  65. from += copy;
  66. bytes -= copy;
  67. }
  68. kunmap(page);
  69. done:
  70. if (skip == iov->iov_len) {
  71. iov++;
  72. skip = 0;
  73. }
  74. i->count -= wanted - bytes;
  75. i->nr_segs -= iov - i->iov;
  76. i->iov = iov;
  77. i->iov_offset = skip;
  78. return wanted - bytes;
  79. }
  80. static size_t copy_page_from_iter_iovec(struct page *page, size_t offset, size_t bytes,
  81. struct iov_iter *i)
  82. {
  83. size_t skip, copy, left, wanted;
  84. const struct iovec *iov;
  85. char __user *buf;
  86. void *kaddr, *to;
  87. if (unlikely(bytes > i->count))
  88. bytes = i->count;
  89. if (unlikely(!bytes))
  90. return 0;
  91. wanted = bytes;
  92. iov = i->iov;
  93. skip = i->iov_offset;
  94. buf = iov->iov_base + skip;
  95. copy = min(bytes, iov->iov_len - skip);
  96. if (!fault_in_pages_readable(buf, copy)) {
  97. kaddr = kmap_atomic(page);
  98. to = kaddr + offset;
  99. /* first chunk, usually the only one */
  100. left = __copy_from_user_inatomic(to, buf, copy);
  101. copy -= left;
  102. skip += copy;
  103. to += copy;
  104. bytes -= copy;
  105. while (unlikely(!left && bytes)) {
  106. iov++;
  107. buf = iov->iov_base;
  108. copy = min(bytes, iov->iov_len);
  109. left = __copy_from_user_inatomic(to, buf, copy);
  110. copy -= left;
  111. skip = copy;
  112. to += copy;
  113. bytes -= copy;
  114. }
  115. if (likely(!bytes)) {
  116. kunmap_atomic(kaddr);
  117. goto done;
  118. }
  119. offset = to - kaddr;
  120. buf += copy;
  121. kunmap_atomic(kaddr);
  122. copy = min(bytes, iov->iov_len - skip);
  123. }
  124. /* Too bad - revert to non-atomic kmap */
  125. kaddr = kmap(page);
  126. to = kaddr + offset;
  127. left = __copy_from_user(to, buf, copy);
  128. copy -= left;
  129. skip += copy;
  130. to += copy;
  131. bytes -= copy;
  132. while (unlikely(!left && bytes)) {
  133. iov++;
  134. buf = iov->iov_base;
  135. copy = min(bytes, iov->iov_len);
  136. left = __copy_from_user(to, buf, copy);
  137. copy -= left;
  138. skip = copy;
  139. to += copy;
  140. bytes -= copy;
  141. }
  142. kunmap(page);
  143. done:
  144. if (skip == iov->iov_len) {
  145. iov++;
  146. skip = 0;
  147. }
  148. i->count -= wanted - bytes;
  149. i->nr_segs -= iov - i->iov;
  150. i->iov = iov;
  151. i->iov_offset = skip;
  152. return wanted - bytes;
  153. }
  154. static size_t __iovec_copy_from_user_inatomic(char *vaddr,
  155. const struct iovec *iov, size_t base, size_t bytes)
  156. {
  157. size_t copied = 0, left = 0;
  158. while (bytes) {
  159. char __user *buf = iov->iov_base + base;
  160. int copy = min(bytes, iov->iov_len - base);
  161. base = 0;
  162. left = __copy_from_user_inatomic(vaddr, buf, copy);
  163. copied += copy;
  164. bytes -= copy;
  165. vaddr += copy;
  166. iov++;
  167. if (unlikely(left))
  168. break;
  169. }
  170. return copied - left;
  171. }
  172. /*
  173. * Copy as much as we can into the page and return the number of bytes which
  174. * were successfully copied. If a fault is encountered then return the number of
  175. * bytes which were copied.
  176. */
  177. static size_t copy_from_user_atomic_iovec(struct page *page,
  178. struct iov_iter *i, unsigned long offset, size_t bytes)
  179. {
  180. char *kaddr;
  181. size_t copied;
  182. kaddr = kmap_atomic(page);
  183. if (likely(i->nr_segs == 1)) {
  184. int left;
  185. char __user *buf = i->iov->iov_base + i->iov_offset;
  186. left = __copy_from_user_inatomic(kaddr + offset, buf, bytes);
  187. copied = bytes - left;
  188. } else {
  189. copied = __iovec_copy_from_user_inatomic(kaddr + offset,
  190. i->iov, i->iov_offset, bytes);
  191. }
  192. kunmap_atomic(kaddr);
  193. return copied;
  194. }
  195. static void advance_iovec(struct iov_iter *i, size_t bytes)
  196. {
  197. BUG_ON(i->count < bytes);
  198. if (likely(i->nr_segs == 1)) {
  199. i->iov_offset += bytes;
  200. i->count -= bytes;
  201. } else {
  202. const struct iovec *iov = i->iov;
  203. size_t base = i->iov_offset;
  204. unsigned long nr_segs = i->nr_segs;
  205. /*
  206. * The !iov->iov_len check ensures we skip over unlikely
  207. * zero-length segments (without overruning the iovec).
  208. */
  209. while (bytes || unlikely(i->count && !iov->iov_len)) {
  210. int copy;
  211. copy = min(bytes, iov->iov_len - base);
  212. BUG_ON(!i->count || i->count < copy);
  213. i->count -= copy;
  214. bytes -= copy;
  215. base += copy;
  216. if (iov->iov_len == base) {
  217. iov++;
  218. nr_segs--;
  219. base = 0;
  220. }
  221. }
  222. i->iov = iov;
  223. i->iov_offset = base;
  224. i->nr_segs = nr_segs;
  225. }
  226. }
  227. /*
  228. * Fault in the first iovec of the given iov_iter, to a maximum length
  229. * of bytes. Returns 0 on success, or non-zero if the memory could not be
  230. * accessed (ie. because it is an invalid address).
  231. *
  232. * writev-intensive code may want this to prefault several iovecs -- that
  233. * would be possible (callers must not rely on the fact that _only_ the
  234. * first iovec will be faulted with the current implementation).
  235. */
  236. int iov_iter_fault_in_readable(struct iov_iter *i, size_t bytes)
  237. {
  238. if (!(i->type & ITER_BVEC)) {
  239. char __user *buf = i->iov->iov_base + i->iov_offset;
  240. bytes = min(bytes, i->iov->iov_len - i->iov_offset);
  241. return fault_in_pages_readable(buf, bytes);
  242. }
  243. return 0;
  244. }
  245. EXPORT_SYMBOL(iov_iter_fault_in_readable);
  246. static unsigned long alignment_iovec(const struct iov_iter *i)
  247. {
  248. const struct iovec *iov = i->iov;
  249. unsigned long res;
  250. size_t size = i->count;
  251. size_t n;
  252. if (!size)
  253. return 0;
  254. res = (unsigned long)iov->iov_base + i->iov_offset;
  255. n = iov->iov_len - i->iov_offset;
  256. if (n >= size)
  257. return res | size;
  258. size -= n;
  259. res |= n;
  260. while (size > (++iov)->iov_len) {
  261. res |= (unsigned long)iov->iov_base | iov->iov_len;
  262. size -= iov->iov_len;
  263. }
  264. res |= (unsigned long)iov->iov_base | size;
  265. return res;
  266. }
  267. void iov_iter_init(struct iov_iter *i, int direction,
  268. const struct iovec *iov, unsigned long nr_segs,
  269. size_t count)
  270. {
  271. /* It will get better. Eventually... */
  272. if (segment_eq(get_fs(), KERNEL_DS))
  273. direction |= ITER_KVEC;
  274. i->type = direction;
  275. i->iov = iov;
  276. i->nr_segs = nr_segs;
  277. i->iov_offset = 0;
  278. i->count = count;
  279. }
  280. EXPORT_SYMBOL(iov_iter_init);
  281. static ssize_t get_pages_iovec(struct iov_iter *i,
  282. struct page **pages, size_t maxsize,
  283. size_t *start)
  284. {
  285. size_t offset = i->iov_offset;
  286. const struct iovec *iov = i->iov;
  287. size_t len;
  288. unsigned long addr;
  289. int n;
  290. int res;
  291. len = iov->iov_len - offset;
  292. if (len > i->count)
  293. len = i->count;
  294. if (len > maxsize)
  295. len = maxsize;
  296. addr = (unsigned long)iov->iov_base + offset;
  297. len += *start = addr & (PAGE_SIZE - 1);
  298. addr &= ~(PAGE_SIZE - 1);
  299. n = (len + PAGE_SIZE - 1) / PAGE_SIZE;
  300. res = get_user_pages_fast(addr, n, (i->type & WRITE) != WRITE, pages);
  301. if (unlikely(res < 0))
  302. return res;
  303. return (res == n ? len : res * PAGE_SIZE) - *start;
  304. }
  305. static ssize_t get_pages_alloc_iovec(struct iov_iter *i,
  306. struct page ***pages, size_t maxsize,
  307. size_t *start)
  308. {
  309. size_t offset = i->iov_offset;
  310. const struct iovec *iov = i->iov;
  311. size_t len;
  312. unsigned long addr;
  313. void *p;
  314. int n;
  315. int res;
  316. len = iov->iov_len - offset;
  317. if (len > i->count)
  318. len = i->count;
  319. if (len > maxsize)
  320. len = maxsize;
  321. addr = (unsigned long)iov->iov_base + offset;
  322. len += *start = addr & (PAGE_SIZE - 1);
  323. addr &= ~(PAGE_SIZE - 1);
  324. n = (len + PAGE_SIZE - 1) / PAGE_SIZE;
  325. p = kmalloc(n * sizeof(struct page *), GFP_KERNEL);
  326. if (!p)
  327. p = vmalloc(n * sizeof(struct page *));
  328. if (!p)
  329. return -ENOMEM;
  330. res = get_user_pages_fast(addr, n, (i->type & WRITE) != WRITE, p);
  331. if (unlikely(res < 0)) {
  332. kvfree(p);
  333. return res;
  334. }
  335. *pages = p;
  336. return (res == n ? len : res * PAGE_SIZE) - *start;
  337. }
  338. static int iov_iter_npages_iovec(const struct iov_iter *i, int maxpages)
  339. {
  340. size_t offset = i->iov_offset;
  341. size_t size = i->count;
  342. const struct iovec *iov = i->iov;
  343. int npages = 0;
  344. int n;
  345. for (n = 0; size && n < i->nr_segs; n++, iov++) {
  346. unsigned long addr = (unsigned long)iov->iov_base + offset;
  347. size_t len = iov->iov_len - offset;
  348. offset = 0;
  349. if (unlikely(!len)) /* empty segment */
  350. continue;
  351. if (len > size)
  352. len = size;
  353. npages += (addr + len + PAGE_SIZE - 1) / PAGE_SIZE
  354. - addr / PAGE_SIZE;
  355. if (npages >= maxpages) /* don't bother going further */
  356. return maxpages;
  357. size -= len;
  358. offset = 0;
  359. }
  360. return min(npages, maxpages);
  361. }
  362. static void memcpy_from_page(char *to, struct page *page, size_t offset, size_t len)
  363. {
  364. char *from = kmap_atomic(page);
  365. memcpy(to, from + offset, len);
  366. kunmap_atomic(from);
  367. }
  368. static void memcpy_to_page(struct page *page, size_t offset, char *from, size_t len)
  369. {
  370. char *to = kmap_atomic(page);
  371. memcpy(to + offset, from, len);
  372. kunmap_atomic(to);
  373. }
  374. static size_t copy_page_to_iter_bvec(struct page *page, size_t offset, size_t bytes,
  375. struct iov_iter *i)
  376. {
  377. size_t skip, copy, wanted;
  378. const struct bio_vec *bvec;
  379. void *kaddr, *from;
  380. if (unlikely(bytes > i->count))
  381. bytes = i->count;
  382. if (unlikely(!bytes))
  383. return 0;
  384. wanted = bytes;
  385. bvec = i->bvec;
  386. skip = i->iov_offset;
  387. copy = min_t(size_t, bytes, bvec->bv_len - skip);
  388. kaddr = kmap_atomic(page);
  389. from = kaddr + offset;
  390. memcpy_to_page(bvec->bv_page, skip + bvec->bv_offset, from, copy);
  391. skip += copy;
  392. from += copy;
  393. bytes -= copy;
  394. while (bytes) {
  395. bvec++;
  396. copy = min(bytes, (size_t)bvec->bv_len);
  397. memcpy_to_page(bvec->bv_page, bvec->bv_offset, from, copy);
  398. skip = copy;
  399. from += copy;
  400. bytes -= copy;
  401. }
  402. kunmap_atomic(kaddr);
  403. if (skip == bvec->bv_len) {
  404. bvec++;
  405. skip = 0;
  406. }
  407. i->count -= wanted - bytes;
  408. i->nr_segs -= bvec - i->bvec;
  409. i->bvec = bvec;
  410. i->iov_offset = skip;
  411. return wanted - bytes;
  412. }
  413. static size_t copy_page_from_iter_bvec(struct page *page, size_t offset, size_t bytes,
  414. struct iov_iter *i)
  415. {
  416. size_t skip, copy, wanted;
  417. const struct bio_vec *bvec;
  418. void *kaddr, *to;
  419. if (unlikely(bytes > i->count))
  420. bytes = i->count;
  421. if (unlikely(!bytes))
  422. return 0;
  423. wanted = bytes;
  424. bvec = i->bvec;
  425. skip = i->iov_offset;
  426. kaddr = kmap_atomic(page);
  427. to = kaddr + offset;
  428. copy = min(bytes, bvec->bv_len - skip);
  429. memcpy_from_page(to, bvec->bv_page, bvec->bv_offset + skip, copy);
  430. to += copy;
  431. skip += copy;
  432. bytes -= copy;
  433. while (bytes) {
  434. bvec++;
  435. copy = min(bytes, (size_t)bvec->bv_len);
  436. memcpy_from_page(to, bvec->bv_page, bvec->bv_offset, copy);
  437. skip = copy;
  438. to += copy;
  439. bytes -= copy;
  440. }
  441. kunmap_atomic(kaddr);
  442. if (skip == bvec->bv_len) {
  443. bvec++;
  444. skip = 0;
  445. }
  446. i->count -= wanted;
  447. i->nr_segs -= bvec - i->bvec;
  448. i->bvec = bvec;
  449. i->iov_offset = skip;
  450. return wanted;
  451. }
  452. static size_t copy_from_user_bvec(struct page *page,
  453. struct iov_iter *i, unsigned long offset, size_t bytes)
  454. {
  455. char *kaddr;
  456. size_t left;
  457. const struct bio_vec *bvec;
  458. size_t base = i->iov_offset;
  459. kaddr = kmap_atomic(page);
  460. for (left = bytes, bvec = i->bvec; left; bvec++, base = 0) {
  461. size_t copy = min(left, bvec->bv_len - base);
  462. if (!bvec->bv_len)
  463. continue;
  464. memcpy_from_page(kaddr + offset, bvec->bv_page,
  465. bvec->bv_offset + base, copy);
  466. offset += copy;
  467. left -= copy;
  468. }
  469. kunmap_atomic(kaddr);
  470. return bytes;
  471. }
  472. static void advance_bvec(struct iov_iter *i, size_t bytes)
  473. {
  474. BUG_ON(i->count < bytes);
  475. if (likely(i->nr_segs == 1)) {
  476. i->iov_offset += bytes;
  477. i->count -= bytes;
  478. } else {
  479. const struct bio_vec *bvec = i->bvec;
  480. size_t base = i->iov_offset;
  481. unsigned long nr_segs = i->nr_segs;
  482. /*
  483. * The !iov->iov_len check ensures we skip over unlikely
  484. * zero-length segments (without overruning the iovec).
  485. */
  486. while (bytes || unlikely(i->count && !bvec->bv_len)) {
  487. int copy;
  488. copy = min(bytes, bvec->bv_len - base);
  489. BUG_ON(!i->count || i->count < copy);
  490. i->count -= copy;
  491. bytes -= copy;
  492. base += copy;
  493. if (bvec->bv_len == base) {
  494. bvec++;
  495. nr_segs--;
  496. base = 0;
  497. }
  498. }
  499. i->bvec = bvec;
  500. i->iov_offset = base;
  501. i->nr_segs = nr_segs;
  502. }
  503. }
  504. static unsigned long alignment_bvec(const struct iov_iter *i)
  505. {
  506. const struct bio_vec *bvec = i->bvec;
  507. unsigned long res;
  508. size_t size = i->count;
  509. size_t n;
  510. if (!size)
  511. return 0;
  512. res = bvec->bv_offset + i->iov_offset;
  513. n = bvec->bv_len - i->iov_offset;
  514. if (n >= size)
  515. return res | size;
  516. size -= n;
  517. res |= n;
  518. while (size > (++bvec)->bv_len) {
  519. res |= bvec->bv_offset | bvec->bv_len;
  520. size -= bvec->bv_len;
  521. }
  522. res |= bvec->bv_offset | size;
  523. return res;
  524. }
  525. static ssize_t get_pages_bvec(struct iov_iter *i,
  526. struct page **pages, size_t maxsize,
  527. size_t *start)
  528. {
  529. const struct bio_vec *bvec = i->bvec;
  530. size_t len = bvec->bv_len - i->iov_offset;
  531. if (len > i->count)
  532. len = i->count;
  533. if (len > maxsize)
  534. len = maxsize;
  535. *start = bvec->bv_offset + i->iov_offset;
  536. get_page(*pages = bvec->bv_page);
  537. return len;
  538. }
  539. static ssize_t get_pages_alloc_bvec(struct iov_iter *i,
  540. struct page ***pages, size_t maxsize,
  541. size_t *start)
  542. {
  543. const struct bio_vec *bvec = i->bvec;
  544. size_t len = bvec->bv_len - i->iov_offset;
  545. if (len > i->count)
  546. len = i->count;
  547. if (len > maxsize)
  548. len = maxsize;
  549. *start = bvec->bv_offset + i->iov_offset;
  550. *pages = kmalloc(sizeof(struct page *), GFP_KERNEL);
  551. if (!*pages)
  552. return -ENOMEM;
  553. get_page(**pages = bvec->bv_page);
  554. return len;
  555. }
  556. static int iov_iter_npages_bvec(const struct iov_iter *i, int maxpages)
  557. {
  558. size_t offset = i->iov_offset;
  559. size_t size = i->count;
  560. const struct bio_vec *bvec = i->bvec;
  561. int npages = 0;
  562. int n;
  563. for (n = 0; size && n < i->nr_segs; n++, bvec++) {
  564. size_t len = bvec->bv_len - offset;
  565. offset = 0;
  566. if (unlikely(!len)) /* empty segment */
  567. continue;
  568. if (len > size)
  569. len = size;
  570. npages++;
  571. if (npages >= maxpages) /* don't bother going further */
  572. return maxpages;
  573. size -= len;
  574. offset = 0;
  575. }
  576. return min(npages, maxpages);
  577. }
  578. size_t copy_page_to_iter(struct page *page, size_t offset, size_t bytes,
  579. struct iov_iter *i)
  580. {
  581. if (i->type & ITER_BVEC)
  582. return copy_page_to_iter_bvec(page, offset, bytes, i);
  583. else
  584. return copy_page_to_iter_iovec(page, offset, bytes, i);
  585. }
  586. EXPORT_SYMBOL(copy_page_to_iter);
  587. size_t copy_page_from_iter(struct page *page, size_t offset, size_t bytes,
  588. struct iov_iter *i)
  589. {
  590. if (i->type & ITER_BVEC)
  591. return copy_page_from_iter_bvec(page, offset, bytes, i);
  592. else
  593. return copy_page_from_iter_iovec(page, offset, bytes, i);
  594. }
  595. EXPORT_SYMBOL(copy_page_from_iter);
  596. size_t iov_iter_copy_from_user_atomic(struct page *page,
  597. struct iov_iter *i, unsigned long offset, size_t bytes)
  598. {
  599. if (i->type & ITER_BVEC)
  600. return copy_from_user_bvec(page, i, offset, bytes);
  601. else
  602. return copy_from_user_atomic_iovec(page, i, offset, bytes);
  603. }
  604. EXPORT_SYMBOL(iov_iter_copy_from_user_atomic);
  605. void iov_iter_advance(struct iov_iter *i, size_t size)
  606. {
  607. if (i->type & ITER_BVEC)
  608. advance_bvec(i, size);
  609. else
  610. advance_iovec(i, size);
  611. }
  612. EXPORT_SYMBOL(iov_iter_advance);
  613. /*
  614. * Return the count of just the current iov_iter segment.
  615. */
  616. size_t iov_iter_single_seg_count(const struct iov_iter *i)
  617. {
  618. if (i->nr_segs == 1)
  619. return i->count;
  620. else if (i->type & ITER_BVEC)
  621. return min(i->count, i->iov->iov_len - i->iov_offset);
  622. else
  623. return min(i->count, i->bvec->bv_len - i->iov_offset);
  624. }
  625. EXPORT_SYMBOL(iov_iter_single_seg_count);
  626. unsigned long iov_iter_alignment(const struct iov_iter *i)
  627. {
  628. if (i->type & ITER_BVEC)
  629. return alignment_bvec(i);
  630. else
  631. return alignment_iovec(i);
  632. }
  633. EXPORT_SYMBOL(iov_iter_alignment);
  634. ssize_t iov_iter_get_pages(struct iov_iter *i,
  635. struct page **pages, size_t maxsize,
  636. size_t *start)
  637. {
  638. if (i->type & ITER_BVEC)
  639. return get_pages_bvec(i, pages, maxsize, start);
  640. else
  641. return get_pages_iovec(i, pages, maxsize, start);
  642. }
  643. EXPORT_SYMBOL(iov_iter_get_pages);
  644. ssize_t iov_iter_get_pages_alloc(struct iov_iter *i,
  645. struct page ***pages, size_t maxsize,
  646. size_t *start)
  647. {
  648. if (i->type & ITER_BVEC)
  649. return get_pages_alloc_bvec(i, pages, maxsize, start);
  650. else
  651. return get_pages_alloc_iovec(i, pages, maxsize, start);
  652. }
  653. EXPORT_SYMBOL(iov_iter_get_pages_alloc);
  654. int iov_iter_npages(const struct iov_iter *i, int maxpages)
  655. {
  656. if (i->type & ITER_BVEC)
  657. return iov_iter_npages_bvec(i, maxpages);
  658. else
  659. return iov_iter_npages_iovec(i, maxpages);
  660. }
  661. EXPORT_SYMBOL(iov_iter_npages);