|
@@ -2133,6 +2133,18 @@ int generic_file_readonly_mmap(struct file * file, struct vm_area_struct * vma)
|
|
|
EXPORT_SYMBOL(generic_file_mmap);
|
|
|
EXPORT_SYMBOL(generic_file_readonly_mmap);
|
|
|
|
|
|
+static struct page *wait_on_page_read(struct page *page)
|
|
|
+{
|
|
|
+ if (!IS_ERR(page)) {
|
|
|
+ wait_on_page_locked(page);
|
|
|
+ if (!PageUptodate(page)) {
|
|
|
+ page_cache_release(page);
|
|
|
+ page = ERR_PTR(-EIO);
|
|
|
+ }
|
|
|
+ }
|
|
|
+ return page;
|
|
|
+}
|
|
|
+
|
|
|
static struct page *__read_cache_page(struct address_space *mapping,
|
|
|
pgoff_t index,
|
|
|
int (*filler)(void *, struct page *),
|
|
@@ -2159,6 +2171,8 @@ repeat:
|
|
|
if (err < 0) {
|
|
|
page_cache_release(page);
|
|
|
page = ERR_PTR(err);
|
|
|
+ } else {
|
|
|
+ page = wait_on_page_read(page);
|
|
|
}
|
|
|
}
|
|
|
return page;
|
|
@@ -2195,6 +2209,10 @@ retry:
|
|
|
if (err < 0) {
|
|
|
page_cache_release(page);
|
|
|
return ERR_PTR(err);
|
|
|
+ } else {
|
|
|
+ page = wait_on_page_read(page);
|
|
|
+ if (IS_ERR(page))
|
|
|
+ return page;
|
|
|
}
|
|
|
out:
|
|
|
mark_page_accessed(page);
|
|
@@ -2202,40 +2220,25 @@ out:
|
|
|
}
|
|
|
|
|
|
/**
|
|
|
- * read_cache_page_async - read into page cache, fill it if needed
|
|
|
+ * read_cache_page - read into page cache, fill it if needed
|
|
|
* @mapping: the page's address_space
|
|
|
* @index: the page index
|
|
|
* @filler: function to perform the read
|
|
|
* @data: first arg to filler(data, page) function, often left as NULL
|
|
|
*
|
|
|
- * Same as read_cache_page, but don't wait for page to become unlocked
|
|
|
- * after submitting it to the filler.
|
|
|
- *
|
|
|
* Read into the page cache. If a page already exists, and PageUptodate() is
|
|
|
- * not set, try to fill the page but don't wait for it to become unlocked.
|
|
|
+ * not set, try to fill the page and wait for it to become unlocked.
|
|
|
*
|
|
|
* If the page does not get brought uptodate, return -EIO.
|
|
|
*/
|
|
|
-struct page *read_cache_page_async(struct address_space *mapping,
|
|
|
+struct page *read_cache_page(struct address_space *mapping,
|
|
|
pgoff_t index,
|
|
|
int (*filler)(void *, struct page *),
|
|
|
void *data)
|
|
|
{
|
|
|
return do_read_cache_page(mapping, index, filler, data, mapping_gfp_mask(mapping));
|
|
|
}
|
|
|
-EXPORT_SYMBOL(read_cache_page_async);
|
|
|
-
|
|
|
-static struct page *wait_on_page_read(struct page *page)
|
|
|
-{
|
|
|
- if (!IS_ERR(page)) {
|
|
|
- wait_on_page_locked(page);
|
|
|
- if (!PageUptodate(page)) {
|
|
|
- page_cache_release(page);
|
|
|
- page = ERR_PTR(-EIO);
|
|
|
- }
|
|
|
- }
|
|
|
- return page;
|
|
|
-}
|
|
|
+EXPORT_SYMBOL(read_cache_page);
|
|
|
|
|
|
/**
|
|
|
* read_cache_page_gfp - read into page cache, using specified page allocation flags.
|
|
@@ -2254,31 +2257,10 @@ struct page *read_cache_page_gfp(struct address_space *mapping,
|
|
|
{
|
|
|
filler_t *filler = (filler_t *)mapping->a_ops->readpage;
|
|
|
|
|
|
- return wait_on_page_read(do_read_cache_page(mapping, index, filler, NULL, gfp));
|
|
|
+ return do_read_cache_page(mapping, index, filler, NULL, gfp);
|
|
|
}
|
|
|
EXPORT_SYMBOL(read_cache_page_gfp);
|
|
|
|
|
|
-/**
|
|
|
- * read_cache_page - read into page cache, fill it if needed
|
|
|
- * @mapping: the page's address_space
|
|
|
- * @index: the page index
|
|
|
- * @filler: function to perform the read
|
|
|
- * @data: first arg to filler(data, page) function, often left as NULL
|
|
|
- *
|
|
|
- * Read into the page cache. If a page already exists, and PageUptodate() is
|
|
|
- * not set, try to fill the page then wait for it to become unlocked.
|
|
|
- *
|
|
|
- * If the page does not get brought uptodate, return -EIO.
|
|
|
- */
|
|
|
-struct page *read_cache_page(struct address_space *mapping,
|
|
|
- pgoff_t index,
|
|
|
- int (*filler)(void *, struct page *),
|
|
|
- void *data)
|
|
|
-{
|
|
|
- return wait_on_page_read(read_cache_page_async(mapping, index, filler, data));
|
|
|
-}
|
|
|
-EXPORT_SYMBOL(read_cache_page);
|
|
|
-
|
|
|
static size_t __iovec_copy_from_user_inatomic(char *vaddr,
|
|
|
const struct iovec *iov, size_t base, size_t bytes)
|
|
|
{
|