|
@@ -86,21 +86,6 @@ static inline void mapping_set_gfp_mask(struct address_space *m, gfp_t mask)
|
|
|
(__force unsigned long)mask;
|
|
|
}
|
|
|
|
|
|
-/*
|
|
|
- * The page cache can be done in larger chunks than
|
|
|
- * one page, because it allows for more efficient
|
|
|
- * throughput (it can then be mapped into user
|
|
|
- * space in smaller chunks for same flexibility).
|
|
|
- *
|
|
|
- * Or rather, it _will_ be done in larger chunks.
|
|
|
- */
|
|
|
-#define PAGE_CACHE_SHIFT PAGE_SHIFT
|
|
|
-#define PAGE_CACHE_SIZE PAGE_SIZE
|
|
|
-#define PAGE_CACHE_MASK PAGE_MASK
|
|
|
-#define PAGE_CACHE_ALIGN(addr) (((addr)+PAGE_CACHE_SIZE-1)&PAGE_CACHE_MASK)
|
|
|
-
|
|
|
-#define page_cache_get(page) get_page(page)
|
|
|
-#define page_cache_release(page) put_page(page)
|
|
|
void release_pages(struct page **pages, int nr, bool cold);
|
|
|
|
|
|
/*
|