|
@@ -59,7 +59,7 @@ void copy_to_user_page(struct vm_area_struct *vma, struct page *page,
|
|
|
unsigned long vaddr, void *dst, const void *src,
|
|
|
unsigned long len)
|
|
|
{
|
|
|
- if (boot_cpu_data.dcache.n_aliases && page_mapped(page) &&
|
|
|
+ if (boot_cpu_data.dcache.n_aliases && page_mapcount(page) &&
|
|
|
test_bit(PG_dcache_clean, &page->flags)) {
|
|
|
void *vto = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK);
|
|
|
memcpy(vto, src, len);
|
|
@@ -78,7 +78,7 @@ void copy_from_user_page(struct vm_area_struct *vma, struct page *page,
|
|
|
unsigned long vaddr, void *dst, const void *src,
|
|
|
unsigned long len)
|
|
|
{
|
|
|
- if (boot_cpu_data.dcache.n_aliases && page_mapped(page) &&
|
|
|
+ if (boot_cpu_data.dcache.n_aliases && page_mapcount(page) &&
|
|
|
test_bit(PG_dcache_clean, &page->flags)) {
|
|
|
void *vfrom = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK);
|
|
|
memcpy(dst, vfrom, len);
|
|
@@ -97,7 +97,7 @@ void copy_user_highpage(struct page *to, struct page *from,
|
|
|
|
|
|
vto = kmap_atomic(to);
|
|
|
|
|
|
- if (boot_cpu_data.dcache.n_aliases && page_mapped(from) &&
|
|
|
+ if (boot_cpu_data.dcache.n_aliases && page_mapcount(from) &&
|
|
|
test_bit(PG_dcache_clean, &from->flags)) {
|
|
|
vfrom = kmap_coherent(from, vaddr);
|
|
|
copy_page(vto, vfrom);
|
|
@@ -153,7 +153,7 @@ void __flush_anon_page(struct page *page, unsigned long vmaddr)
|
|
|
unsigned long addr = (unsigned long) page_address(page);
|
|
|
|
|
|
if (pages_do_alias(addr, vmaddr)) {
|
|
|
- if (boot_cpu_data.dcache.n_aliases && page_mapped(page) &&
|
|
|
+ if (boot_cpu_data.dcache.n_aliases && page_mapcount(page) &&
|
|
|
test_bit(PG_dcache_clean, &page->flags)) {
|
|
|
void *kaddr;
|
|
|
|