|
@@ -798,29 +798,19 @@ static int dax_insert_mapping(struct address_space *mapping,
|
|
|
.sector = to_sector(bh, mapping->host),
|
|
|
.size = bh->b_size,
|
|
|
};
|
|
|
- int error;
|
|
|
void *ret;
|
|
|
void *entry = *entryp;
|
|
|
|
|
|
- i_mmap_lock_read(mapping);
|
|
|
-
|
|
|
- if (dax_map_atomic(bdev, &dax) < 0) {
|
|
|
- error = PTR_ERR(dax.addr);
|
|
|
- goto out;
|
|
|
- }
|
|
|
+ if (dax_map_atomic(bdev, &dax) < 0)
|
|
|
+ return PTR_ERR(dax.addr);
|
|
|
dax_unmap_atomic(bdev, &dax);
|
|
|
|
|
|
ret = dax_insert_mapping_entry(mapping, vmf, entry, dax.sector);
|
|
|
- if (IS_ERR(ret)) {
|
|
|
- error = PTR_ERR(ret);
|
|
|
- goto out;
|
|
|
- }
|
|
|
+ if (IS_ERR(ret))
|
|
|
+ return PTR_ERR(ret);
|
|
|
*entryp = ret;
|
|
|
|
|
|
- error = vm_insert_mixed(vma, vaddr, dax.pfn);
|
|
|
- out:
|
|
|
- i_mmap_unlock_read(mapping);
|
|
|
- return error;
|
|
|
+ return vm_insert_mixed(vma, vaddr, dax.pfn);
|
|
|
}
|
|
|
|
|
|
/**
|
|
@@ -1058,8 +1048,6 @@ int __dax_pmd_fault(struct vm_area_struct *vma, unsigned long address,
|
|
|
truncate_pagecache_range(inode, lstart, lend);
|
|
|
}
|
|
|
|
|
|
- i_mmap_lock_read(mapping);
|
|
|
-
|
|
|
if (!write && !buffer_mapped(&bh)) {
|
|
|
spinlock_t *ptl;
|
|
|
pmd_t entry;
|
|
@@ -1148,8 +1136,6 @@ int __dax_pmd_fault(struct vm_area_struct *vma, unsigned long address,
|
|
|
}
|
|
|
|
|
|
out:
|
|
|
- i_mmap_unlock_read(mapping);
|
|
|
-
|
|
|
return result;
|
|
|
|
|
|
fallback:
|