|
@@ -2094,14 +2094,14 @@ int nommu_shrink_inode_mappings(struct inode *inode, size_t size,
|
|
|
high = (size + PAGE_SIZE - 1) >> PAGE_SHIFT;
|
|
|
|
|
|
down_write(&nommu_region_sem);
|
|
|
- i_mmap_lock_write(inode->i_mapping);
|
|
|
+ i_mmap_lock_read(inode->i_mapping);
|
|
|
|
|
|
/* search for VMAs that fall within the dead zone */
|
|
|
vma_interval_tree_foreach(vma, &inode->i_mapping->i_mmap, low, high) {
|
|
|
/* found one - only interested if it's shared out of the page
|
|
|
* cache */
|
|
|
if (vma->vm_flags & VM_SHARED) {
|
|
|
- i_mmap_unlock_write(inode->i_mapping);
|
|
|
+ i_mmap_unlock_read(inode->i_mapping);
|
|
|
up_write(&nommu_region_sem);
|
|
|
return -ETXTBSY; /* not quite true, but near enough */
|
|
|
}
|
|
@@ -2113,8 +2113,7 @@ int nommu_shrink_inode_mappings(struct inode *inode, size_t size,
|
|
|
* we don't check for any regions that start beyond the EOF as there
|
|
|
* shouldn't be any
|
|
|
*/
|
|
|
- vma_interval_tree_foreach(vma, &inode->i_mapping->i_mmap,
|
|
|
- 0, ULONG_MAX) {
|
|
|
+ vma_interval_tree_foreach(vma, &inode->i_mapping->i_mmap, 0, ULONG_MAX) {
|
|
|
if (!(vma->vm_flags & VM_SHARED))
|
|
|
continue;
|
|
|
|
|
@@ -2129,7 +2128,7 @@ int nommu_shrink_inode_mappings(struct inode *inode, size_t size,
|
|
|
}
|
|
|
}
|
|
|
|
|
|
- i_mmap_unlock_write(inode->i_mapping);
|
|
|
+ i_mmap_unlock_read(inode->i_mapping);
|
|
|
up_write(&nommu_region_sem);
|
|
|
return 0;
|
|
|
}
|