aboutsummaryrefslogtreecommitdiff
path: root/mm/memory.c
diff options
context:
space:
mode:
authorAl Viro <viro@zeniv.linux.org.uk>2012-03-05 13:41:15 -0500
committerAl Viro <viro@zeniv.linux.org.uk>2012-03-20 21:39:51 -0400
commit6e8bb0193af3f308ef22817a5560422d33e58b90 (patch)
tree6001421c8d389bd00b18e0510e3f6c9130f9f80b /mm/memory.c
parent853f5e264018113b1f96f05551b07a74b836c7fc (diff)
VM: make unmap_vmas() return void
same story - nobody uses it and it's been pointless since "mm: Remove i_mmap_lock lockbreak" went in. Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
Diffstat (limited to 'mm/memory.c')
-rw-r--r--mm/memory.c6
1 files changed, 1 insertions, 5 deletions
diff --git a/mm/memory.c b/mm/memory.c
index cfb57b007a6..016c67587ef 100644
--- a/mm/memory.c
+++ b/mm/memory.c
@@ -1316,8 +1316,6 @@ static void unmap_page_range(struct mmu_gather *tlb,
* @nr_accounted: Place number of unmapped pages in vm-accountable vma's here
* @details: details of nonlinear truncation or shared cache invalidation
*
- * Returns the end address of the unmapping (restart addr if interrupted).
- *
* Unmap all pages in the vma list.
*
* Only addresses between `start' and `end' will be unmapped.
@@ -1329,7 +1327,7 @@ static void unmap_page_range(struct mmu_gather *tlb,
* ensure that any thus-far unmapped pages are flushed before unmap_vmas()
* drops the lock and schedules.
*/
-unsigned long unmap_vmas(struct mmu_gather *tlb,
+void unmap_vmas(struct mmu_gather *tlb,
struct vm_area_struct *vma, unsigned long start_addr,
unsigned long end_addr, unsigned long *nr_accounted,
struct zap_details *details)
@@ -1372,11 +1370,9 @@ unsigned long unmap_vmas(struct mmu_gather *tlb,
} else
unmap_page_range(tlb, vma, start, end, details);
}
- start = end;
}
mmu_notifier_invalidate_range_end(mm, start_addr, end_addr);
- return start; /* which is now the end (or restart) address */
}
/**