mm: kill mm_wr_locked from unmap_vmas() and unmap_single_vma()

Kill mm_wr_locked since commit f8e97613fe ("mm: convert VM_PFNMAP
tracking to pfnmap_track() + pfnmap_untrack()") remove the user.

Link: https://lkml.kernel.org/r/20251104085709.2688433-1-wangkefeng.wang@huawei.com
Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com>
Reviewed-by: Lorenzo Stoakes <lorenzo.stoakes@oracle.com>
Reviewed-by: Liam R. Howlett <Liam.Howlett@oracle.com>
Acked-by: David Hildenbrand (Red Hat) <david@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
This commit is contained in:
Kefeng Wang
2025-11-04 16:57:09 +08:00
committed by Andrew Morton
parent 3b12a53b64
commit 340b59816b
5 changed files with 9 additions and 15 deletions

View File

@@ -2480,7 +2480,7 @@ static inline void zap_vma_pages(struct vm_area_struct *vma)
} }
void unmap_vmas(struct mmu_gather *tlb, struct ma_state *mas, void unmap_vmas(struct mmu_gather *tlb, struct ma_state *mas,
struct vm_area_struct *start_vma, unsigned long start, struct vm_area_struct *start_vma, unsigned long start,
unsigned long end, unsigned long tree_end, bool mm_wr_locked); unsigned long end, unsigned long tree_end);
struct mmu_notifier_range; struct mmu_notifier_range;

View File

@@ -2023,8 +2023,7 @@ void unmap_page_range(struct mmu_gather *tlb,
static void unmap_single_vma(struct mmu_gather *tlb, static void unmap_single_vma(struct mmu_gather *tlb,
struct vm_area_struct *vma, unsigned long start_addr, struct vm_area_struct *vma, unsigned long start_addr,
unsigned long end_addr, unsigned long end_addr, struct zap_details *details)
struct zap_details *details, bool mm_wr_locked)
{ {
unsigned long start = max(vma->vm_start, start_addr); unsigned long start = max(vma->vm_start, start_addr);
unsigned long end; unsigned long end;
@@ -2070,7 +2069,6 @@ static void unmap_single_vma(struct mmu_gather *tlb,
* @start_addr: virtual address at which to start unmapping * @start_addr: virtual address at which to start unmapping
* @end_addr: virtual address at which to end unmapping * @end_addr: virtual address at which to end unmapping
* @tree_end: The maximum index to check * @tree_end: The maximum index to check
* @mm_wr_locked: lock flag
* *
* Unmap all pages in the vma list. * Unmap all pages in the vma list.
* *
@@ -2085,8 +2083,7 @@ static void unmap_single_vma(struct mmu_gather *tlb,
*/ */
void unmap_vmas(struct mmu_gather *tlb, struct ma_state *mas, void unmap_vmas(struct mmu_gather *tlb, struct ma_state *mas,
struct vm_area_struct *vma, unsigned long start_addr, struct vm_area_struct *vma, unsigned long start_addr,
unsigned long end_addr, unsigned long tree_end, unsigned long end_addr, unsigned long tree_end)
bool mm_wr_locked)
{ {
struct mmu_notifier_range range; struct mmu_notifier_range range;
struct zap_details details = { struct zap_details details = {
@@ -2102,8 +2099,7 @@ void unmap_vmas(struct mmu_gather *tlb, struct ma_state *mas,
unsigned long start = start_addr; unsigned long start = start_addr;
unsigned long end = end_addr; unsigned long end = end_addr;
hugetlb_zap_begin(vma, &start, &end); hugetlb_zap_begin(vma, &start, &end);
unmap_single_vma(tlb, vma, start, end, &details, unmap_single_vma(tlb, vma, start, end, &details);
mm_wr_locked);
hugetlb_zap_end(vma, &details); hugetlb_zap_end(vma, &details);
vma = mas_find(mas, tree_end - 1); vma = mas_find(mas, tree_end - 1);
} while (vma && likely(!xa_is_zero(vma))); } while (vma && likely(!xa_is_zero(vma)));
@@ -2139,7 +2135,7 @@ void zap_page_range_single_batched(struct mmu_gather *tlb,
* unmap 'address-end' not 'range.start-range.end' as range * unmap 'address-end' not 'range.start-range.end' as range
* could have been expanded for hugetlb pmd sharing. * could have been expanded for hugetlb pmd sharing.
*/ */
unmap_single_vma(tlb, vma, address, end, details, false); unmap_single_vma(tlb, vma, address, end, details);
mmu_notifier_invalidate_range_end(&range); mmu_notifier_invalidate_range_end(&range);
if (is_vm_hugetlb_page(vma)) { if (is_vm_hugetlb_page(vma)) {
/* /*

View File

@@ -1274,7 +1274,7 @@ void exit_mmap(struct mm_struct *mm)
tlb_gather_mmu_fullmm(&tlb, mm); tlb_gather_mmu_fullmm(&tlb, mm);
/* update_hiwater_rss(mm) here? but nobody should be looking */ /* update_hiwater_rss(mm) here? but nobody should be looking */
/* Use ULONG_MAX here to ensure all VMAs in the mm are unmapped */ /* Use ULONG_MAX here to ensure all VMAs in the mm are unmapped */
unmap_vmas(&tlb, &vmi.mas, vma, 0, ULONG_MAX, ULONG_MAX, false); unmap_vmas(&tlb, &vmi.mas, vma, 0, ULONG_MAX, ULONG_MAX);
mmap_read_unlock(mm); mmap_read_unlock(mm);
/* /*

View File

@@ -483,8 +483,7 @@ void unmap_region(struct ma_state *mas, struct vm_area_struct *vma,
tlb_gather_mmu(&tlb, mm); tlb_gather_mmu(&tlb, mm);
update_hiwater_rss(mm); update_hiwater_rss(mm);
unmap_vmas(&tlb, mas, vma, vma->vm_start, vma->vm_end, vma->vm_end, unmap_vmas(&tlb, mas, vma, vma->vm_start, vma->vm_end, vma->vm_end);
/* mm_wr_locked = */ true);
mas_set(mas, vma->vm_end); mas_set(mas, vma->vm_end);
free_pgtables(&tlb, mas, vma, prev ? prev->vm_end : FIRST_USER_ADDRESS, free_pgtables(&tlb, mas, vma, prev ? prev->vm_end : FIRST_USER_ADDRESS,
next ? next->vm_start : USER_PGTABLES_CEILING, next ? next->vm_start : USER_PGTABLES_CEILING,
@@ -1228,7 +1227,7 @@ static inline void vms_clear_ptes(struct vma_munmap_struct *vms,
tlb_gather_mmu(&tlb, vms->vma->vm_mm); tlb_gather_mmu(&tlb, vms->vma->vm_mm);
update_hiwater_rss(vms->vma->vm_mm); update_hiwater_rss(vms->vma->vm_mm);
unmap_vmas(&tlb, mas_detach, vms->vma, vms->start, vms->end, unmap_vmas(&tlb, mas_detach, vms->vma, vms->start, vms->end,
vms->vma_count, mm_wr_locked); vms->vma_count);
mas_set(mas_detach, 1); mas_set(mas_detach, 1);
/* start and end may be different if there is no prev or next vma. */ /* start and end may be different if there is no prev or next vma. */

View File

@@ -848,8 +848,7 @@ static inline void update_hiwater_vm(struct mm_struct *mm)
static inline void unmap_vmas(struct mmu_gather *tlb, struct ma_state *mas, static inline void unmap_vmas(struct mmu_gather *tlb, struct ma_state *mas,
struct vm_area_struct *vma, unsigned long start_addr, struct vm_area_struct *vma, unsigned long start_addr,
unsigned long end_addr, unsigned long tree_end, unsigned long end_addr, unsigned long tree_end)
bool mm_wr_locked)
{ {
} }