diff mbox series

[v3] mm/rmap: do not add fully unmapped large folio to deferred split list

Message ID 20240424224614.477221-1-zi.yan@sent.com (mailing list archive)
State New
Headers show
Series [v3] mm/rmap: do not add fully unmapped large folio to deferred split list | expand

Commit Message

Zi Yan April 24, 2024, 10:46 p.m. UTC
From: Zi Yan <ziy@nvidia.com>

In __folio_remove_rmap(), a large folio is added to deferred split list
if any page in a folio loses its final mapping. It is possible that
the folio is unmapped fully, but it is unnecessary to add the folio
to deferred split list at all. Fix it by checking folio->_nr_pages_mapped
before adding a folio to deferred split list. If the folio is already
on the deferred split list, it will be skipped. This issue applies to
both PTE-mapped THP and mTHP.

Commit 98046944a159 ("mm: huge_memory: add the missing
folio_test_pmd_mappable() for THP split statistics") tried to exclude
mTHP deferred split stats from THP_DEFERRED_SPLIT_PAGE, but it does not
fix the above issue. A fully unmapped PTE-mapped order-9 THP was still
added to deferred split list and counted as THP_DEFERRED_SPLIT_PAGE,
since nr is 512 (non zero), level is RMAP_LEVEL_PTE, and inside
deferred_split_folio() the order-9 folio is folio_test_pmd_mappable().
However, this miscount was present even earlier due to implementation,
since PTEs are unmapped individually and first PTE unmapping adds the THP
into the deferred split list.

With commit b06dc281aa99 ("mm/rmap: introduce
folio_remove_rmap_[pte|ptes|pmd]()"), kernel is able to unmap PTE-mapped
folios in one shot without causing the miscount, hence this patch.

Signed-off-by: Zi Yan <ziy@nvidia.com>
Reviewed-by: Yang Shi <shy828301@gmail.com>
---
 mm/rmap.c | 7 ++++---
 1 file changed, 4 insertions(+), 3 deletions(-)


base-commit: 2541ee5668b019c486dd3e815114130e35c1495d

Comments

Lance Yang April 25, 2024, 3:45 a.m. UTC | #1
Hey Zi,

On Thu, Apr 25, 2024 at 6:46 AM Zi Yan <zi.yan@sent.com> wrote:
>
> From: Zi Yan <ziy@nvidia.com>
>
> In __folio_remove_rmap(), a large folio is added to deferred split list
> if any page in a folio loses its final mapping. It is possible that
> the folio is unmapped fully, but it is unnecessary to add the folio

Agreed. If a folio is fully unmapped, then that's unnecessary to add
to the deferred split list.

> to deferred split list at all. Fix it by checking folio->_nr_pages_mapped
> before adding a folio to deferred split list. If the folio is already
> on the deferred split list, it will be skipped. This issue applies to
> both PTE-mapped THP and mTHP.
>
> Commit 98046944a159 ("mm: huge_memory: add the missing
> folio_test_pmd_mappable() for THP split statistics") tried to exclude
> mTHP deferred split stats from THP_DEFERRED_SPLIT_PAGE, but it does not
> fix the above issue. A fully unmapped PTE-mapped order-9 THP was still
> added to deferred split list and counted as THP_DEFERRED_SPLIT_PAGE,
> since nr is 512 (non zero), level is RMAP_LEVEL_PTE, and inside
> deferred_split_folio() the order-9 folio is folio_test_pmd_mappable().
> However, this miscount was present even earlier due to implementation,
> since PTEs are unmapped individually and first PTE unmapping adds the THP
> into the deferred split list.
>
> With commit b06dc281aa99 ("mm/rmap: introduce
> folio_remove_rmap_[pte|ptes|pmd]()"), kernel is able to unmap PTE-mapped
> folios in one shot without causing the miscount, hence this patch.
>
> Signed-off-by: Zi Yan <ziy@nvidia.com>
> Reviewed-by: Yang Shi <shy828301@gmail.com>
> ---
>  mm/rmap.c | 7 ++++---
>  1 file changed, 4 insertions(+), 3 deletions(-)
>
> diff --git a/mm/rmap.c b/mm/rmap.c
> index a7913a454028..2809348add7b 100644
> --- a/mm/rmap.c
> +++ b/mm/rmap.c
> @@ -1553,9 +1553,10 @@ static __always_inline void __folio_remove_rmap(struct folio *folio,
>                  * page of the folio is unmapped and at least one page
>                  * is still mapped.
>                  */
> -               if (folio_test_large(folio) && folio_test_anon(folio))
> -                       if (level == RMAP_LEVEL_PTE || nr < nr_pmdmapped)
> -                               deferred_split_folio(folio);
> +               if (folio_test_large(folio) && folio_test_anon(folio) &&
> +                   ((level == RMAP_LEVEL_PTE && atomic_read(mapped)) ||
> +                    (level == RMAP_LEVEL_PMD && nr < nr_pmdmapped)))

Perhaps we only need to check the mapcount?

IIUC, if a large folio that was PMD/PTE mapped is fully unmapped here,
then folio_mapcount() will return 0.

-               if (folio_test_large(folio) && folio_test_anon(folio))
-                       if (level == RMAP_LEVEL_PTE || nr < nr_pmdmapped)
-                               deferred_split_folio(folio);
+               if (folio_test_large(folio) && folio_test_anon(folio) &&
+                   folio_mapcount(folio))
+                       deferred_split_folio(folio);

Thanks,
Lance




> +                       deferred_split_folio(folio);
>         }
>
>         /*
>
> base-commit: 2541ee5668b019c486dd3e815114130e35c1495d
> --
> 2.43.0
>
David Hildenbrand April 25, 2024, 7:19 a.m. UTC | #2
On 25.04.24 00:46, Zi Yan wrote:
> From: Zi Yan <ziy@nvidia.com>
> 
> In __folio_remove_rmap(), a large folio is added to deferred split list
> if any page in a folio loses its final mapping. It is possible that
> the folio is unmapped fully, but it is unnecessary to add the folio
> to deferred split list at all. Fix it by checking folio->_nr_pages_mapped
> before adding a folio to deferred split list. If the folio is already
> on the deferred split list, it will be skipped. This issue applies to
> both PTE-mapped THP and mTHP.
> 
> Commit 98046944a159 ("mm: huge_memory: add the missing
> folio_test_pmd_mappable() for THP split statistics") tried to exclude
> mTHP deferred split stats from THP_DEFERRED_SPLIT_PAGE, but it does not
> fix the above issue. A fully unmapped PTE-mapped order-9 THP was still

Once again: your patch won't fix it either.

> added to deferred split list and counted as THP_DEFERRED_SPLIT_PAGE,
> since nr is 512 (non zero), level is RMAP_LEVEL_PTE, and inside
> deferred_split_folio() the order-9 folio is folio_test_pmd_mappable().
> However, this miscount was present even earlier due to implementation,
> since PTEs are unmapped individually and first PTE unmapping adds the THP
> into the deferred split list.

It will still be present. Just less frequently.

> 
> With commit b06dc281aa99 ("mm/rmap: introduce
> folio_remove_rmap_[pte|ptes|pmd]()"), kernel is able to unmap PTE-mapped
> folios in one shot without causing the miscount, hence this patch.
> 
> Signed-off-by: Zi Yan <ziy@nvidia.com>
> Reviewed-by: Yang Shi <shy828301@gmail.com>
> ---
>   mm/rmap.c | 7 ++++---
>   1 file changed, 4 insertions(+), 3 deletions(-)
> 
> diff --git a/mm/rmap.c b/mm/rmap.c
> index a7913a454028..2809348add7b 100644
> --- a/mm/rmap.c
> +++ b/mm/rmap.c
> @@ -1553,9 +1553,10 @@ static __always_inline void __folio_remove_rmap(struct folio *folio,
>   		 * page of the folio is unmapped and at least one page
>   		 * is still mapped.
>   		 */
> -		if (folio_test_large(folio) && folio_test_anon(folio))
> -			if (level == RMAP_LEVEL_PTE || nr < nr_pmdmapped)
> -				deferred_split_folio(folio);
> +		if (folio_test_large(folio) && folio_test_anon(folio) &&
> +		    ((level == RMAP_LEVEL_PTE && atomic_read(mapped)) ||
> +		     (level == RMAP_LEVEL_PMD && nr < nr_pmdmapped)))
> +			deferred_split_folio(folio);

Please refrain from posting a new patch before the discussion on the old 
one is done.

See my comments on v2 why optimizing out the function call is a 
reasonable thing to do *where we cannot batch* and the misaccounting 
will still happen. But it can be done independently.
David Hildenbrand April 25, 2024, 7:21 a.m. UTC | #3
On 25.04.24 05:45, Lance Yang wrote:
> Hey Zi,
> 
> On Thu, Apr 25, 2024 at 6:46 AM Zi Yan <zi.yan@sent.com> wrote:
>>
>> From: Zi Yan <ziy@nvidia.com>
>>
>> In __folio_remove_rmap(), a large folio is added to deferred split list
>> if any page in a folio loses its final mapping. It is possible that
>> the folio is unmapped fully, but it is unnecessary to add the folio
> 
> Agreed. If a folio is fully unmapped, then that's unnecessary to add
> to the deferred split list.
> 
>> to deferred split list at all. Fix it by checking folio->_nr_pages_mapped
>> before adding a folio to deferred split list. If the folio is already
>> on the deferred split list, it will be skipped. This issue applies to
>> both PTE-mapped THP and mTHP.
>>
>> Commit 98046944a159 ("mm: huge_memory: add the missing
>> folio_test_pmd_mappable() for THP split statistics") tried to exclude
>> mTHP deferred split stats from THP_DEFERRED_SPLIT_PAGE, but it does not
>> fix the above issue. A fully unmapped PTE-mapped order-9 THP was still
>> added to deferred split list and counted as THP_DEFERRED_SPLIT_PAGE,
>> since nr is 512 (non zero), level is RMAP_LEVEL_PTE, and inside
>> deferred_split_folio() the order-9 folio is folio_test_pmd_mappable().
>> However, this miscount was present even earlier due to implementation,
>> since PTEs are unmapped individually and first PTE unmapping adds the THP
>> into the deferred split list.
>>
>> With commit b06dc281aa99 ("mm/rmap: introduce
>> folio_remove_rmap_[pte|ptes|pmd]()"), kernel is able to unmap PTE-mapped
>> folios in one shot without causing the miscount, hence this patch.
>>
>> Signed-off-by: Zi Yan <ziy@nvidia.com>
>> Reviewed-by: Yang Shi <shy828301@gmail.com>
>> ---
>>   mm/rmap.c | 7 ++++---
>>   1 file changed, 4 insertions(+), 3 deletions(-)
>>
>> diff --git a/mm/rmap.c b/mm/rmap.c
>> index a7913a454028..2809348add7b 100644
>> --- a/mm/rmap.c
>> +++ b/mm/rmap.c
>> @@ -1553,9 +1553,10 @@ static __always_inline void __folio_remove_rmap(struct folio *folio,
>>                   * page of the folio is unmapped and at least one page
>>                   * is still mapped.
>>                   */
>> -               if (folio_test_large(folio) && folio_test_anon(folio))
>> -                       if (level == RMAP_LEVEL_PTE || nr < nr_pmdmapped)
>> -                               deferred_split_folio(folio);
>> +               if (folio_test_large(folio) && folio_test_anon(folio) &&
>> +                   ((level == RMAP_LEVEL_PTE && atomic_read(mapped)) ||
>> +                    (level == RMAP_LEVEL_PMD && nr < nr_pmdmapped)))
> 
> Perhaps we only need to check the mapcount?
> 
> IIUC, if a large folio that was PMD/PTE mapped is fully unmapped here,
> then folio_mapcount() will return 0.

See discussion on v1. folio_large_mapcount() would achieve the same 
without another folio_test_large() check, but in the context of this 
patch it doesn't really matter.
Lance Yang April 25, 2024, 7:27 a.m. UTC | #4
On Thu, Apr 25, 2024 at 3:21 PM David Hildenbrand <david@redhat.com> wrote:
>
> On 25.04.24 05:45, Lance Yang wrote:
> > Hey Zi,
> >
> > On Thu, Apr 25, 2024 at 6:46 AM Zi Yan <zi.yan@sent.com> wrote:
> >>
> >> From: Zi Yan <ziy@nvidia.com>
> >>
> >> In __folio_remove_rmap(), a large folio is added to deferred split list
> >> if any page in a folio loses its final mapping. It is possible that
> >> the folio is unmapped fully, but it is unnecessary to add the folio
> >
> > Agreed. If a folio is fully unmapped, then that's unnecessary to add
> > to the deferred split list.
> >
> >> to deferred split list at all. Fix it by checking folio->_nr_pages_mapped
> >> before adding a folio to deferred split list. If the folio is already
> >> on the deferred split list, it will be skipped. This issue applies to
> >> both PTE-mapped THP and mTHP.
> >>
> >> Commit 98046944a159 ("mm: huge_memory: add the missing
> >> folio_test_pmd_mappable() for THP split statistics") tried to exclude
> >> mTHP deferred split stats from THP_DEFERRED_SPLIT_PAGE, but it does not
> >> fix the above issue. A fully unmapped PTE-mapped order-9 THP was still
> >> added to deferred split list and counted as THP_DEFERRED_SPLIT_PAGE,
> >> since nr is 512 (non zero), level is RMAP_LEVEL_PTE, and inside
> >> deferred_split_folio() the order-9 folio is folio_test_pmd_mappable().
> >> However, this miscount was present even earlier due to implementation,
> >> since PTEs are unmapped individually and first PTE unmapping adds the THP
> >> into the deferred split list.
> >>
> >> With commit b06dc281aa99 ("mm/rmap: introduce
> >> folio_remove_rmap_[pte|ptes|pmd]()"), kernel is able to unmap PTE-mapped
> >> folios in one shot without causing the miscount, hence this patch.
> >>
> >> Signed-off-by: Zi Yan <ziy@nvidia.com>
> >> Reviewed-by: Yang Shi <shy828301@gmail.com>
> >> ---
> >>   mm/rmap.c | 7 ++++---
> >>   1 file changed, 4 insertions(+), 3 deletions(-)
> >>
> >> diff --git a/mm/rmap.c b/mm/rmap.c
> >> index a7913a454028..2809348add7b 100644
> >> --- a/mm/rmap.c
> >> +++ b/mm/rmap.c
> >> @@ -1553,9 +1553,10 @@ static __always_inline void __folio_remove_rmap(struct folio *folio,
> >>                   * page of the folio is unmapped and at least one page
> >>                   * is still mapped.
> >>                   */
> >> -               if (folio_test_large(folio) && folio_test_anon(folio))
> >> -                       if (level == RMAP_LEVEL_PTE || nr < nr_pmdmapped)
> >> -                               deferred_split_folio(folio);
> >> +               if (folio_test_large(folio) && folio_test_anon(folio) &&
> >> +                   ((level == RMAP_LEVEL_PTE && atomic_read(mapped)) ||
> >> +                    (level == RMAP_LEVEL_PMD && nr < nr_pmdmapped)))
> >
> > Perhaps we only need to check the mapcount?
> >
> > IIUC, if a large folio that was PMD/PTE mapped is fully unmapped here,
> > then folio_mapcount() will return 0.
>
> See discussion on v1. folio_large_mapcount() would achieve the same
> without another folio_test_large() check, but in the context of this
> patch it doesn't really matter.

Got it. Thanks for pointing that out!
I'll take a closer look at the discussion in v1.

Thanks,
Lance


>
> --
> Cheers,
>
> David / dhildenb
>
David Hildenbrand April 25, 2024, 7:29 a.m. UTC | #5
On 25.04.24 09:27, Lance Yang wrote:
> On Thu, Apr 25, 2024 at 3:21 PM David Hildenbrand <david@redhat.com> wrote:
>>
>> On 25.04.24 05:45, Lance Yang wrote:
>>> Hey Zi,
>>>
>>> On Thu, Apr 25, 2024 at 6:46 AM Zi Yan <zi.yan@sent.com> wrote:
>>>>
>>>> From: Zi Yan <ziy@nvidia.com>
>>>>
>>>> In __folio_remove_rmap(), a large folio is added to deferred split list
>>>> if any page in a folio loses its final mapping. It is possible that
>>>> the folio is unmapped fully, but it is unnecessary to add the folio
>>>
>>> Agreed. If a folio is fully unmapped, then that's unnecessary to add
>>> to the deferred split list.
>>>
>>>> to deferred split list at all. Fix it by checking folio->_nr_pages_mapped
>>>> before adding a folio to deferred split list. If the folio is already
>>>> on the deferred split list, it will be skipped. This issue applies to
>>>> both PTE-mapped THP and mTHP.
>>>>
>>>> Commit 98046944a159 ("mm: huge_memory: add the missing
>>>> folio_test_pmd_mappable() for THP split statistics") tried to exclude
>>>> mTHP deferred split stats from THP_DEFERRED_SPLIT_PAGE, but it does not
>>>> fix the above issue. A fully unmapped PTE-mapped order-9 THP was still
>>>> added to deferred split list and counted as THP_DEFERRED_SPLIT_PAGE,
>>>> since nr is 512 (non zero), level is RMAP_LEVEL_PTE, and inside
>>>> deferred_split_folio() the order-9 folio is folio_test_pmd_mappable().
>>>> However, this miscount was present even earlier due to implementation,
>>>> since PTEs are unmapped individually and first PTE unmapping adds the THP
>>>> into the deferred split list.
>>>>
>>>> With commit b06dc281aa99 ("mm/rmap: introduce
>>>> folio_remove_rmap_[pte|ptes|pmd]()"), kernel is able to unmap PTE-mapped
>>>> folios in one shot without causing the miscount, hence this patch.
>>>>
>>>> Signed-off-by: Zi Yan <ziy@nvidia.com>
>>>> Reviewed-by: Yang Shi <shy828301@gmail.com>
>>>> ---
>>>>    mm/rmap.c | 7 ++++---
>>>>    1 file changed, 4 insertions(+), 3 deletions(-)
>>>>
>>>> diff --git a/mm/rmap.c b/mm/rmap.c
>>>> index a7913a454028..2809348add7b 100644
>>>> --- a/mm/rmap.c
>>>> +++ b/mm/rmap.c
>>>> @@ -1553,9 +1553,10 @@ static __always_inline void __folio_remove_rmap(struct folio *folio,
>>>>                    * page of the folio is unmapped and at least one page
>>>>                    * is still mapped.
>>>>                    */
>>>> -               if (folio_test_large(folio) && folio_test_anon(folio))
>>>> -                       if (level == RMAP_LEVEL_PTE || nr < nr_pmdmapped)
>>>> -                               deferred_split_folio(folio);
>>>> +               if (folio_test_large(folio) && folio_test_anon(folio) &&
>>>> +                   ((level == RMAP_LEVEL_PTE && atomic_read(mapped)) ||
>>>> +                    (level == RMAP_LEVEL_PMD && nr < nr_pmdmapped)))
>>>
>>> Perhaps we only need to check the mapcount?
>>>
>>> IIUC, if a large folio that was PMD/PTE mapped is fully unmapped here,
>>> then folio_mapcount() will return 0.
>>
>> See discussion on v1. folio_large_mapcount() would achieve the same
>> without another folio_test_large() check, but in the context of this
>> patch it doesn't really matter.
> 
> Got it. Thanks for pointing that out!
> I'll take a closer look at the discussion in v1.

Forgot to add: as long as the large mapcount patches are not upstream, 
folio_large_mapcount() would be expensive. So this patch can be added 
independent of the other stuff.
Lance Yang April 25, 2024, 7:35 a.m. UTC | #6
On Thu, Apr 25, 2024 at 3:29 PM David Hildenbrand <david@redhat.com> wrote:
>
> On 25.04.24 09:27, Lance Yang wrote:
> > On Thu, Apr 25, 2024 at 3:21 PM David Hildenbrand <david@redhat.com> wrote:
> >>
> >> On 25.04.24 05:45, Lance Yang wrote:
> >>> Hey Zi,
> >>>
> >>> On Thu, Apr 25, 2024 at 6:46 AM Zi Yan <zi.yan@sent.com> wrote:
> >>>>
> >>>> From: Zi Yan <ziy@nvidia.com>
> >>>>
> >>>> In __folio_remove_rmap(), a large folio is added to deferred split list
> >>>> if any page in a folio loses its final mapping. It is possible that
> >>>> the folio is unmapped fully, but it is unnecessary to add the folio
> >>>
> >>> Agreed. If a folio is fully unmapped, then that's unnecessary to add
> >>> to the deferred split list.
> >>>
> >>>> to deferred split list at all. Fix it by checking folio->_nr_pages_mapped
> >>>> before adding a folio to deferred split list. If the folio is already
> >>>> on the deferred split list, it will be skipped. This issue applies to
> >>>> both PTE-mapped THP and mTHP.
> >>>>
> >>>> Commit 98046944a159 ("mm: huge_memory: add the missing
> >>>> folio_test_pmd_mappable() for THP split statistics") tried to exclude
> >>>> mTHP deferred split stats from THP_DEFERRED_SPLIT_PAGE, but it does not
> >>>> fix the above issue. A fully unmapped PTE-mapped order-9 THP was still
> >>>> added to deferred split list and counted as THP_DEFERRED_SPLIT_PAGE,
> >>>> since nr is 512 (non zero), level is RMAP_LEVEL_PTE, and inside
> >>>> deferred_split_folio() the order-9 folio is folio_test_pmd_mappable().
> >>>> However, this miscount was present even earlier due to implementation,
> >>>> since PTEs are unmapped individually and first PTE unmapping adds the THP
> >>>> into the deferred split list.
> >>>>
> >>>> With commit b06dc281aa99 ("mm/rmap: introduce
> >>>> folio_remove_rmap_[pte|ptes|pmd]()"), kernel is able to unmap PTE-mapped
> >>>> folios in one shot without causing the miscount, hence this patch.
> >>>>
> >>>> Signed-off-by: Zi Yan <ziy@nvidia.com>
> >>>> Reviewed-by: Yang Shi <shy828301@gmail.com>
> >>>> ---
> >>>>    mm/rmap.c | 7 ++++---
> >>>>    1 file changed, 4 insertions(+), 3 deletions(-)
> >>>>
> >>>> diff --git a/mm/rmap.c b/mm/rmap.c
> >>>> index a7913a454028..2809348add7b 100644
> >>>> --- a/mm/rmap.c
> >>>> +++ b/mm/rmap.c
> >>>> @@ -1553,9 +1553,10 @@ static __always_inline void __folio_remove_rmap(struct folio *folio,
> >>>>                    * page of the folio is unmapped and at least one page
> >>>>                    * is still mapped.
> >>>>                    */
> >>>> -               if (folio_test_large(folio) && folio_test_anon(folio))
> >>>> -                       if (level == RMAP_LEVEL_PTE || nr < nr_pmdmapped)
> >>>> -                               deferred_split_folio(folio);
> >>>> +               if (folio_test_large(folio) && folio_test_anon(folio) &&
> >>>> +                   ((level == RMAP_LEVEL_PTE && atomic_read(mapped)) ||
> >>>> +                    (level == RMAP_LEVEL_PMD && nr < nr_pmdmapped)))
> >>>
> >>> Perhaps we only need to check the mapcount?
> >>>
> >>> IIUC, if a large folio that was PMD/PTE mapped is fully unmapped here,
> >>> then folio_mapcount() will return 0.
> >>
> >> See discussion on v1. folio_large_mapcount() would achieve the same
> >> without another folio_test_large() check, but in the context of this
> >> patch it doesn't really matter.
> >
> > Got it. Thanks for pointing that out!
> > I'll take a closer look at the discussion in v1.
>
> Forgot to add: as long as the large mapcount patches are not upstream,
> folio_large_mapcount() would be expensive. So this patch can be added
> independent of the other stuff.

Thanks for clarifying!
Lance

>
> --
> Cheers,
>
> David / dhildenb
>
Zi Yan April 25, 2024, 2:53 p.m. UTC | #7
On 25 Apr 2024, at 3:19, David Hildenbrand wrote:

> On 25.04.24 00:46, Zi Yan wrote:
>> From: Zi Yan <ziy@nvidia.com>
>>
>> In __folio_remove_rmap(), a large folio is added to deferred split list
>> if any page in a folio loses its final mapping. It is possible that
>> the folio is unmapped fully, but it is unnecessary to add the folio
>> to deferred split list at all. Fix it by checking folio->_nr_pages_mapped
>> before adding a folio to deferred split list. If the folio is already
>> on the deferred split list, it will be skipped. This issue applies to
>> both PTE-mapped THP and mTHP.
>>
>> Commit 98046944a159 ("mm: huge_memory: add the missing
>> folio_test_pmd_mappable() for THP split statistics") tried to exclude
>> mTHP deferred split stats from THP_DEFERRED_SPLIT_PAGE, but it does not
>> fix the above issue. A fully unmapped PTE-mapped order-9 THP was still
>
> Once again: your patch won't fix it either.
>
>> added to deferred split list and counted as THP_DEFERRED_SPLIT_PAGE,
>> since nr is 512 (non zero), level is RMAP_LEVEL_PTE, and inside
>> deferred_split_folio() the order-9 folio is folio_test_pmd_mappable().
>> However, this miscount was present even earlier due to implementation,
>> since PTEs are unmapped individually and first PTE unmapping adds the THP
>> into the deferred split list.
>
> It will still be present. Just less frequently.

OK. Let me reread the email exchanges between you and Yang and clarify
the details in the commit log.

>
>>
>> With commit b06dc281aa99 ("mm/rmap: introduce
>> folio_remove_rmap_[pte|ptes|pmd]()"), kernel is able to unmap PTE-mapped
>> folios in one shot without causing the miscount, hence this patch.
>>
>> Signed-off-by: Zi Yan <ziy@nvidia.com>
>> Reviewed-by: Yang Shi <shy828301@gmail.com>
>> ---
>>   mm/rmap.c | 7 ++++---
>>   1 file changed, 4 insertions(+), 3 deletions(-)
>>
>> diff --git a/mm/rmap.c b/mm/rmap.c
>> index a7913a454028..2809348add7b 100644
>> --- a/mm/rmap.c
>> +++ b/mm/rmap.c
>> @@ -1553,9 +1553,10 @@ static __always_inline void __folio_remove_rmap(struct folio *folio,
>>   		 * page of the folio is unmapped and at least one page
>>   		 * is still mapped.
>>   		 */
>> -		if (folio_test_large(folio) && folio_test_anon(folio))
>> -			if (level == RMAP_LEVEL_PTE || nr < nr_pmdmapped)
>> -				deferred_split_folio(folio);
>> +		if (folio_test_large(folio) && folio_test_anon(folio) &&
>> +		    ((level == RMAP_LEVEL_PTE && atomic_read(mapped)) ||
>> +		     (level == RMAP_LEVEL_PMD && nr < nr_pmdmapped)))
>> +			deferred_split_folio(folio);
>
> Please refrain from posting a new patch before the discussion on the old one is done.
>
> See my comments on v2 why optimizing out the function call is a reasonable thing to do *where we cannot batch* and the misaccounting will still happen. But it can be done independently.

Got it. Will keep the deferred list checking here and send a new one with commit
log changes too.

Thank you for the reviews.


--
Best Regards,
Yan, Zi
David Hildenbrand April 25, 2024, 3:15 p.m. UTC | #8
On 25.04.24 16:53, Zi Yan wrote:
> On 25 Apr 2024, at 3:19, David Hildenbrand wrote:
> 
>> On 25.04.24 00:46, Zi Yan wrote:
>>> From: Zi Yan <ziy@nvidia.com>
>>>
>>> In __folio_remove_rmap(), a large folio is added to deferred split list
>>> if any page in a folio loses its final mapping. It is possible that
>>> the folio is unmapped fully, but it is unnecessary to add the folio
>>> to deferred split list at all. Fix it by checking folio->_nr_pages_mapped
>>> before adding a folio to deferred split list. If the folio is already
>>> on the deferred split list, it will be skipped. This issue applies to
>>> both PTE-mapped THP and mTHP.
>>>
>>> Commit 98046944a159 ("mm: huge_memory: add the missing
>>> folio_test_pmd_mappable() for THP split statistics") tried to exclude
>>> mTHP deferred split stats from THP_DEFERRED_SPLIT_PAGE, but it does not
>>> fix the above issue. A fully unmapped PTE-mapped order-9 THP was still
>>
>> Once again: your patch won't fix it either.
>>
>>> added to deferred split list and counted as THP_DEFERRED_SPLIT_PAGE,
>>> since nr is 512 (non zero), level is RMAP_LEVEL_PTE, and inside
>>> deferred_split_folio() the order-9 folio is folio_test_pmd_mappable().
>>> However, this miscount was present even earlier due to implementation,
>>> since PTEs are unmapped individually and first PTE unmapping adds the THP
>>> into the deferred split list.
>>
>> It will still be present. Just less frequently.
> 
> OK. Let me reread the email exchanges between you and Yang and clarify
> the details in the commit log.

Likely something like:

--
In __folio_remove_rmap(), a large folio is added to deferred split list
if any page in a folio loses its final mapping. But, it is possible that
the folio is now fully unmapped and adding it to the deferred split list 
is unnecessary.

For PMD-mapped THPs, that was not really an issue, because removing the 
last PMD mapping in the absence of PTE mappings would not have added the 
folio to the deferred split queue.

However, for PTE-mapped THPs, which are now more prominent due to mTHP, 
we will always end up adding them to the deferred split queue.

One side effect of this is that we will frequently increase the 
THP_DEFERRED_SPLIT_PAGE stat for PTE-mapped THP, making it look like we 
frequently get many partially mapped folios -- although we are simply
unmapping the whole thing stepwise.

Core-mm will now try batch-unmapping consecutive PTEs of PTE-mapped THPs 
where possible. If we're lucky, we unmap the whole thing in one go and 
can avoid adding the folio to the deferred split queue, reducing the 
THP_DEFERRED_SPLIT_PAGE noise.

But there will still be noise when we cannot batch-unmap a complete 
PTE-mapped folio in one go -- or where this type of batching is not 
implemented yet.
--

Feel free to reuse what you consider reasonable.
Zi Yan April 25, 2024, 3:16 p.m. UTC | #9
On 25 Apr 2024, at 11:15, David Hildenbrand wrote:

> On 25.04.24 16:53, Zi Yan wrote:
>> On 25 Apr 2024, at 3:19, David Hildenbrand wrote:
>>
>>> On 25.04.24 00:46, Zi Yan wrote:
>>>> From: Zi Yan <ziy@nvidia.com>
>>>>
>>>> In __folio_remove_rmap(), a large folio is added to deferred split list
>>>> if any page in a folio loses its final mapping. It is possible that
>>>> the folio is unmapped fully, but it is unnecessary to add the folio
>>>> to deferred split list at all. Fix it by checking folio->_nr_pages_mapped
>>>> before adding a folio to deferred split list. If the folio is already
>>>> on the deferred split list, it will be skipped. This issue applies to
>>>> both PTE-mapped THP and mTHP.
>>>>
>>>> Commit 98046944a159 ("mm: huge_memory: add the missing
>>>> folio_test_pmd_mappable() for THP split statistics") tried to exclude
>>>> mTHP deferred split stats from THP_DEFERRED_SPLIT_PAGE, but it does not
>>>> fix the above issue. A fully unmapped PTE-mapped order-9 THP was still
>>>
>>> Once again: your patch won't fix it either.
>>>
>>>> added to deferred split list and counted as THP_DEFERRED_SPLIT_PAGE,
>>>> since nr is 512 (non zero), level is RMAP_LEVEL_PTE, and inside
>>>> deferred_split_folio() the order-9 folio is folio_test_pmd_mappable().
>>>> However, this miscount was present even earlier due to implementation,
>>>> since PTEs are unmapped individually and first PTE unmapping adds the THP
>>>> into the deferred split list.
>>>
>>> It will still be present. Just less frequently.
>>
>> OK. Let me reread the email exchanges between you and Yang and clarify
>> the details in the commit log.
>
> Likely something like:
>
> --
> In __folio_remove_rmap(), a large folio is added to deferred split list
> if any page in a folio loses its final mapping. But, it is possible that
> the folio is now fully unmapped and adding it to the deferred split list is unnecessary.
>
> For PMD-mapped THPs, that was not really an issue, because removing the last PMD mapping in the absence of PTE mappings would not have added the folio to the deferred split queue.
>
> However, for PTE-mapped THPs, which are now more prominent due to mTHP, we will always end up adding them to the deferred split queue.
>
> One side effect of this is that we will frequently increase the THP_DEFERRED_SPLIT_PAGE stat for PTE-mapped THP, making it look like we frequently get many partially mapped folios -- although we are simply
> unmapping the whole thing stepwise.
>
> Core-mm will now try batch-unmapping consecutive PTEs of PTE-mapped THPs where possible. If we're lucky, we unmap the whole thing in one go and can avoid adding the folio to the deferred split queue, reducing the THP_DEFERRED_SPLIT_PAGE noise.
>
> But there will still be noise when we cannot batch-unmap a complete PTE-mapped folio in one go -- or where this type of batching is not implemented yet.
> --
>
> Feel free to reuse what you consider reasonable.

Sure. Thank you a lot for drafting it!

--
Best Regards,
Yan, Zi
Yang Shi April 25, 2024, 3:49 p.m. UTC | #10
On Thu, Apr 25, 2024 at 7:53 AM Zi Yan <ziy@nvidia.com> wrote:
>
> On 25 Apr 2024, at 3:19, David Hildenbrand wrote:
>
> > On 25.04.24 00:46, Zi Yan wrote:
> >> From: Zi Yan <ziy@nvidia.com>
> >>
> >> In __folio_remove_rmap(), a large folio is added to deferred split list
> >> if any page in a folio loses its final mapping. It is possible that
> >> the folio is unmapped fully, but it is unnecessary to add the folio
> >> to deferred split list at all. Fix it by checking folio->_nr_pages_mapped
> >> before adding a folio to deferred split list. If the folio is already
> >> on the deferred split list, it will be skipped. This issue applies to
> >> both PTE-mapped THP and mTHP.
> >>
> >> Commit 98046944a159 ("mm: huge_memory: add the missing
> >> folio_test_pmd_mappable() for THP split statistics") tried to exclude
> >> mTHP deferred split stats from THP_DEFERRED_SPLIT_PAGE, but it does not
> >> fix the above issue. A fully unmapped PTE-mapped order-9 THP was still
> >
> > Once again: your patch won't fix it either.
> >
> >> added to deferred split list and counted as THP_DEFERRED_SPLIT_PAGE,
> >> since nr is 512 (non zero), level is RMAP_LEVEL_PTE, and inside
> >> deferred_split_folio() the order-9 folio is folio_test_pmd_mappable().
> >> However, this miscount was present even earlier due to implementation,
> >> since PTEs are unmapped individually and first PTE unmapping adds the THP
> >> into the deferred split list.
> >
> > It will still be present. Just less frequently.
>
> OK. Let me reread the email exchanges between you and Yang and clarify
> the details in the commit log.

There are still some places which may unmap PTE-mapped THP in page
granularity, for example, migration.

>
> >
> >>
> >> With commit b06dc281aa99 ("mm/rmap: introduce
> >> folio_remove_rmap_[pte|ptes|pmd]()"), kernel is able to unmap PTE-mapped
> >> folios in one shot without causing the miscount, hence this patch.
> >>
> >> Signed-off-by: Zi Yan <ziy@nvidia.com>
> >> Reviewed-by: Yang Shi <shy828301@gmail.com>
> >> ---
> >>   mm/rmap.c | 7 ++++---
> >>   1 file changed, 4 insertions(+), 3 deletions(-)
> >>
> >> diff --git a/mm/rmap.c b/mm/rmap.c
> >> index a7913a454028..2809348add7b 100644
> >> --- a/mm/rmap.c
> >> +++ b/mm/rmap.c
> >> @@ -1553,9 +1553,10 @@ static __always_inline void __folio_remove_rmap(struct folio *folio,
> >>               * page of the folio is unmapped and at least one page
> >>               * is still mapped.
> >>               */
> >> -            if (folio_test_large(folio) && folio_test_anon(folio))
> >> -                    if (level == RMAP_LEVEL_PTE || nr < nr_pmdmapped)
> >> -                            deferred_split_folio(folio);
> >> +            if (folio_test_large(folio) && folio_test_anon(folio) &&
> >> +                ((level == RMAP_LEVEL_PTE && atomic_read(mapped)) ||
> >> +                 (level == RMAP_LEVEL_PMD && nr < nr_pmdmapped)))
> >> +                    deferred_split_folio(folio);
> >
> > Please refrain from posting a new patch before the discussion on the old one is done.
> >
> > See my comments on v2 why optimizing out the function call is a reasonable thing to do *where we cannot batch* and the misaccounting will still happen. But it can be done independently.
>
> Got it. Will keep the deferred list checking here and send a new one with commit
> log changes too.
>
> Thank you for the reviews.
>
>
> --
> Best Regards,
> Yan, Zi
diff mbox series

Patch

diff --git a/mm/rmap.c b/mm/rmap.c
index a7913a454028..2809348add7b 100644
--- a/mm/rmap.c
+++ b/mm/rmap.c
@@ -1553,9 +1553,10 @@  static __always_inline void __folio_remove_rmap(struct folio *folio,
 		 * page of the folio is unmapped and at least one page
 		 * is still mapped.
 		 */
-		if (folio_test_large(folio) && folio_test_anon(folio))
-			if (level == RMAP_LEVEL_PTE || nr < nr_pmdmapped)
-				deferred_split_folio(folio);
+		if (folio_test_large(folio) && folio_test_anon(folio) &&
+		    ((level == RMAP_LEVEL_PTE && atomic_read(mapped)) ||
+		     (level == RMAP_LEVEL_PMD && nr < nr_pmdmapped)))
+			deferred_split_folio(folio);
 	}
 
 	/*