2023-08-30 17:31:07 +02:00
|
|
|
==============================
|
|
|
|
Unevictable LRU Infrastructure
|
|
|
|
==============================
|
|
|
|
|
|
|
|
.. contents:: :local:
|
|
|
|
|
|
|
|
|
|
|
|
Introduction
|
|
|
|
============
|
|
|
|
|
|
|
|
This document describes the Linux memory manager's "Unevictable LRU"
|
|
|
|
infrastructure and the use of this to manage several types of "unevictable"
|
|
|
|
folios.
|
|
|
|
|
|
|
|
The document attempts to provide the overall rationale behind this mechanism
|
|
|
|
and the rationale for some of the design decisions that drove the
|
|
|
|
implementation. The latter design rationale is discussed in the context of an
|
|
|
|
implementation description. Admittedly, one can obtain the implementation
|
|
|
|
details - the "what does it do?" - by reading the code. One hopes that the
|
|
|
|
descriptions below add value by provide the answer to "why does it do that?".
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
The Unevictable LRU
|
|
|
|
===================
|
|
|
|
|
|
|
|
The Unevictable LRU facility adds an additional LRU list to track unevictable
|
|
|
|
folios and to hide these folios from vmscan. This mechanism is based on a patch
|
|
|
|
by Larry Woodman of Red Hat to address several scalability problems with folio
|
|
|
|
reclaim in Linux. The problems have been observed at customer sites on large
|
|
|
|
memory x86_64 systems.
|
|
|
|
|
|
|
|
To illustrate this with an example, a non-NUMA x86_64 platform with 128GB of
|
|
|
|
main memory will have over 32 million 4k pages in a single node. When a large
|
|
|
|
fraction of these pages are not evictable for any reason [see below], vmscan
|
|
|
|
will spend a lot of time scanning the LRU lists looking for the small fraction
|
|
|
|
of pages that are evictable. This can result in a situation where all CPUs are
|
|
|
|
spending 100% of their time in vmscan for hours or days on end, with the system
|
|
|
|
completely unresponsive.
|
|
|
|
|
|
|
|
The unevictable list addresses the following classes of unevictable pages:
|
|
|
|
|
|
|
|
* Those owned by ramfs.
|
|
|
|
|
2023-10-24 12:59:35 +02:00
|
|
|
* Those owned by tmpfs with the noswap mount option.
|
|
|
|
|
2023-08-30 17:31:07 +02:00
|
|
|
* Those mapped into SHM_LOCK'd shared memory regions.
|
|
|
|
|
|
|
|
* Those mapped into VM_LOCKED [mlock()ed] VMAs.
|
|
|
|
|
|
|
|
The infrastructure may also be able to handle other conditions that make pages
|
|
|
|
unevictable, either by definition or by circumstance, in the future.
|
|
|
|
|
|
|
|
|
|
|
|
The Unevictable LRU Folio List
|
|
|
|
------------------------------
|
|
|
|
|
|
|
|
The Unevictable LRU folio list is a lie. It was never an LRU-ordered
|
|
|
|
list, but a companion to the LRU-ordered anonymous and file, active and
|
|
|
|
inactive folio lists; and now it is not even a folio list. But following
|
|
|
|
familiar convention, here in this document and in the source, we often
|
|
|
|
imagine it as a fifth LRU folio list.
|
|
|
|
|
|
|
|
The Unevictable LRU infrastructure consists of an additional, per-node, LRU list
|
|
|
|
called the "unevictable" list and an associated folio flag, PG_unevictable, to
|
|
|
|
indicate that the folio is being managed on the unevictable list.
|
|
|
|
|
|
|
|
The PG_unevictable flag is analogous to, and mutually exclusive with, the
|
|
|
|
PG_active flag in that it indicates on which LRU list a folio resides when
|
|
|
|
PG_lru is set.
|
|
|
|
|
|
|
|
The Unevictable LRU infrastructure maintains unevictable folios as if they were
|
|
|
|
on an additional LRU list for a few reasons:
|
|
|
|
|
|
|
|
(1) We get to "treat unevictable folios just like we treat other folios in the
|
|
|
|
system - which means we get to use the same code to manipulate them, the
|
|
|
|
same code to isolate them (for migrate, etc.), the same code to keep track
|
|
|
|
of the statistics, etc..." [Rik van Riel]
|
|
|
|
|
|
|
|
(2) We want to be able to migrate unevictable folios between nodes for memory
|
|
|
|
defragmentation, workload management and memory hotplug. The Linux kernel
|
|
|
|
can only migrate folios that it can successfully isolate from the LRU
|
|
|
|
lists (or "Movable" pages: outside of consideration here). If we were to
|
|
|
|
maintain folios elsewhere than on an LRU-like list, where they can be
|
|
|
|
detected by folio_isolate_lru(), we would prevent their migration.
|
|
|
|
|
|
|
|
The unevictable list does not differentiate between file-backed and
|
|
|
|
anonymous, swap-backed folios. This differentiation is only important
|
|
|
|
while the folios are, in fact, evictable.
|
|
|
|
|
|
|
|
The unevictable list benefits from the "arrayification" of the per-node LRU
|
|
|
|
lists and statistics originally proposed and posted by Christoph Lameter.
|
|
|
|
|
|
|
|
|
|
|
|
Memory Control Group Interaction
|
|
|
|
--------------------------------
|
|
|
|
|
|
|
|
The unevictable LRU facility interacts with the memory control group [aka
|
|
|
|
memory controller; see Documentation/admin-guide/cgroup-v1/memory.rst] by
|
|
|
|
extending the lru_list enum.
|
|
|
|
|
|
|
|
The memory controller data structure automatically gets a per-node unevictable
|
|
|
|
list as a result of the "arrayification" of the per-node LRU lists (one per
|
|
|
|
lru_list enum element). The memory controller tracks the movement of pages to
|
|
|
|
and from the unevictable list.
|
|
|
|
|
|
|
|
When a memory control group comes under memory pressure, the controller will
|
|
|
|
not attempt to reclaim pages on the unevictable list. This has a couple of
|
|
|
|
effects:
|
|
|
|
|
|
|
|
(1) Because the pages are "hidden" from reclaim on the unevictable list, the
|
|
|
|
reclaim process can be more efficient, dealing only with pages that have a
|
|
|
|
chance of being reclaimed.
|
|
|
|
|
|
|
|
(2) On the other hand, if too many of the pages charged to the control group
|
|
|
|
are unevictable, the evictable portion of the working set of the tasks in
|
|
|
|
the control group may not fit into the available memory. This can cause
|
|
|
|
the control group to thrash or to OOM-kill tasks.
|
|
|
|
|
|
|
|
|
|
|
|
.. _mark_addr_space_unevict:
|
|
|
|
|
|
|
|
Marking Address Spaces Unevictable
|
|
|
|
----------------------------------
|
|
|
|
|
|
|
|
For facilities such as ramfs none of the pages attached to the address space
|
|
|
|
may be evicted. To prevent eviction of any such pages, the AS_UNEVICTABLE
|
|
|
|
address space flag is provided, and this can be manipulated by a filesystem
|
|
|
|
using a number of wrapper functions:
|
|
|
|
|
|
|
|
* ``void mapping_set_unevictable(struct address_space *mapping);``
|
|
|
|
|
|
|
|
Mark the address space as being completely unevictable.
|
|
|
|
|
|
|
|
* ``void mapping_clear_unevictable(struct address_space *mapping);``
|
|
|
|
|
|
|
|
Mark the address space as being evictable.
|
|
|
|
|
|
|
|
* ``int mapping_unevictable(struct address_space *mapping);``
|
|
|
|
|
|
|
|
Query the address space, and return true if it is completely
|
|
|
|
unevictable.
|
|
|
|
|
|
|
|
These are currently used in three places in the kernel:
|
|
|
|
|
|
|
|
(1) By ramfs to mark the address spaces of its inodes when they are created,
|
|
|
|
and this mark remains for the life of the inode.
|
|
|
|
|
|
|
|
(2) By SYSV SHM to mark SHM_LOCK'd address spaces until SHM_UNLOCK is called.
|
|
|
|
Note that SHM_LOCK is not required to page in the locked pages if they're
|
|
|
|
swapped out; the application must touch the pages manually if it wants to
|
|
|
|
ensure they're in memory.
|
|
|
|
|
|
|
|
(3) By the i915 driver to mark pinned address space until it's unpinned. The
|
|
|
|
amount of unevictable memory marked by i915 driver is roughly the bounded
|
|
|
|
object size in debugfs/dri/0/i915_gem_objects.
|
|
|
|
|
|
|
|
|
|
|
|
Detecting Unevictable Pages
|
|
|
|
---------------------------
|
|
|
|
|
|
|
|
The function folio_evictable() in mm/internal.h determines whether a folio is
|
|
|
|
evictable or not using the query function outlined above [see section
|
|
|
|
:ref:`Marking address spaces unevictable <mark_addr_space_unevict>`]
|
|
|
|
to check the AS_UNEVICTABLE flag.
|
|
|
|
|
|
|
|
For address spaces that are so marked after being populated (as SHM regions
|
|
|
|
might be), the lock action (e.g. SHM_LOCK) can be lazy, and need not populate
|
|
|
|
the page tables for the region as does, for example, mlock(), nor need it make
|
|
|
|
any special effort to push any pages in the SHM_LOCK'd area to the unevictable
|
|
|
|
list. Instead, vmscan will do this if and when it encounters the folios during
|
|
|
|
a reclamation scan.
|
|
|
|
|
|
|
|
On an unlock action (such as SHM_UNLOCK), the unlocker (e.g. shmctl()) must scan
|
|
|
|
the pages in the region and "rescue" them from the unevictable list if no other
|
|
|
|
condition is keeping them unevictable. If an unevictable region is destroyed,
|
|
|
|
the pages are also "rescued" from the unevictable list in the process of
|
|
|
|
freeing them.
|
|
|
|
|
|
|
|
folio_evictable() also checks for mlocked folios by calling
|
|
|
|
folio_test_mlocked(), which is set when a folio is faulted into a
|
|
|
|
VM_LOCKED VMA, or found in a VMA being VM_LOCKED.
|
|
|
|
|
|
|
|
|
|
|
|
Vmscan's Handling of Unevictable Folios
|
|
|
|
---------------------------------------
|
|
|
|
|
|
|
|
If unevictable folios are culled in the fault path, or moved to the unevictable
|
|
|
|
list at mlock() or mmap() time, vmscan will not encounter the folios until they
|
|
|
|
have become evictable again (via munlock() for example) and have been "rescued"
|
|
|
|
from the unevictable list. However, there may be situations where we decide,
|
|
|
|
for the sake of expediency, to leave an unevictable folio on one of the regular
|
|
|
|
active/inactive LRU lists for vmscan to deal with. vmscan checks for such
|
|
|
|
folios in all of the shrink_{active|inactive|page}_list() functions and will
|
|
|
|
"cull" such folios that it encounters: that is, it diverts those folios to the
|
|
|
|
unevictable list for the memory cgroup and node being scanned.
|
|
|
|
|
|
|
|
There may be situations where a folio is mapped into a VM_LOCKED VMA,
|
|
|
|
but the folio does not have the mlocked flag set. Such folios will make
|
|
|
|
it all the way to shrink_active_list() or shrink_page_list() where they
|
|
|
|
will be detected when vmscan walks the reverse map in folio_referenced()
|
|
|
|
or try_to_unmap(). The folio is culled to the unevictable list when it
|
|
|
|
is released by the shrinker.
|
|
|
|
|
|
|
|
To "cull" an unevictable folio, vmscan simply puts the folio back on
|
|
|
|
the LRU list using folio_putback_lru() - the inverse operation to
|
|
|
|
folio_isolate_lru() - after dropping the folio lock. Because the
|
|
|
|
condition which makes the folio unevictable may change once the folio
|
|
|
|
is unlocked, __pagevec_lru_add_fn() will recheck the unevictable state
|
|
|
|
of a folio before placing it on the unevictable list.
|
|
|
|
|
|
|
|
|
|
|
|
MLOCKED Pages
|
|
|
|
=============
|
|
|
|
|
|
|
|
The unevictable folio list is also useful for mlock(), in addition to ramfs and
|
|
|
|
SYSV SHM. Note that mlock() is only available in CONFIG_MMU=y situations; in
|
|
|
|
NOMMU situations, all mappings are effectively mlocked.
|
|
|
|
|
|
|
|
|
|
|
|
History
|
|
|
|
-------
|
|
|
|
|
|
|
|
The "Unevictable mlocked Pages" infrastructure is based on work originally
|
|
|
|
posted by Nick Piggin in an RFC patch entitled "mm: mlocked pages off LRU".
|
|
|
|
Nick posted his patch as an alternative to a patch posted by Christoph Lameter
|
|
|
|
to achieve the same objective: hiding mlocked pages from vmscan.
|
|
|
|
|
|
|
|
In Nick's patch, he used one of the struct page LRU list link fields as a count
|
|
|
|
of VM_LOCKED VMAs that map the page (Rik van Riel had the same idea three years
|
|
|
|
earlier). But this use of the link field for a count prevented the management
|
|
|
|
of the pages on an LRU list, and thus mlocked pages were not migratable as
|
|
|
|
isolate_lru_page() could not detect them, and the LRU list link field was not
|
|
|
|
available to the migration subsystem.
|
|
|
|
|
|
|
|
Nick resolved this by putting mlocked pages back on the LRU list before
|
|
|
|
attempting to isolate them, thus abandoning the count of VM_LOCKED VMAs. When
|
|
|
|
Nick's patch was integrated with the Unevictable LRU work, the count was
|
|
|
|
replaced by walking the reverse map when munlocking, to determine whether any
|
|
|
|
other VM_LOCKED VMAs still mapped the page.
|
|
|
|
|
|
|
|
However, walking the reverse map for each page when munlocking was ugly and
|
|
|
|
inefficient, and could lead to catastrophic contention on a file's rmap lock,
|
|
|
|
when many processes which had it mlocked were trying to exit. In 5.18, the
|
|
|
|
idea of keeping mlock_count in Unevictable LRU list link field was revived and
|
|
|
|
put to work, without preventing the migration of mlocked pages. This is why
|
|
|
|
the "Unevictable LRU list" cannot be a linked list of pages now; but there was
|
|
|
|
no use for that linked list anyway - though its size is maintained for meminfo.
|
|
|
|
|
|
|
|
|
|
|
|
Basic Management
|
|
|
|
----------------
|
|
|
|
|
|
|
|
mlocked pages - pages mapped into a VM_LOCKED VMA - are a class of unevictable
|
|
|
|
pages. When such a page has been "noticed" by the memory management subsystem,
|
|
|
|
the page is marked with the PG_mlocked flag. This can be manipulated using the
|
|
|
|
PageMlocked() functions.
|
|
|
|
|
|
|
|
A PG_mlocked page will be placed on the unevictable list when it is added to
|
|
|
|
the LRU. Such pages can be "noticed" by memory management in several places:
|
|
|
|
|
|
|
|
(1) in the mlock()/mlock2()/mlockall() system call handlers;
|
|
|
|
|
|
|
|
(2) in the mmap() system call handler when mmapping a region with the
|
|
|
|
MAP_LOCKED flag;
|
|
|
|
|
|
|
|
(3) mmapping a region in a task that has called mlockall() with the MCL_FUTURE
|
|
|
|
flag;
|
|
|
|
|
|
|
|
(4) in the fault path and when a VM_LOCKED stack segment is expanded; or
|
|
|
|
|
|
|
|
(5) as mentioned above, in vmscan:shrink_page_list() when attempting to
|
|
|
|
reclaim a page in a VM_LOCKED VMA by folio_referenced() or try_to_unmap().
|
|
|
|
|
|
|
|
mlocked pages become unlocked and rescued from the unevictable list when:
|
|
|
|
|
|
|
|
(1) mapped in a range unlocked via the munlock()/munlockall() system calls;
|
|
|
|
|
|
|
|
(2) munmap()'d out of the last VM_LOCKED VMA that maps the page, including
|
|
|
|
unmapping at task exit;
|
|
|
|
|
|
|
|
(3) when the page is truncated from the last VM_LOCKED VMA of an mmapped file;
|
|
|
|
or
|
|
|
|
|
|
|
|
(4) before a page is COW'd in a VM_LOCKED VMA.
|
|
|
|
|
|
|
|
|
|
|
|
mlock()/mlock2()/mlockall() System Call Handling
|
|
|
|
------------------------------------------------
|
|
|
|
|
|
|
|
mlock(), mlock2() and mlockall() system call handlers proceed to mlock_fixup()
|
|
|
|
for each VMA in the range specified by the call. In the case of mlockall(),
|
|
|
|
this is the entire active address space of the task. Note that mlock_fixup()
|
|
|
|
is used for both mlocking and munlocking a range of memory. A call to mlock()
|
|
|
|
an already VM_LOCKED VMA, or to munlock() a VMA that is not VM_LOCKED, is
|
|
|
|
treated as a no-op and mlock_fixup() simply returns.
|
|
|
|
|
|
|
|
If the VMA passes some filtering as described in "Filtering Special VMAs"
|
|
|
|
below, mlock_fixup() will attempt to merge the VMA with its neighbors or split
|
|
|
|
off a subset of the VMA if the range does not cover the entire VMA. Any pages
|
|
|
|
already present in the VMA are then marked as mlocked by mlock_folio() via
|
|
|
|
mlock_pte_range() via walk_page_range() via mlock_vma_pages_range().
|
|
|
|
|
|
|
|
Before returning from the system call, do_mlock() or mlockall() will call
|
|
|
|
__mm_populate() to fault in the remaining pages via get_user_pages() and to
|
|
|
|
mark those pages as mlocked as they are faulted.
|
|
|
|
|
|
|
|
Note that the VMA being mlocked might be mapped with PROT_NONE. In this case,
|
|
|
|
get_user_pages() will be unable to fault in the pages. That's okay. If pages
|
|
|
|
do end up getting faulted into this VM_LOCKED VMA, they will be handled in the
|
|
|
|
fault path - which is also how mlock2()'s MLOCK_ONFAULT areas are handled.
|
|
|
|
|
|
|
|
For each PTE (or PMD) being faulted into a VMA, the page add rmap function
|
|
|
|
calls mlock_vma_folio(), which calls mlock_folio() when the VMA is VM_LOCKED
|
|
|
|
(unless it is a PTE mapping of a part of a transparent huge page). Or when
|
|
|
|
it is a newly allocated anonymous page, folio_add_lru_vma() calls
|
|
|
|
mlock_new_folio() instead: similar to mlock_folio(), but can make better
|
|
|
|
judgments, since this page is held exclusively and known not to be on LRU yet.
|
|
|
|
|
|
|
|
mlock_folio() sets PG_mlocked immediately, then places the page on the CPU's
|
|
|
|
mlock folio batch, to batch up the rest of the work to be done under lru_lock by
|
|
|
|
__mlock_folio(). __mlock_folio() sets PG_unevictable, initializes mlock_count
|
|
|
|
and moves the page to unevictable state ("the unevictable LRU", but with
|
|
|
|
mlock_count in place of LRU threading). Or if the page was already PG_lru
|
|
|
|
and PG_unevictable and PG_mlocked, it simply increments the mlock_count.
|
|
|
|
|
|
|
|
But in practice that may not work ideally: the page may not yet be on an LRU, or
|
|
|
|
it may have been temporarily isolated from LRU. In such cases the mlock_count
|
|
|
|
field cannot be touched, but will be set to 0 later when __munlock_folio()
|
|
|
|
returns the page to "LRU". Races prohibit mlock_count from being set to 1 then:
|
|
|
|
rather than risk stranding a page indefinitely as unevictable, always err with
|
|
|
|
mlock_count on the low side, so that when munlocked the page will be rescued to
|
|
|
|
an evictable LRU, then perhaps be mlocked again later if vmscan finds it in a
|
|
|
|
VM_LOCKED VMA.
|
|
|
|
|
|
|
|
|
|
|
|
Filtering Special VMAs
|
|
|
|
----------------------
|
|
|
|
|
|
|
|
mlock_fixup() filters several classes of "special" VMAs:
|
|
|
|
|
|
|
|
1) VMAs with VM_IO or VM_PFNMAP set are skipped entirely. The pages behind
|
|
|
|
these mappings are inherently pinned, so we don't need to mark them as
|
|
|
|
mlocked. In any case, most of the pages have no struct page in which to so
|
|
|
|
mark the page. Because of this, get_user_pages() will fail for these VMAs,
|
|
|
|
so there is no sense in attempting to visit them.
|
|
|
|
|
|
|
|
2) VMAs mapping hugetlbfs page are already effectively pinned into memory. We
|
|
|
|
neither need nor want to mlock() these pages. But __mm_populate() includes
|
|
|
|
hugetlbfs ranges, allocating the huge pages and populating the PTEs.
|
|
|
|
|
|
|
|
3) VMAs with VM_DONTEXPAND are generally userspace mappings of kernel pages,
|
|
|
|
such as the VDSO page, relay channel pages, etc. These pages are inherently
|
|
|
|
unevictable and are not managed on the LRU lists. __mm_populate() includes
|
|
|
|
these ranges, populating the PTEs if not already populated.
|
|
|
|
|
|
|
|
4) VMAs with VM_MIXEDMAP set are not marked VM_LOCKED, but __mm_populate()
|
|
|
|
includes these ranges, populating the PTEs if not already populated.
|
|
|
|
|
|
|
|
Note that for all of these special VMAs, mlock_fixup() does not set the
|
|
|
|
VM_LOCKED flag. Therefore, we won't have to deal with them later during
|
|
|
|
munlock(), munmap() or task exit. Neither does mlock_fixup() account these
|
|
|
|
VMAs against the task's "locked_vm".
|
|
|
|
|
|
|
|
|
|
|
|
munlock()/munlockall() System Call Handling
|
|
|
|
-------------------------------------------
|
|
|
|
|
|
|
|
The munlock() and munlockall() system calls are handled by the same
|
|
|
|
mlock_fixup() function as mlock(), mlock2() and mlockall() system calls are.
|
|
|
|
If called to munlock an already munlocked VMA, mlock_fixup() simply returns.
|
|
|
|
Because of the VMA filtering discussed above, VM_LOCKED will not be set in
|
|
|
|
any "special" VMAs. So, those VMAs will be ignored for munlock.
|
|
|
|
|
|
|
|
If the VMA is VM_LOCKED, mlock_fixup() again attempts to merge or split off the
|
|
|
|
specified range. All pages in the VMA are then munlocked by munlock_folio() via
|
|
|
|
mlock_pte_range() via walk_page_range() via mlock_vma_pages_range() - the same
|
|
|
|
function used when mlocking a VMA range, with new flags for the VMA indicating
|
|
|
|
that it is munlock() being performed.
|
|
|
|
|
|
|
|
munlock_folio() uses the mlock pagevec to batch up work to be done
|
|
|
|
under lru_lock by __munlock_folio(). __munlock_folio() decrements the
|
|
|
|
folio's mlock_count, and when that reaches 0 it clears the mlocked flag
|
|
|
|
and clears the unevictable flag, moving the folio from unevictable state
|
|
|
|
to the inactive LRU.
|
|
|
|
|
|
|
|
But in practice that may not work ideally: the folio may not yet have reached
|
|
|
|
"the unevictable LRU", or it may have been temporarily isolated from it. In
|
|
|
|
those cases its mlock_count field is unusable and must be assumed to be 0: so
|
|
|
|
that the folio will be rescued to an evictable LRU, then perhaps be mlocked
|
|
|
|
again later if vmscan finds it in a VM_LOCKED VMA.
|
|
|
|
|
|
|
|
|
|
|
|
Migrating MLOCKED Pages
|
|
|
|
-----------------------
|
|
|
|
|
|
|
|
A page that is being migrated has been isolated from the LRU lists and is held
|
|
|
|
locked across unmapping of the page, updating the page's address space entry
|
|
|
|
and copying the contents and state, until the page table entry has been
|
|
|
|
replaced with an entry that refers to the new page. Linux supports migration
|
|
|
|
of mlocked pages and other unevictable pages. PG_mlocked is cleared from the
|
|
|
|
the old page when it is unmapped from the last VM_LOCKED VMA, and set when the
|
|
|
|
new page is mapped in place of migration entry in a VM_LOCKED VMA. If the page
|
|
|
|
was unevictable because mlocked, PG_unevictable follows PG_mlocked; but if the
|
|
|
|
page was unevictable for other reasons, PG_unevictable is copied explicitly.
|
|
|
|
|
|
|
|
Note that page migration can race with mlocking or munlocking of the same page.
|
|
|
|
There is mostly no problem since page migration requires unmapping all PTEs of
|
|
|
|
the old page (including munlock where VM_LOCKED), then mapping in the new page
|
|
|
|
(including mlock where VM_LOCKED). The page table locks provide sufficient
|
|
|
|
synchronization.
|
|
|
|
|
|
|
|
However, since mlock_vma_pages_range() starts by setting VM_LOCKED on a VMA,
|
|
|
|
before mlocking any pages already present, if one of those pages were migrated
|
|
|
|
before mlock_pte_range() reached it, it would get counted twice in mlock_count.
|
|
|
|
To prevent that, mlock_vma_pages_range() temporarily marks the VMA as VM_IO,
|
|
|
|
so that mlock_vma_folio() will skip it.
|
|
|
|
|
|
|
|
To complete page migration, we place the old and new pages back onto the LRU
|
|
|
|
afterwards. The "unneeded" page - old page on success, new page on failure -
|
|
|
|
is freed when the reference count held by the migration process is released.
|
|
|
|
|
|
|
|
|
|
|
|
Compacting MLOCKED Pages
|
|
|
|
------------------------
|
|
|
|
|
|
|
|
The memory map can be scanned for compactable regions and the default behavior
|
|
|
|
is to let unevictable pages be moved. /proc/sys/vm/compact_unevictable_allowed
|
|
|
|
controls this behavior (see Documentation/admin-guide/sysctl/vm.rst). The work
|
|
|
|
of compaction is mostly handled by the page migration code and the same work
|
|
|
|
flow as described in Migrating MLOCKED Pages will apply.
|
|
|
|
|
|
|
|
|
|
|
|
MLOCKING Transparent Huge Pages
|
|
|
|
-------------------------------
|
|
|
|
|
|
|
|
A transparent huge page is represented by a single entry on an LRU list.
|
|
|
|
Therefore, we can only make unevictable an entire compound page, not
|
|
|
|
individual subpages.
|
|
|
|
|
|
|
|
If a user tries to mlock() part of a huge page, and no user mlock()s the
|
|
|
|
whole of the huge page, we want the rest of the page to be reclaimable.
|
|
|
|
|
|
|
|
We cannot just split the page on partial mlock() as split_huge_page() can
|
|
|
|
fail and a new intermittent failure mode for the syscall is undesirable.
|
|
|
|
|
|
|
|
We handle this by keeping PTE-mlocked huge pages on evictable LRU lists:
|
|
|
|
the PMD on the border of a VM_LOCKED VMA will be split into a PTE table.
|
|
|
|
|
|
|
|
This way the huge page is accessible for vmscan. Under memory pressure the
|
|
|
|
page will be split, subpages which belong to VM_LOCKED VMAs will be moved
|
|
|
|
to the unevictable LRU and the rest can be reclaimed.
|
|
|
|
|
|
|
|
/proc/meminfo's Unevictable and Mlocked amounts do not include those parts
|
|
|
|
of a transparent huge page which are mapped only by PTEs in VM_LOCKED VMAs.
|
|
|
|
|
|
|
|
|
|
|
|
mmap(MAP_LOCKED) System Call Handling
|
|
|
|
-------------------------------------
|
|
|
|
|
|
|
|
In addition to the mlock(), mlock2() and mlockall() system calls, an application
|
|
|
|
can request that a region of memory be mlocked by supplying the MAP_LOCKED flag
|
|
|
|
to the mmap() call. There is one important and subtle difference here, though.
|
|
|
|
mmap() + mlock() will fail if the range cannot be faulted in (e.g. because
|
|
|
|
mm_populate fails) and returns with ENOMEM while mmap(MAP_LOCKED) will not fail.
|
|
|
|
The mmaped area will still have properties of the locked area - pages will not
|
|
|
|
get swapped out - but major page faults to fault memory in might still happen.
|
|
|
|
|
|
|
|
Furthermore, any mmap() call or brk() call that expands the heap by a task
|
|
|
|
that has previously called mlockall() with the MCL_FUTURE flag will result
|
|
|
|
in the newly mapped memory being mlocked. Before the unevictable/mlock
|
|
|
|
changes, the kernel simply called make_pages_present() to allocate pages
|
|
|
|
and populate the page table.
|
|
|
|
|
|
|
|
To mlock a range of memory under the unevictable/mlock infrastructure,
|
|
|
|
the mmap() handler and task address space expansion functions call
|
|
|
|
populate_vma_page_range() specifying the vma and the address range to mlock.
|
|
|
|
|
|
|
|
|
|
|
|
munmap()/exit()/exec() System Call Handling
|
|
|
|
-------------------------------------------
|
|
|
|
|
|
|
|
When unmapping an mlocked region of memory, whether by an explicit call to
|
|
|
|
munmap() or via an internal unmap from exit() or exec() processing, we must
|
|
|
|
munlock the pages if we're removing the last VM_LOCKED VMA that maps the pages.
|
|
|
|
Before the unevictable/mlock changes, mlocking did not mark the pages in any
|
|
|
|
way, so unmapping them required no processing.
|
|
|
|
|
|
|
|
For each PTE (or PMD) being unmapped from a VMA, page_remove_rmap() calls
|
|
|
|
munlock_vma_folio(), which calls munlock_folio() when the VMA is VM_LOCKED
|
|
|
|
(unless it was a PTE mapping of a part of a transparent huge page).
|
|
|
|
|
|
|
|
munlock_folio() uses the mlock pagevec to batch up work to be done
|
|
|
|
under lru_lock by __munlock_folio(). __munlock_folio() decrements the
|
|
|
|
folio's mlock_count, and when that reaches 0 it clears the mlocked flag
|
|
|
|
and clears the unevictable flag, moving the folio from unevictable state
|
|
|
|
to the inactive LRU.
|
|
|
|
|
|
|
|
But in practice that may not work ideally: the folio may not yet have reached
|
|
|
|
"the unevictable LRU", or it may have been temporarily isolated from it. In
|
|
|
|
those cases its mlock_count field is unusable and must be assumed to be 0: so
|
|
|
|
that the folio will be rescued to an evictable LRU, then perhaps be mlocked
|
|
|
|
again later if vmscan finds it in a VM_LOCKED VMA.
|
|
|
|
|
|
|
|
|
|
|
|
Truncating MLOCKED Pages
|
|
|
|
------------------------
|
|
|
|
|
|
|
|
File truncation or hole punching forcibly unmaps the deleted pages from
|
|
|
|
userspace; truncation even unmaps and deletes any private anonymous pages
|
|
|
|
which had been Copied-On-Write from the file pages now being truncated.
|
|
|
|
|
|
|
|
Mlocked pages can be munlocked and deleted in this way: like with munmap(),
|
|
|
|
for each PTE (or PMD) being unmapped from a VMA, page_remove_rmap() calls
|
|
|
|
munlock_vma_folio(), which calls munlock_folio() when the VMA is VM_LOCKED
|
|
|
|
(unless it was a PTE mapping of a part of a transparent huge page).
|
|
|
|
|
|
|
|
However, if there is a racing munlock(), since mlock_vma_pages_range() starts
|
|
|
|
munlocking by clearing VM_LOCKED from a VMA, before munlocking all the pages
|
|
|
|
present, if one of those pages were unmapped by truncation or hole punch before
|
|
|
|
mlock_pte_range() reached it, it would not be recognized as mlocked by this VMA,
|
|
|
|
and would not be counted out of mlock_count. In this rare case, a page may
|
|
|
|
still appear as PG_mlocked after it has been fully unmapped: and it is left to
|
|
|
|
release_pages() (or __page_cache_release()) to clear it and update statistics
|
|
|
|
before freeing (this event is counted in /proc/vmstat unevictable_pgs_cleared,
|
|
|
|
which is usually 0).
|
|
|
|
|
|
|
|
|
|
|
|
Page Reclaim in shrink_*_list()
|
|
|
|
-------------------------------
|
|
|
|
|
|
|
|
vmscan's shrink_active_list() culls any obviously unevictable pages -
|
|
|
|
i.e. !page_evictable(page) pages - diverting those to the unevictable list.
|
|
|
|
However, shrink_active_list() only sees unevictable pages that made it onto the
|
|
|
|
active/inactive LRU lists. Note that these pages do not have PG_unevictable
|
|
|
|
set - otherwise they would be on the unevictable list and shrink_active_list()
|
|
|
|
would never see them.
|
|
|
|
|
|
|
|
Some examples of these unevictable pages on the LRU lists are:
|
|
|
|
|
|
|
|
(1) ramfs pages that have been placed on the LRU lists when first allocated.
|
|
|
|
|
|
|
|
(2) SHM_LOCK'd shared memory pages. shmctl(SHM_LOCK) does not attempt to
|
|
|
|
allocate or fault in the pages in the shared memory region. This happens
|
|
|
|
when an application accesses the page the first time after SHM_LOCK'ing
|
|
|
|
the segment.
|
|
|
|
|
|
|
|
(3) pages still mapped into VM_LOCKED VMAs, which should be marked mlocked,
|
|
|
|
but events left mlock_count too low, so they were munlocked too early.
|
|
|
|
|
|
|
|
vmscan's shrink_inactive_list() and shrink_page_list() also divert obviously
|
|
|
|
unevictable pages found on the inactive lists to the appropriate memory cgroup
|
|
|
|
and node unevictable list.
|
|
|
|
|
|
|
|
rmap's folio_referenced_one(), called via vmscan's shrink_active_list() or
|
|
|
|
shrink_page_list(), and rmap's try_to_unmap_one() called via shrink_page_list(),
|
|
|
|
check for (3) pages still mapped into VM_LOCKED VMAs, and call mlock_vma_folio()
|
|
|
|
to correct them. Such pages are culled to the unevictable list when released
|
|
|
|
by the shrinker.
|