123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382 |
- .. _hugetlbpage:
- =============
- HugeTLB Pages
- =============
- Overview
- ========
- The intent of this file is to give a brief summary of hugetlbpage support in
- the Linux kernel. This support is built on top of multiple page size support
- that is provided by most modern architectures. For example, x86 CPUs normally
- support 4K and 2M (1G if architecturally supported) page sizes, ia64
- architecture supports multiple page sizes 4K, 8K, 64K, 256K, 1M, 4M, 16M,
- 256M and ppc64 supports 4K and 16M. A TLB is a cache of virtual-to-physical
- translations. Typically this is a very scarce resource on processor.
- Operating systems try to make best use of limited number of TLB resources.
- This optimization is more critical now as bigger and bigger physical memories
- (several GBs) are more readily available.
- Users can use the huge page support in Linux kernel by either using the mmap
- system call or standard SYSV shared memory system calls (shmget, shmat).
- First the Linux kernel needs to be built with the CONFIG_HUGETLBFS
- (present under "File systems") and CONFIG_HUGETLB_PAGE (selected
- automatically when CONFIG_HUGETLBFS is selected) configuration
- options.
- The ``/proc/meminfo`` file provides information about the total number of
- persistent hugetlb pages in the kernel's huge page pool. It also displays
- default huge page size and information about the number of free, reserved
- and surplus huge pages in the pool of huge pages of default size.
- The huge page size is needed for generating the proper alignment and
- size of the arguments to system calls that map huge page regions.
- The output of ``cat /proc/meminfo`` will include lines like::
- HugePages_Total: uuu
- HugePages_Free: vvv
- HugePages_Rsvd: www
- HugePages_Surp: xxx
- Hugepagesize: yyy kB
- Hugetlb: zzz kB
- where:
- HugePages_Total
- is the size of the pool of huge pages.
- HugePages_Free
- is the number of huge pages in the pool that are not yet
- allocated.
- HugePages_Rsvd
- is short for "reserved," and is the number of huge pages for
- which a commitment to allocate from the pool has been made,
- but no allocation has yet been made. Reserved huge pages
- guarantee that an application will be able to allocate a
- huge page from the pool of huge pages at fault time.
- HugePages_Surp
- is short for "surplus," and is the number of huge pages in
- the pool above the value in ``/proc/sys/vm/nr_hugepages``. The
- maximum number of surplus huge pages is controlled by
- ``/proc/sys/vm/nr_overcommit_hugepages``.
- Hugepagesize
- is the default hugepage size (in Kb).
- Hugetlb
- is the total amount of memory (in kB), consumed by huge
- pages of all sizes.
- If huge pages of different sizes are in use, this number
- will exceed HugePages_Total \* Hugepagesize. To get more
- detailed information, please, refer to
- ``/sys/kernel/mm/hugepages`` (described below).
- ``/proc/filesystems`` should also show a filesystem of type "hugetlbfs"
- configured in the kernel.
- ``/proc/sys/vm/nr_hugepages`` indicates the current number of "persistent" huge
- pages in the kernel's huge page pool. "Persistent" huge pages will be
- returned to the huge page pool when freed by a task. A user with root
- privileges can dynamically allocate more or free some persistent huge pages
- by increasing or decreasing the value of ``nr_hugepages``.
- Pages that are used as huge pages are reserved inside the kernel and cannot
- be used for other purposes. Huge pages cannot be swapped out under
- memory pressure.
- Once a number of huge pages have been pre-allocated to the kernel huge page
- pool, a user with appropriate privilege can use either the mmap system call
- or shared memory system calls to use the huge pages. See the discussion of
- :ref:`Using Huge Pages <using_huge_pages>`, below.
- The administrator can allocate persistent huge pages on the kernel boot
- command line by specifying the "hugepages=N" parameter, where 'N' = the
- number of huge pages requested. This is the most reliable method of
- allocating huge pages as memory has not yet become fragmented.
- Some platforms support multiple huge page sizes. To allocate huge pages
- of a specific size, one must precede the huge pages boot command parameters
- with a huge page size selection parameter "hugepagesz=<size>". <size> must
- be specified in bytes with optional scale suffix [kKmMgG]. The default huge
- page size may be selected with the "default_hugepagesz=<size>" boot parameter.
- When multiple huge page sizes are supported, ``/proc/sys/vm/nr_hugepages``
- indicates the current number of pre-allocated huge pages of the default size.
- Thus, one can use the following command to dynamically allocate/deallocate
- default sized persistent huge pages::
- echo 20 > /proc/sys/vm/nr_hugepages
- This command will try to adjust the number of default sized huge pages in the
- huge page pool to 20, allocating or freeing huge pages, as required.
- On a NUMA platform, the kernel will attempt to distribute the huge page pool
- over all the set of allowed nodes specified by the NUMA memory policy of the
- task that modifies ``nr_hugepages``. The default for the allowed nodes--when the
- task has default memory policy--is all on-line nodes with memory. Allowed
- nodes with insufficient available, contiguous memory for a huge page will be
- silently skipped when allocating persistent huge pages. See the
- :ref:`discussion below <mem_policy_and_hp_alloc>`
- of the interaction of task memory policy, cpusets and per node attributes
- with the allocation and freeing of persistent huge pages.
- The success or failure of huge page allocation depends on the amount of
- physically contiguous memory that is present in system at the time of the
- allocation attempt. If the kernel is unable to allocate huge pages from
- some nodes in a NUMA system, it will attempt to make up the difference by
- allocating extra pages on other nodes with sufficient available contiguous
- memory, if any.
- System administrators may want to put this command in one of the local rc
- init files. This will enable the kernel to allocate huge pages early in
- the boot process when the possibility of getting physical contiguous pages
- is still very high. Administrators can verify the number of huge pages
- actually allocated by checking the sysctl or meminfo. To check the per node
- distribution of huge pages in a NUMA system, use::
- cat /sys/devices/system/node/node*/meminfo | fgrep Huge
- ``/proc/sys/vm/nr_overcommit_hugepages`` specifies how large the pool of
- huge pages can grow, if more huge pages than ``/proc/sys/vm/nr_hugepages`` are
- requested by applications. Writing any non-zero value into this file
- indicates that the hugetlb subsystem is allowed to try to obtain that
- number of "surplus" huge pages from the kernel's normal page pool, when the
- persistent huge page pool is exhausted. As these surplus huge pages become
- unused, they are freed back to the kernel's normal page pool.
- When increasing the huge page pool size via ``nr_hugepages``, any existing
- surplus pages will first be promoted to persistent huge pages. Then, additional
- huge pages will be allocated, if necessary and if possible, to fulfill
- the new persistent huge page pool size.
- The administrator may shrink the pool of persistent huge pages for
- the default huge page size by setting the ``nr_hugepages`` sysctl to a
- smaller value. The kernel will attempt to balance the freeing of huge pages
- across all nodes in the memory policy of the task modifying ``nr_hugepages``.
- Any free huge pages on the selected nodes will be freed back to the kernel's
- normal page pool.
- Caveat: Shrinking the persistent huge page pool via ``nr_hugepages`` such that
- it becomes less than the number of huge pages in use will convert the balance
- of the in-use huge pages to surplus huge pages. This will occur even if
- the number of surplus pages would exceed the overcommit value. As long as
- this condition holds--that is, until ``nr_hugepages+nr_overcommit_hugepages`` is
- increased sufficiently, or the surplus huge pages go out of use and are freed--
- no more surplus huge pages will be allowed to be allocated.
- With support for multiple huge page pools at run-time available, much of
- the huge page userspace interface in ``/proc/sys/vm`` has been duplicated in
- sysfs.
- The ``/proc`` interfaces discussed above have been retained for backwards
- compatibility. The root huge page control directory in sysfs is::
- /sys/kernel/mm/hugepages
- For each huge page size supported by the running kernel, a subdirectory
- will exist, of the form::
- hugepages-${size}kB
- Inside each of these directories, the same set of files will exist::
- nr_hugepages
- nr_hugepages_mempolicy
- nr_overcommit_hugepages
- free_hugepages
- resv_hugepages
- surplus_hugepages
- which function as described above for the default huge page-sized case.
- .. _mem_policy_and_hp_alloc:
- Interaction of Task Memory Policy with Huge Page Allocation/Freeing
- ===================================================================
- Whether huge pages are allocated and freed via the ``/proc`` interface or
- the ``/sysfs`` interface using the ``nr_hugepages_mempolicy`` attribute, the
- NUMA nodes from which huge pages are allocated or freed are controlled by the
- NUMA memory policy of the task that modifies the ``nr_hugepages_mempolicy``
- sysctl or attribute. When the ``nr_hugepages`` attribute is used, mempolicy
- is ignored.
- The recommended method to allocate or free huge pages to/from the kernel
- huge page pool, using the ``nr_hugepages`` example above, is::
- numactl --interleave <node-list> echo 20 \
- >/proc/sys/vm/nr_hugepages_mempolicy
- or, more succinctly::
- numactl -m <node-list> echo 20 >/proc/sys/vm/nr_hugepages_mempolicy
- This will allocate or free ``abs(20 - nr_hugepages)`` to or from the nodes
- specified in <node-list>, depending on whether number of persistent huge pages
- is initially less than or greater than 20, respectively. No huge pages will be
- allocated nor freed on any node not included in the specified <node-list>.
- When adjusting the persistent hugepage count via ``nr_hugepages_mempolicy``, any
- memory policy mode--bind, preferred, local or interleave--may be used. The
- resulting effect on persistent huge page allocation is as follows:
- #. Regardless of mempolicy mode [see
- :ref:`Documentation/admin-guide/mm/numa_memory_policy.rst <numa_memory_policy>`],
- persistent huge pages will be distributed across the node or nodes
- specified in the mempolicy as if "interleave" had been specified.
- However, if a node in the policy does not contain sufficient contiguous
- memory for a huge page, the allocation will not "fallback" to the nearest
- neighbor node with sufficient contiguous memory. To do this would cause
- undesirable imbalance in the distribution of the huge page pool, or
- possibly, allocation of persistent huge pages on nodes not allowed by
- the task's memory policy.
- #. One or more nodes may be specified with the bind or interleave policy.
- If more than one node is specified with the preferred policy, only the
- lowest numeric id will be used. Local policy will select the node where
- the task is running at the time the nodes_allowed mask is constructed.
- For local policy to be deterministic, the task must be bound to a cpu or
- cpus in a single node. Otherwise, the task could be migrated to some
- other node at any time after launch and the resulting node will be
- indeterminate. Thus, local policy is not very useful for this purpose.
- Any of the other mempolicy modes may be used to specify a single node.
- #. The nodes allowed mask will be derived from any non-default task mempolicy,
- whether this policy was set explicitly by the task itself or one of its
- ancestors, such as numactl. This means that if the task is invoked from a
- shell with non-default policy, that policy will be used. One can specify a
- node list of "all" with numactl --interleave or --membind [-m] to achieve
- interleaving over all nodes in the system or cpuset.
- #. Any task mempolicy specified--e.g., using numactl--will be constrained by
- the resource limits of any cpuset in which the task runs. Thus, there will
- be no way for a task with non-default policy running in a cpuset with a
- subset of the system nodes to allocate huge pages outside the cpuset
- without first moving to a cpuset that contains all of the desired nodes.
- #. Boot-time huge page allocation attempts to distribute the requested number
- of huge pages over all on-lines nodes with memory.
- Per Node Hugepages Attributes
- =============================
- A subset of the contents of the root huge page control directory in sysfs,
- described above, will be replicated under each the system device of each
- NUMA node with memory in::
- /sys/devices/system/node/node[0-9]*/hugepages/
- Under this directory, the subdirectory for each supported huge page size
- contains the following attribute files::
- nr_hugepages
- free_hugepages
- surplus_hugepages
- The free\_' and surplus\_' attribute files are read-only. They return the number
- of free and surplus [overcommitted] huge pages, respectively, on the parent
- node.
- The ``nr_hugepages`` attribute returns the total number of huge pages on the
- specified node. When this attribute is written, the number of persistent huge
- pages on the parent node will be adjusted to the specified value, if sufficient
- resources exist, regardless of the task's mempolicy or cpuset constraints.
- Note that the number of overcommit and reserve pages remain global quantities,
- as we don't know until fault time, when the faulting task's mempolicy is
- applied, from which node the huge page allocation will be attempted.
- .. _using_huge_pages:
- Using Huge Pages
- ================
- If the user applications are going to request huge pages using mmap system
- call, then it is required that system administrator mount a file system of
- type hugetlbfs::
- mount -t hugetlbfs \
- -o uid=<value>,gid=<value>,mode=<value>,pagesize=<value>,size=<value>,\
- min_size=<value>,nr_inodes=<value> none /mnt/huge
- This command mounts a (pseudo) filesystem of type hugetlbfs on the directory
- ``/mnt/huge``. Any file created on ``/mnt/huge`` uses huge pages.
- The ``uid`` and ``gid`` options sets the owner and group of the root of the
- file system. By default the ``uid`` and ``gid`` of the current process
- are taken.
- The ``mode`` option sets the mode of root of file system to value & 01777.
- This value is given in octal. By default the value 0755 is picked.
- If the platform supports multiple huge page sizes, the ``pagesize`` option can
- be used to specify the huge page size and associated pool. ``pagesize``
- is specified in bytes. If ``pagesize`` is not specified the platform's
- default huge page size and associated pool will be used.
- The ``size`` option sets the maximum value of memory (huge pages) allowed
- for that filesystem (``/mnt/huge``). The ``size`` option can be specified
- in bytes, or as a percentage of the specified huge page pool (``nr_hugepages``).
- The size is rounded down to HPAGE_SIZE boundary.
- The ``min_size`` option sets the minimum value of memory (huge pages) allowed
- for the filesystem. ``min_size`` can be specified in the same way as ``size``,
- either bytes or a percentage of the huge page pool.
- At mount time, the number of huge pages specified by ``min_size`` are reserved
- for use by the filesystem.
- If there are not enough free huge pages available, the mount will fail.
- As huge pages are allocated to the filesystem and freed, the reserve count
- is adjusted so that the sum of allocated and reserved huge pages is always
- at least ``min_size``.
- The option ``nr_inodes`` sets the maximum number of inodes that ``/mnt/huge``
- can use.
- If the ``size``, ``min_size`` or ``nr_inodes`` option is not provided on
- command line then no limits are set.
- For ``pagesize``, ``size``, ``min_size`` and ``nr_inodes`` options, you can
- use [G|g]/[M|m]/[K|k] to represent giga/mega/kilo.
- For example, size=2K has the same meaning as size=2048.
- While read system calls are supported on files that reside on hugetlb
- file systems, write system calls are not.
- Regular chown, chgrp, and chmod commands (with right permissions) could be
- used to change the file attributes on hugetlbfs.
- Also, it is important to note that no such mount command is required if
- applications are going to use only shmat/shmget system calls or mmap with
- MAP_HUGETLB. For an example of how to use mmap with MAP_HUGETLB see
- :ref:`map_hugetlb <map_hugetlb>` below.
- Users who wish to use hugetlb memory via shared memory segment should be
- members of a supplementary group and system admin needs to configure that gid
- into ``/proc/sys/vm/hugetlb_shm_group``. It is possible for same or different
- applications to use any combination of mmaps and shm* calls, though the mount of
- filesystem will be required for using mmap calls without MAP_HUGETLB.
- Syscalls that operate on memory backed by hugetlb pages only have their lengths
- aligned to the native page size of the processor; they will normally fail with
- errno set to EINVAL or exclude hugetlb pages that extend beyond the length if
- not hugepage aligned. For example, munmap(2) will fail if memory is backed by
- a hugetlb page and the length is smaller than the hugepage size.
- Examples
- ========
- .. _map_hugetlb:
- ``map_hugetlb``
- see tools/testing/selftests/vm/map_hugetlb.c
- ``hugepage-shm``
- see tools/testing/selftests/vm/hugepage-shm.c
- ``hugepage-mmap``
- see tools/testing/selftests/vm/hugepage-mmap.c
- The `libhugetlbfs`_ library provides a wide range of userspace tools
- to help with huge page usability, environment setup, and control.
- .. _libhugetlbfs: https://github.com/libhugetlbfs/libhugetlbfs
|