// SPDX-License-Identifier: GPL-2.0 /* * Optimize vmemmap pages associated with HugeTLB * * Copyright (c) 2020, Bytedance. All rights reserved. * * Author: Muchun Song * * See Documentation/vm/vmemmap_dedup.rst */ #define pr_fmt(fmt) "HugeTLB: " fmt #include #include "hugetlb_vmemmap.h" /* * There are a lot of struct page structures associated with each HugeTLB page. * For tail pages, the value of compound_head is the same. So we can reuse first * page of head page structures. We map the virtual addresses of all the pages * of tail page structures to the head page struct, and then free these page * frames. Therefore, we need to reserve one pages as vmemmap areas. */ #define RESERVE_VMEMMAP_NR 1U #define RESERVE_VMEMMAP_SIZE (RESERVE_VMEMMAP_NR << PAGE_SHIFT) enum vmemmap_optimize_mode { VMEMMAP_OPTIMIZE_OFF, VMEMMAP_OPTIMIZE_ON, }; DEFINE_STATIC_KEY_MAYBE(CONFIG_HUGETLB_PAGE_OPTIMIZE_VMEMMAP_DEFAULT_ON, hugetlb_optimize_vmemmap_key); EXPORT_SYMBOL(hugetlb_optimize_vmemmap_key); static enum vmemmap_optimize_mode vmemmap_optimize_mode = IS_ENABLED(CONFIG_HUGETLB_PAGE_OPTIMIZE_VMEMMAP_DEFAULT_ON); static void vmemmap_optimize_mode_switch(enum vmemmap_optimize_mode to) { if (vmemmap_optimize_mode == to) return; if (to == VMEMMAP_OPTIMIZE_OFF) static_branch_dec(&hugetlb_optimize_vmemmap_key); else static_branch_inc(&hugetlb_optimize_vmemmap_key); WRITE_ONCE(vmemmap_optimize_mode, to); } static int __init hugetlb_vmemmap_early_param(char *buf) { bool enable; enum vmemmap_optimize_mode mode; if (kstrtobool(buf, &enable)) return -EINVAL; mode = enable ? VMEMMAP_OPTIMIZE_ON : VMEMMAP_OPTIMIZE_OFF; vmemmap_optimize_mode_switch(mode); return 0; } early_param("hugetlb_free_vmemmap", hugetlb_vmemmap_early_param); /* * Previously discarded vmemmap pages will be allocated and remapping * after this function returns zero. */ int hugetlb_vmemmap_alloc(struct hstate *h, struct page *head) { int ret; unsigned long vmemmap_addr = (unsigned long)head; unsigned long vmemmap_end, vmemmap_reuse, vmemmap_pages; if (!HPageVmemmapOptimized(head)) return 0; vmemmap_addr += RESERVE_VMEMMAP_SIZE; vmemmap_pages = hugetlb_optimize_vmemmap_pages(h); vmemmap_end = vmemmap_addr + (vmemmap_pages << PAGE_SHIFT); vmemmap_reuse = vmemmap_addr - PAGE_SIZE; /* * The pages which the vmemmap virtual address range [@vmemmap_addr, * @vmemmap_end) are mapped to are freed to the buddy allocator, and * the range is mapped to the page which @vmemmap_reuse is mapped to. * When a HugeTLB page is freed to the buddy allocator, previously * discarded vmemmap pages must be allocated and remapping. */ ret = vmemmap_remap_alloc(vmemmap_addr, vmemmap_end, vmemmap_reuse, GFP_KERNEL | __GFP_NORETRY | __GFP_THISNODE); if (!ret) { ClearHPageVmemmapOptimized(head); static_branch_dec(&hugetlb_optimize_vmemmap_key); } return ret; } void hugetlb_vmemmap_free(struct hstate *h, struct page *head) { unsigned long vmemmap_addr = (unsigned long)head; unsigned long vmemmap_end, vmemmap_reuse, vmemmap_pages; vmemmap_pages = hugetlb_optimize_vmemmap_pages(h); if (!vmemmap_pages) return; if (READ_ONCE(vmemmap_optimize_mode) == VMEMMAP_OPTIMIZE_OFF) return; static_branch_inc(&hugetlb_optimize_vmemmap_key); vmemmap_addr += RESERVE_VMEMMAP_SIZE; vmemmap_end = vmemmap_addr + (vmemmap_pages << PAGE_SHIFT); vmemmap_reuse = vmemmap_addr - PAGE_SIZE; /* * Remap the vmemmap virtual address range [@vmemmap_addr, @vmemmap_end) * to the page which @vmemmap_reuse is mapped to, then free the pages * which the range [@vmemmap_addr, @vmemmap_end] is mapped to. */ if (vmemmap_remap_free(vmemmap_addr, vmemmap_end, vmemmap_reuse)) static_branch_dec(&hugetlb_optimize_vmemmap_key); else SetHPageVmemmapOptimized(head); } void __init hugetlb_vmemmap_init(struct hstate *h) { unsigned int nr_pages = pages_per_huge_page(h); unsigned int vmemmap_pages; /* * There are only (RESERVE_VMEMMAP_SIZE / sizeof(struct page)) struct * page structs that can be used when CONFIG_HUGETLB_PAGE_OPTIMIZE_VMEMMAP, * so add a BUILD_BUG_ON to catch invalid usage of the tail struct page. */ BUILD_BUG_ON(__NR_USED_SUBPAGE >= RESERVE_VMEMMAP_SIZE / sizeof(struct page)); if (!is_power_of_2(sizeof(struct page))) { pr_warn_once("cannot optimize vmemmap pages because \"struct page\" crosses page boundaries\n"); static_branch_disable(&hugetlb_optimize_vmemmap_key); return; } vmemmap_pages = (nr_pages * sizeof(struct page)) >> PAGE_SHIFT; /* * The head page is not to be freed to buddy allocator, the other tail * pages will map to the head page, so they can be freed. * * Could RESERVE_VMEMMAP_NR be greater than @vmemmap_pages? It is true * on some architectures (e.g. aarch64). See Documentation/arm64/ * hugetlbpage.rst for more details. */ if (likely(vmemmap_pages > RESERVE_VMEMMAP_NR)) h->optimize_vmemmap_pages = vmemmap_pages - RESERVE_VMEMMAP_NR; pr_info("can optimize %d vmemmap pages for %s\n", h->optimize_vmemmap_pages, h->name); } #ifdef CONFIG_PROC_SYSCTL static int hugetlb_optimize_vmemmap_handler(struct ctl_table *table, int write, void *buffer, size_t *length, loff_t *ppos) { int ret; enum vmemmap_optimize_mode mode; static DEFINE_MUTEX(sysctl_mutex); if (write && !capable(CAP_SYS_ADMIN)) return -EPERM; mutex_lock(&sysctl_mutex); mode = vmemmap_optimize_mode; table->data = &mode; ret = proc_dointvec_minmax(table, write, buffer, length, ppos); if (write && !ret) vmemmap_optimize_mode_switch(mode); mutex_unlock(&sysctl_mutex); return ret; } static struct ctl_table hugetlb_vmemmap_sysctls[] = { { .procname = "hugetlb_optimize_vmemmap", .maxlen = sizeof(enum vmemmap_optimize_mode), .mode = 0644, .proc_handler = hugetlb_optimize_vmemmap_handler, .extra1 = SYSCTL_ZERO, .extra2 = SYSCTL_ONE, }, { } }; static __init int hugetlb_vmemmap_sysctls_init(void) { /* * If "memory_hotplug.memmap_on_memory" is enabled or "struct page" * crosses page boundaries, the vmemmap pages cannot be optimized. */ if (!mhp_memmap_on_memory() && is_power_of_2(sizeof(struct page))) register_sysctl_init("vm", hugetlb_vmemmap_sysctls); return 0; } late_initcall(hugetlb_vmemmap_sysctls_init); #endif /* CONFIG_PROC_SYSCTL */