提交 eb761d65 编写于 作者: D Daniel Jordan 提交者: Xie XiuQi

mm: parallelize deferred struct page initialization within each node

hulk inclusion
category: feature
bugzilla: 13228
CVE: NA
---------------------------

Deferred struct page initialization currently runs one thread per node,
but this is a bottleneck during boot on big machines, so use ktask
within each pgdatinit thread to parallelize the struct page
initialization, allowing the system to take better advantage of its
memory bandwidth.

Because the system is not fully up yet and most CPUs are idle, use more
than the default maximum number of ktask threads.  The kernel doesn't
know the memory bandwidth of a given system to get the most efficient
number of threads, so there's some guesswork involved.  In testing, a
reasonable value turned out to be about a quarter of the CPUs on the
node.

__free_pages_core used to increase the zone's managed page count by the
number of pages being freed.  To accommodate multiple threads, however,
account the number of freed pages with an atomic shared across the ktask
threads and bump the managed page count with it after ktask is finished.

Test:    Boot the machine with deferred struct page init three times

Machine: Intel(R) Xeon(R) CPU E5-2699 v4 @ 2.20GHz, 88 CPUs, 503G memory,
         2 sockets

kernel                   speedup   max time per   stdev
                                   node (ms)

baseline (4.15-rc2)                        5860     8.6
ktask                      9.56x            613    12.4
Signed-off-by: NDaniel Jordan <daniel.m.jordan@oracle.com>
Signed-off-by: NHongbo Yao <yaohongbo@huawei.com>
Reviewed-by: NXie XiuQi <xiexiuqi@huawei.com>
Tested-by: NHongbo Yao <yaohongbo@huawei.com>
Signed-off-by: NYang Yingliang <yangyingliang@huawei.com>
上级 228e4183
......@@ -66,6 +66,7 @@
#include <linux/ftrace.h>
#include <linux/lockdep.h>
#include <linux/nmi.h>
#include <linux/ktask.h>
#include <asm/sections.h>
#include <asm/tlbflush.h>
......@@ -1295,7 +1296,6 @@ static void __init __free_pages_boot_core(struct page *page, unsigned int order)
__ClearPageReserved(p);
set_page_count(p, 0);
page_zone(page)->managed_pages += nr_pages;
set_page_refcounted(page);
__free_pages(page, order);
}
......@@ -1359,6 +1359,8 @@ void __init __free_pages_bootmem(struct page *page, unsigned long pfn,
{
if (early_page_uninitialised(pfn))
return;
page_zone(page)->managed_pages += 1UL << order;
return __free_pages_boot_core(page, order);
}
......@@ -1497,23 +1499,31 @@ deferred_pfn_valid(int nid, unsigned long pfn,
return true;
}
struct deferred_args {
int nid;
int zid;
atomic64_t nr_pages;
};
/*
* Free pages to buddy allocator. Try to free aligned pages in
* pageblock_nr_pages sizes.
*/
static void __init deferred_free_pages(int nid, int zid, unsigned long pfn,
unsigned long end_pfn)
static int __init deferred_free_pages(int nid, int zid, unsigned long pfn,
unsigned long end_pfn)
{
struct mminit_pfnnid_cache nid_init_state = { };
unsigned long nr_pgmask = pageblock_nr_pages - 1;
unsigned long nr_free = 0;
unsigned long nr_free = 0, nr_pages = 0;
for (; pfn < end_pfn; pfn++) {
if (!deferred_pfn_valid(nid, pfn, &nid_init_state)) {
deferred_free_range(pfn - nr_free, nr_free);
nr_pages += nr_free;
nr_free = 0;
} else if (!(pfn & nr_pgmask)) {
deferred_free_range(pfn - nr_free, nr_free);
nr_pages += nr_free;
nr_free = 1;
touch_nmi_watchdog();
} else {
......@@ -1522,16 +1532,27 @@ static void __init deferred_free_pages(int nid, int zid, unsigned long pfn,
}
/* Free the last block of pages to allocator */
deferred_free_range(pfn - nr_free, nr_free);
nr_pages += nr_free;
return nr_pages;
}
static int __init deferred_free_chunk(unsigned long pfn, unsigned long end_pfn,
struct deferred_args *args)
{
unsigned long nr_pages = deferred_free_pages(args->nid, args->zid, pfn,
end_pfn);
atomic64_add(nr_pages, &args->nr_pages);
return KTASK_RETURN_SUCCESS;
}
/*
* Initialize struct pages. We minimize pfn page lookups and scheduler checks
* by performing it only once every pageblock_nr_pages.
* Return number of pages initialized.
* Return number of pages initialized in deferred_args.
*/
static unsigned long __init deferred_init_pages(int nid, int zid,
unsigned long pfn,
unsigned long end_pfn)
static int __init deferred_init_pages(int nid, int zid, unsigned long pfn,
unsigned long end_pfn)
{
struct mminit_pfnnid_cache nid_init_state = { };
unsigned long nr_pgmask = pageblock_nr_pages - 1;
......@@ -1551,7 +1572,17 @@ static unsigned long __init deferred_init_pages(int nid, int zid,
__init_single_page(page, pfn, zid, nid);
nr_pages++;
}
return (nr_pages);
return nr_pages;
}
static int __init deferred_init_chunk(unsigned long pfn, unsigned long end_pfn,
struct deferred_args *args)
{
unsigned long nr_pages = deferred_init_pages(args->nid, args->zid, pfn,
end_pfn);
atomic64_add(nr_pages, &args->nr_pages);
return KTASK_RETURN_SUCCESS;
}
/* Initialise remaining memory on a node */
......@@ -1560,13 +1591,15 @@ static int __init deferred_init_memmap(void *data)
pg_data_t *pgdat = data;
int nid = pgdat->node_id;
unsigned long start = jiffies;
unsigned long nr_pages = 0;
unsigned long nr_init = 0, nr_free = 0;
unsigned long spfn, epfn, first_init_pfn, flags;
phys_addr_t spa, epa;
int zid;
struct zone *zone;
const struct cpumask *cpumask = cpumask_of_node(pgdat->node_id);
u64 i;
unsigned long nr_node_cpus;
struct ktask_node kn;
/* Bind memory initialisation thread to a local node if possible */
if (!cpumask_empty(cpumask))
......@@ -1580,6 +1613,14 @@ static int __init deferred_init_memmap(void *data)
return 0;
}
/*
* We'd like to know the memory bandwidth of the chip to calculate the
* most efficient number of threads to start, but we can't. In
* testing, a good value for a variety of systems was a quarter of the
* CPUs on the node.
*/
nr_node_cpus = DIV_ROUND_UP(cpumask_weight(cpumask), 4);
/* Sanity check boundaries */
BUG_ON(pgdat->first_deferred_pfn < pgdat->node_start_pfn);
BUG_ON(pgdat->first_deferred_pfn > pgdat_end_pfn(pgdat));
......@@ -1600,21 +1641,46 @@ static int __init deferred_init_memmap(void *data)
* page in __free_one_page()).
*/
for_each_free_mem_range(i, nid, MEMBLOCK_NONE, &spa, &epa, NULL) {
struct deferred_args args = { nid, zid, ATOMIC64_INIT(0) };
DEFINE_KTASK_CTL(ctl, deferred_init_chunk, &args,
KTASK_PTE_MINCHUNK);
ktask_ctl_set_max_threads(&ctl, nr_node_cpus);
spfn = max_t(unsigned long, first_init_pfn, PFN_UP(spa));
epfn = min_t(unsigned long, zone_end_pfn(zone), PFN_DOWN(epa));
nr_pages += deferred_init_pages(nid, zid, spfn, epfn);
kn.kn_start = (void *)spfn;
kn.kn_task_size = (spfn < epfn) ? epfn - spfn : 0;
kn.kn_nid = nid;
(void) ktask_run_numa(&kn, 1, &ctl);
nr_init += atomic64_read(&args.nr_pages);
}
for_each_free_mem_range(i, nid, MEMBLOCK_NONE, &spa, &epa, NULL) {
struct deferred_args args = { nid, zid, ATOMIC64_INIT(0) };
DEFINE_KTASK_CTL(ctl, deferred_free_chunk, &args,
KTASK_PTE_MINCHUNK);
ktask_ctl_set_max_threads(&ctl, nr_node_cpus);
spfn = max_t(unsigned long, first_init_pfn, PFN_UP(spa));
epfn = min_t(unsigned long, zone_end_pfn(zone), PFN_DOWN(epa));
deferred_free_pages(nid, zid, spfn, epfn);
kn.kn_start = (void *)spfn;
kn.kn_task_size = (spfn < epfn) ? epfn - spfn : 0;
kn.kn_nid = nid;
(void) ktask_run_numa(&kn, 1, &ctl);
nr_free += atomic64_read(&args.nr_pages);
}
pgdat_resize_unlock(pgdat, &flags);
/* Sanity check that the next zone really is unpopulated */
WARN_ON(++zid < MAX_NR_ZONES && populated_zone(++zone));
VM_BUG_ON(nr_init != nr_free);
zone->managed_pages += nr_free;
pr_info("node %d initialised, %lu pages in %ums\n", nid, nr_pages,
pr_info("node %d initialised, %lu pages in %ums\n", nid, nr_free,
jiffies_to_msecs(jiffies - start));
pgdat_init_report_one_done();
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册