2019-05-19 15:08:55 +03:00
// SPDX-License-Identifier: GPL-2.0-only
2008-07-24 08:26:49 +04:00
/*
* mm_init . c - Memory initialisation verification and debugging
*
* Copyright 2008 IBM Corporation , 2008
* Author Mel Gorman < mel @ csn . ul . ie >
*
*/
# include <linux/kernel.h>
# include <linux/init.h>
2008-07-24 08:27:39 +04:00
# include <linux/kobject.h>
2011-10-16 10:01:52 +04:00
# include <linux/export.h>
2013-07-04 02:02:44 +04:00
# include <linux/memory.h>
# include <linux/notifier.h>
2015-07-01 00:57:05 +03:00
# include <linux/sched.h>
2020-08-07 09:23:15 +03:00
# include <linux/mman.h>
2008-07-24 08:26:51 +04:00
# include "internal.h"
2008-07-24 08:26:49 +04:00
2008-07-24 08:27:39 +04:00
# ifdef CONFIG_DEBUG_MEMORY_INIT
2015-02-13 02:00:12 +03:00
int __meminitdata mminit_loglevel ;
2008-07-24 08:26:49 +04:00
2008-07-24 08:26:52 +04:00
/* The zonelists are simply reported, validation is manual. */
2015-02-13 02:00:09 +03:00
void __init mminit_verify_zonelist ( void )
2008-07-24 08:26:52 +04:00
{
int nid ;
if ( mminit_loglevel < MMINIT_VERIFY )
return ;
for_each_online_node ( nid ) {
pg_data_t * pgdat = NODE_DATA ( nid ) ;
struct zone * zone ;
struct zoneref * z ;
struct zonelist * zonelist ;
int i , listid , zoneid ;
2020-04-07 06:08:36 +03:00
BUILD_BUG_ON ( MAX_ZONELISTS > 2 ) ;
2008-07-24 08:26:52 +04:00
for ( i = 0 ; i < MAX_ZONELISTS * MAX_NR_ZONES ; i + + ) {
/* Identify the zone and nodelist */
zoneid = i % MAX_NR_ZONES ;
listid = i / MAX_NR_ZONES ;
zonelist = & pgdat - > node_zonelists [ listid ] ;
zone = & pgdat - > node_zones [ zoneid ] ;
if ( ! populated_zone ( zone ) )
continue ;
/* Print information about the zonelist */
printk ( KERN_DEBUG " mminit::zonelist %s %d:%s = " ,
listid > 0 ? " thisnode " : " general " , nid ,
zone - > name ) ;
/* Iterate the zonelist */
2018-08-22 07:53:32 +03:00
for_each_zone_zonelist ( zone , z , zonelist , zoneid )
pr_cont ( " %d:%s " , zone_to_nid ( zone ) , zone - > name ) ;
2016-03-18 00:19:50 +03:00
pr_cont ( " \n " ) ;
2008-07-24 08:26:52 +04:00
}
}
}
2008-07-24 08:26:51 +04:00
void __init mminit_verify_pageflags_layout ( void )
{
int shift , width ;
unsigned long or_mask , add_mask ;
shift = 8 * sizeof ( unsigned long ) ;
2020-06-02 07:52:49 +03:00
width = shift - SECTIONS_WIDTH - NODES_WIDTH - ZONES_WIDTH
mm: multi-gen LRU: groundwork
Evictable pages are divided into multiple generations for each lruvec.
The youngest generation number is stored in lrugen->max_seq for both
anon and file types as they are aged on an equal footing. The oldest
generation numbers are stored in lrugen->min_seq[] separately for anon
and file types as clean file pages can be evicted regardless of swap
constraints. These three variables are monotonically increasing.
Generation numbers are truncated into order_base_2(MAX_NR_GENS+1) bits
in order to fit into the gen counter in folio->flags. Each truncated
generation number is an index to lrugen->lists[]. The sliding window
technique is used to track at least MIN_NR_GENS and at most
MAX_NR_GENS generations. The gen counter stores a value within [1,
MAX_NR_GENS] while a page is on one of lrugen->lists[]. Otherwise it
stores 0.
There are two conceptually independent procedures: "the aging", which
produces young generations, and "the eviction", which consumes old
generations. They form a closed-loop system, i.e., "the page reclaim".
Both procedures can be invoked from userspace for the purposes of working
set estimation and proactive reclaim. These techniques are commonly used
to optimize job scheduling (bin packing) in data centers [1][2].
To avoid confusion, the terms "hot" and "cold" will be applied to the
multi-gen LRU, as a new convention; the terms "active" and "inactive" will
be applied to the active/inactive LRU, as usual.
The protection of hot pages and the selection of cold pages are based
on page access channels and patterns. There are two access channels:
one through page tables and the other through file descriptors. The
protection of the former channel is by design stronger because:
1. The uncertainty in determining the access patterns of the former
channel is higher due to the approximation of the accessed bit.
2. The cost of evicting the former channel is higher due to the TLB
flushes required and the likelihood of encountering the dirty bit.
3. The penalty of underprotecting the former channel is higher because
applications usually do not prepare themselves for major page
faults like they do for blocked I/O. E.g., GUI applications
commonly use dedicated I/O threads to avoid blocking rendering
threads.
There are also two access patterns: one with temporal locality and the
other without. For the reasons listed above, the former channel is
assumed to follow the former pattern unless VM_SEQ_READ or VM_RAND_READ is
present; the latter channel is assumed to follow the latter pattern unless
outlying refaults have been observed [3][4].
The next patch will address the "outlying refaults". Three macros, i.e.,
LRU_REFS_WIDTH, LRU_REFS_PGOFF and LRU_REFS_MASK, used later are added in
this patch to make the entire patchset less diffy.
A page is added to the youngest generation on faulting. The aging needs
to check the accessed bit at least twice before handing this page over to
the eviction. The first check takes care of the accessed bit set on the
initial fault; the second check makes sure this page has not been used
since then. This protocol, AKA second chance, requires a minimum of two
generations, hence MIN_NR_GENS.
[1] https://dl.acm.org/doi/10.1145/3297858.3304053
[2] https://dl.acm.org/doi/10.1145/3503222.3507731
[3] https://lwn.net/Articles/495543/
[4] https://lwn.net/Articles/815342/
Link: https://lkml.kernel.org/r/20220918080010.2920238-6-yuzhao@google.com
Signed-off-by: Yu Zhao <yuzhao@google.com>
Acked-by: Brian Geffon <bgeffon@google.com>
Acked-by: Jan Alexander Steffens (heftig) <heftig@archlinux.org>
Acked-by: Oleksandr Natalenko <oleksandr@natalenko.name>
Acked-by: Steven Barrett <steven@liquorix.net>
Acked-by: Suleiman Souhlal <suleiman@google.com>
Tested-by: Daniel Byrne <djbyrne@mtu.edu>
Tested-by: Donald Carr <d@chaos-reins.com>
Tested-by: Holger Hoffstätte <holger@applied-asynchrony.com>
Tested-by: Konstantin Kharlamov <Hi-Angel@yandex.ru>
Tested-by: Shuang Zhai <szhai2@cs.rochester.edu>
Tested-by: Sofia Trinh <sofia.trinh@edi.works>
Tested-by: Vaibhav Jain <vaibhav@linux.ibm.com>
Cc: Andi Kleen <ak@linux.intel.com>
Cc: Aneesh Kumar K.V <aneesh.kumar@linux.ibm.com>
Cc: Barry Song <baohua@kernel.org>
Cc: Catalin Marinas <catalin.marinas@arm.com>
Cc: Dave Hansen <dave.hansen@linux.intel.com>
Cc: Hillf Danton <hdanton@sina.com>
Cc: Jens Axboe <axboe@kernel.dk>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Jonathan Corbet <corbet@lwn.net>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Mel Gorman <mgorman@suse.de>
Cc: Miaohe Lin <linmiaohe@huawei.com>
Cc: Michael Larabel <Michael@MichaelLarabel.com>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Mike Rapoport <rppt@kernel.org>
Cc: Mike Rapoport <rppt@linux.ibm.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Qi Zheng <zhengqi.arch@bytedance.com>
Cc: Tejun Heo <tj@kernel.org>
Cc: Vlastimil Babka <vbabka@suse.cz>
Cc: Will Deacon <will@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
2022-09-18 11:00:02 +03:00
- LAST_CPUPID_SHIFT - KASAN_TAG_WIDTH - LRU_GEN_WIDTH - LRU_REFS_WIDTH ;
2008-07-24 08:26:51 +04:00
mminit_dprintk ( MMINIT_TRACE , " pageflags_layout_widths " ,
mm: multi-gen LRU: groundwork
Evictable pages are divided into multiple generations for each lruvec.
The youngest generation number is stored in lrugen->max_seq for both
anon and file types as they are aged on an equal footing. The oldest
generation numbers are stored in lrugen->min_seq[] separately for anon
and file types as clean file pages can be evicted regardless of swap
constraints. These three variables are monotonically increasing.
Generation numbers are truncated into order_base_2(MAX_NR_GENS+1) bits
in order to fit into the gen counter in folio->flags. Each truncated
generation number is an index to lrugen->lists[]. The sliding window
technique is used to track at least MIN_NR_GENS and at most
MAX_NR_GENS generations. The gen counter stores a value within [1,
MAX_NR_GENS] while a page is on one of lrugen->lists[]. Otherwise it
stores 0.
There are two conceptually independent procedures: "the aging", which
produces young generations, and "the eviction", which consumes old
generations. They form a closed-loop system, i.e., "the page reclaim".
Both procedures can be invoked from userspace for the purposes of working
set estimation and proactive reclaim. These techniques are commonly used
to optimize job scheduling (bin packing) in data centers [1][2].
To avoid confusion, the terms "hot" and "cold" will be applied to the
multi-gen LRU, as a new convention; the terms "active" and "inactive" will
be applied to the active/inactive LRU, as usual.
The protection of hot pages and the selection of cold pages are based
on page access channels and patterns. There are two access channels:
one through page tables and the other through file descriptors. The
protection of the former channel is by design stronger because:
1. The uncertainty in determining the access patterns of the former
channel is higher due to the approximation of the accessed bit.
2. The cost of evicting the former channel is higher due to the TLB
flushes required and the likelihood of encountering the dirty bit.
3. The penalty of underprotecting the former channel is higher because
applications usually do not prepare themselves for major page
faults like they do for blocked I/O. E.g., GUI applications
commonly use dedicated I/O threads to avoid blocking rendering
threads.
There are also two access patterns: one with temporal locality and the
other without. For the reasons listed above, the former channel is
assumed to follow the former pattern unless VM_SEQ_READ or VM_RAND_READ is
present; the latter channel is assumed to follow the latter pattern unless
outlying refaults have been observed [3][4].
The next patch will address the "outlying refaults". Three macros, i.e.,
LRU_REFS_WIDTH, LRU_REFS_PGOFF and LRU_REFS_MASK, used later are added in
this patch to make the entire patchset less diffy.
A page is added to the youngest generation on faulting. The aging needs
to check the accessed bit at least twice before handing this page over to
the eviction. The first check takes care of the accessed bit set on the
initial fault; the second check makes sure this page has not been used
since then. This protocol, AKA second chance, requires a minimum of two
generations, hence MIN_NR_GENS.
[1] https://dl.acm.org/doi/10.1145/3297858.3304053
[2] https://dl.acm.org/doi/10.1145/3503222.3507731
[3] https://lwn.net/Articles/495543/
[4] https://lwn.net/Articles/815342/
Link: https://lkml.kernel.org/r/20220918080010.2920238-6-yuzhao@google.com
Signed-off-by: Yu Zhao <yuzhao@google.com>
Acked-by: Brian Geffon <bgeffon@google.com>
Acked-by: Jan Alexander Steffens (heftig) <heftig@archlinux.org>
Acked-by: Oleksandr Natalenko <oleksandr@natalenko.name>
Acked-by: Steven Barrett <steven@liquorix.net>
Acked-by: Suleiman Souhlal <suleiman@google.com>
Tested-by: Daniel Byrne <djbyrne@mtu.edu>
Tested-by: Donald Carr <d@chaos-reins.com>
Tested-by: Holger Hoffstätte <holger@applied-asynchrony.com>
Tested-by: Konstantin Kharlamov <Hi-Angel@yandex.ru>
Tested-by: Shuang Zhai <szhai2@cs.rochester.edu>
Tested-by: Sofia Trinh <sofia.trinh@edi.works>
Tested-by: Vaibhav Jain <vaibhav@linux.ibm.com>
Cc: Andi Kleen <ak@linux.intel.com>
Cc: Aneesh Kumar K.V <aneesh.kumar@linux.ibm.com>
Cc: Barry Song <baohua@kernel.org>
Cc: Catalin Marinas <catalin.marinas@arm.com>
Cc: Dave Hansen <dave.hansen@linux.intel.com>
Cc: Hillf Danton <hdanton@sina.com>
Cc: Jens Axboe <axboe@kernel.dk>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Jonathan Corbet <corbet@lwn.net>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Mel Gorman <mgorman@suse.de>
Cc: Miaohe Lin <linmiaohe@huawei.com>
Cc: Michael Larabel <Michael@MichaelLarabel.com>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Mike Rapoport <rppt@kernel.org>
Cc: Mike Rapoport <rppt@linux.ibm.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Qi Zheng <zhengqi.arch@bytedance.com>
Cc: Tejun Heo <tj@kernel.org>
Cc: Vlastimil Babka <vbabka@suse.cz>
Cc: Will Deacon <will@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
2022-09-18 11:00:02 +03:00
" Section %d Node %d Zone %d Lastcpupid %d Kasantag %d Gen %d Tier %d Flags %d \n " ,
2008-07-24 08:26:51 +04:00
SECTIONS_WIDTH ,
NODES_WIDTH ,
ZONES_WIDTH ,
2013-10-07 14:29:20 +04:00
LAST_CPUPID_WIDTH ,
2020-06-02 07:52:49 +03:00
KASAN_TAG_WIDTH ,
mm: multi-gen LRU: groundwork
Evictable pages are divided into multiple generations for each lruvec.
The youngest generation number is stored in lrugen->max_seq for both
anon and file types as they are aged on an equal footing. The oldest
generation numbers are stored in lrugen->min_seq[] separately for anon
and file types as clean file pages can be evicted regardless of swap
constraints. These three variables are monotonically increasing.
Generation numbers are truncated into order_base_2(MAX_NR_GENS+1) bits
in order to fit into the gen counter in folio->flags. Each truncated
generation number is an index to lrugen->lists[]. The sliding window
technique is used to track at least MIN_NR_GENS and at most
MAX_NR_GENS generations. The gen counter stores a value within [1,
MAX_NR_GENS] while a page is on one of lrugen->lists[]. Otherwise it
stores 0.
There are two conceptually independent procedures: "the aging", which
produces young generations, and "the eviction", which consumes old
generations. They form a closed-loop system, i.e., "the page reclaim".
Both procedures can be invoked from userspace for the purposes of working
set estimation and proactive reclaim. These techniques are commonly used
to optimize job scheduling (bin packing) in data centers [1][2].
To avoid confusion, the terms "hot" and "cold" will be applied to the
multi-gen LRU, as a new convention; the terms "active" and "inactive" will
be applied to the active/inactive LRU, as usual.
The protection of hot pages and the selection of cold pages are based
on page access channels and patterns. There are two access channels:
one through page tables and the other through file descriptors. The
protection of the former channel is by design stronger because:
1. The uncertainty in determining the access patterns of the former
channel is higher due to the approximation of the accessed bit.
2. The cost of evicting the former channel is higher due to the TLB
flushes required and the likelihood of encountering the dirty bit.
3. The penalty of underprotecting the former channel is higher because
applications usually do not prepare themselves for major page
faults like they do for blocked I/O. E.g., GUI applications
commonly use dedicated I/O threads to avoid blocking rendering
threads.
There are also two access patterns: one with temporal locality and the
other without. For the reasons listed above, the former channel is
assumed to follow the former pattern unless VM_SEQ_READ or VM_RAND_READ is
present; the latter channel is assumed to follow the latter pattern unless
outlying refaults have been observed [3][4].
The next patch will address the "outlying refaults". Three macros, i.e.,
LRU_REFS_WIDTH, LRU_REFS_PGOFF and LRU_REFS_MASK, used later are added in
this patch to make the entire patchset less diffy.
A page is added to the youngest generation on faulting. The aging needs
to check the accessed bit at least twice before handing this page over to
the eviction. The first check takes care of the accessed bit set on the
initial fault; the second check makes sure this page has not been used
since then. This protocol, AKA second chance, requires a minimum of two
generations, hence MIN_NR_GENS.
[1] https://dl.acm.org/doi/10.1145/3297858.3304053
[2] https://dl.acm.org/doi/10.1145/3503222.3507731
[3] https://lwn.net/Articles/495543/
[4] https://lwn.net/Articles/815342/
Link: https://lkml.kernel.org/r/20220918080010.2920238-6-yuzhao@google.com
Signed-off-by: Yu Zhao <yuzhao@google.com>
Acked-by: Brian Geffon <bgeffon@google.com>
Acked-by: Jan Alexander Steffens (heftig) <heftig@archlinux.org>
Acked-by: Oleksandr Natalenko <oleksandr@natalenko.name>
Acked-by: Steven Barrett <steven@liquorix.net>
Acked-by: Suleiman Souhlal <suleiman@google.com>
Tested-by: Daniel Byrne <djbyrne@mtu.edu>
Tested-by: Donald Carr <d@chaos-reins.com>
Tested-by: Holger Hoffstätte <holger@applied-asynchrony.com>
Tested-by: Konstantin Kharlamov <Hi-Angel@yandex.ru>
Tested-by: Shuang Zhai <szhai2@cs.rochester.edu>
Tested-by: Sofia Trinh <sofia.trinh@edi.works>
Tested-by: Vaibhav Jain <vaibhav@linux.ibm.com>
Cc: Andi Kleen <ak@linux.intel.com>
Cc: Aneesh Kumar K.V <aneesh.kumar@linux.ibm.com>
Cc: Barry Song <baohua@kernel.org>
Cc: Catalin Marinas <catalin.marinas@arm.com>
Cc: Dave Hansen <dave.hansen@linux.intel.com>
Cc: Hillf Danton <hdanton@sina.com>
Cc: Jens Axboe <axboe@kernel.dk>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Jonathan Corbet <corbet@lwn.net>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Mel Gorman <mgorman@suse.de>
Cc: Miaohe Lin <linmiaohe@huawei.com>
Cc: Michael Larabel <Michael@MichaelLarabel.com>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Mike Rapoport <rppt@kernel.org>
Cc: Mike Rapoport <rppt@linux.ibm.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Qi Zheng <zhengqi.arch@bytedance.com>
Cc: Tejun Heo <tj@kernel.org>
Cc: Vlastimil Babka <vbabka@suse.cz>
Cc: Will Deacon <will@kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
2022-09-18 11:00:02 +03:00
LRU_GEN_WIDTH ,
LRU_REFS_WIDTH ,
2008-07-24 08:26:51 +04:00
NR_PAGEFLAGS ) ;
mminit_dprintk ( MMINIT_TRACE , " pageflags_layout_shifts " ,
2020-06-02 07:52:49 +03:00
" Section %d Node %d Zone %d Lastcpupid %d Kasantag %d \n " ,
2008-07-24 08:26:51 +04:00
SECTIONS_SHIFT ,
NODES_SHIFT ,
2013-02-23 04:34:47 +04:00
ZONES_SHIFT ,
2020-06-02 07:52:49 +03:00
LAST_CPUPID_SHIFT ,
KASAN_TAG_WIDTH ) ;
2013-02-23 04:34:47 +04:00
mminit_dprintk ( MMINIT_TRACE , " pageflags_layout_pgshifts " ,
2020-06-02 07:52:49 +03:00
" Section %lu Node %lu Zone %lu Lastcpupid %lu Kasantag %lu \n " ,
2008-07-24 08:26:51 +04:00
( unsigned long ) SECTIONS_PGSHIFT ,
( unsigned long ) NODES_PGSHIFT ,
2013-02-23 04:34:47 +04:00
( unsigned long ) ZONES_PGSHIFT ,
2020-06-02 07:52:49 +03:00
( unsigned long ) LAST_CPUPID_PGSHIFT ,
( unsigned long ) KASAN_TAG_PGSHIFT ) ;
2013-02-23 04:34:47 +04:00
mminit_dprintk ( MMINIT_TRACE , " pageflags_layout_nodezoneid " ,
" Node/Zone ID: %lu -> %lu \n " ,
( unsigned long ) ( ZONEID_PGOFF + ZONEID_SHIFT ) ,
( unsigned long ) ZONEID_PGOFF ) ;
2008-07-24 08:26:51 +04:00
mminit_dprintk ( MMINIT_TRACE , " pageflags_layout_usage " ,
2013-02-23 04:34:47 +04:00
" location: %d -> %d layout %d -> %d unused %d -> %d page-flags \n " ,
2008-07-24 08:26:51 +04:00
shift , width , width , NR_PAGEFLAGS , NR_PAGEFLAGS , 0 ) ;
# ifdef NODE_NOT_IN_PAGE_FLAGS
mminit_dprintk ( MMINIT_TRACE , " pageflags_layout_nodeflags " ,
" Node not in page flags " ) ;
# endif
2013-10-07 14:29:20 +04:00
# ifdef LAST_CPUPID_NOT_IN_PAGE_FLAGS
2013-02-23 04:34:47 +04:00
mminit_dprintk ( MMINIT_TRACE , " pageflags_layout_nodeflags " ,
2013-10-07 14:29:20 +04:00
" Last cpupid not in page flags " ) ;
2013-02-23 04:34:47 +04:00
# endif
2008-07-24 08:26:51 +04:00
if ( SECTIONS_WIDTH ) {
shift - = SECTIONS_WIDTH ;
BUG_ON ( shift ! = SECTIONS_PGSHIFT ) ;
}
if ( NODES_WIDTH ) {
shift - = NODES_WIDTH ;
BUG_ON ( shift ! = NODES_PGSHIFT ) ;
}
if ( ZONES_WIDTH ) {
shift - = ZONES_WIDTH ;
BUG_ON ( shift ! = ZONES_PGSHIFT ) ;
}
/* Check for bitmask overlaps */
or_mask = ( ZONES_MASK < < ZONES_PGSHIFT ) |
( NODES_MASK < < NODES_PGSHIFT ) |
( SECTIONS_MASK < < SECTIONS_PGSHIFT ) ;
add_mask = ( ZONES_MASK < < ZONES_PGSHIFT ) +
( NODES_MASK < < NODES_PGSHIFT ) +
( SECTIONS_MASK < < SECTIONS_PGSHIFT ) ;
BUG_ON ( or_mask ! = add_mask ) ;
}
2008-07-24 08:26:49 +04:00
static __init int set_mminit_loglevel ( char * str )
{
get_option ( & str , & mminit_loglevel ) ;
return 0 ;
}
early_param ( " mminit_loglevel " , set_mminit_loglevel ) ;
2008-07-24 08:27:39 +04:00
# endif /* CONFIG_DEBUG_MEMORY_INIT */
2008-07-24 08:27:39 +04:00
struct kobject * mm_kobj ;
EXPORT_SYMBOL_GPL ( mm_kobj ) ;
2013-07-04 02:02:44 +04:00
# ifdef CONFIG_SMP
s32 vm_committed_as_batch = 32 ;
2020-08-07 09:23:15 +03:00
void mm_compute_batch ( int overcommit_policy )
2013-07-04 02:02:44 +04:00
{
u64 memsized_batch ;
s32 nr = num_present_cpus ( ) ;
s32 batch = max_t ( s32 , nr * 2 , 32 ) ;
2020-08-07 09:23:15 +03:00
unsigned long ram_pages = totalram_pages ( ) ;
/*
* For policy OVERCOMMIT_NEVER , set batch size to 0.4 % of
* ( total memory / # cpus ) , and lift it to 25 % for other policies
* to easy the possible lock contention for percpu_counter
* vm_committed_as , while the max limit is INT_MAX
*/
if ( overcommit_policy = = OVERCOMMIT_NEVER )
memsized_batch = min_t ( u64 , ram_pages / nr / 256 , INT_MAX ) ;
else
memsized_batch = min_t ( u64 , ram_pages / nr / 4 , INT_MAX ) ;
2013-07-04 02:02:44 +04:00
vm_committed_as_batch = max_t ( s32 , memsized_batch , batch ) ;
}
static int __meminit mm_compute_batch_notifier ( struct notifier_block * self ,
unsigned long action , void * arg )
{
switch ( action ) {
case MEM_ONLINE :
case MEM_OFFLINE :
2020-08-07 09:23:15 +03:00
mm_compute_batch ( sysctl_overcommit_memory ) ;
2020-12-15 06:15:00 +03:00
break ;
2013-07-04 02:02:44 +04:00
default :
break ;
}
return NOTIFY_OK ;
}
static int __init mm_compute_batch_init ( void )
{
2020-08-07 09:23:15 +03:00
mm_compute_batch ( sysctl_overcommit_memory ) ;
2022-09-23 06:33:47 +03:00
hotplug_memory_notifier ( mm_compute_batch_notifier , MM_COMPUTE_BATCH_PRI ) ;
2013-07-04 02:02:44 +04:00
return 0 ;
}
__initcall ( mm_compute_batch_init ) ;
# endif
2008-07-24 08:27:39 +04:00
static int __init mm_sysfs_init ( void )
{
mm_kobj = kobject_create_and_add ( " mm " , kernel_kobj ) ;
if ( ! mm_kobj )
return - ENOMEM ;
return 0 ;
}
2014-01-28 05:06:55 +04:00
postcore_initcall ( mm_sysfs_init ) ;