1. 18 9月, 2021 2 次提交
  2. 17 9月, 2021 1 次提交
  3. 16 9月, 2021 1 次提交
  4. 15 9月, 2021 1 次提交
    • L
      memblock: introduce saner 'memblock_free_ptr()' interface · 77e02cf5
      Linus Torvalds 提交于
      The boot-time allocation interface for memblock is a mess, with
      'memblock_alloc()' returning a virtual pointer, but then you are
      supposed to free it with 'memblock_free()' that takes a _physical_
      address.
      
      Not only is that all kinds of strange and illogical, but it actually
      causes bugs, when people then use it like a normal allocation function,
      and it fails spectacularly on a NULL pointer:
      
         https://lore.kernel.org/all/20210912140820.GD25450@xsang-OptiPlex-9020/
      
      or just random memory corruption if the debug checks don't catch it:
      
         https://lore.kernel.org/all/61ab2d0c-3313-aaab-514c-e15b7aa054a0@suse.cz/
      
      I really don't want to apply patches that treat the symptoms, when the
      fundamental cause is this horribly confusing interface.
      
      I started out looking at just automating a sane replacement sequence,
      but because of this mix or virtual and physical addresses, and because
      people have used the "__pa()" macro that can take either a regular
      kernel pointer, or just the raw "unsigned long" address, it's all quite
      messy.
      
      So this just introduces a new saner interface for freeing a virtual
      address that was allocated using 'memblock_alloc()', and that was kept
      as a regular kernel pointer.  And then it converts a couple of users
      that are obvious and easy to test, including the 'xbc_nodes' case in
      lib/bootconfig.c that caused problems.
      Reported-by: Nkernel test robot <oliver.sang@intel.com>
      Fixes: 40caa127 ("init: bootconfig: Remove all bootconfig data when the init memory is removed")
      Cc: Steven Rostedt <rostedt@goodmis.org>
      Cc: Mike Rapoport <rppt@kernel.org>
      Cc: Andrew Morton <akpm@linux-foundation.org>
      Cc: Ingo Molnar <mingo@kernel.org>
      Cc: Masami Hiramatsu <mhiramat@kernel.org>
      Cc: Vlastimil Babka <vbabka@suse.cz>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      77e02cf5
  5. 14 9月, 2021 6 次提交
  6. 13 9月, 2021 1 次提交
  7. 11 9月, 2021 3 次提交
    • T
      Documentation: core-api/cpuhotplug: Rewrite the API section · c9871c80
      Thomas Gleixner 提交于
      Dave stumbled over the incomplete and confusing documentation of the CPU
      hotplug API.
      
      Rewrite it, add the missing function documentations and correct the
      existing ones.
      Reported-by: NDave Chinner <david@fromorbit.com>
      Signed-off-by: NThomas Gleixner <tglx@linutronix.de>
      Link: https://lore.kernel.org/r/20210909123212.489059409@linutronix.de
      c9871c80
    • S
      cpu/hotplug: Remove deprecated CPU-hotplug functions. · 8c854303
      Sebastian Andrzej Siewior 提交于
      No users in tree use the deprecated CPU-hotplug functions anymore.
      
      Remove them.
      Signed-off-by: NSebastian Andrzej Siewior <bigeasy@linutronix.de>
      Signed-off-by: NThomas Gleixner <tglx@linutronix.de>
      Link: https://lore.kernel.org/r/20210803141621.780504-39-bigeasy@linutronix.de
      8c854303
    • Y
      bpf, mm: Fix lockdep warning triggered by stack_map_get_build_id_offset() · 2f1aaf3e
      Yonghong Song 提交于
      Currently the bpf selftest "get_stack_raw_tp" triggered the warning:
      
        [ 1411.304463] WARNING: CPU: 3 PID: 140 at include/linux/mmap_lock.h:164 find_vma+0x47/0xa0
        [ 1411.304469] Modules linked in: bpf_testmod(O) [last unloaded: bpf_testmod]
        [ 1411.304476] CPU: 3 PID: 140 Comm: systemd-journal Tainted: G        W  O      5.14.0+ #53
        [ 1411.304479] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS rel-1.14.0-0-g155821a1990b-prebuilt.qemu.org 04/01/2014
        [ 1411.304481] RIP: 0010:find_vma+0x47/0xa0
        [ 1411.304484] Code: de 48 89 ef e8 ba f5 fe ff 48 85 c0 74 2e 48 83 c4 08 5b 5d c3 48 8d bf 28 01 00 00 be ff ff ff ff e8 2d 9f d8 00 85 c0 75 d4 <0f> 0b 48 89 de 48 8
        [ 1411.304487] RSP: 0018:ffffabd440403db8 EFLAGS: 00010246
        [ 1411.304490] RAX: 0000000000000000 RBX: 00007f00ad80a0e0 RCX: 0000000000000000
        [ 1411.304492] RDX: 0000000000000001 RSI: ffffffff9776b144 RDI: ffffffff977e1b0e
        [ 1411.304494] RBP: ffff9cf5c2f50000 R08: ffff9cf5c3eb25d8 R09: 00000000fffffffe
        [ 1411.304496] R10: 0000000000000001 R11: 00000000ef974e19 R12: ffff9cf5c39ae0e0
        [ 1411.304498] R13: 0000000000000000 R14: 0000000000000000 R15: ffff9cf5c39ae0e0
        [ 1411.304501] FS:  00007f00ae754780(0000) GS:ffff9cf5fba00000(0000) knlGS:0000000000000000
        [ 1411.304504] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
        [ 1411.304506] CR2: 000000003e34343c CR3: 0000000103a98005 CR4: 0000000000370ee0
        [ 1411.304508] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
        [ 1411.304510] DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400
        [ 1411.304512] Call Trace:
        [ 1411.304517]  stack_map_get_build_id_offset+0x17c/0x260
        [ 1411.304528]  __bpf_get_stack+0x18f/0x230
        [ 1411.304541]  bpf_get_stack_raw_tp+0x5a/0x70
        [ 1411.305752] RAX: 0000000000000000 RBX: 5541f689495641d7 RCX: 0000000000000000
        [ 1411.305756] RDX: 0000000000000001 RSI: ffffffff9776b144 RDI: ffffffff977e1b0e
        [ 1411.305758] RBP: ffff9cf5c02b2f40 R08: ffff9cf5ca7606c0 R09: ffffcbd43ee02c04
        [ 1411.306978]  bpf_prog_32007c34f7726d29_bpf_prog1+0xaf/0xd9c
        [ 1411.307861] R10: 0000000000000001 R11: 0000000000000044 R12: ffff9cf5c2ef60e0
        [ 1411.307865] R13: 0000000000000005 R14: 0000000000000000 R15: ffff9cf5c2ef6108
        [ 1411.309074]  bpf_trace_run2+0x8f/0x1a0
        [ 1411.309891] FS:  00007ff485141700(0000) GS:ffff9cf5fae00000(0000) knlGS:0000000000000000
        [ 1411.309896] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
        [ 1411.311221]  syscall_trace_enter.isra.20+0x161/0x1f0
        [ 1411.311600] CR2: 00007ff48514d90e CR3: 0000000107114001 CR4: 0000000000370ef0
        [ 1411.312291]  do_syscall_64+0x15/0x80
        [ 1411.312941] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
        [ 1411.313803]  entry_SYSCALL_64_after_hwframe+0x44/0xae
        [ 1411.314223] DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400
        [ 1411.315082] RIP: 0033:0x7f00ad80a0e0
        [ 1411.315626] Call Trace:
        [ 1411.315632]  stack_map_get_build_id_offset+0x17c/0x260
      
      To reproduce, first build `test_progs` binary:
      
        make -C tools/testing/selftests/bpf -j60
      
      and then run the binary at tools/testing/selftests/bpf directory:
      
        ./test_progs -t get_stack_raw_tp
      
      The warning is due to commit 5b78ed24 ("mm/pagemap: add mmap_assert_locked()
      annotations to find_vma*()") which added mmap_assert_locked() in find_vma()
      function. The mmap_assert_locked() function asserts that mm->mmap_lock needs
      to be held. But this is not the case for bpf_get_stack() or bpf_get_stackid()
      helper (kernel/bpf/stackmap.c), which uses mmap_read_trylock_non_owner()
      instead. Since mm->mmap_lock is not held in bpf_get_stack[id]() use case,
      the above warning is emitted during test run.
      
      This patch fixed the issue by (1). using mmap_read_trylock() instead of
      mmap_read_trylock_non_owner() to satisfy lockdep checking in find_vma(), and
      (2). droping lockdep for mmap_lock right before the irq_work_queue(). The
      function mmap_read_trylock_non_owner() is also removed since after this
      patch nobody calls it any more.
      
      Fixes: 5b78ed24 ("mm/pagemap: add mmap_assert_locked() annotations to find_vma*()")
      Suggested-by: NJason Gunthorpe <jgg@ziepe.ca>
      Signed-off-by: NYonghong Song <yhs@fb.com>
      Signed-off-by: NDaniel Borkmann <daniel@iogearbox.net>
      Reviewed-by: NLiam R. Howlett <Liam.Howlett@oracle.com>
      Cc: Luigi Rizzo <lrizzo@google.com>
      Cc: Jason Gunthorpe <jgg@ziepe.ca>
      Cc: linux-mm@kvack.org
      Link: https://lore.kernel.org/bpf/20210909155000.1610299-1-yhs@fb.com
      2f1aaf3e
  8. 10 9月, 2021 1 次提交
  9. 09 9月, 2021 24 次提交
    • E
      net/af_unix: fix a data-race in unix_dgram_poll · 04f08eb4
      Eric Dumazet 提交于
      syzbot reported another data-race in af_unix [1]
      
      Lets change __skb_insert() to use WRITE_ONCE() when changing
      skb head qlen.
      
      Also, change unix_dgram_poll() to use lockless version
      of unix_recvq_full()
      
      It is verry possible we can switch all/most unix_recvq_full()
      to the lockless version, this will be done in a future kernel version.
      
      [1] HEAD commit: 8596e589
      
      BUG: KCSAN: data-race in skb_queue_tail / unix_dgram_poll
      
      write to 0xffff88814eeb24e0 of 4 bytes by task 25815 on cpu 0:
       __skb_insert include/linux/skbuff.h:1938 [inline]
       __skb_queue_before include/linux/skbuff.h:2043 [inline]
       __skb_queue_tail include/linux/skbuff.h:2076 [inline]
       skb_queue_tail+0x80/0xa0 net/core/skbuff.c:3264
       unix_dgram_sendmsg+0xff2/0x1600 net/unix/af_unix.c:1850
       sock_sendmsg_nosec net/socket.c:703 [inline]
       sock_sendmsg net/socket.c:723 [inline]
       ____sys_sendmsg+0x360/0x4d0 net/socket.c:2392
       ___sys_sendmsg net/socket.c:2446 [inline]
       __sys_sendmmsg+0x315/0x4b0 net/socket.c:2532
       __do_sys_sendmmsg net/socket.c:2561 [inline]
       __se_sys_sendmmsg net/socket.c:2558 [inline]
       __x64_sys_sendmmsg+0x53/0x60 net/socket.c:2558
       do_syscall_x64 arch/x86/entry/common.c:50 [inline]
       do_syscall_64+0x3d/0x90 arch/x86/entry/common.c:80
       entry_SYSCALL_64_after_hwframe+0x44/0xae
      
      read to 0xffff88814eeb24e0 of 4 bytes by task 25834 on cpu 1:
       skb_queue_len include/linux/skbuff.h:1869 [inline]
       unix_recvq_full net/unix/af_unix.c:194 [inline]
       unix_dgram_poll+0x2bc/0x3e0 net/unix/af_unix.c:2777
       sock_poll+0x23e/0x260 net/socket.c:1288
       vfs_poll include/linux/poll.h:90 [inline]
       ep_item_poll fs/eventpoll.c:846 [inline]
       ep_send_events fs/eventpoll.c:1683 [inline]
       ep_poll fs/eventpoll.c:1798 [inline]
       do_epoll_wait+0x6ad/0xf00 fs/eventpoll.c:2226
       __do_sys_epoll_wait fs/eventpoll.c:2238 [inline]
       __se_sys_epoll_wait fs/eventpoll.c:2233 [inline]
       __x64_sys_epoll_wait+0xf6/0x120 fs/eventpoll.c:2233
       do_syscall_x64 arch/x86/entry/common.c:50 [inline]
       do_syscall_64+0x3d/0x90 arch/x86/entry/common.c:80
       entry_SYSCALL_64_after_hwframe+0x44/0xae
      
      value changed: 0x0000001b -> 0x00000001
      
      Reported by Kernel Concurrency Sanitizer on:
      CPU: 1 PID: 25834 Comm: syz-executor.1 Tainted: G        W         5.14.0-syzkaller #0
      Hardware name: Google Google Compute Engine/Google Compute Engine, BIOS Google 01/01/2011
      
      Fixes: 86b18aaa ("skbuff: fix a data race in skb_queue_len()")
      Cc: Qian Cai <cai@lca.pw>
      Signed-off-by: NEric Dumazet <edumazet@google.com>
      Signed-off-by: NDavid S. Miller <davem@davemloft.net>
      04f08eb4
    • L
      mmap_lock: change trace and locking order · 10994316
      Liam Howlett 提交于
      Print to the trace log before releasing the lock to avoid racing with
      other trace log printers of the same lock type.
      
      Link: https://lkml.kernel.org/r/20210903022041.1843024-1-Liam.Howlett@oracle.comSigned-off-by: NLiam R. Howlett <Liam.Howlett@oracle.com>
      Suggested-by: NSteven Rostedt (VMware) <rostedt@goodmis.org>
      Reviewed-by: NMatthew Wilcox (Oracle) <willy@infradead.org>
      Cc: Michel Lespinasse <walken.cr@gmail.com>
      Cc: Vlastimil Babka <vbabka@suse.cz>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      10994316
    • L
      mm/hugetlb: initialize hugetlb_usage in mm_init · 13db8c50
      Liu Zixian 提交于
      After fork, the child process will get incorrect (2x) hugetlb_usage.  If
      a process uses 5 2MB hugetlb pages in an anonymous mapping,
      
      	HugetlbPages:	   10240 kB
      
      and then forks, the child will show,
      
      	HugetlbPages:	   20480 kB
      
      The reason for double the amount is because hugetlb_usage will be copied
      from the parent and then increased when we copy page tables from parent
      to child.  Child will have 2x actual usage.
      
      Fix this by adding hugetlb_count_init in mm_init.
      
      Link: https://lkml.kernel.org/r/20210826071742.877-1-liuzixian4@huawei.com
      Fixes: 5d317b2b ("mm: hugetlb: proc: add HugetlbPages field to /proc/PID/status")
      Signed-off-by: NLiu Zixian <liuzixian4@huawei.com>
      Reviewed-by: NNaoya Horiguchi <naoya.horiguchi@nec.com>
      Reviewed-by: NMike Kravetz <mike.kravetz@oracle.com>
      Cc: <stable@vger.kernel.org>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      13db8c50
    • N
      compiler_attributes.h: move __compiletime_{error|warning} · b83a9084
      Nick Desaulniers 提交于
      Clang 14 will add support for __attribute__((__error__(""))) and
      __attribute__((__warning__(""))). To make use of these in
      __compiletime_error and __compiletime_warning (as used by BUILD_BUG and
      friends) for newer clang and detect/fallback for older versions of
      clang, move these to compiler_attributes.h and guard them with
      __has_attribute preprocessor guards.
      
      Link: https://reviews.llvm.org/D106030
      Link: https://bugs.llvm.org/show_bug.cgi?id=16428
      Link: https://github.com/ClangBuiltLinux/linux/issues/1173Signed-off-by: NNick Desaulniers <ndesaulniers@google.com>
      Reviewed-by: NNathan Chancellor <nathan@kernel.org>
      Reviewed-by: NKees Cook <keescook@chromium.org>
      [Reworded, landed in Clang 14]
      Signed-off-by: NMiguel Ojeda <ojeda@kernel.org>
      b83a9084
    • A
      arch: remove compat_alloc_user_space · a7a08b27
      Arnd Bergmann 提交于
      All users of compat_alloc_user_space() and copy_in_user() have been
      removed from the kernel, only a few functions in sparc remain that can be
      changed to calling arch_copy_in_user() instead.
      
      Link: https://lkml.kernel.org/r/20210727144859.4150043-7-arnd@kernel.orgSigned-off-by: NArnd Bergmann <arnd@arndb.de>
      Reviewed-by: NChristoph Hellwig <hch@lst.de>
      Cc: Al Viro <viro@zeniv.linux.org.uk>
      Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
      Cc: Borislav Petkov <bp@alien8.de>
      Cc: Catalin Marinas <catalin.marinas@arm.com>
      Cc: Christian Borntraeger <borntraeger@de.ibm.com>
      Cc: Christoph Hellwig <hch@infradead.org>
      Cc: "David S. Miller" <davem@davemloft.net>
      Cc: Eric Biederman <ebiederm@xmission.com>
      Cc: Feng Tang <feng.tang@intel.com>
      Cc: Heiko Carstens <hca@linux.ibm.com>
      Cc: Helge Deller <deller@gmx.de>
      Cc: "H. Peter Anvin" <hpa@zytor.com>
      Cc: Ingo Molnar <mingo@redhat.com>
      Cc: "James E.J. Bottomley" <James.Bottomley@HansenPartnership.com>
      Cc: Michael Ellerman <mpe@ellerman.id.au>
      Cc: Paul Mackerras <paulus@samba.org>
      Cc: Thomas Bogendoerfer <tsbogend@alpha.franken.de>
      Cc: Thomas Gleixner <tglx@linutronix.de>
      Cc: Vasily Gorbik <gor@linux.ibm.com>
      Cc: Will Deacon <will@kernel.org>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      a7a08b27
    • A
      compat: remove some compat entry points · 59ab844e
      Arnd Bergmann 提交于
      These are all handled correctly when calling the native system call entry
      point, so remove the special cases.
      
      Link: https://lkml.kernel.org/r/20210727144859.4150043-6-arnd@kernel.orgSigned-off-by: NArnd Bergmann <arnd@arndb.de>
      Reviewed-by: NChristoph Hellwig <hch@lst.de>
      Cc: Al Viro <viro@zeniv.linux.org.uk>
      Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
      Cc: Borislav Petkov <bp@alien8.de>
      Cc: Catalin Marinas <catalin.marinas@arm.com>
      Cc: Christian Borntraeger <borntraeger@de.ibm.com>
      Cc: Christoph Hellwig <hch@infradead.org>
      Cc: "David S. Miller" <davem@davemloft.net>
      Cc: Eric Biederman <ebiederm@xmission.com>
      Cc: Feng Tang <feng.tang@intel.com>
      Cc: Heiko Carstens <hca@linux.ibm.com>
      Cc: Helge Deller <deller@gmx.de>
      Cc: "H. Peter Anvin" <hpa@zytor.com>
      Cc: Ingo Molnar <mingo@redhat.com>
      Cc: "James E.J. Bottomley" <James.Bottomley@HansenPartnership.com>
      Cc: Michael Ellerman <mpe@ellerman.id.au>
      Cc: Paul Mackerras <paulus@samba.org>
      Cc: Thomas Bogendoerfer <tsbogend@alpha.franken.de>
      Cc: Thomas Gleixner <tglx@linutronix.de>
      Cc: Vasily Gorbik <gor@linux.ibm.com>
      Cc: Will Deacon <will@kernel.org>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      59ab844e
    • A
      mm: simplify compat numa syscalls · e130242d
      Arnd Bergmann 提交于
      The compat implementations for mbind, get_mempolicy, set_mempolicy and
      migrate_pages are just there to handle the subtly different layout of
      bitmaps on 32-bit hosts.
      
      The compat implementation however lacks some of the checks that are
      present in the native one, in particular for checking that the extra bits
      are all zero when user space has a larger mask size than the kernel.
      Worse, those extra bits do not get cleared when copying in or out of the
      kernel, which can lead to incorrect data as well.
      
      Unify the implementation to handle the compat bitmap layout directly in
      the get_nodes() and copy_nodes_to_user() helpers.  Splitting out the
      get_bitmap() helper from get_nodes() also helps readability of the native
      case.
      
      On x86, two additional problems are addressed by this: compat tasks can
      pass a bitmap at the end of a mapping, causing a fault when reading across
      the page boundary for a 64-bit word.  x32 tasks might also run into
      problems with get_mempolicy corrupting data when an odd number of 32-bit
      words gets passed.
      
      On parisc the migrate_pages() system call apparently had the wrong calling
      convention, as big-endian architectures expect the words inside of a
      bitmap to be swapped.  This is not a problem though since parisc has no
      NUMA support.
      
      [arnd@arndb.de: fix mempolicy crash]
        Link: https://lkml.kernel.org/r/20210730143417.3700653-1-arnd@kernel.org
        Link: https://lore.kernel.org/lkml/YQPLG20V3dmOfq3a@osiris/
      
      Link: https://lkml.kernel.org/r/20210727144859.4150043-5-arnd@kernel.orgSigned-off-by: NArnd Bergmann <arnd@arndb.de>
      Reviewed-by: NChristoph Hellwig <hch@lst.de>
      Cc: Al Viro <viro@zeniv.linux.org.uk>
      Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
      Cc: Borislav Petkov <bp@alien8.de>
      Cc: Catalin Marinas <catalin.marinas@arm.com>
      Cc: Christian Borntraeger <borntraeger@de.ibm.com>
      Cc: Christoph Hellwig <hch@infradead.org>
      Cc: "David S. Miller" <davem@davemloft.net>
      Cc: Eric Biederman <ebiederm@xmission.com>
      Cc: Feng Tang <feng.tang@intel.com>
      Cc: Heiko Carstens <hca@linux.ibm.com>
      Cc: Helge Deller <deller@gmx.de>
      Cc: "H. Peter Anvin" <hpa@zytor.com>
      Cc: Ingo Molnar <mingo@redhat.com>
      Cc: "James E.J. Bottomley" <James.Bottomley@HansenPartnership.com>
      Cc: Michael Ellerman <mpe@ellerman.id.au>
      Cc: Paul Mackerras <paulus@samba.org>
      Cc: Thomas Bogendoerfer <tsbogend@alpha.franken.de>
      Cc: Thomas Gleixner <tglx@linutronix.de>
      Cc: Vasily Gorbik <gor@linux.ibm.com>
      Cc: Will Deacon <will@kernel.org>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      e130242d
    • T
      pid: cleanup the stale comment mentioning pidmap_init(). · 5b91a75b
      Takahiro Itazuri 提交于
      pidmap_init() has already been replaced with pid_idr_init() in the commit
      95846ecf ("pid: replace pid bitmap implementation with IDR API").
      Cleanup the stale comment which still mentions it.
      
      Link: https://lkml.kernel.org/r/20210714120713.19825-1-itazur@amazon.comSigned-off-by: NTakahiro Itazuri <itazur@amazon.com>
      Cc: Kuniyuki Iwashima <kuniyu@amazon.co.jp>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      5b91a75b
    • N
      fs/epoll: use a per-cpu counter for user's watches count · 1e1c1583
      Nicholas Piggin 提交于
      This counter tracks the number of watches a user has, to compare against
      the 'max_user_watches' limit. This causes a scalability bottleneck on
      SPECjbb2015 on large systems as there is only one user. Changing to a
      per-cpu counter increases throughput of the benchmark by about 30% on a
      16-socket, > 1000 thread system.
      
      [rdunlap@infradead.org: fix build errors in kernel/user.c when CONFIG_EPOLL=n]
      [npiggin@gmail.com: move ifdefs into wrapper functions, slightly improve panic message]
        Link: https://lkml.kernel.org/r/1628051945.fens3r99ox.astroid@bobo.none
      [akpm@linux-foundation.org: tweak user_epoll_alloc(), per Guenter]
        Link: https://lkml.kernel.org/r/20210804191421.GA1900577@roeck-us.net
      
      Link: https://lkml.kernel.org/r/20210802032013.2751916-1-npiggin@gmail.comSigned-off-by: NNicholas Piggin <npiggin@gmail.com>
      Reported-by: NAnton Blanchard <anton@ozlabs.org>
      Cc: Alexander Viro <viro@zeniv.linux.org.uk>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      1e1c1583
    • D
      units: add the HZ macros · e2c77032
      Daniel Lezcano 提交于
      The macros for the unit conversion for frequency are duplicated in
      different places.
      
      Provide these macros in the 'units' header, so they can be reused.
      
      Link: https://lkml.kernel.org/r/20210816114732.1834145-3-daniel.lezcano@linaro.orgSigned-off-by: NDaniel Lezcano <daniel.lezcano@linaro.org>
      Reviewed-by: NChristian Eggers <ceggers@arri.de>
      Reviewed-by: NAndy Shevchenko <andriy.shevchenko@linux.intel.com>
      Cc: Chanwoo Choi <cw00.choi@samsung.com>
      Cc: Guenter Roeck <linux@roeck-us.net>
      Cc: Jonathan Cameron <jic23@kernel.org>
      Cc: Jonathan Cameron <Jonathan.Cameron@huawei.com>
      Cc: Kyungmin Park <kyungmin.park@samsung.com>
      Cc: Lars-Peter Clausen <lars@metafoo.de>
      Cc: Lukasz Luba <lukasz.luba@arm.com>
      Cc: Maxime Coquelin <mcoquelin.stm32@gmail.com>
      Cc: Miquel Raynal <miquel.raynal@bootlin.com>
      Cc: MyungJoo Ham <myungjoo.ham@samsung.com>
      Cc: Peter Meerwald <pmeerw@pmeerw.net>
      Cc: "Rafael J. Wysocki" <rafael@kernel.org>
      Cc: Zhang Rui <rui.zhang@intel.com>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      e2c77032
    • D
      units: change from 'L' to 'UL' · c9221919
      Daniel Lezcano 提交于
      Patch series "Add Hz macros", v3.
      
      There are multiple definitions of the HZ_PER_MHZ or HZ_PER_KHZ in the
      different drivers.  Instead of duplicating this definition again and
      again, add one in the units.h header to be reused in all the place the
      redefiniton occurs.
      
      At the same time, change the type of the Watts, as they can not be
      negative.
      
      This patch (of 10):
      
      The users of the macros are safe to be assigned with an unsigned instead
      of signed as the variables using them are themselves unsigned.
      
      Link: https://lkml.kernel.org/r/20210816114732.1834145-1-daniel.lezcano@linaro.org
      Link: https://lkml.kernel.org/r/20210816114732.1834145-2-daniel.lezcano@linaro.orgSigned-off-by: NDaniel Lezcano <daniel.lezcano@linaro.org>
      Cc: Andy Shevchenko <andriy.shevchenko@linux.intel.com>
      Cc: Jonathan Cameron <jic23@kernel.org>
      Cc: Christian Eggers <ceggers@arri.de>
      Cc: Lukasz Luba <lukasz.luba@arm.com>
      Cc: MyungJoo Ham <myungjoo.ham@samsung.com>
      Cc: Kyungmin Park <kyungmin.park@samsung.com>
      Cc: Lars-Peter Clausen <lars@metafoo.de>
      Cc: Peter Meerwald <pmeerw@pmeerw.net>
      Cc: Zhang Rui <rui.zhang@intel.com>
      Cc: Guenter Roeck <linux@roeck-us.net>
      Cc: Miquel Raynal <miquel.raynal@bootlin.com>
      Cc: Maxime Coquelin <mcoquelin.stm32@gmail.com>
      Cc: "Rafael J. Wysocki" <rafael@kernel.org>
      Cc: Daniel Lezcano <daniel.lezcano@linaro.org>
      Cc: Chanwoo Choi <cw00.choi@samsung.com>
      Cc: Jonathan Cameron <Jonathan.Cameron@huawei.com>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      c9221919
    • A
    • S
      mm/damon: implement a debugfs-based user space interface · 4bc05954
      SeongJae Park 提交于
      DAMON is designed to be used by kernel space code such as the memory
      management subsystems, and therefore it provides only kernel space API.
      That said, letting the user space control DAMON could provide some
      benefits to them.  For example, it will allow user space to analyze their
      specific workloads and make their own special optimizations.
      
      For such cases, this commit implements a simple DAMON application kernel
      module, namely 'damon-dbgfs', which merely wraps the DAMON api and exports
      those to the user space via the debugfs.
      
      'damon-dbgfs' exports three files, ``attrs``, ``target_ids``, and
      ``monitor_on`` under its debugfs directory, ``<debugfs>/damon/``.
      
      Attributes
      ----------
      
      Users can read and write the ``sampling interval``, ``aggregation
      interval``, ``regions update interval``, and min/max number of monitoring
      target regions by reading from and writing to the ``attrs`` file.  For
      example, below commands set those values to 5 ms, 100 ms, 1,000 ms, 10,
      1000 and check it again::
      
          # cd <debugfs>/damon
          # echo 5000 100000 1000000 10 1000 > attrs
          # cat attrs
          5000 100000 1000000 10 1000
      
      Target IDs
      ----------
      
      Some types of address spaces supports multiple monitoring target.  For
      example, the virtual memory address spaces monitoring can have multiple
      processes as the monitoring targets.  Users can set the targets by writing
      relevant id values of the targets to, and get the ids of the current
      targets by reading from the ``target_ids`` file.  In case of the virtual
      address spaces monitoring, the values should be pids of the monitoring
      target processes.  For example, below commands set processes having pids
      42 and 4242 as the monitoring targets and check it again::
      
          # cd <debugfs>/damon
          # echo 42 4242 > target_ids
          # cat target_ids
          42 4242
      
      Note that setting the target ids doesn't start the monitoring.
      
      Turning On/Off
      --------------
      
      Setting the files as described above doesn't incur effect unless you
      explicitly start the monitoring.  You can start, stop, and check the
      current status of the monitoring by writing to and reading from the
      ``monitor_on`` file.  Writing ``on`` to the file starts the monitoring of
      the targets with the attributes.  Writing ``off`` to the file stops those.
      DAMON also stops if every targets are invalidated (in case of the virtual
      memory monitoring, target processes are invalidated when terminated).
      Below example commands turn on, off, and check the status of DAMON::
      
          # cd <debugfs>/damon
          # echo on > monitor_on
          # echo off > monitor_on
          # cat monitor_on
          off
      
      Please note that you cannot write to the above-mentioned debugfs files
      while the monitoring is turned on.  If you write to the files while DAMON
      is running, an error code such as ``-EBUSY`` will be returned.
      
      [akpm@linux-foundation.org: remove unneeded "alloc failed" printks]
      [akpm@linux-foundation.org: replace macro with static inline]
      
      Link: https://lkml.kernel.org/r/20210716081449.22187-8-sj38.park@gmail.comSigned-off-by: NSeongJae Park <sjpark@amazon.de>
      Reviewed-by: NLeonard Foerster <foersleo@amazon.de>
      Reviewed-by: NFernand Sieber <sieberf@amazon.com>
      Cc: Alexander Shishkin <alexander.shishkin@linux.intel.com>
      Cc: Amit Shah <amit@kernel.org>
      Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
      Cc: Brendan Higgins <brendanhiggins@google.com>
      Cc: David Hildenbrand <david@redhat.com>
      Cc: David Rientjes <rientjes@google.com>
      Cc: David Woodhouse <dwmw@amazon.com>
      Cc: Fan Du <fan.du@intel.com>
      Cc: Greg Kroah-Hartman <greg@kroah.com>
      Cc: Greg Thelen <gthelen@google.com>
      Cc: Ingo Molnar <mingo@redhat.com>
      Cc: Joe Perches <joe@perches.com>
      Cc: Jonathan Cameron <Jonathan.Cameron@huawei.com>
      Cc: Jonathan Corbet <corbet@lwn.net>
      Cc: Marco Elver <elver@google.com>
      Cc: Markus Boehme <markubo@amazon.de>
      Cc: Maximilian Heyne <mheyne@amazon.de>
      Cc: Mel Gorman <mgorman@suse.de>
      Cc: Minchan Kim <minchan@kernel.org>
      Cc: Namhyung Kim <namhyung@kernel.org>
      Cc: Peter Zijlstra <peterz@infradead.org>
      Cc: Rik van Riel <riel@surriel.com>
      Cc: Shakeel Butt <shakeelb@google.com>
      Cc: Shuah Khan <shuah@kernel.org>
      Cc: Steven Rostedt (VMware) <rostedt@goodmis.org>
      Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
      Cc: Vlastimil Babka <vbabka@suse.cz>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      4bc05954
    • S
      mm/damon: implement primitives for the virtual memory address spaces · 3f49584b
      SeongJae Park 提交于
      This commit introduces a reference implementation of the address space
      specific low level primitives for the virtual address space, so that users
      of DAMON can easily monitor the data accesses on virtual address spaces of
      specific processes by simply configuring the implementation to be used by
      DAMON.
      
      The low level primitives for the fundamental access monitoring are defined
      in two parts:
      
      1. Identification of the monitoring target address range for the address
         space.
      2. Access check of specific address range in the target space.
      
      The reference implementation for the virtual address space does the works
      as below.
      
      PTE Accessed-bit Based Access Check
      -----------------------------------
      
      The implementation uses PTE Accessed-bit for basic access checks.  That
      is, it clears the bit for the next sampling target page and checks whether
      it is set again after one sampling period.  This could disturb the reclaim
      logic.  DAMON uses ``PG_idle`` and ``PG_young`` page flags to solve the
      conflict, as Idle page tracking does.
      
      VMA-based Target Address Range Construction
      -------------------------------------------
      
      Only small parts in the super-huge virtual address space of the processes
      are mapped to physical memory and accessed.  Thus, tracking the unmapped
      address regions is just wasteful.  However, because DAMON can deal with
      some level of noise using the adaptive regions adjustment mechanism,
      tracking every mapping is not strictly required but could even incur a
      high overhead in some cases.  That said, too huge unmapped areas inside
      the monitoring target should be removed to not take the time for the
      adaptive mechanism.
      
      For the reason, this implementation converts the complex mappings to three
      distinct regions that cover every mapped area of the address space.  Also,
      the two gaps between the three regions are the two biggest unmapped areas
      in the given address space.  The two biggest unmapped areas would be the
      gap between the heap and the uppermost mmap()-ed region, and the gap
      between the lowermost mmap()-ed region and the stack in most of the cases.
      Because these gaps are exceptionally huge in usual address spaces,
      excluding these will be sufficient to make a reasonable trade-off.  Below
      shows this in detail::
      
          <heap>
          <BIG UNMAPPED REGION 1>
          <uppermost mmap()-ed region>
          (small mmap()-ed regions and munmap()-ed regions)
          <lowermost mmap()-ed region>
          <BIG UNMAPPED REGION 2>
          <stack>
      
      [akpm@linux-foundation.org: mm/damon/vaddr.c needs highmem.h for kunmap_atomic()]
      [sjpark@amazon.de: remove unnecessary PAGE_EXTENSION setup]
        Link: https://lkml.kernel.org/r/20210806095153.6444-2-sj38.park@gmail.com
      [sjpark@amazon.de: safely walk page table]
        Link: https://lkml.kernel.org/r/20210831161800.29419-1-sj38.park@gmail.com
      
      Link: https://lkml.kernel.org/r/20210716081449.22187-6-sj38.park@gmail.comSigned-off-by: NSeongJae Park <sjpark@amazon.de>
      Reviewed-by: NLeonard Foerster <foersleo@amazon.de>
      Reviewed-by: NFernand Sieber <sieberf@amazon.com>
      Acked-by: NShakeel Butt <shakeelb@google.com>
      Cc: Alexander Shishkin <alexander.shishkin@linux.intel.com>
      Cc: Amit Shah <amit@kernel.org>
      Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
      Cc: Brendan Higgins <brendanhiggins@google.com>
      Cc: David Hildenbrand <david@redhat.com>
      Cc: David Rientjes <rientjes@google.com>
      Cc: David Woodhouse <dwmw@amazon.com>
      Cc: Fan Du <fan.du@intel.com>
      Cc: Greg Kroah-Hartman <greg@kroah.com>
      Cc: Greg Thelen <gthelen@google.com>
      Cc: Ingo Molnar <mingo@redhat.com>
      Cc: Joe Perches <joe@perches.com>
      Cc: Jonathan Cameron <Jonathan.Cameron@huawei.com>
      Cc: Jonathan Corbet <corbet@lwn.net>
      Cc: Marco Elver <elver@google.com>
      Cc: Markus Boehme <markubo@amazon.de>
      Cc: Maximilian Heyne <mheyne@amazon.de>
      Cc: Mel Gorman <mgorman@suse.de>
      Cc: Minchan Kim <minchan@kernel.org>
      Cc: Namhyung Kim <namhyung@kernel.org>
      Cc: Peter Zijlstra <peterz@infradead.org>
      Cc: Rik van Riel <riel@surriel.com>
      Cc: Shuah Khan <shuah@kernel.org>
      Cc: Steven Rostedt (VMware) <rostedt@goodmis.org>
      Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
      Cc: Vlastimil Babka <vbabka@suse.cz>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      3f49584b
    • S
      mm/idle_page_tracking: make PG_idle reusable · 1c676e0d
      SeongJae Park 提交于
      PG_idle and PG_young allow the two PTE Accessed bit users, Idle Page
      Tracking and the reclaim logic concurrently work while not interfering
      with each other.  That is, when they need to clear the Accessed bit, they
      set PG_young to represent the previous state of the bit, respectively.
      And when they need to read the bit, if the bit is cleared, they further
      read the PG_young to know whether the other has cleared the bit meanwhile
      or not.
      
      For yet another user of the PTE Accessed bit, we could add another page
      flag, or extend the mechanism to use the flags.  For the DAMON usecase,
      however, we don't need to do that just yet.  IDLE_PAGE_TRACKING and DAMON
      are mutually exclusive, so there's only ever going to be one user of the
      current set of flags.
      
      In this commit, we split out the CONFIG options to allow for the use of
      PG_young and PG_idle outside of idle page tracking.
      
      In the next commit, DAMON's reference implementation of the virtual memory
      address space monitoring primitives will use it.
      
      [sjpark@amazon.de: set PAGE_EXTENSION for non-64BIT]
        Link: https://lkml.kernel.org/r/20210806095153.6444-1-sj38.park@gmail.com
      [akpm@linux-foundation.org: tweak Kconfig text]
      [sjpark@amazon.de: hide PAGE_IDLE_FLAG from users]
        Link: https://lkml.kernel.org/r/20210813081238.34705-1-sj38.park@gmail.com
      
      Link: https://lkml.kernel.org/r/20210716081449.22187-5-sj38.park@gmail.comSigned-off-by: NSeongJae Park <sjpark@amazon.de>
      Reviewed-by: NShakeel Butt <shakeelb@google.com>
      Reviewed-by: NFernand Sieber <sieberf@amazon.com>
      Cc: Alexander Shishkin <alexander.shishkin@linux.intel.com>
      Cc: Amit Shah <amit@kernel.org>
      Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
      Cc: Brendan Higgins <brendanhiggins@google.com>
      Cc: David Hildenbrand <david@redhat.com>
      Cc: David Rientjes <rientjes@google.com>
      Cc: David Woodhouse <dwmw@amazon.com>
      Cc: Fan Du <fan.du@intel.com>
      Cc: Greg Kroah-Hartman <greg@kroah.com>
      Cc: Greg Thelen <gthelen@google.com>
      Cc: Ingo Molnar <mingo@redhat.com>
      Cc: Joe Perches <joe@perches.com>
      Cc: Jonathan Cameron <Jonathan.Cameron@huawei.com>
      Cc: Jonathan Corbet <corbet@lwn.net>
      Cc: Leonard Foerster <foersleo@amazon.de>
      Cc: Marco Elver <elver@google.com>
      Cc: Markus Boehme <markubo@amazon.de>
      Cc: Maximilian Heyne <mheyne@amazon.de>
      Cc: Mel Gorman <mgorman@suse.de>
      Cc: Minchan Kim <minchan@kernel.org>
      Cc: Namhyung Kim <namhyung@kernel.org>
      Cc: Peter Zijlstra <peterz@infradead.org>
      Cc: Rik van Riel <riel@surriel.com>
      Cc: Shuah Khan <shuah@kernel.org>
      Cc: Steven Rostedt (VMware) <rostedt@goodmis.org>
      Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
      Cc: Vlastimil Babka <vbabka@suse.cz>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      1c676e0d
    • S
      mm/damon: adaptively adjust regions · b9a6ac4e
      SeongJae Park 提交于
      Even somehow the initial monitoring target regions are well constructed to
      fulfill the assumption (pages in same region have similar access
      frequencies), the data access pattern can be dynamically changed.  This
      will result in low monitoring quality.  To keep the assumption as much as
      possible, DAMON adaptively merges and splits each region based on their
      access frequency.
      
      For each ``aggregation interval``, it compares the access frequencies of
      adjacent regions and merges those if the frequency difference is small.
      Then, after it reports and clears the aggregated access frequency of each
      region, it splits each region into two or three regions if the total
      number of regions will not exceed the user-specified maximum number of
      regions after the split.
      
      In this way, DAMON provides its best-effort quality and minimal overhead
      while keeping the upper-bound overhead that users set.
      
      Link: https://lkml.kernel.org/r/20210716081449.22187-4-sj38.park@gmail.comSigned-off-by: NSeongJae Park <sjpark@amazon.de>
      Reviewed-by: NLeonard Foerster <foersleo@amazon.de>
      Reviewed-by: NFernand Sieber <sieberf@amazon.com>
      Acked-by: NShakeel Butt <shakeelb@google.com>
      Cc: Alexander Shishkin <alexander.shishkin@linux.intel.com>
      Cc: Amit Shah <amit@kernel.org>
      Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
      Cc: Brendan Higgins <brendanhiggins@google.com>
      Cc: David Hildenbrand <david@redhat.com>
      Cc: David Rientjes <rientjes@google.com>
      Cc: David Woodhouse <dwmw@amazon.com>
      Cc: Fan Du <fan.du@intel.com>
      Cc: Greg Kroah-Hartman <greg@kroah.com>
      Cc: Greg Thelen <gthelen@google.com>
      Cc: Ingo Molnar <mingo@redhat.com>
      Cc: Joe Perches <joe@perches.com>
      Cc: Jonathan Cameron <Jonathan.Cameron@huawei.com>
      Cc: Jonathan Corbet <corbet@lwn.net>
      Cc: Marco Elver <elver@google.com>
      Cc: Markus Boehme <markubo@amazon.de>
      Cc: Maximilian Heyne <mheyne@amazon.de>
      Cc: Mel Gorman <mgorman@suse.de>
      Cc: Minchan Kim <minchan@kernel.org>
      Cc: Namhyung Kim <namhyung@kernel.org>
      Cc: Peter Zijlstra <peterz@infradead.org>
      Cc: Rik van Riel <riel@surriel.com>
      Cc: Shuah Khan <shuah@kernel.org>
      Cc: Steven Rostedt (VMware) <rostedt@goodmis.org>
      Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
      Cc: Vlastimil Babka <vbabka@suse.cz>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      b9a6ac4e
    • S
      mm/damon/core: implement region-based sampling · f23b8eee
      SeongJae Park 提交于
      To avoid the unbounded increase of the overhead, DAMON groups adjacent
      pages that are assumed to have the same access frequencies into a
      region.  As long as the assumption (pages in a region have the same
      access frequencies) is kept, only one page in the region is required to
      be checked.  Thus, for each ``sampling interval``,
      
       1. the 'prepare_access_checks' primitive picks one page in each region,
       2. waits for one ``sampling interval``,
       3. checks whether the page is accessed meanwhile, and
       4. increases the access count of the region if so.
      
      Therefore, the monitoring overhead is controllable by adjusting the
      number of regions.  DAMON allows both the underlying primitives and user
      callbacks to adjust regions for the trade-off.  In other words, this
      commit makes DAMON to use not only time-based sampling but also
      space-based sampling.
      
      This scheme, however, cannot preserve the quality of the output if the
      assumption is not guaranteed.  Next commit will address this problem.
      
      Link: https://lkml.kernel.org/r/20210716081449.22187-3-sj38.park@gmail.comSigned-off-by: NSeongJae Park <sjpark@amazon.de>
      Reviewed-by: NLeonard Foerster <foersleo@amazon.de>
      Reviewed-by: NFernand Sieber <sieberf@amazon.com>
      Acked-by: NShakeel Butt <shakeelb@google.com>
      Cc: Alexander Shishkin <alexander.shishkin@linux.intel.com>
      Cc: Amit Shah <amit@kernel.org>
      Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
      Cc: Brendan Higgins <brendanhiggins@google.com>
      Cc: David Hildenbrand <david@redhat.com>
      Cc: David Rientjes <rientjes@google.com>
      Cc: David Woodhouse <dwmw@amazon.com>
      Cc: Fan Du <fan.du@intel.com>
      Cc: Greg Kroah-Hartman <greg@kroah.com>
      Cc: Greg Thelen <gthelen@google.com>
      Cc: Ingo Molnar <mingo@redhat.com>
      Cc: Joe Perches <joe@perches.com>
      Cc: Jonathan Cameron <Jonathan.Cameron@huawei.com>
      Cc: Jonathan Corbet <corbet@lwn.net>
      Cc: Marco Elver <elver@google.com>
      Cc: Markus Boehme <markubo@amazon.de>
      Cc: Maximilian Heyne <mheyne@amazon.de>
      Cc: Mel Gorman <mgorman@suse.de>
      Cc: Minchan Kim <minchan@kernel.org>
      Cc: Namhyung Kim <namhyung@kernel.org>
      Cc: Peter Zijlstra <peterz@infradead.org>
      Cc: Rik van Riel <riel@surriel.com>
      Cc: Shuah Khan <shuah@kernel.org>
      Cc: Steven Rostedt (VMware) <rostedt@goodmis.org>
      Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
      Cc: Vlastimil Babka <vbabka@suse.cz>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      f23b8eee
    • S
      mm: introduce Data Access MONitor (DAMON) · 2224d848
      SeongJae Park 提交于
      Patch series "Introduce Data Access MONitor (DAMON)", v34.
      
      Introduction
      ============
      
      DAMON is a data access monitoring framework for the Linux kernel.  The
      core mechanisms of DAMON called 'region based sampling' and 'adaptive
      regions adjustment' (refer to 'mechanisms.rst' in the 11th patch of this
      patchset for the detail) make it
      
      - accurate (The monitored information is useful for DRAM level memory
        management.  It might not appropriate for Cache-level accuracy,
        though.),
      
      - light-weight (The monitoring overhead is low enough to be applied
        online while making no impact on the performance of the target
        workloads.), and
      
      - scalable (the upper-bound of the instrumentation overhead is
        controllable regardless of the size of target workloads.).
      
      Using this framework, therefore, several memory management mechanisms such
      as reclamation and THP can be optimized to aware real data access
      patterns.  Experimental access pattern aware memory management
      optimization works that incurring high instrumentation overhead will be
      able to have another try.
      
      Though DAMON is for kernel subsystems, it can be easily exposed to the
      user space by writing a DAMON-wrapper kernel subsystem.  Then, user space
      users who have some special workloads will be able to write personalized
      tools or applications for deeper understanding and specialized
      optimizations of their systems.
      
      DAMON is also merged in two public Amazon Linux kernel trees that based on
      v5.4.y[1] and v5.10.y[2].
      
      [1] https://github.com/amazonlinux/linux/tree/amazon-5.4.y/master/mm/damon
      [2] https://github.com/amazonlinux/linux/tree/amazon-5.10.y/master/mm/damon
      
      The userspace tool[1] is available, released under GPLv2, and actively
      being maintained.  I am also planning to implement another basic user
      interface in perf[2].  Also, the basic test suite for DAMON is available
      under GPLv2[3].
      
      [1] https://github.com/awslabs/damo
      [2] https://lore.kernel.org/linux-mm/20210107120729.22328-1-sjpark@amazon.com/
      [3] https://github.com/awslabs/damon-tests
      
      Long-term Plan
      --------------
      
      DAMON is a part of a project called Data Access-aware Operating System
      (DAOS).  As the name implies, I want to improve the performance and
      efficiency of systems using fine-grained data access patterns.  The
      optimizations are for both kernel and user spaces.  I will therefore
      modify or create kernel subsystems, export some of those to user space and
      implement user space library / tools.  Below shows the layers and
      components for the project.
      
          ---------------------------------------------------------------------------
          Primitives:     PTE Accessed bit, PG_idle, rmap, (Intel CMT), ...
          Framework:      DAMON
          Features:       DAMOS, virtual addr, physical addr, ...
          Applications:   DAMON-debugfs, (DARC), ...
          ^^^^^^^^^^^^^^^^^^^^^^^    KERNEL SPACE    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
      
          Raw Interface:  debugfs, (sysfs), (damonfs), tracepoints, (sys_damon), ...
      
          vvvvvvvvvvvvvvvvvvvvvvv    USER SPACE      vvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvv
          Library:        (libdamon), ...
          Tools:          DAMO, (perf), ...
          ---------------------------------------------------------------------------
      
      The components in parentheses or marked as '...' are not implemented yet
      but in the future plan.  IOW, those are the TODO tasks of DAOS project.
      For more detail, please refer to the plans:
      https://lore.kernel.org/linux-mm/20201202082731.24828-1-sjpark@amazon.com/
      
      Evaluations
      ===========
      
      We evaluated DAMON's overhead, monitoring quality and usefulness using 24
      realistic workloads on my QEMU/KVM based virtual machine running a kernel
      that v24 DAMON patchset is applied.
      
      DAMON is lightweight.  It increases system memory usage by 0.39% and slows
      target workloads down by 1.16%.
      
      DAMON is accurate and useful for memory management optimizations.  An
      experimental DAMON-based operation scheme for THP, namely 'ethp', removes
      76.15% of THP memory overheads while preserving 51.25% of THP speedup.
      Another experimental DAMON-based 'proactive reclamation' implementation,
      'prcl', reduces 93.38% of residential sets and 23.63% of system memory
      footprint while incurring only 1.22% runtime overhead in the best case
      (parsec3/freqmine).
      
      NOTE that the experimental THP optimization and proactive reclamation are
      not for production but only for proof of concepts.
      
      Please refer to the official document[1] or "Documentation/admin-guide/mm:
      Add a document for DAMON" patch in this patchset for detailed evaluation
      setup and results.
      
      [1] https://damonitor.github.io/doc/html/latest-damon/admin-guide/mm/damon/eval.html
      
      Real-world User Story
      =====================
      
      In summary, DAMON has used on production systems and proved its usefulness.
      
      DAMON as a profiler
      -------------------
      
      We analyzed characteristics of a large scale production systems of our
      customers using DAMON.  The systems utilize 70GB DRAM and 36 CPUs.  From
      this, we were able to find interesting things below.
      
      There were obviously different access pattern under idle workload and
      active workload.  Under the idle workload, it accessed large memory
      regions with low frequency, while the active workload accessed small
      memory regions with high freuqnecy.
      
      DAMON found a 7GB memory region that showing obviously high access
      frequency under the active workload.  We believe this is the
      performance-effective working set and need to be protected.
      
      There was a 4KB memory region that showing highest access frequency under
      not only active but also idle workloads.  We think this must be a hottest
      code section like thing that should never be paged out.
      
      For this analysis, DAMON used only 0.3-1% of single CPU time.  Because we
      used recording-based analysis, it consumed about 3-12 MB of disk space per
      20 minutes.  This is only small amount of disk space, but we can further
      reduce the disk usage by using non-recording-based DAMON features.  I'd
      like to argue that only DAMON can do such detailed analysis (finding 4KB
      highest region in 70GB memory) with the light overhead.
      
      DAMON as a system optimization tool
      -----------------------------------
      
      We also found below potential performance problems on the systems and made
      DAMON-based solutions.
      
      The system doesn't want to make the workload suffer from the page
      reclamation and thus it utilizes enough DRAM but no swap device.  However,
      we found the system is actively reclaiming file-backed pages, because the
      system has intensive file IO.  The file IO turned out to be not
      performance critical for the workload, but the customer wanted to ensure
      performance critical file-backed pages like code section to not mistakenly
      be evicted.
      
      Using direct IO should or `mlock()` would be a straightforward solution,
      but modifying the user space code is not easy for the customer.
      Alternatively, we could use DAMON-based operation scheme[1].  By using it,
      we can ask DAMON to track access frequency of each region and make
      'process_madvise(MADV_WILLNEED)[2]' call for regions having specific size
      and access frequency for a time interval.
      
      We also found the system is having high number of TLB misses.  We tried
      'always' THP enabled policy and it greatly reduced TLB misses, but the
      page reclamation also been more frequent due to the THP internal
      fragmentation caused memory bloat.  We could try another DAMON-based
      operation scheme that applies 'MADV_HUGEPAGE' to memory regions having
      >=2MB size and high access frequency, while applying 'MADV_NOHUGEPAGE' to
      regions having <2MB size and low access frequency.
      
      We do not own the systems so we only reported the analysis results and
      possible optimization solutions to the customers.  The customers satisfied
      about the analysis results and promised to try the optimization guides.
      
      [1] https://lore.kernel.org/linux-mm/20201006123931.5847-1-sjpark@amazon.com/
      [2] https://lore.kernel.org/linux-api/20200622192900.22757-4-minchan@kernel.org/
      
      Comparison with Idle Page Tracking
      ==================================
      
      Idle Page Tracking allows users to set and read idleness of pages using a
      bitmap file which represents each page with each bit of the file.  One
      recommended usage of it is working set size detection.  Users can do that
      by
      
          1. find PFN of each page for workloads in interest,
          2. set all the pages as idle by doing writes to the bitmap file,
          3. wait until the workload accesses its working set, and
          4. read the idleness of the pages again and count pages became not idle.
      
      NOTE: While Idle Page Tracking is for user space users, DAMON is primarily
      designed for kernel subsystems though it can easily exposed to the user
      space.  Hence, this section only assumes such user space use of DAMON.
      
      For what use cases Idle Page Tracking would be better?
      ------------------------------------------------------
      
      1. Flexible usecases other than hotness monitoring.
      
      Because Idle Page Tracking allows users to control the primitive (Page
      idleness) by themselves, Idle Page Tracking users can do anything they
      want.  Meanwhile, DAMON is primarily designed to monitor the hotness of
      each memory region.  For this, DAMON asks users to provide sampling
      interval and aggregation interval.  For the reason, there could be some
      use case that using Idle Page Tracking is simpler.
      
      2. Physical memory monitoring.
      
      Idle Page Tracking receives PFN range as input, so natively supports
      physical memory monitoring.
      
      DAMON is designed to be extensible for multiple address spaces and use
      cases by implementing and using primitives for the given use case.
      Therefore, by theory, DAMON has no limitation in the type of target
      address space as long as primitives for the given address space exists.
      However, the default primitives introduced by this patchset supports only
      virtual address spaces.
      
      Therefore, for physical memory monitoring, you should implement your own
      primitives and use it, or simply use Idle Page Tracking.
      
      Nonetheless, RFC patchsets[1] for the physical memory address space
      primitives is already available.  It also supports user memory same to
      Idle Page Tracking.
      
      [1] https://lore.kernel.org/linux-mm/20200831104730.28970-1-sjpark@amazon.com/
      
      For what use cases DAMON is better?
      -----------------------------------
      
      1. Hotness Monitoring.
      
      Idle Page Tracking let users know only if a page frame is accessed or not.
      For hotness check, the user should write more code and use more memory.
      DAMON do that by itself.
      
      2. Low Monitoring Overhead
      
      DAMON receives user's monitoring request with one step and then provide
      the results.  So, roughly speaking, DAMON require only O(1) user/kernel
      context switches.
      
      In case of Idle Page Tracking, however, because the interface receives
      contiguous page frames, the number of user/kernel context switches
      increases as the monitoring target becomes complex and huge.  As a result,
      the context switch overhead could be not negligible.
      
      Moreover, DAMON is born to handle with the monitoring overhead.  Because
      the core mechanism is pure logical, Idle Page Tracking users might be able
      to implement the mechanism on their own, but it would be time consuming
      and the user/kernel context switching will still more frequent than that
      of DAMON.  Also, the kernel subsystems cannot use the logic in this case.
      
      3. Page granularity working set size detection.
      
      Until v22 of this patchset, this was categorized as the thing Idle Page
      Tracking could do better, because DAMON basically maintains additional
      metadata for each of the monitoring target regions.  So, in the page
      granularity working set size detection use case, DAMON would incur (number
      of monitoring target pages * size of metadata) memory overhead.  Size of
      the single metadata item is about 54 bytes, so assuming 4KB pages, about
      1.3% of monitoring target pages will be additionally used.
      
      All essential metadata for Idle Page Tracking are embedded in 'struct
      page' and page table entries.  Therefore, in this use case, only one
      counter variable for working set size accounting is required if Idle Page
      Tracking is used.
      
      There are more details to consider, but roughly speaking, this is true in
      most cases.
      
      However, the situation changed from v23.  Now DAMON supports arbitrary
      types of monitoring targets, which don't use the metadata.  Using that,
      DAMON can do the working set size detection with no additional space
      overhead but less user-kernel context switch.  A first draft for the
      implementation of monitoring primitives for this usage is available in a
      DAMON development tree[1].  An RFC patchset for it based on this patchset
      will also be available soon.
      
      Since v24, the arbitrary type support is dropped from this patchset
      because this patchset doesn't introduce real use of the type.  You can
      still get it from the DAMON development tree[2], though.
      
      [1] https://github.com/sjp38/linux/tree/damon/pgidle_hack
      [2] https://github.com/sjp38/linux/tree/damon/master
      
      4. More future usecases
      
      While Idle Page Tracking has tight coupling with base primitives (PG_Idle
      and page table Accessed bits), DAMON is designed to be extensible for many
      use cases and address spaces.  If you need some special address type or
      want to use special h/w access check primitives, you can write your own
      primitives for that and configure DAMON to use those.  Therefore, if your
      use case could be changed a lot in future, using DAMON could be better.
      
      Can I use both Idle Page Tracking and DAMON?
      --------------------------------------------
      
      Yes, though using them concurrently for overlapping memory regions could
      result in interference to each other.  Nevertheless, such use case would
      be rare or makes no sense at all.  Even in the case, the noise would bot
      be really significant.  So, you can choose whatever you want depending on
      the characteristics of your use cases.
      
      More Information
      ================
      
      We prepared a showcase web site[1] that you can get more information.
      There are
      
      - the official documentations[2],
      - the heatmap format dynamic access pattern of various realistic workloads for
        heap area[3], mmap()-ed area[4], and stack[5] area,
      - the dynamic working set size distribution[6] and chronological working set
        size changes[7], and
      - the latest performance test results[8].
      
      [1] https://damonitor.github.io/_index
      [2] https://damonitor.github.io/doc/html/latest-damon
      [3] https://damonitor.github.io/test/result/visual/latest/rec.heatmap.0.png.html
      [4] https://damonitor.github.io/test/result/visual/latest/rec.heatmap.1.png.html
      [5] https://damonitor.github.io/test/result/visual/latest/rec.heatmap.2.png.html
      [6] https://damonitor.github.io/test/result/visual/latest/rec.wss_sz.png.html
      [7] https://damonitor.github.io/test/result/visual/latest/rec.wss_time.png.html
      [8] https://damonitor.github.io/test/result/perf/latest/html/index.html
      
      Baseline and Complete Git Trees
      ===============================
      
      The patches are based on the latest -mm tree, specifically
      v5.14-rc1-mmots-2021-07-15-18-47 of https://github.com/hnaz/linux-mm.  You can
      also clone the complete git tree:
      
          $ git clone git://github.com/sjp38/linux -b damon/patches/v34
      
      The web is also available:
      https://github.com/sjp38/linux/releases/tag/damon/patches/v34
      
      Development Trees
      -----------------
      
      There are a couple of trees for entire DAMON patchset series and features
      for future release.
      
      - For latest release: https://github.com/sjp38/linux/tree/damon/master
      - For next release: https://github.com/sjp38/linux/tree/damon/next
      
      Long-term Support Trees
      -----------------------
      
      For people who want to test DAMON but using LTS kernels, there are another
      couple of trees based on two latest LTS kernels respectively and
      containing the 'damon/master' backports.
      
      - For v5.4.y: https://github.com/sjp38/linux/tree/damon/for-v5.4.y
      - For v5.10.y: https://github.com/sjp38/linux/tree/damon/for-v5.10.y
      
      Amazon Linux Kernel Trees
      -------------------------
      
      DAMON is also merged in two public Amazon Linux kernel trees that based on
      v5.4.y[1] and v5.10.y[2].
      
      [1] https://github.com/amazonlinux/linux/tree/amazon-5.4.y/master/mm/damon
      [2] https://github.com/amazonlinux/linux/tree/amazon-5.10.y/master/mm/damon
      
      Git Tree for Diff of Patches
      ============================
      
      For easy review of diff between different versions of each patch, I
      prepared a git tree containing all versions of the DAMON patchset series:
      https://github.com/sjp38/damon-patches
      
      You can clone it and use 'diff' for easy review of changes between
      different versions of the patchset.  For example:
      
          $ git clone https://github.com/sjp38/damon-patches && cd damon-patches
          $ diff -u damon/v33 damon/v34
      
      Sequence Of Patches
      ===================
      
      First three patches implement the core logics of DAMON.  The 1st patch
      introduces basic sampling based hotness monitoring for arbitrary types of
      targets.  Following two patches implement the core mechanisms for control
      of overhead and accuracy, namely regions based sampling (patch 2) and
      adaptive regions adjustment (patch 3).
      
      Now the essential parts of DAMON is complete, but it cannot work unless
      someone provides monitoring primitives for a specific use case.  The
      following two patches make it just work for virtual address spaces
      monitoring.  The 4th patch makes 'PG_idle' can be used by DAMON and the
      5th patch implements the virtual memory address space specific monitoring
      primitives using page table Accessed bits and the 'PG_idle' page flag.
      
      Now DAMON just works for virtual address space monitoring via the kernel
      space api.  To let the user space users can use DAMON, following four
      patches add interfaces for them.  The 6th patch adds a tracepoint for
      monitoring results.  The 7th patch implements a DAMON application kernel
      module, namely damon-dbgfs, that simply wraps DAMON and exposes DAMON
      interface to the user space via the debugfs interface.  The 8th patch
      further exports pid of monitoring thread (kdamond) to user space for
      easier cpu usage accounting, and the 9th patch makes the debugfs interface
      to support multiple contexts.
      
      Three patches for maintainability follows.  The 10th patch adds
      documentations for both the user space and the kernel space.  The 11th
      patch provides unit tests (based on the kunit) while the 12th patch adds
      user space tests (based on the kselftest).
      
      Finally, the last patch (13th) updates the MAINTAINERS file.
      
      This patch (of 13):
      
      DAMON is a data access monitoring framework for the Linux kernel.  The
      core mechanisms of DAMON make it
      
       - accurate (the monitoring output is useful enough for DRAM level
         performance-centric memory management; It might be inappropriate for
         CPU cache levels, though),
       - light-weight (the monitoring overhead is normally low enough to be
         applied online), and
       - scalable (the upper-bound of the overhead is in constant range
         regardless of the size of target workloads).
      
      Using this framework, hence, we can easily write efficient kernel space
      data access monitoring applications.  For example, the kernel's memory
      management mechanisms can make advanced decisions using this.
      Experimental data access aware optimization works that incurring high
      access monitoring overhead could again be implemented on top of this.
      
      Due to its simple and flexible interface, providing user space interface
      would be also easy.  Then, user space users who have some special
      workloads can write personalized applications for better understanding and
      optimizations of their workloads and systems.
      
      ===
      
      Nevertheless, this commit is defining and implementing only basic access
      check part without the overhead-accuracy handling core logic.  The basic
      access check is as below.
      
      The output of DAMON says what memory regions are how frequently accessed
      for a given duration.  The resolution of the access frequency is
      controlled by setting ``sampling interval`` and ``aggregation interval``.
      In detail, DAMON checks access to each page per ``sampling interval`` and
      aggregates the results.  In other words, counts the number of the accesses
      to each region.  After each ``aggregation interval`` passes, DAMON calls
      callback functions that previously registered by users so that users can
      read the aggregated results and then clears the results.  This can be
      described in below simple pseudo-code::
      
          init()
          while monitoring_on:
              for page in monitoring_target:
                  if accessed(page):
                      nr_accesses[page] += 1
              if time() % aggregation_interval == 0:
                  for callback in user_registered_callbacks:
                      callback(monitoring_target, nr_accesses)
                  for page in monitoring_target:
                      nr_accesses[page] = 0
              if time() % update_interval == 0:
                  update()
              sleep(sampling interval)
      
      The target regions constructed at the beginning of the monitoring and
      updated after each ``regions_update_interval``, because the target regions
      could be dynamically changed (e.g., mmap() or memory hotplug).  The
      monitoring overhead of this mechanism will arbitrarily increase as the
      size of the target workload grows.
      
      The basic monitoring primitives for actual access check and dynamic target
      regions construction aren't in the core part of DAMON.  Instead, it allows
      users to implement their own primitives that are optimized for their use
      case and configure DAMON to use those.  In other words, users cannot use
      current version of DAMON without some additional works.
      
      Following commits will implement the core mechanisms for the
      overhead-accuracy control and default primitives implementations.
      
      Link: https://lkml.kernel.org/r/20210716081449.22187-1-sj38.park@gmail.com
      Link: https://lkml.kernel.org/r/20210716081449.22187-2-sj38.park@gmail.comSigned-off-by: NSeongJae Park <sjpark@amazon.de>
      Reviewed-by: NLeonard Foerster <foersleo@amazon.de>
      Reviewed-by: NFernand Sieber <sieberf@amazon.com>
      Acked-by: NShakeel Butt <shakeelb@google.com>
      Cc: Jonathan Cameron <Jonathan.Cameron@huawei.com>
      Cc: Alexander Shishkin <alexander.shishkin@linux.intel.com>
      Cc: Amit Shah <amit@kernel.org>
      Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
      Cc: Jonathan Corbet <corbet@lwn.net>
      Cc: David Hildenbrand <david@redhat.com>
      Cc: David Woodhouse <dwmw@amazon.com>
      Cc: Marco Elver <elver@google.com>
      Cc: Fan Du <fan.du@intel.com>
      Cc: Greg Kroah-Hartman <greg@kroah.com>
      Cc: Greg Thelen <gthelen@google.com>
      Cc: Joe Perches <joe@perches.com>
      Cc: Mel Gorman <mgorman@suse.de>
      Cc: Maximilian Heyne <mheyne@amazon.de>
      Cc: Minchan Kim <minchan@kernel.org>
      Cc: Ingo Molnar <mingo@redhat.com>
      Cc: Namhyung Kim <namhyung@kernel.org>
      Cc: Peter Zijlstra <peterz@infradead.org>
      Cc: Rik van Riel <riel@surriel.com>
      Cc: David Rientjes <rientjes@google.com>
      Cc: Steven Rostedt (VMware) <rostedt@goodmis.org>
      Cc: Shuah Khan <shuah@kernel.org>
      Cc: Vlastimil Babka <vbabka@suse.cz>
      Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
      Cc: Brendan Higgins <brendanhiggins@google.com>
      Cc: Markus Boehme <markubo@amazon.de>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      2224d848
    • M
      mm: introduce PAGEFLAGS_MASK to replace ((1UL << NR_PAGEFLAGS) - 1) · 41c961b9
      Muchun Song 提交于
      Instead of hard-coding ((1UL << NR_PAGEFLAGS) - 1) everywhere, introducing
      PAGEFLAGS_MASK to make the code clear to get the page flags.
      
      Link: https://lkml.kernel.org/r/20210819150712.59948-1-songmuchun@bytedance.comSigned-off-by: NMuchun Song <songmuchun@bytedance.com>
      Reviewed-by: NRoman Gushchin <guro@fb.com>
      Acked-by: NJohannes Weiner <hannes@cmpxchg.org>
      Reviewed-by: NShakeel Butt <shakeelb@google.com>
      Cc: Michal Hocko <mhocko@kernel.org>
      Cc: Vladimir Davydov <vdavydov.dev@gmail.com>
      Cc: Matthew Wilcox (Oracle) <willy@infradead.org>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      41c961b9
    • S
      highmem: don't disable preemption on RT in kmap_atomic() · 51386120
      Sebastian Andrzej Siewior 提交于
      kmap_atomic() disables preemption and pagefaults for historical reasons.
      The conversion to kmap_local(), which only disables migration, cannot be
      done wholesale because quite some call sites need to be updated to
      accommodate with the changed semantics.
      
      On PREEMPT_RT enabled kernels the kmap_atomic() semantics are problematic
      due to the implicit disabling of preemption which makes it impossible to
      acquire 'sleeping' spinlocks within the kmap atomic sections.
      
      PREEMPT_RT replaces the preempt_disable() with a migrate_disable() for
      more than a decade.  It could be argued that this is a justification to do
      this unconditionally, but PREEMPT_RT covers only a limited number of
      architectures and it disables some functionality which limits the coverage
      further.
      
      Limit the replacement to PREEMPT_RT for now.
      
      Link: https://lkml.kernel.org/r/20210810091116.pocdmaatdcogvdso@linutronix.deSigned-off-by: NSebastian Andrzej Siewior <bigeasy@linutronix.de>
      Acked-by: NVlastimil Babka <vbabka@suse.cz>
      Cc: Thomas Gleixner <tglx@linutronix.de>
      Cc: Peter Zijlstra <peterz@infradead.org>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      51386120
    • C
      mm: move ioremap_page_range to vmalloc.c · 82a70ce0
      Christoph Hellwig 提交于
      Patch series "small ioremap cleanups".
      
      The first patch moves a little code around the vmalloc/ioremap boundary
      following a bigger move by Nick earlier.  The second enforces
      non-executable mapping on ioremap just like we do for vmap.  No driver
      currently uses executable mappings anyway, as they should.
      
      This patch (of 2):
      
      This keeps it together with the implementation, and to remove the
      vmap_range wrapper.
      
      Link: https://lkml.kernel.org/r/20210824091259.1324527-1-hch@lst.de
      Link: https://lkml.kernel.org/r/20210824091259.1324527-2-hch@lst.deSigned-off-by: NChristoph Hellwig <hch@lst.de>
      Reviewed-by: NNicholas Piggin <npiggin@gmail.com>
      Cc: Peter Zijlstra <peterz@infradead.org>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      82a70ce0
    • M
      mm: remove redundant compound_head() calling · fe3df441
      Muchun Song 提交于
      There is a READ_ONCE() in the macro of compound_head(), which will prevent
      compiler from optimizing the code when there are more than once calling of
      it in a function.  Remove the redundant calling of compound_head() from
      page_to_index() and page_add_file_rmap() for better code generation.
      
      Link: https://lkml.kernel.org/r/20210811101431.83940-1-songmuchun@bytedance.comSigned-off-by: NMuchun Song <songmuchun@bytedance.com>
      Reviewed-by: NDavid Howells <dhowells@redhat.com>
      Cc: Matthew Wilcox (Oracle) <willy@infradead.org>
      Cc: William Kucharski <william.kucharski@oracle.com>
      Cc: Kirill A. Shutemov <kirill.shutemov@linux.intel.com>
      Cc: Johannes Weiner <hannes@cmpxchg.org>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      fe3df441
    • D
      mm/memory_hotplug: improved dynamic memory group aware "auto-movable" online policy · 3fcebf90
      David Hildenbrand 提交于
      Currently, the "auto-movable" online policy does not allow for hotplugged
      KERNEL (ZONE_NORMAL) memory to increase the amount of MOVABLE memory we
      can have, primarily, because there is no coordiantion across memory
      devices and we don't want to create zone-imbalances accidentially when
      unplugging memory.
      
      However, within a single memory device it's different.  Let's allow for
      KERNEL memory within a dynamic memory group to allow for more MOVABLE
      within the same memory group.  The only thing we have to take care of is
      that the managing driver avoids zone imbalances by unplugging MOVABLE
      memory first, otherwise there can be corner cases where unplug of memory
      could result in (accidential) zone imbalances.
      
      virtio-mem is the only user of dynamic memory groups and recently added
      support for prioritizing unplug of ZONE_MOVABLE over ZONE_NORMAL, so we
      don't need a new toggle to enable it for dynamic memory groups.
      
      We limit this handling to dynamic memory groups, because:
      
      * We want to keep the runtime overhead for collecting stats when
        onlining a single memory block small.  We tend to have only a handful of
        dynamic memory groups, but we can have quite some static memory groups
        (e.g., 256 DIMMs).
      
      * It doesn't make too much sense for static memory groups, as we try
        onlining all applicable memory blocks either completely to ZONE_MOVABLE
        or not.  In ordinary operation, we won't have a mixture of zones within
        a static memory group.
      
      When adding memory to a dynamic memory group, we'll first online memory to
      ZONE_MOVABLE as long as early KERNEL memory allows for it.  Then, we'll
      online the next unit(s) to ZONE_NORMAL, until we can online the next
      unit(s) to ZONE_MOVABLE.
      
      For a simple virtio-mem device with a MOVABLE:KERNEL ratio of 3:1, it will
      result in a layout like:
      
        [M][M][M][M][M][M][M][M][N][M][M][M][N][M][M][M]...
        ^ movable memory due to early kernel memory
      			   ^ allows for more movable memory ...
      			      ^-----^ ... here
      				       ^ allows for more movable memory ...
      				          ^-----^ ... here
      
      While the created layout is sub-optimal when it comes to contiguous zones,
      it gives us the maximum flexibility when dynamically growing/shrinking a
      device; we can grow small VMs really big in small steps, and still shrink
      reliably to e.g., 1/4 of the maximum VM size in this example, removing
      full memory blocks along with meta data more reliably.
      
      Mark dynamic memory groups in the xarray such that we can efficiently
      iterate over them when collecting stats.  In usual setups, we have one
      virtio-mem device per NUMA node, and usually only a small number of NUMA
      nodes.
      
      Note: for now, there seems to be no compelling reason to make this
      behavior configurable.
      
      Link: https://lkml.kernel.org/r/20210806124715.17090-10-david@redhat.comSigned-off-by: NDavid Hildenbrand <david@redhat.com>
      Cc: Anshuman Khandual <anshuman.khandual@arm.com>
      Cc: Dan Williams <dan.j.williams@intel.com>
      Cc: Dave Hansen <dave.hansen@linux.intel.com>
      Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
      Cc: Hui Zhu <teawater@gmail.com>
      Cc: Jason Wang <jasowang@redhat.com>
      Cc: Len Brown <lenb@kernel.org>
      Cc: Marek Kedzierski <mkedzier@redhat.com>
      Cc: "Michael S. Tsirkin" <mst@redhat.com>
      Cc: Michal Hocko <mhocko@kernel.org>
      Cc: Mike Rapoport <rppt@kernel.org>
      Cc: Oscar Salvador <osalvador@suse.de>
      Cc: Pankaj Gupta <pankaj.gupta.linux@gmail.com>
      Cc: Pavel Tatashin <pasha.tatashin@soleen.com>
      Cc: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
      Cc: "Rafael J. Wysocki" <rjw@rjwysocki.net>
      Cc: Vitaly Kuznetsov <vkuznets@redhat.com>
      Cc: Vlastimil Babka <vbabka@suse.cz>
      Cc: Wei Yang <richard.weiyang@linux.alibaba.com>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      3fcebf90
    • D
      mm/memory_hotplug: memory group aware "auto-movable" online policy · 445fcf7c
      David Hildenbrand 提交于
      Use memory groups to improve our "auto-movable" onlining policy:
      
      1. For static memory groups (e.g., a DIMM), online a memory block MOVABLE
         only if all other memory blocks in the group are either MOVABLE or could
         be onlined MOVABLE. A DIMM will either be MOVABLE or not, not a mixture.
      
      2. For dynamic memory groups (e.g., a virtio-mem device), online a
         memory block MOVABLE only if all other memory blocks inside the
         current unit are either MOVABLE or could be onlined MOVABLE. For a
         virtio-mem device with a device block size with 512 MiB, all 128 MiB
         memory blocks wihin a 512 MiB unit will either be MOVABLE or not, not
         a mixture.
      
      We have to pass the memory group to zone_for_pfn_range() to take the
      memory group into account.
      
      Note: for now, there seems to be no compelling reason to make this
      behavior configurable.
      
      Link: https://lkml.kernel.org/r/20210806124715.17090-9-david@redhat.comSigned-off-by: NDavid Hildenbrand <david@redhat.com>
      Cc: Anshuman Khandual <anshuman.khandual@arm.com>
      Cc: Dan Williams <dan.j.williams@intel.com>
      Cc: Dave Hansen <dave.hansen@linux.intel.com>
      Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
      Cc: Hui Zhu <teawater@gmail.com>
      Cc: Jason Wang <jasowang@redhat.com>
      Cc: Len Brown <lenb@kernel.org>
      Cc: Marek Kedzierski <mkedzier@redhat.com>
      Cc: "Michael S. Tsirkin" <mst@redhat.com>
      Cc: Michal Hocko <mhocko@kernel.org>
      Cc: Mike Rapoport <rppt@kernel.org>
      Cc: Oscar Salvador <osalvador@suse.de>
      Cc: Pankaj Gupta <pankaj.gupta.linux@gmail.com>
      Cc: Pavel Tatashin <pasha.tatashin@soleen.com>
      Cc: Rafael J. Wysocki <rafael.j.wysocki@intel.com>
      Cc: "Rafael J. Wysocki" <rjw@rjwysocki.net>
      Cc: Vitaly Kuznetsov <vkuznets@redhat.com>
      Cc: Vlastimil Babka <vbabka@suse.cz>
      Cc: Wei Yang <richard.weiyang@linux.alibaba.com>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      445fcf7c