1. 08 1月, 2020 1 次提交
    • S
      char/random: silence a lockdep splat with printk() · 1b710b1b
      Sergey Senozhatsky 提交于
      Sergey didn't like the locking order,
      
      uart_port->lock  ->  tty_port->lock
      
      uart_write (uart_port->lock)
        __uart_start
          pl011_start_tx
            pl011_tx_chars
              uart_write_wakeup
                tty_port_tty_wakeup
                  tty_port_default
                    tty_port_tty_get (tty_port->lock)
      
      but those code is so old, and I have no clue how to de-couple it after
      checking other locks in the splat. There is an onging effort to make all
      printk() as deferred, so until that happens, workaround it for now as a
      short-term fix.
      
      LTP: starting iogen01 (export LTPROOT; rwtest -N iogen01 -i 120s -s
      read,write -Da -Dv -n 2 500b:$TMPDIR/doio.f1.$$
      1000b:$TMPDIR/doio.f2.$$)
      WARNING: possible circular locking dependency detected
      ------------------------------------------------------
      doio/49441 is trying to acquire lock:
      ffff008b7cff7290 (&(&zone->lock)->rlock){..-.}, at: rmqueue+0x138/0x2050
      
      but task is already holding lock:
      60ff000822352818 (&pool->lock/1){-.-.}, at: start_flush_work+0xd8/0x3f0
      
        which lock already depends on the new lock.
      
        the existing dependency chain (in reverse order) is:
      
        -> #4 (&pool->lock/1){-.-.}:
             lock_acquire+0x320/0x360
             _raw_spin_lock+0x64/0x80
             __queue_work+0x4b4/0xa10
             queue_work_on+0xac/0x11c
             tty_schedule_flip+0x84/0xbc
             tty_flip_buffer_push+0x1c/0x28
             pty_write+0x98/0xd0
             n_tty_write+0x450/0x60c
             tty_write+0x338/0x474
             __vfs_write+0x88/0x214
             vfs_write+0x12c/0x1a4
             redirected_tty_write+0x90/0xdc
             do_loop_readv_writev+0x140/0x180
             do_iter_write+0xe0/0x10c
             vfs_writev+0x134/0x1cc
             do_writev+0xbc/0x130
             __arm64_sys_writev+0x58/0x8c
             el0_svc_handler+0x170/0x240
             el0_sync_handler+0x150/0x250
             el0_sync+0x164/0x180
      
        -> #3 (&(&port->lock)->rlock){-.-.}:
             lock_acquire+0x320/0x360
             _raw_spin_lock_irqsave+0x7c/0x9c
             tty_port_tty_get+0x24/0x60
             tty_port_default_wakeup+0x1c/0x3c
             tty_port_tty_wakeup+0x34/0x40
             uart_write_wakeup+0x28/0x44
             pl011_tx_chars+0x1b8/0x270
             pl011_start_tx+0x24/0x70
             __uart_start+0x5c/0x68
             uart_write+0x164/0x1c8
             do_output_char+0x33c/0x348
             n_tty_write+0x4bc/0x60c
             tty_write+0x338/0x474
             redirected_tty_write+0xc0/0xdc
             do_loop_readv_writev+0x140/0x180
             do_iter_write+0xe0/0x10c
             vfs_writev+0x134/0x1cc
             do_writev+0xbc/0x130
             __arm64_sys_writev+0x58/0x8c
             el0_svc_handler+0x170/0x240
             el0_sync_handler+0x150/0x250
             el0_sync+0x164/0x180
      
        -> #2 (&port_lock_key){-.-.}:
             lock_acquire+0x320/0x360
             _raw_spin_lock+0x64/0x80
             pl011_console_write+0xec/0x2cc
             console_unlock+0x794/0x96c
             vprintk_emit+0x260/0x31c
             vprintk_default+0x54/0x7c
             vprintk_func+0x218/0x254
             printk+0x7c/0xa4
             register_console+0x734/0x7b0
             uart_add_one_port+0x734/0x834
             pl011_register_port+0x6c/0xac
             sbsa_uart_probe+0x234/0x2ec
             platform_drv_probe+0xd4/0x124
             really_probe+0x250/0x71c
             driver_probe_device+0xb4/0x200
             __device_attach_driver+0xd8/0x188
             bus_for_each_drv+0xbc/0x110
             __device_attach+0x120/0x220
             device_initial_probe+0x20/0x2c
             bus_probe_device+0x54/0x100
             device_add+0xae8/0xc2c
             platform_device_add+0x278/0x3b8
             platform_device_register_full+0x238/0x2ac
             acpi_create_platform_device+0x2dc/0x3a8
             acpi_bus_attach+0x390/0x3cc
             acpi_bus_attach+0x108/0x3cc
             acpi_bus_attach+0x108/0x3cc
             acpi_bus_attach+0x108/0x3cc
             acpi_bus_scan+0x7c/0xb0
             acpi_scan_init+0xe4/0x304
             acpi_init+0x100/0x114
             do_one_initcall+0x348/0x6a0
             do_initcall_level+0x190/0x1fc
             do_basic_setup+0x34/0x4c
             kernel_init_freeable+0x19c/0x260
             kernel_init+0x18/0x338
             ret_from_fork+0x10/0x18
      
        -> #1 (console_owner){-...}:
             lock_acquire+0x320/0x360
             console_lock_spinning_enable+0x6c/0x7c
             console_unlock+0x4f8/0x96c
             vprintk_emit+0x260/0x31c
             vprintk_default+0x54/0x7c
             vprintk_func+0x218/0x254
             printk+0x7c/0xa4
             get_random_u64+0x1c4/0x1dc
             shuffle_pick_tail+0x40/0xac
             __free_one_page+0x424/0x710
             free_one_page+0x70/0x120
             __free_pages_ok+0x61c/0xa94
             __free_pages_core+0x1bc/0x294
             memblock_free_pages+0x38/0x48
             __free_pages_memory+0xcc/0xfc
             __free_memory_core+0x70/0x78
             free_low_memory_core_early+0x148/0x18c
             memblock_free_all+0x18/0x54
             mem_init+0xb4/0x17c
             mm_init+0x14/0x38
             start_kernel+0x19c/0x530
      
        -> #0 (&(&zone->lock)->rlock){..-.}:
             validate_chain+0xf6c/0x2e2c
             __lock_acquire+0x868/0xc2c
             lock_acquire+0x320/0x360
             _raw_spin_lock+0x64/0x80
             rmqueue+0x138/0x2050
             get_page_from_freelist+0x474/0x688
             __alloc_pages_nodemask+0x3b4/0x18dc
             alloc_pages_current+0xd0/0xe0
             alloc_slab_page+0x2b4/0x5e0
             new_slab+0xc8/0x6bc
             ___slab_alloc+0x3b8/0x640
             kmem_cache_alloc+0x4b4/0x588
             __debug_object_init+0x778/0x8b4
             debug_object_init_on_stack+0x40/0x50
             start_flush_work+0x16c/0x3f0
             __flush_work+0xb8/0x124
             flush_work+0x20/0x30
             xlog_cil_force_lsn+0x88/0x204 [xfs]
             xfs_log_force_lsn+0x128/0x1b8 [xfs]
             xfs_file_fsync+0x3c4/0x488 [xfs]
             vfs_fsync_range+0xb0/0xd0
             generic_write_sync+0x80/0xa0 [xfs]
             xfs_file_buffered_aio_write+0x66c/0x6e4 [xfs]
             xfs_file_write_iter+0x1a0/0x218 [xfs]
             __vfs_write+0x1cc/0x214
             vfs_write+0x12c/0x1a4
             ksys_write+0xb0/0x120
             __arm64_sys_write+0x54/0x88
             el0_svc_handler+0x170/0x240
             el0_sync_handler+0x150/0x250
             el0_sync+0x164/0x180
      
             other info that might help us debug this:
      
       Chain exists of:
         &(&zone->lock)->rlock --> &(&port->lock)->rlock --> &pool->lock/1
      
       Possible unsafe locking scenario:
      
             CPU0                    CPU1
             ----                    ----
        lock(&pool->lock/1);
                                     lock(&(&port->lock)->rlock);
                                     lock(&pool->lock/1);
        lock(&(&zone->lock)->rlock);
      
                      *** DEADLOCK ***
      
      4 locks held by doio/49441:
       #0: a0ff00886fc27408 (sb_writers#8){.+.+}, at: vfs_write+0x118/0x1a4
       #1: 8fff00080810dfe0 (&xfs_nondir_ilock_class){++++}, at:
      xfs_ilock+0x2a8/0x300 [xfs]
       #2: ffff9000129f2390 (rcu_read_lock){....}, at:
      rcu_lock_acquire+0x8/0x38
       #3: 60ff000822352818 (&pool->lock/1){-.-.}, at:
      start_flush_work+0xd8/0x3f0
      
                     stack backtrace:
      CPU: 48 PID: 49441 Comm: doio Tainted: G        W
      Hardware name: HPE Apollo 70             /C01_APACHE_MB         , BIOS
      L50_5.13_1.11 06/18/2019
      Call trace:
       dump_backtrace+0x0/0x248
       show_stack+0x20/0x2c
       dump_stack+0xe8/0x150
       print_circular_bug+0x368/0x380
       check_noncircular+0x28c/0x294
       validate_chain+0xf6c/0x2e2c
       __lock_acquire+0x868/0xc2c
       lock_acquire+0x320/0x360
       _raw_spin_lock+0x64/0x80
       rmqueue+0x138/0x2050
       get_page_from_freelist+0x474/0x688
       __alloc_pages_nodemask+0x3b4/0x18dc
       alloc_pages_current+0xd0/0xe0
       alloc_slab_page+0x2b4/0x5e0
       new_slab+0xc8/0x6bc
       ___slab_alloc+0x3b8/0x640
       kmem_cache_alloc+0x4b4/0x588
       __debug_object_init+0x778/0x8b4
       debug_object_init_on_stack+0x40/0x50
       start_flush_work+0x16c/0x3f0
       __flush_work+0xb8/0x124
       flush_work+0x20/0x30
       xlog_cil_force_lsn+0x88/0x204 [xfs]
       xfs_log_force_lsn+0x128/0x1b8 [xfs]
       xfs_file_fsync+0x3c4/0x488 [xfs]
       vfs_fsync_range+0xb0/0xd0
       generic_write_sync+0x80/0xa0 [xfs]
       xfs_file_buffered_aio_write+0x66c/0x6e4 [xfs]
       xfs_file_write_iter+0x1a0/0x218 [xfs]
       __vfs_write+0x1cc/0x214
       vfs_write+0x12c/0x1a4
       ksys_write+0xb0/0x120
       __arm64_sys_write+0x54/0x88
       el0_svc_handler+0x170/0x240
       el0_sync_handler+0x150/0x250
       el0_sync+0x164/0x180
      Reviewed-by: NSergey Senozhatsky <sergey.senozhatsky.work@gmail.com>
      Signed-off-by: NQian Cai <cai@lca.pw>
      Link: https://lore.kernel.org/r/1573679785-21068-1-git-send-email-cai@lca.pwSigned-off-by: NTheodore Ts'o <tytso@mit.edu>
      1b710b1b
  2. 18 12月, 2019 1 次提交
  3. 17 11月, 2019 1 次提交
  4. 23 10月, 2019 1 次提交
  5. 03 10月, 2019 1 次提交
    • B
      char/random: Add a newline at the end of the file · 3fd57e7a
      Borislav Petkov 提交于
      On Tue, Oct 01, 2019 at 10:14:40AM -0700, Linus Torvalds wrote:
      > The previous state of the file didn't have that 0xa at the end, so you get that
      >
      >
      >   -EXPORT_SYMBOL_GPL(add_bootloader_randomness);
      >   \ No newline at end of file
      >   +EXPORT_SYMBOL_GPL(add_bootloader_randomness);
      >
      > which is "the '-' line doesn't have a newline, the '+' line does" marker.
      
      Aaha, that makes total sense, thanks for explaining. Oh well, let's fix
      it then so that people don't scratch heads like me.
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      3fd57e7a
  6. 30 9月, 2019 1 次提交
    • L
      random: try to actively add entropy rather than passively wait for it · 50ee7529
      Linus Torvalds 提交于
      For 5.3 we had to revert a nice ext4 IO pattern improvement, because it
      caused a bootup regression due to lack of entropy at bootup together
      with arguably broken user space that was asking for secure random
      numbers when it really didn't need to.
      
      See commit 72dbcf72 (Revert "ext4: make __ext4_get_inode_loc plug").
      
      This aims to solve the issue by actively generating entropy noise using
      the CPU cycle counter when waiting for the random number generator to
      initialize.  This only works when you have a high-frequency time stamp
      counter available, but that's the case on all modern x86 CPU's, and on
      most other modern CPU's too.
      
      What we do is to generate jitter entropy from the CPU cycle counter
      under a somewhat complex load: calling the scheduler while also
      guaranteeing a certain amount of timing noise by also triggering a
      timer.
      
      I'm sure we can tweak this, and that people will want to look at other
      alternatives, but there's been a number of papers written on jitter
      entropy, and this should really be fairly conservative by crediting one
      bit of entropy for every timer-induced jump in the cycle counter.  Not
      because the timer itself would be all that unpredictable, but because
      the interaction between the timer and the loop is going to be.
      
      Even if (and perhaps particularly if) the timer actually happens on
      another CPU, the cacheline interaction between the loop that reads the
      cycle counter and the timer itself firing is going to add perturbations
      to the cycle counter values that get mixed into the entropy pool.
      
      As Thomas pointed out, with a modern out-of-order CPU, even quite simple
      loops show a fair amount of hard-to-predict timing variability even in
      the absense of external interrupts.  But this tries to take that further
      by actually having a fairly complex interaction.
      
      This is not going to solve the entropy issue for architectures that have
      no CPU cycle counter, but it's not clear how (and if) that is solvable,
      and the hardware in question is largely starting to be irrelevant.  And
      by doing this we can at least avoid some of the even more contentious
      approaches (like making the entropy waiting time out in order to avoid
      the possibly unbounded waiting).
      
      Cc: Ahmed Darwish <darwish.07@gmail.com>
      Cc: Thomas Gleixner <tglx@linutronix.de>
      Cc: Theodore Ts'o <tytso@mit.edu>
      Cc: Nicholas Mc Guire <hofrat@opentech.at>
      Cc: Andy Lutomirski <luto@kernel.org>
      Cc: Kees Cook <keescook@chromium.org>
      Cc: Willy Tarreau <w@1wt.eu>
      Cc: Alexander E. Patrakov <patrakov@gmail.com>
      Cc: Lennart Poettering <mzxreary@0pointer.de>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      50ee7529
  7. 09 9月, 2019 1 次提交
    • S
      random: Use wait_event_freezable() in add_hwgenerator_randomness() · 59b56948
      Stephen Boyd 提交于
      Sebastian reports that after commit ff296293 ("random: Support freezable
      kthreads in add_hwgenerator_randomness()") we can call might_sleep() when the
      task state is TASK_INTERRUPTIBLE (state=1). This leads to the following warning.
      
       do not call blocking ops when !TASK_RUNNING; state=1 set at [<00000000349d1489>] prepare_to_wait_event+0x5a/0x180
       WARNING: CPU: 0 PID: 828 at kernel/sched/core.c:6741 __might_sleep+0x6f/0x80
       Modules linked in:
      
       CPU: 0 PID: 828 Comm: hwrng Not tainted 5.3.0-rc7-next-20190903+ #46
       RIP: 0010:__might_sleep+0x6f/0x80
      
       Call Trace:
        kthread_freezable_should_stop+0x1b/0x60
        add_hwgenerator_randomness+0xdd/0x130
        hwrng_fillfn+0xbf/0x120
        kthread+0x10c/0x140
        ret_from_fork+0x27/0x50
      
      We shouldn't call kthread_freezable_should_stop() from deep within the
      wait_event code because the task state is still set as
      TASK_INTERRUPTIBLE instead of TASK_RUNNING and
      kthread_freezable_should_stop() will try to call into the freezer with
      the task in the wrong state. Use wait_event_freezable() instead so that
      it calls schedule() in the right place and tries to enter the freezer
      when the task state is TASK_RUNNING instead.
      Reported-by: NSebastian Andrzej Siewior <bigeasy@linutronix.de>
      Tested-by: NSebastian Andrzej Siewior <bigeasy@linutronix.de>
      Cc: Keerthy <j-keerthy@ti.com>
      Fixes: ff296293 ("random: Support freezable kthreads in add_hwgenerator_randomness()")
      Signed-off-by: NStephen Boyd <swboyd@chromium.org>
      Signed-off-by: NHerbert Xu <herbert@gondor.apana.org.au>
      59b56948
  8. 23 8月, 2019 1 次提交
    • H
      fdt: add support for rng-seed · 428826f5
      Hsin-Yi Wang 提交于
      Introducing a chosen node, rng-seed, which is an entropy that can be
      passed to kernel called very early to increase initial device
      randomness. Bootloader should provide this entropy and the value is
      read from /chosen/rng-seed in DT.
      
      Obtain of_fdt_crc32 for CRC check after early_init_dt_scan_nodes(),
      since early_init_dt_scan_chosen() would modify fdt to erase rng-seed.
      
      Add a new interface add_bootloader_randomness() for rng-seed use case.
      Depends on whether the seed is trustworthy, rng seed would be passed to
      add_hwgenerator_randomness(). Otherwise it would be passed to
      add_device_randomness(). Decision is controlled by kernel config
      RANDOM_TRUST_BOOTLOADER.
      Signed-off-by: NHsin-Yi Wang <hsinyi@chromium.org>
      Reviewed-by: NStephen Boyd <swboyd@chromium.org>
      Reviewed-by: NRob Herring <robh@kernel.org>
      Reviewed-by: Theodore Ts'o <tytso@mit.edu> # drivers/char/random.c
      Signed-off-by: NWill Deacon <will@kernel.org>
      428826f5
  9. 22 8月, 2019 1 次提交
    • S
      random: Support freezable kthreads in add_hwgenerator_randomness() · ff296293
      Stephen Boyd 提交于
      The kthread calling this function is freezable after commit 03a3bb7a
      ("hwrng: core - Freeze khwrng thread during suspend") is applied.
      Unfortunately, this function uses wait_event_interruptible() but doesn't
      check for the kthread being woken up by the fake freezer signal. When a
      user suspends the system, this kthread will wake up and if it fails the
      entropy size check it will immediately go back to sleep and not go into
      the freezer. Eventually, suspend will fail because the task never froze
      and a warning message like this may appear:
      
       PM: suspend entry (deep)
       Filesystems sync: 0.000 seconds
       Freezing user space processes ... (elapsed 0.001 seconds) done.
       OOM killer disabled.
       Freezing remaining freezable tasks ...
       Freezing of tasks failed after 20.003 seconds (1 tasks refusing to freeze, wq_busy=0):
       hwrng           R  running task        0   289      2 0x00000020
       [<c08c64c4>] (__schedule) from [<c08c6a10>] (schedule+0x3c/0xc0)
       [<c08c6a10>] (schedule) from [<c05dbd8c>] (add_hwgenerator_randomness+0xb0/0x100)
       [<c05dbd8c>] (add_hwgenerator_randomness) from [<bf1803c8>] (hwrng_fillfn+0xc0/0x14c [rng_core])
       [<bf1803c8>] (hwrng_fillfn [rng_core]) from [<c015abec>] (kthread+0x134/0x148)
       [<c015abec>] (kthread) from [<c01010e8>] (ret_from_fork+0x14/0x2c)
      
      Check for a freezer signal here and skip adding any randomness if the
      task wakes up because it was frozen. This should make the kthread freeze
      properly and suspend work again.
      
      Fixes: 03a3bb7a ("hwrng: core - Freeze khwrng thread during suspend")
      Reported-by: NKeerthy <j-keerthy@ti.com>
      Tested-by: NKeerthy <j-keerthy@ti.com>
      Signed-off-by: NStephen Boyd <swboyd@chromium.org>
      Signed-off-by: NHerbert Xu <herbert@gondor.apana.org.au>
      ff296293
  10. 26 5月, 2019 1 次提交
  11. 20 4月, 2019 4 次提交
    • S
      random: add a spinlock_t to struct batched_entropy · b7d5dc21
      Sebastian Andrzej Siewior 提交于
      The per-CPU variable batched_entropy_uXX is protected by get_cpu_var().
      This is just a preempt_disable() which ensures that the variable is only
      from the local CPU. It does not protect against users on the same CPU
      from another context. It is possible that a preemptible context reads
      slot 0 and then an interrupt occurs and the same value is read again.
      
      The above scenario is confirmed by lockdep if we add a spinlock:
      | ================================
      | WARNING: inconsistent lock state
      | 5.1.0-rc3+ #42 Not tainted
      | --------------------------------
      | inconsistent {SOFTIRQ-ON-W} -> {IN-SOFTIRQ-W} usage.
      | ksoftirqd/9/56 [HC0[0]:SC1[1]:HE0:SE0] takes:
      | (____ptrval____) (batched_entropy_u32.lock){+.?.}, at: get_random_u32+0x3e/0xe0
      | {SOFTIRQ-ON-W} state was registered at:
      |   _raw_spin_lock+0x2a/0x40
      |   get_random_u32+0x3e/0xe0
      |   new_slab+0x15c/0x7b0
      |   ___slab_alloc+0x492/0x620
      |   __slab_alloc.isra.73+0x53/0xa0
      |   kmem_cache_alloc_node+0xaf/0x2a0
      |   copy_process.part.41+0x1e1/0x2370
      |   _do_fork+0xdb/0x6d0
      |   kernel_thread+0x20/0x30
      |   kthreadd+0x1ba/0x220
      |   ret_from_fork+0x3a/0x50
      …
      | other info that might help us debug this:
      |  Possible unsafe locking scenario:
      |
      |        CPU0
      |        ----
      |   lock(batched_entropy_u32.lock);
      |   <Interrupt>
      |     lock(batched_entropy_u32.lock);
      |
      |  *** DEADLOCK ***
      |
      | stack backtrace:
      | Call Trace:
      …
      |  kmem_cache_alloc_trace+0x20e/0x270
      |  ipmi_alloc_recv_msg+0x16/0x40
      …
      |  __do_softirq+0xec/0x48d
      |  run_ksoftirqd+0x37/0x60
      |  smpboot_thread_fn+0x191/0x290
      |  kthread+0xfe/0x130
      |  ret_from_fork+0x3a/0x50
      
      Add a spinlock_t to the batched_entropy data structure and acquire the
      lock while accessing it. Acquire the lock with disabled interrupts
      because this function may be used from interrupt context.
      
      Remove the batched_entropy_reset_lock lock. Now that we have a lock for
      the data scructure, we can access it from a remote CPU.
      Signed-off-by: NSebastian Andrzej Siewior <bigeasy@linutronix.de>
      Signed-off-by: NTheodore Ts'o <tytso@mit.edu>
      b7d5dc21
    • G
      random: document get_random_int() family · 92e507d2
      George Spelvin 提交于
      Explain what these functions are for and when they offer
      an advantage over get_random_bytes().
      
      (We still need documentation on rng_is_initialized(), the
      random_ready_callback system, and early boot in general.)
      Signed-off-by: NGeorge Spelvin <lkml@sdf.org>
      Signed-off-by: NTheodore Ts'o <tytso@mit.edu>
      92e507d2
    • J
      random: fix CRNG initialization when random.trust_cpu=1 · fe6f1a6a
      Jon DeVree 提交于
      When the system boots with random.trust_cpu=1 it doesn't initialize the
      per-NUMA CRNGs because it skips the rest of the CRNG startup code. This
      means that the code from 1e7f583a ("random: make /dev/urandom scalable
      for silly userspace programs") is not used when random.trust_cpu=1.
      
      crash> dmesg | grep random:
      [    0.000000] random: get_random_bytes called from start_kernel+0x94/0x530 with crng_init=0
      [    0.314029] random: crng done (trusting CPU's manufacturer)
      crash> print crng_node_pool
      $6 = (struct crng_state **) 0x0
      
      After adding the missing call to numa_crng_init() the per-NUMA CRNGs are
      initialized again:
      
      crash> dmesg | grep random:
      [    0.000000] random: get_random_bytes called from start_kernel+0x94/0x530 with crng_init=0
      [    0.314031] random: crng done (trusting CPU's manufacturer)
      crash> print crng_node_pool
      $1 = (struct crng_state **) 0xffff9a915f4014a0
      
      The call to invalidate_batched_entropy() was also missing. This is
      important for architectures like PPC and S390 which only have the
      arch_get_random_seed_* functions.
      
      Fixes: 39a8883a ("random: add a config option to trust the CPU's hwrng")
      Signed-off-by: NJon DeVree <nuxi@vault24.org>
      Signed-off-by: NTheodore Ts'o <tytso@mit.edu>
      fe6f1a6a
    • K
      random: move rand_initialize() earlier · d5553523
      Kees Cook 提交于
      Right now rand_initialize() is run as an early_initcall(), but it only
      depends on timekeeping_init() (for mixing ktime_get_real() into the
      pools). However, the call to boot_init_stack_canary() for stack canary
      initialization runs earlier, which triggers a warning at boot:
      
      random: get_random_bytes called from start_kernel+0x357/0x548 with crng_init=0
      
      Instead, this moves rand_initialize() to after timekeeping_init(), and moves
      canary initialization here as well.
      
      Note that this warning may still remain for machines that do not have
      UEFI RNG support (which initializes the RNG pools during setup_arch()),
      or for x86 machines without RDRAND (or booting without "random.trust=on"
      or CONFIG_RANDOM_TRUST_CPU=y).
      Signed-off-by: NKees Cook <keescook@chromium.org>
      Signed-off-by: NTheodore Ts'o <tytso@mit.edu>
      d5553523
  12. 17 4月, 2019 4 次提交
  13. 20 11月, 2018 1 次提交
    • E
      crypto: chacha20-generic - refactor to allow varying number of rounds · 1ca1b917
      Eric Biggers 提交于
      In preparation for adding XChaCha12 support, rename/refactor
      chacha20-generic to support different numbers of rounds.  The
      justification for needing XChaCha12 support is explained in more detail
      in the patch "crypto: chacha - add XChaCha12 support".
      
      The only difference between ChaCha{8,12,20} are the number of rounds
      itself; all other parts of the algorithm are the same.  Therefore,
      remove the "20" from all definitions, structures, functions, files, etc.
      that will be shared by all ChaCha versions.
      
      Also make ->setkey() store the round count in the chacha_ctx (previously
      chacha20_ctx).  The generic code then passes the round count through to
      chacha_block().  There will be a ->setkey() function for each explicitly
      allowed round count; the encrypt/decrypt functions will be the same.  I
      decided not to do it the opposite way (same ->setkey() function for all
      round counts, with different encrypt/decrypt functions) because that
      would have required more boilerplate code in architecture-specific
      implementations of ChaCha and XChaCha.
      Reviewed-by: NArd Biesheuvel <ard.biesheuvel@linaro.org>
      Acked-by: NMartin Willi <martin@strongswan.org>
      Signed-off-by: NEric Biggers <ebiggers@google.com>
      Signed-off-by: NHerbert Xu <herbert@gondor.apana.org.au>
      1ca1b917
  14. 21 9月, 2018 1 次提交
    • E
      crypto: chacha20 - Fix chacha20_block() keystream alignment (again) · a5e9f557
      Eric Biggers 提交于
      In commit 9f480fae ("crypto: chacha20 - Fix keystream alignment for
      chacha20_block()"), I had missed that chacha20_block() can be called
      directly on the buffer passed to get_random_bytes(), which can have any
      alignment.  So, while my commit didn't break anything, it didn't fully
      solve the alignment problems.
      
      Revert my solution and just update chacha20_block() to use
      put_unaligned_le32(), so the output buffer need not be aligned.
      This is simpler, and on many CPUs it's the same speed.
      
      But, I kept the 'tmp' buffers in extract_crng_user() and
      _get_random_bytes() 4-byte aligned, since that alignment is actually
      needed for _crng_backtrack_protect() too.
      Reported-by: NStephan Müller <smueller@chronox.de>
      Cc: Theodore Ts'o <tytso@mit.edu>
      Signed-off-by: NEric Biggers <ebiggers@google.com>
      Signed-off-by: NHerbert Xu <herbert@gondor.apana.org.au>
      a5e9f557
  15. 02 9月, 2018 1 次提交
  16. 03 8月, 2018 1 次提交
    • J
      random: Make crng state queryable · 9a47249d
      Jason A. Donenfeld 提交于
      It is very useful to be able to know whether or not get_random_bytes_wait
      / wait_for_random_bytes is going to block or not, or whether plain
      get_random_bytes is going to return good randomness or bad randomness.
      
      The particular use case is for mitigating certain attacks in WireGuard.
      A handshake packet arrives and is queued up. Elsewhere a worker thread
      takes items from the queue and processes them. In replying to these
      items, it needs to use some random data, and it has to be good random
      data. If we simply block until we can have good randomness, then it's
      possible for an attacker to fill the queue up with packets waiting to be
      processed. Upon realizing the queue is full, WireGuard will detect that
      it's under a denial of service attack, and behave accordingly. A better
      approach is just to drop incoming handshake packets if the crng is not
      yet initialized.
      
      This patch, therefore, makes that information directly accessible.
      Signed-off-by: NJason A. Donenfeld <Jason@zx2c4.com>
      Signed-off-by: NTheodore Ts'o <tytso@mit.edu>
      9a47249d
  17. 25 7月, 2018 2 次提交
    • I
      random: remove preempt disabled region · b34fbaa9
      Ingo Molnar 提交于
      No need to keep preemption disabled across the whole function.
      
      mix_pool_bytes() uses a spin_lock() to protect the pool and there are
      other places like write_pool() whhich invoke mix_pool_bytes() without
      disabling preemption.
      credit_entropy_bits() is invoked from other places like
      add_hwgenerator_randomness() without disabling preemption.
      
      Before commit 95b709b6 ("random: drop trickle mode") the function
      used __this_cpu_inc_return() which would require disabled preemption.
      The preempt_disable() section was added in commit 43d5d3018c37 ("[PATCH]
      random driver preempt robustness", history tree).  It was claimed that
      the code relied on "vt_ioctl() being called under BKL".
      
      Cc: "Theodore Ts'o" <tytso@mit.edu>
      Signed-off-by: NIngo Molnar <mingo@elte.hu>
      Signed-off-by: NThomas Gleixner <tglx@linutronix.de>
      [bigeasy: enhance the commit message]
      Signed-off-by: NSebastian Andrzej Siewior <bigeasy@linutronix.de>
      Signed-off-by: NTheodore Ts'o <tytso@mit.edu>
      b34fbaa9
    • T
      random: add a config option to trust the CPU's hwrng · 39a8883a
      Theodore Ts'o 提交于
      This gives the user building their own kernel (or a Linux
      distribution) the option of deciding whether or not to trust the CPU's
      hardware random number generator (e.g., RDRAND for x86 CPU's) as being
      correctly implemented and not having a back door introduced (perhaps
      courtesy of a Nation State's law enforcement or intelligence
      agencies).
      
      This will prevent getrandom(2) from blocking, if there is a
      willingness to trust the CPU manufacturer.
      Signed-off-by: NTheodore Ts'o <tytso@mit.edu>
      39a8883a
  18. 18 7月, 2018 3 次提交
    • T
      random: Return nbytes filled from hw RNG · 753d433b
      Tobin C. Harding 提交于
      Currently the function get_random_bytes_arch() has return value 'void'.
      If the hw RNG fails we currently fall back to using get_random_bytes().
      This defeats the purpose of requesting random material from the hw RNG
      in the first place.
      
      There are currently no intree users of get_random_bytes_arch().
      
      Only get random bytes from the hw RNG, make function return the number
      of bytes retrieved from the hw RNG.
      Acked-by: NTheodore Ts'o <tytso@mit.edu>
      Reviewed-by: NSteven Rostedt (VMware) <rostedt@goodmis.org>
      Signed-off-by: NTobin C. Harding <me@tobin.cc>
      Signed-off-by: NTheodore Ts'o <tytso@mit.edu>
      753d433b
    • T
      random: Fix whitespace pre random-bytes work · 8ddd6efa
      Tobin C. Harding 提交于
      There are a couple of whitespace issues around the function
      get_random_bytes_arch().  In preparation for patching this function
      let's clean them up.
      Acked-by: NTheodore Ts'o <tytso@mit.edu>
      Signed-off-by: NTobin C. Harding <me@tobin.cc>
      Signed-off-by: NTheodore Ts'o <tytso@mit.edu>
      8ddd6efa
    • T
      random: mix rdrand with entropy sent in from userspace · 81e69df3
      Theodore Ts'o 提交于
      Fedora has integrated the jitter entropy daemon to work around slow
      boot problems, especially on VM's that don't support virtio-rng:
      
          https://bugzilla.redhat.com/show_bug.cgi?id=1572944
      
      It's understandable why they did this, but the Jitter entropy daemon
      works fundamentally on the principle: "the CPU microarchitecture is
      **so** complicated and we can't figure it out, so it *must* be
      random".  Yes, it uses statistical tests to "prove" it is secure, but
      AES_ENCRYPT(NSA_KEY, COUNTER++) will also pass statistical tests with
      flying colors.
      
      So if RDRAND is available, mix it into entropy submitted from
      userspace.  It can't hurt, and if you believe the NSA has backdoored
      RDRAND, then they probably have enough details about the Intel
      microarchitecture that they can reverse engineer how the Jitter
      entropy daemon affects the microarchitecture, and attack its output
      stream.  And if RDRAND is in fact an honest DRNG, it will immeasurably
      improve on what the Jitter entropy daemon might produce.
      
      This also provides some protection against someone who is able to read
      or set the entropy seed file.
      Signed-off-by: NTheodore Ts'o <tytso@mit.edu>
      Cc: stable@vger.kernel.org
      Cc: Arnd Bergmann <arnd@arndb.de>
      81e69df3
  19. 29 6月, 2018 1 次提交
    • L
      Revert changes to convert to ->poll_mask() and aio IOCB_CMD_POLL · a11e1d43
      Linus Torvalds 提交于
      The poll() changes were not well thought out, and completely
      unexplained.  They also caused a huge performance regression, because
      "->poll()" was no longer a trivial file operation that just called down
      to the underlying file operations, but instead did at least two indirect
      calls.
      
      Indirect calls are sadly slow now with the Spectre mitigation, but the
      performance problem could at least be largely mitigated by changing the
      "->get_poll_head()" operation to just have a per-file-descriptor pointer
      to the poll head instead.  That gets rid of one of the new indirections.
      
      But that doesn't fix the new complexity that is completely unwarranted
      for the regular case.  The (undocumented) reason for the poll() changes
      was some alleged AIO poll race fixing, but we don't make the common case
      slower and more complex for some uncommon special case, so this all
      really needs way more explanations and most likely a fundamental
      redesign.
      
      [ This revert is a revert of about 30 different commits, not reverted
        individually because that would just be unnecessarily messy  - Linus ]
      
      Cc: Al Viro <viro@zeniv.linux.org.uk>
      Cc: Christoph Hellwig <hch@lst.de>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      a11e1d43
  20. 26 5月, 2018 1 次提交
  21. 25 4月, 2018 2 次提交
  22. 14 4月, 2018 5 次提交
  23. 01 3月, 2018 3 次提交
  24. 12 2月, 2018 1 次提交
    • L
      vfs: do bulk POLL* -> EPOLL* replacement · a9a08845
      Linus Torvalds 提交于
      This is the mindless scripted replacement of kernel use of POLL*
      variables as described by Al, done by this script:
      
          for V in IN OUT PRI ERR RDNORM RDBAND WRNORM WRBAND HUP RDHUP NVAL MSG; do
              L=`git grep -l -w POLL$V | grep -v '^t' | grep -v /um/ | grep -v '^sa' | grep -v '/poll.h$'|grep -v '^D'`
              for f in $L; do sed -i "-es/^\([^\"]*\)\(\<POLL$V\>\)/\\1E\\2/" $f; done
          done
      
      with de-mangling cleanups yet to come.
      
      NOTE! On almost all architectures, the EPOLL* constants have the same
      values as the POLL* constants do.  But they keyword here is "almost".
      For various bad reasons they aren't the same, and epoll() doesn't
      actually work quite correctly in some cases due to this on Sparc et al.
      
      The next patch from Al will sort out the final differences, and we
      should be all done.
      Scripted-by: NAl Viro <viro@zeniv.linux.org.uk>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      a9a08845