1. 31 3月, 2018 24 次提交
  2. 23 3月, 2018 11 次提交
  3. 16 3月, 2018 5 次提交
    • A
      crypto: arm64/sha256-neon - play nice with CONFIG_PREEMPT kernels · 1a3713c7
      Ard Biesheuvel 提交于
      Tweak the SHA256 update routines to invoke the SHA256 block transform
      block by block, to avoid excessive scheduling delays caused by the
      NEON algorithm running with preemption disabled.
      
      Also, remove a stale comment which no longer applies now that kernel
      mode NEON is actually disallowed in some contexts.
      Signed-off-by: NArd Biesheuvel <ard.biesheuvel@linaro.org>
      Signed-off-by: NHerbert Xu <herbert@gondor.apana.org.au>
      1a3713c7
    • A
      crypto: arm64/aes-blk - add 4 way interleave to CBC-MAC encrypt path · 870c163a
      Ard Biesheuvel 提交于
      CBC MAC is strictly sequential, and so the current AES code simply
      processes the input one block at a time. However, we are about to add
      yield support, which adds a bit of overhead, and which we prefer to
      align with other modes in terms of granularity (i.e., it is better to
      have all routines yield every 64 bytes and not have an exception for
      CBC MAC which yields every 16 bytes)
      
      So unroll the loop by 4. We still cannot perform the AES algorithm in
      parallel, but we can at least merge the loads and stores.
      Signed-off-by: NArd Biesheuvel <ard.biesheuvel@linaro.org>
      Signed-off-by: NHerbert Xu <herbert@gondor.apana.org.au>
      870c163a
    • A
      crypto: arm64/aes-blk - add 4 way interleave to CBC encrypt path · a8f8a69e
      Ard Biesheuvel 提交于
      CBC encryption is strictly sequential, and so the current AES code
      simply processes the input one block at a time. However, we are
      about to add yield support, which adds a bit of overhead, and which
      we prefer to align with other modes in terms of granularity (i.e.,
      it is better to have all routines yield every 64 bytes and not have
      an exception for CBC encrypt which yields every 16 bytes)
      
      So unroll the loop by 4. We still cannot perform the AES algorithm in
      parallel, but we can at least merge the loads and stores.
      Signed-off-by: NArd Biesheuvel <ard.biesheuvel@linaro.org>
      Signed-off-by: NHerbert Xu <herbert@gondor.apana.org.au>
      a8f8a69e
    • A
      crypto: arm64/aes-blk - remove configurable interleave · 55868b45
      Ard Biesheuvel 提交于
      The AES block mode implementation using Crypto Extensions or plain NEON
      was written before real hardware existed, and so its interleave factor
      was made build time configurable (as well as an option to instantiate
      all interleaved sequences inline rather than as subroutines)
      
      We ended up using INTERLEAVE=4 with inlining disabled for both flavors
      of the core AES routines, so let's stick with that, and remove the option
      to configure this at build time. This makes the code easier to modify,
      which is nice now that we're adding yield support.
      Signed-off-by: NArd Biesheuvel <ard.biesheuvel@linaro.org>
      Signed-off-by: NHerbert Xu <herbert@gondor.apana.org.au>
      55868b45
    • A
      crypto: arm64/chacha20 - move kernel mode neon en/disable into loop · 4bf7e7a1
      Ard Biesheuvel 提交于
      When kernel mode NEON was first introduced on arm64, the preserve and
      restore of the userland NEON state was completely unoptimized, and
      involved saving all registers on each call to kernel_neon_begin(),
      and restoring them on each call to kernel_neon_end(). For this reason,
      the NEON crypto code that was introduced at the time keeps the NEON
      enabled throughout the execution of the crypto API methods, which may
      include calls back into the crypto API that could result in memory
      allocation or other actions that we should avoid when running with
      preemption disabled.
      
      Since then, we have optimized the kernel mode NEON handling, which now
      restores lazily (upon return to userland), and so the preserve action
      is only costly the first time it is called after entering the kernel.
      
      So let's put the kernel_neon_begin() and kernel_neon_end() calls around
      the actual invocations of the NEON crypto code, and run the remainder of
      the code with kernel mode NEON disabled (and preemption enabled)
      Signed-off-by: NArd Biesheuvel <ard.biesheuvel@linaro.org>
      Signed-off-by: NHerbert Xu <herbert@gondor.apana.org.au>
      4bf7e7a1