1. 23 2月, 2015 1 次提交
    • B
      x86/alternatives: Add instruction padding · 4332195c
      Borislav Petkov 提交于
      Up until now we have always paid attention to make sure the length of
      the new instruction replacing the old one is at least less or equal to
      the length of the old instruction. If the new instruction is longer, at
      the time it replaces the old instruction it will overwrite the beginning
      of the next instruction in the kernel image and cause your pants to
      catch fire.
      
      So instead of having to pay attention, teach the alternatives framework
      to pad shorter old instructions with NOPs at buildtime - but only in the
      case when
      
        len(old instruction(s)) < len(new instruction(s))
      
      and add nothing in the >= case. (In that case we do add_nops() when
      patching).
      
      This way the alternatives user shouldn't have to care about instruction
      sizes and simply use the macros.
      
      Add asm ALTERNATIVE* flavor macros too, while at it.
      
      Also, we need to save the pad length in a separate struct alt_instr
      member for NOP optimization and the way to do that reliably is to carry
      the pad length instead of trying to detect whether we're looking at
      single-byte NOPs or at pathological instruction offsets like e9 90 90 90
      90, for example, which is a valid instruction.
      
      Thanks to Michael Matz for the great help with toolchain questions.
      Signed-off-by: NBorislav Petkov <bp@suse.de>
      4332195c
  2. 14 2月, 2015 1 次提交
    • A
      x86_64: kasan: add interceptors for memset/memmove/memcpy functions · 393f203f
      Andrey Ryabinin 提交于
      Recently instrumentation of builtin functions calls was removed from GCC
      5.0.  To check the memory accessed by such functions, userspace asan
      always uses interceptors for them.
      
      So now we should do this as well.  This patch declares
      memset/memmove/memcpy as weak symbols.  In mm/kasan/kasan.c we have our
      own implementation of those functions which checks memory before accessing
      it.
      
      Default memset/memmove/memcpy now now always have aliases with '__'
      prefix.  For files that built without kasan instrumentation (e.g.
      mm/slub.c) original mem* replaced (via #define) with prefixed variants,
      cause we don't want to check memory accesses there.
      Signed-off-by: NAndrey Ryabinin <a.ryabinin@samsung.com>
      Cc: Dmitry Vyukov <dvyukov@google.com>
      Cc: Konstantin Serebryany <kcc@google.com>
      Cc: Dmitry Chernenkov <dmitryc@google.com>
      Signed-off-by: NAndrey Konovalov <adech.fo@gmail.com>
      Cc: Yuri Gribov <tetra2005@gmail.com>
      Cc: Konstantin Khlebnikov <koct9i@gmail.com>
      Cc: Sasha Levin <sasha.levin@oracle.com>
      Cc: Christoph Lameter <cl@linux.com>
      Cc: Joonsoo Kim <iamjoonsoo.kim@lge.com>
      Cc: Dave Hansen <dave.hansen@intel.com>
      Cc: Andi Kleen <andi@firstfloor.org>
      Cc: Ingo Molnar <mingo@elte.hu>
      Cc: Thomas Gleixner <tglx@linutronix.de>
      Cc: "H. Peter Anvin" <hpa@zytor.com>
      Cc: Christoph Lameter <cl@linux.com>
      Cc: Pekka Enberg <penberg@kernel.org>
      Cc: David Rientjes <rientjes@google.com>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      393f203f
  3. 15 4月, 2013 1 次提交
  4. 27 1月, 2012 2 次提交
  5. 18 5月, 2011 1 次提交
  6. 02 5月, 2011 1 次提交
  7. 24 8月, 2010 1 次提交
    • M
      x86, mem: Optimize memcpy by avoiding memory false dependece · 59daa706
      Ma Ling 提交于
      All read operations after allocation stage can run speculatively,
      all write operation will run in program order, and if addresses are
      different read may run before older write operation, otherwise wait
      until write commit. However CPU don't check each address bit,
      so read could fail to recognize different address even they
      are in different page.For example if rsi is 0xf004, rdi is 0xe008,
      in following operation there will generate big performance latency.
      1. movq (%rsi),	%rax
      2. movq %rax,	(%rdi)
      3. movq 8(%rsi), %rax
      4. movq %rax,	8(%rdi)
      
      If %rsi and rdi were in really the same meory page, there are TRUE
      read-after-write dependence because instruction 2 write 0x008 and
      instruction 3 read 0x00c, the two address are overlap partially.
      Actually there are in different page and no any issues,
      but without checking each address bit CPU could think they are
      in the same page, and instruction 3 have to wait for instruction 2
      to write data into cache from write buffer, then load data from cache,
      the cost time read spent is equal to mfence instruction. We may avoid it by
      tuning operation sequence as follow.
      
      1. movq 8(%rsi), %rax
      2. movq %rax,	8(%rdi)
      3. movq (%rsi),	%rax
      4. movq %rax,	(%rdi)
      
      Instruction 3 read 0x004, instruction 2 write address 0x010, no any
      dependence.  At last on Core2 we gain 1.83x speedup compared with
      original instruction sequence.  In this patch we first handle small
      size(less 20bytes), then jump to different copy mode. Based on our
      micro-benchmark small bytes from 1 to 127 bytes, we got up to 2X
      improvement, and up to 1.5X improvement for 1024 bytes on Corei7.  (We
      use our micro-benchmark, and will do further test according to your
      requirment)
      Signed-off-by: NMa Ling <ling.ma@intel.com>
      LKML-Reference: <1277753065-18610-1-git-send-email-ling.ma@intel.com>
      Signed-off-by: NH. Peter Anvin <hpa@zytor.com>
      59daa706
  8. 08 7月, 2010 1 次提交
    • H
      x86, alternatives: Use 16-bit numbers for cpufeature index · 83a7a2ad
      H. Peter Anvin 提交于
      We already have cpufeature indicies above 255, so use a 16-bit number
      for the alternatives index.  This consumes a padding field and so
      doesn't add any size, but it means that abusing the padding field to
      create assembly errors on overflow no longer works.  We can retain the
      test simply by redirecting it to the .discard section, however.
      
      [ v3: updated to include open-coded locations ]
      Signed-off-by: NH. Peter Anvin <hpa@linux.intel.com>
      LKML-Reference: <tip-f88731e3068f9d1392ba71cc9f50f035d26a0d4f@git.kernel.org>
      Signed-off-by: NH. Peter Anvin <hpa@zytor.com>
      83a7a2ad
  9. 30 12月, 2009 1 次提交
    • J
      x86-64: Modify memcpy()/memset() alternatives mechanism · 7269e881
      Jan Beulich 提交于
      In order to avoid unnecessary chains of branches, rather than
      implementing memcpy()/memset()'s access to their alternative
      implementations via a jump, patch the (larger) original function
      directly.
      
      The memcpy() part of this is slightly subtle: while alternative
      instruction patching does itself use memcpy(), with the
      replacement block being less than 64-bytes in size the main loop
      of the original function doesn't get used for copying memcpy_c()
      over memcpy(), and hence we can safely write over its beginning.
      
      Also note that the CFI annotations are fine for both variants of
      each of the functions.
      Signed-off-by: NJan Beulich <jbeulich@novell.com>
      Cc: Nick Piggin <npiggin@suse.de>
      Cc: Linus Torvalds <torvalds@linux-foundation.org>
      Cc: Andrew Morton <akpm@linux-foundation.org>
      LKML-Reference: <4B2BB8D30200007800026AF2@vpn.id2.novell.com>
      Signed-off-by: NIngo Molnar <mingo@elte.hu>
      7269e881
  10. 12 3月, 2009 2 次提交
  11. 11 10月, 2007 2 次提交
  12. 12 8月, 2007 1 次提交
  13. 04 10月, 2006 1 次提交
  14. 26 9月, 2006 1 次提交
  15. 05 2月, 2006 1 次提交
  16. 15 11月, 2005 1 次提交
  17. 17 4月, 2005 1 次提交
    • L
      Linux-2.6.12-rc2 · 1da177e4
      Linus Torvalds 提交于
      Initial git repository build. I'm not bothering with the full history,
      even though we have it. We can create a separate "historical" git
      archive of that later if we want to, and in the meantime it's about
      3.2GB when imported into git - space that would just make the early
      git days unnecessarily complicated, when we don't have a lot of good
      infrastructure for it.
      
      Let it rip!
      1da177e4