1. 13 4月, 2010 3 次提交
  2. 11 12月, 2009 2 次提交
  3. 09 12月, 2009 1 次提交
  4. 27 11月, 2009 1 次提交
  5. 07 11月, 2009 1 次提交
  6. 30 5月, 2009 1 次提交
  7. 09 2月, 2009 2 次提交
    • D
      sparc64: Fix probe_kernel_{read,write}(). · aeb39876
      David S. Miller 提交于
      This is based upon a report from Chris Torek and his initial patch.
      From Chris's report:
      
      --------------------
      This came up in testing kgdb, using the built-in tests -- turn
      on CONFIG_KGDB_TESTS, then
      
          echo V1 > /sys/module/kgdbts/parameters/kgdbts
      
      -- but it would affect using kgdb if you were debugging and looking
      at bad pointers.
      --------------------
      
      When we get a copy_{from,to}_user() request and the %asi is set to
      something other than ASI_AIUS (which is userspace) then we branch off
      to a routine called memcpy_user_stub().  It just does a straight
      memcpy since we are copying from kernel to kernel in this case.
      
      The logic was that since source and destination are both kernel
      pointers we don't need to have exception checks.
      
      But for what probe_kernel_{read,write}() is trying to do, we have to
      have the checks, otherwise things like kgdb bad kernel pointer
      accesses don't do the right thing.
      Signed-off-by: NDavid S. Miller <davem@davemloft.net>
      aeb39876
    • D
      sparc64: Kill .fixup section bloat. · 40bdac7d
      David S. Miller 提交于
      This is an implementation of a suggestion made by Chris Torek:
      --------------------
      Something else I noticed in passing: the EX and EX_LD/EX_ST macros
      scattered throughout the various .S files make a fair bit of .fixup
      code, all of which does the same thing.  At the cost of one symbol
      in copy_in_user.S, you could just have one common two-instruction
      retl-and-mov-1 fixup that they all share.
      --------------------
      
      The following is with a defconfig build:
      
         text	   data	    bss	    dec	    hex	filename
      3972767	 344024	 584449	4901240	 4ac978	vmlinux.orig
      39688877	 344024	 584449	4897360	 4aba50	vmlinux
      Signed-off-by: NDavid S. Miller <davem@davemloft.net>
      40bdac7d
  8. 18 1月, 2009 1 次提交
    • S
      sparc64: fix modpost failure · 62dfcd33
      Sam Ravnborg 提交于
      Previously PeeCeeI.o was a library but it
      was always pulled in due to insw and friends being exported
      (at least for a modular kernel).
      
      But this resulted in modpost failures if there where no in-kernel
      users because then insw & friends were not linked in.
      
      Fix this by including PeeCeeI.o in the kernel unconditionally.
      
      The only drawback for this solution is that a nonmodular kernel
      will always include insw & friends no matter if they are in use or not.
      Reported-by: NMeelis Roos <mroos@linux.ee>
      Signed-off-by: NSam Ravnborg <sam@ravnborg.org>
      Signed-off-by: NDavid S. Miller <davem@davemloft.net>
      62dfcd33
  9. 09 1月, 2009 2 次提交
  10. 09 12月, 2008 3 次提交
  11. 05 12月, 2008 2 次提交
  12. 20 5月, 2008 1 次提交
  13. 29 4月, 2008 1 次提交
  14. 01 2月, 2008 1 次提交
  15. 27 8月, 2007 1 次提交
  16. 25 7月, 2007 1 次提交
    • A
      [SPARC32]: Fix bug in sparc optimized memset. · f61698e6
      Alexander Shmelev 提交于
      Sparc optimized memset (arch/sparc/lib/memset.S) does not fill last
      byte of the memory area, if area size is less than 8 bytes and start
      address is not word (4-bytes) aligned.
      
      Here is code chunk where bug located:
      /* %o0 - memory address, %o1 - size, %g3 - value */
      8:
           add    %o0, 1, %o0
          subcc    %o1, 1, %o1
          bne,a    8b
           stb %g3, [%o0 - 1]
      
      This code should write byte every loop iteration, but last time delay
      instruction stb is not executed because branch instruction sets
      "annul" bit.
      
      Patch replaces bne,a by bne instruction.
      
      Error can be reproduced by simple kernel module:
      
      --------------------
      #include <linux/module.h>
      #include <linux/config.h>
      #include <linux/kernel.h>
      #include <linux/errno.h>
      #include <string.h>
      
      static void do_memset(void **p, int size)
      {
              memset(p, 0x00, size);
      }
      
      static int __init memset_test_init(void)
      {
          char fooc[8];
          int *fooi;
          memset(fooc, 0xba, sizeof(fooc));
      
          do_memset((void**)(fooc + 3), 1);
      
          fooi = (int*) fooc;
          printk("%08X %08X\n", fooi[0], fooi[1]);
      
          return -1;
      }
      
      static void __exit memset_test_cleanup(void)
      {
          return;
      }
      
      module_init(memset_test_init);
      module_exit(memset_test_cleanup);
      
      MODULE_LICENSE("GPL");
      EXPORT_NO_SYMBOLS;
      --------------------
      Signed-off-by: NAlexander Shmelev <ashmelev@task.sun.mcst.ru>
      Signed-off-by: NDavid S. Miller <davem@davemloft.net>
      f61698e6
  17. 31 5月, 2007 1 次提交
  18. 29 5月, 2007 1 次提交
  19. 09 5月, 2007 1 次提交
  20. 27 3月, 2007 1 次提交
  21. 18 12月, 2006 1 次提交
  22. 09 10月, 2006 1 次提交
    • A
      [PATCH] sparc32 rwlock fix · 7a39f522
      Al Viro 提交于
      read_trylock() is broken on sparc32 (doesn't build and didn't work
      right, actually).  Proposed fix:
      
       - make "writer holds lock" distinguishable from "reader tries to grab
         lock"
      
       - have __raw_read_trylock() try to acquire the mutex (in LSB of lock),
         terminating spin if we see that there's writer holding it.  Then do
         the rest as we do in read_lock().
      
      Thanks to Ingo for discussion...
      Signed-off-by: NAl Viro <viro@zeniv.linux.org.uk>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      7a39f522
  23. 30 9月, 2006 1 次提交
  24. 01 7月, 2006 1 次提交
  25. 26 6月, 2006 1 次提交
  26. 13 12月, 2005 1 次提交
  27. 14 11月, 2005 3 次提交
  28. 12 9月, 2005 1 次提交
  29. 11 9月, 2005 1 次提交
    • I
      [PATCH] spinlock consolidation · fb1c8f93
      Ingo Molnar 提交于
      This patch (written by me and also containing many suggestions of Arjan van
      de Ven) does a major cleanup of the spinlock code.  It does the following
      things:
      
       - consolidates and enhances the spinlock/rwlock debugging code
      
       - simplifies the asm/spinlock.h files
      
       - encapsulates the raw spinlock type and moves generic spinlock
         features (such as ->break_lock) into the generic code.
      
       - cleans up the spinlock code hierarchy to get rid of the spaghetti.
      
      Most notably there's now only a single variant of the debugging code,
      located in lib/spinlock_debug.c.  (previously we had one SMP debugging
      variant per architecture, plus a separate generic one for UP builds)
      
      Also, i've enhanced the rwlock debugging facility, it will now track
      write-owners.  There is new spinlock-owner/CPU-tracking on SMP builds too.
      All locks have lockup detection now, which will work for both soft and hard
      spin/rwlock lockups.
      
      The arch-level include files now only contain the minimally necessary
      subset of the spinlock code - all the rest that can be generalized now
      lives in the generic headers:
      
       include/asm-i386/spinlock_types.h       |   16
       include/asm-x86_64/spinlock_types.h     |   16
      
      I have also split up the various spinlock variants into separate files,
      making it easier to see which does what. The new layout is:
      
         SMP                         |  UP
         ----------------------------|-----------------------------------
         asm/spinlock_types_smp.h    |  linux/spinlock_types_up.h
         linux/spinlock_types.h      |  linux/spinlock_types.h
         asm/spinlock_smp.h          |  linux/spinlock_up.h
         linux/spinlock_api_smp.h    |  linux/spinlock_api_up.h
         linux/spinlock.h            |  linux/spinlock.h
      
      /*
       * here's the role of the various spinlock/rwlock related include files:
       *
       * on SMP builds:
       *
       *  asm/spinlock_types.h: contains the raw_spinlock_t/raw_rwlock_t and the
       *                        initializers
       *
       *  linux/spinlock_types.h:
       *                        defines the generic type and initializers
       *
       *  asm/spinlock.h:       contains the __raw_spin_*()/etc. lowlevel
       *                        implementations, mostly inline assembly code
       *
       *   (also included on UP-debug builds:)
       *
       *  linux/spinlock_api_smp.h:
       *                        contains the prototypes for the _spin_*() APIs.
       *
       *  linux/spinlock.h:     builds the final spin_*() APIs.
       *
       * on UP builds:
       *
       *  linux/spinlock_type_up.h:
       *                        contains the generic, simplified UP spinlock type.
       *                        (which is an empty structure on non-debug builds)
       *
       *  linux/spinlock_types.h:
       *                        defines the generic type and initializers
       *
       *  linux/spinlock_up.h:
       *                        contains the __raw_spin_*()/etc. version of UP
       *                        builds. (which are NOPs on non-debug, non-preempt
       *                        builds)
       *
       *   (included on UP-non-debug builds:)
       *
       *  linux/spinlock_api_up.h:
       *                        builds the _spin_*() APIs.
       *
       *  linux/spinlock.h:     builds the final spin_*() APIs.
       */
      
      All SMP and UP architectures are converted by this patch.
      
      arm, i386, ia64, ppc, ppc64, s390/s390x, x64 was build-tested via
      crosscompilers.  m32r, mips, sh, sparc, have not been tested yet, but should
      be mostly fine.
      
      From: Grant Grundler <grundler@parisc-linux.org>
      
        Booted and lightly tested on a500-44 (64-bit, SMP kernel, dual CPU).
        Builds 32-bit SMP kernel (not booted or tested).  I did not try to build
        non-SMP kernels.  That should be trivial to fix up later if necessary.
      
        I converted bit ops atomic_hash lock to raw_spinlock_t.  Doing so avoids
        some ugly nesting of linux/*.h and asm/*.h files.  Those particular locks
        are well tested and contained entirely inside arch specific code.  I do NOT
        expect any new issues to arise with them.
      
       If someone does ever need to use debug/metrics with them, then they will
        need to unravel this hairball between spinlocks, atomic ops, and bit ops
        that exist only because parisc has exactly one atomic instruction: LDCW
        (load and clear word).
      
      From: "Luck, Tony" <tony.luck@intel.com>
      
         ia64 fix
      Signed-off-by: NIngo Molnar <mingo@elte.hu>
      Signed-off-by: NArjan van de Ven <arjanv@infradead.org>
      Signed-off-by: NGrant Grundler <grundler@parisc-linux.org>
      Cc: Matthew Wilcox <willy@debian.org>
      Signed-off-by: NHirokazu Takata <takata@linux-m32r.org>
      Signed-off-by: NMikael Pettersson <mikpe@csd.uu.se>
      Signed-off-by: NBenoit Boissinot <benoit.boissinot@ens-lyon.org>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      fb1c8f93
  30. 10 9月, 2005 1 次提交