1. 31 1月, 2007 3 次提交
  2. 30 1月, 2007 1 次提交
  3. 27 1月, 2007 15 次提交
  4. 25 1月, 2007 1 次提交
  5. 23 1月, 2007 3 次提交
  6. 22 1月, 2007 7 次提交
  7. 18 1月, 2007 2 次提交
  8. 13 1月, 2007 1 次提交
  9. 12 1月, 2007 2 次提交
    • D
      [PATCH] Revert bd_mount_mutex back to a semaphore · f73ca1b7
      David Chinner 提交于
      Revert bd_mount_mutex back to a semaphore so that xfs_freeze -f /mnt/newtest;
      xfs_freeze -u /mnt/newtest works safely and doesn't produce lockdep warnings.
      
      (XFS unlocks the semaphore from a different task, by design.  The mutex
      code warns about this)
      Signed-off-by: NDave Chinner <dgc@sgi.com>
      Cc: Ingo Molnar <mingo@elte.hu>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      f73ca1b7
    • T
      [PATCH] NFS: Fix race in nfs_release_page() · e3db7691
      Trond Myklebust 提交于
          NFS: Fix race in nfs_release_page()
      
          invalidate_inode_pages2() may find the dirty bit has been set on a page
          owing to the fact that the page may still be mapped after it was locked.
          Only after the call to unmap_mapping_range() are we sure that the page
          can no longer be dirtied.
          In order to fix this, NFS has hooked the releasepage() method and tries
          to write the page out between the call to unmap_mapping_range() and the
          call to remove_mapping(). This, however leads to deadlocks in the page
          reclaim code, where the page may be locked without holding a reference
          to the inode or dentry.
      
          Fix is to add a new address_space_operation, launder_page(), which will
          attempt to write out a dirty page without releasing the page lock.
      Signed-off-by: NTrond Myklebust <Trond.Myklebust@netapp.com>
      
          Also, the bare SetPageDirty() can skew all sort of accounting leading to
          other nasties.
      
      [akpm@osdl.org: cleanup]
      Signed-off-by: NPeter Zijlstra <a.p.zijlstra@chello.nl>
      Cc: Trond Myklebust <Trond.Myklebust@netapp.com>
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      e3db7691
  10. 11 1月, 2007 1 次提交
  11. 10 1月, 2007 2 次提交
  12. 07 1月, 2007 1 次提交
    • L
      Revert "[PATCH] binfmt_elf: randomize PIE binaries (2nd try)" · 90cb28e8
      Linus Torvalds 提交于
      This reverts commit 59287c09.
      
      Hugh Dickins reports that it causes random failures on x86 with SuSE
      10.2, and points out
      
        "Isn't that randomization, anywhere from 0x10000 to ELF_ET_DYN_BASE,
         sure to place the ET_DYN from time to time just where the comment
         says it's trying to avoid? I assume that somehow results in the error
         reported."
      
      (where the comment in question is the existing comment in the source
      code about mmap/brk clashes).
      Suggested-by: NHugh Dickins <hugh@veritas.com>
      Acked-by: NMarcus Meissner <meissner@suse.de>
      Cc: Andrew Morton <akpm@osdl.org>
      Cc: Andi Kleen <ak@suse.de>
      Cc: Ingo Molnar <mingo@elte.hu>
      Cc: Dave Jones <davej@codemonkey.org.uk>
      Cc: Arjan van de Ven <arjan@linux.intel.com>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      90cb28e8
  13. 06 1月, 2007 1 次提交
    • E
      [PATCH] fix garbage instead of zeroes in UFS · d63b7090
      Evgeniy Dushistov 提交于
      Looks like this is the problem, which point Al Viro some time ago:
      
      ufs's get_block callback allocates 16k of disk at a time, and links that
      entire 16k into the file's metadata.  But because get_block is called for only
      a single buffer_head (a 2k buffer_head in this case?) we are only able to tell
      the VFS that this 2k is buffer_new().
      
      So when ufs_getfrag_block() is later called to map some more data in the file,
      and when that data resides within the remaining 14k of this fragment,
      ufs_getfrag_block() will incorrectly return a !buffer_new() buffer_head.
      
      I don't see _right_ way to do nullification of whole block, if use inode
      page cache, some pages may be outside of inode limits (inode size), and
      will be lost; if use blockdev page cache it is possible to zero real data,
      if later inode page cache will be used.
      
      The simpliest way, as can I see usage of block device page cache, but not only
      mark dirty, but also sync it during "nullification".  I use my simple tests
      collection, which I used for check that create,open,write,read,close works on
      ufs, and I see that this patch makes ufs code 18% slower then before.
      Signed-off-by: NAndrew Morton <akpm@osdl.org>
      Signed-off-by: NLinus Torvalds <torvalds@osdl.org>
      d63b7090