1. 26 5月, 2011 4 次提交
  2. 25 5月, 2011 2 次提交
    • P
      mm: now that all old mmu_gather code is gone, remove the storage · 1c395176
      Peter Zijlstra 提交于
      Fold all the mmu_gather rework patches into one for submission
      Signed-off-by: NPeter Zijlstra <a.p.zijlstra@chello.nl>
      Reported-by: NHugh Dickins <hughd@google.com>
      Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
      Cc: David Miller <davem@davemloft.net>
      Cc: Martin Schwidefsky <schwidefsky@de.ibm.com>
      Cc: Russell King <rmk@arm.linux.org.uk>
      Cc: Paul Mundt <lethal@linux-sh.org>
      Cc: Jeff Dike <jdike@addtoit.com>
      Cc: Richard Weinberger <richard@nod.at>
      Cc: Tony Luck <tony.luck@intel.com>
      Cc: KAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com>
      Cc: Mel Gorman <mel@csn.ul.ie>
      Cc: KOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com>
      Cc: Nick Piggin <npiggin@kernel.dk>
      Cc: Namhyung Kim <namhyung@gmail.com>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      1c395176
    • D
      arch, mm: filter disallowed nodes from arch specific show_mem functions · 7bf02ea2
      David Rientjes 提交于
      Architectures that implement their own show_mem() function did not pass
      the filter argument to show_free_areas() to appropriately avoid emitting
      the state of nodes that are disallowed in the current context.  This patch
      now passes the filter argument to show_free_areas() so those nodes are now
      avoided.
      
      This patch also removes the show_free_areas() wrapper around
      __show_free_areas() and converts existing callers to pass an empty filter.
      
      ia64 emits additional information for each node, so skip_free_areas_zone()
      must be made global to filter disallowed nodes and it is converted to use
      a nid argument rather than a zone for this use case.
      Signed-off-by: NDavid Rientjes <rientjes@google.com>
      Cc: Russell King <linux@arm.linux.org.uk>
      Cc: Tony Luck <tony.luck@intel.com>
      Cc: Fenghua Yu <fenghua.yu@intel.com>
      Cc: Kyle McMartin <kyle@mcmartin.ca>
      Cc: Helge Deller <deller@gmx.de>
      Cc: James Bottomley <jejb@parisc-linux.org>
      Cc: "David S. Miller" <davem@davemloft.net>
      Cc: Guan Xuetao <gxt@mprc.pku.edu.cn>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      7bf02ea2
  3. 21 5月, 2011 1 次提交
  4. 16 5月, 2011 1 次提交
  5. 12 5月, 2011 2 次提交
  6. 04 5月, 2011 1 次提交
  7. 28 4月, 2011 1 次提交
  8. 14 4月, 2011 1 次提交
  9. 02 4月, 2011 1 次提交
  10. 31 3月, 2011 1 次提交
  11. 25 3月, 2011 1 次提交
  12. 10 3月, 2011 1 次提交
  13. 09 3月, 2011 1 次提交
  14. 08 3月, 2011 1 次提交
  15. 24 2月, 2011 3 次提交
    • R
      ARM: move L1_CACHE_SHIFT_6 to mm/Kconfig · 5637a126
      Russell King 提交于
      Move L1_CACHE_SHIFT related options together, rather than spreading them
      across two separate Kconfig files.
      Signed-off-by: NRussell King <rmk+kernel@arm.linux.org.uk>
      5637a126
    • N
      ARM: 6639/1: allow highmem on SMP platforms without h/w TLB ops broadcast · aaa50048
      Nicolas Pitre 提交于
      In commit e616c591, highmem support was
      deactivated for SMP platforms without hardware TLB ops broadcast because
      usage of kmap_high_get() requires that IRQs be disabled when kmap_lock
      is locked which is incompatible with the IPI mechanism used by the
      software TLB ops broadcast invoked through flush_all_zero_pkmaps().
      
      The reason for kmap_high_get() is to ensure that the currently kmap'd
      page usage count does not decrease to zero while we're using its
      existing virtual mapping in an atomic context.  With a VIVT cache this
      is essential to do due to cache coherency issues, but with a VIPT cache
      this is only an optimization so not to pay the price of establishing a
      second mapping if an existing one can be used.  However, on VIPT
      platforms without hardware TLB maintenance we can give up on that
      optimization in order to be able to use highmem.
      
      From ARMv7 onwards the TLB ops are broadcasted in hardware, so let's
      disable ARCH_NEEDS_KMAP_HIGH_GET only when CONFIG_SMP and
      CONFIG_CPU_TLB_V6 are defined.
      Signed-off-by: NNicolas Pitre <nicolas.pitre@linaro.org>
      Tested-by: NSaeed Bishara <saeed.bishara@gmail.com>
      Signed-off-by: NRussell King <rmk+kernel@arm.linux.org.uk>
      aaa50048
    • R
      ARM: DMA: top-down allocation in DMA coherent region · 459c1517
      Russell King 提交于
      Achieve better usage of the DMA coherent region by doing top-down
      allocation rather than bottom up.  If we ask for a 128kB allocation,
      this will be aligned to 128kB and satisfied from the very bottom
      address.  If we then ask for a 600kB allocation, this will be aligned
      to 1MB, and we will have a 896kB hole.
      
      Performing top-down allocation resolves this by allocating the 128kB
      at the very top, and then the 600kB can come in below it without any
      unnecessary wastage.
      
      This problem was reported by Janusz Krzysztofik, who had 2 x 128kB +
      1 x 640kB allocations which wouldn't fit into 1MB.
      Tested-by: NJanusz Krzysztofik <jkrzyszt@tis.icnet.pl>
      Signed-off-by: NRussell King <rmk+kernel@arm.linux.org.uk>
      459c1517
  16. 23 2月, 2011 1 次提交
  17. 22 2月, 2011 3 次提交
  18. 19 2月, 2011 2 次提交
  19. 15 2月, 2011 3 次提交
  20. 10 2月, 2011 2 次提交
  21. 03 2月, 2011 7 次提交