1. 07 8月, 2013 2 次提交
    • C
      9b6846ce
    • C
      tile: support LSI MEGARAID SAS HBA hybrid dma_ops · 803c874a
      Chris Metcalf 提交于
      The LSI MEGARAID SAS HBA suffers from the problem where it can do
      64-bit DMA to streaming buffers but not to consistent buffers.
      In other words, 64-bit DMA is used for disk data transfers and 32-bit
      DMA must be used for control message transfers. According to LSI,
      the firmware is not fully functional yet. This change implements a
      kind of hybrid dma_ops to support this.
      
      Note that on most other platforms, the 64-bit DMA addressing space is the
      same as the 32-bit DMA space and they overlap the physical memory space.
      No special arrangement is needed to support this kind of mixed DMA
      capability.  On TILE-Gx, the 64-bit DMA space is completely separate
      from the 32-bit DMA space.  Due to the use of the IOMMU, the 64-bit DMA
      space doesn't overlap the physical memory space.  On the other hand,
      the 32-bit DMA space overlaps the physical memory space under 4GB.
      The separate address spaces make it necessary to have separate dma_ops.
      Signed-off-by: NChris Metcalf <cmetcalf@tilera.com>
      803c874a
  2. 19 7月, 2012 3 次提交
    • C
      tile pci: enable IOMMU to support DMA for legacy devices · 41bb38fc
      Chris Metcalf 提交于
      This change uses the TRIO IOMMU to map the PCI DMA space and physical
      memory at different addresses.  We also now use the dma_mapping_ops
      to provide support for non-PCI DMA, PCIe DMA (64-bit) and legacy PCI
      DMA (32-bit).  We use the kernel's software I/O TLB framework
      (i.e. bounce buffers) for the legacy 32-bit PCI device support since
      there are a limited number of TLB entries in the IOMMU and it is
      non-trivial to handle indexing, searching, matching, etc.  For 32-bit
      devices the performance impact of bounce buffers should not be a concern.
      Signed-off-by: NChris Metcalf <cmetcalf@tilera.com>
      41bb38fc
    • C
      arch/tile: enable ZONE_DMA for tilegx · eef015c8
      Chris Metcalf 提交于
      This is required for PCI root complex legacy support and USB OHCI root
      complex support.  With this change tilegx now supports allocating memory
      whose PA fits in 32 bits.
      Signed-off-by: NChris Metcalf <cmetcalf@tilera.com>
      eef015c8
    • C
      tilegx pci: support I/O to arbitrarily-cached pages · bbaa22c3
      Chris Metcalf 提交于
      The tilegx PCI root complex support (currently only in linux-next)
      is limited to pages that are homed on cached in the default manner,
      i.e. "hash-for-home".  This change supports delivery of I/O data to
      pages that are cached in other ways (locally on a particular core,
      uncached, user-managed incoherent, etc.).
      
      A large part of the change is supporting flushing pages from cache
      on particular homes so that we can transition the data that we are
      delivering to or from the device appropriately.  The new homecache_finv*
      routines handle this.
      
      Some changes to page_table_range_init() were also required to make
      the fixmap code work correctly on tilegx; it hadn't been used there
      before.
      
      We also remove some stub mark_caches_evicted_*() routines that
      were just no-ops anyway.
      Signed-off-by: NChris Metcalf <cmetcalf@tilera.com>
      bbaa22c3
  3. 04 12月, 2011 1 次提交
  4. 05 5月, 2011 1 次提交
  5. 11 3月, 2011 1 次提交
    • C
      arch/tile: support 4KB page size as well as 64KB · 76c567fb
      Chris Metcalf 提交于
      The Tilera architecture traditionally supports 64KB page sizes
      to improve TLB utilization and improve performance when the
      hardware is being used primarily to run a single application.
      
      For more generic server scenarios, it can be beneficial to run
      with 4KB page sizes, so this commit allows that to be specified
      (by modifying the arch/tile/include/hv/pagesize.h header).
      
      As part of this change, we also re-worked the PTE management
      slightly so that PTE writes all go through a __set_pte() function
      where we can do some additional validation.  The set_pte_order()
      function was eliminated since the "order" argument wasn't being used.
      
      One bug uncovered was in the PCI DMA code, which wasn't properly
      flushing the specified range.  This was benign with 64KB pages,
      but with 4KB pages we were getting some larger flushes wrong.
      
      The per-cpu memory reservation code also needed updating to
      conform with the newer percpu stuff; before it always chose 64KB,
      and that was always correct, but with 4KB granularity we now have
      to pay closer attention and reserve the amount of memory that will
      be requested when the percpu code starts allocating.
      Signed-off-by: NChris Metcalf <cmetcalf@tilera.com>
      76c567fb
  6. 05 6月, 2010 2 次提交