1. 06 4月, 2016 2 次提交
  2. 18 3月, 2016 1 次提交
    • J
      mm: introduce page reference manipulation functions · fe896d18
      Joonsoo Kim 提交于
      The success of CMA allocation largely depends on the success of
      migration and key factor of it is page reference count.  Until now, page
      reference is manipulated by direct calling atomic functions so we cannot
      follow up who and where manipulate it.  Then, it is hard to find actual
      reason of CMA allocation failure.  CMA allocation should be guaranteed
      to succeed so finding offending place is really important.
      
      In this patch, call sites where page reference is manipulated are
      converted to introduced wrapper function.  This is preparation step to
      add tracepoint to each page reference manipulation function.  With this
      facility, we can easily find reason of CMA allocation failure.  There is
      no functional change in this patch.
      
      In addition, this patch also converts reference read sites.  It will
      help a second step that renames page._count to something else and
      prevents later attempt to direct access to it (Suggested by Andrew).
      Signed-off-by: NJoonsoo Kim <iamjoonsoo.kim@lge.com>
      Acked-by: NMichal Nazarewicz <mina86@mina86.com>
      Acked-by: NVlastimil Babka <vbabka@suse.cz>
      Cc: Minchan Kim <minchan@kernel.org>
      Cc: Mel Gorman <mgorman@techsingularity.net>
      Cc: "Kirill A. Shutemov" <kirill.shutemov@linux.intel.com>
      Cc: Sergey Senozhatsky <sergey.senozhatsky.work@gmail.com>
      Cc: Steven Rostedt <rostedt@goodmis.org>
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      fe896d18
  3. 17 2月, 2016 1 次提交
  4. 22 12月, 2015 1 次提交
  5. 14 12月, 2015 3 次提交
  6. 06 12月, 2015 3 次提交
  7. 24 11月, 2015 1 次提交
  8. 16 10月, 2015 1 次提交
    • J
      drivers/net/intel: use napi_complete_done() · 32b3e08f
      Jesse Brandeburg 提交于
      As per Eric Dumazet's previous patches:
      (see commit (24d2e4a5) - tg3: use napi_complete_done())
      
      Quoting verbatim:
      Using napi_complete_done() instead of napi_complete() allows
      us to use /sys/class/net/ethX/gro_flush_timeout
      
      GRO layer can aggregate more packets if the flush is delayed a bit,
      without having to set too big coalescing parameters that impact
      latencies.
      </end quote>
      
      Tested
      configuration: low latency via ethtool -C ethx adaptive-rx off
      				rx-usecs 10 adaptive-tx off tx-usecs 15
      workload: streaming rx using netperf TCP_MAERTS
      
      igb:
      MIGRATED TCP MAERTS TEST from 0.0.0.0 (0.0.0.0) port 0 AF_INET to 10.0.0.1 () port 0 AF_INET : demo
      ...
      Interim result:  941.48 10^6bits/s over 1.000 seconds ending at 1440193171.589
      
      Alignment      Offset         Bytes    Bytes       Recvs   Bytes    Sends
      Local  Remote  Local  Remote  Xfered   Per                 Per
      Recv   Send    Recv   Send             Recv (avg)          Send (avg)
          8       8      0       0 1176930056  1475.36    797726   16384.00  71905
      
      MIGRATED TCP MAERTS TEST from 0.0.0.0 (0.0.0.0) port 0 AF_INET to 10.0.0.1 () port 0 AF_INET : demo
      ...
      Interim result:  941.49 10^6bits/s over 0.997 seconds ending at 1440193142.763
      
      Alignment      Offset         Bytes    Bytes       Recvs   Bytes    Sends
      Local  Remote  Local  Remote  Xfered   Per                 Per
      Recv   Send    Recv   Send             Recv (avg)          Send (avg)
          8       8      0       0 1175182320  50476.00     23282   16384.00  71816
      
      i40e:
      Hard to test because the traffic is incoming so fast (24Gb/s) that GRO
      always receives 87kB, even at the highest interrupt rate.
      
      Other drivers were only compile tested.
      Signed-off-by: NJesse Brandeburg <jesse.brandeburg@intel.com>
      Tested-by: NAndrew Bowers <andrewx.bowers@intel.com>
      Signed-off-by: NJeff Kirsher <jeffrey.t.kirsher@intel.com>
      32b3e08f
  9. 14 10月, 2015 1 次提交
  10. 23 9月, 2015 1 次提交
  11. 16 9月, 2015 2 次提交
  12. 22 8月, 2015 1 次提交
    • M
      mm: make page pfmemalloc check more robust · 2f064f34
      Michal Hocko 提交于
      Commit c48a11c7 ("netvm: propagate page->pfmemalloc to skb") added
      checks for page->pfmemalloc to __skb_fill_page_desc():
      
              if (page->pfmemalloc && !page->mapping)
                      skb->pfmemalloc = true;
      
      It assumes page->mapping == NULL implies that page->pfmemalloc can be
      trusted.  However, __delete_from_page_cache() can set set page->mapping
      to NULL and leave page->index value alone.  Due to being in union, a
      non-zero page->index will be interpreted as true page->pfmemalloc.
      
      So the assumption is invalid if the networking code can see such a page.
      And it seems it can.  We have encountered this with a NFS over loopback
      setup when such a page is attached to a new skbuf.  There is no copying
      going on in this case so the page confuses __skb_fill_page_desc which
      interprets the index as pfmemalloc flag and the network stack drops
      packets that have been allocated using the reserves unless they are to
      be queued on sockets handling the swapping which is the case here and
      that leads to hangs when the nfs client waits for a response from the
      server which has been dropped and thus never arrive.
      
      The struct page is already heavily packed so rather than finding another
      hole to put it in, let's do a trick instead.  We can reuse the index
      again but define it to an impossible value (-1UL).  This is the page
      index so it should never see the value that large.  Replace all direct
      users of page->pfmemalloc by page_is_pfmemalloc which will hide this
      nastiness from unspoiled eyes.
      
      The information will get lost if somebody wants to use page->index
      obviously but that was the case before and the original code expected
      that the information should be persisted somewhere else if that is
      really needed (e.g.  what SLAB and SLUB do).
      
      [akpm@linux-foundation.org: fix blooper in slub]
      Fixes: c48a11c7 ("netvm: propagate page->pfmemalloc to skb")
      Signed-off-by: NMichal Hocko <mhocko@suse.com>
      Debugged-by: NVlastimil Babka <vbabka@suse.com>
      Debugged-by: NJiri Bohac <jbohac@suse.com>
      Cc: Eric Dumazet <eric.dumazet@gmail.com>
      Cc: David Miller <davem@davemloft.net>
      Acked-by: NMel Gorman <mgorman@suse.de>
      Cc: <stable@vger.kernel.org>	[3.6+]
      Signed-off-by: NAndrew Morton <akpm@linux-foundation.org>
      Signed-off-by: NLinus Torvalds <torvalds@linux-foundation.org>
      2f064f34
  13. 18 6月, 2015 1 次提交
  14. 04 5月, 2015 1 次提交
  15. 15 4月, 2015 4 次提交
  16. 03 3月, 2015 3 次提交
  17. 23 1月, 2015 2 次提交
  18. 14 1月, 2015 1 次提交
  19. 12 12月, 2014 1 次提交
    • A
      fm10k/igb/ixgbe: Use dma_rmb on Rx descriptor reads · 124b74c1
      Alexander Duyck 提交于
      This change makes it so that dma_rmb is used when reading the Rx
      descriptor.  The advantage of dma_rmb is that it allows for a much
      lower cost barrier on x86, powerpc, arm, and arm64 architectures than a
      traditional memory barrier when dealing with reads that only have to
      synchronize to coherent memory.
      
      In addition I have updated the code so that it just checks to see if any
      bits have been set instead of just the DD bit since the DD bit will always
      be set as a part of a descriptor write-back so we just need to check for a
      non-zero value being present at that memory location rather than just
      checking for any specific bit.  This allows the code itself to appear much
      cleaner and allows the compiler more room to optimize.
      
      Cc: Matthew Vick <matthew.vick@intel.com>
      Cc: Don Skidmore <donald.c.skidmore@intel.com>
      Acked-by: NJeff Kirsher <jeffrey.t.kirsher@intel.com>
      Signed-off-by: NAlexander Duyck <alexander.h.duyck@redhat.com>
      Signed-off-by: NDavid S. Miller <davem@davemloft.net>
      124b74c1
  20. 11 12月, 2014 1 次提交
  21. 09 12月, 2014 1 次提交
  22. 12 11月, 2014 1 次提交
  23. 15 10月, 2014 1 次提交
  24. 11 10月, 2014 1 次提交
  25. 23 9月, 2014 4 次提交
    • A
      fm10k: Add support for PTP · a211e013
      Alexander Duyck 提交于
      This change adds support for the Linux PTP Hardware clock and timestamping
      functionality provided by the hardware.  There are actually two cases that
      this timestamping is meant to support.
      
      The first case would be an ordinary clock scenario.  In this configuration
      the host interface does not have access to BAR 4.  However all of the host
      interfaces should be locked into the same boundary clock region and as such
      they are all on the same clock anyway.  With this being the case they can
      synchronize among themselves and only need to adjust the offset since they
      are all on the same clock with the same frequency.
      
      The second case is a boundary clock scenario.  This is a special case and
      would require both BAR 4 access, and a means of presenting a netdev per
      boundary region.  The current plan is to use DSA at some point in the
      future to provide these interfaces, but the DSA portion is still under
      development.
      Signed-off-by: NAlexander Duyck <alexander.h.duyck@intel.com>
      Acked-by: NRichard Cochran <richardcochran@gmail.com>
      Signed-off-by: NJeff Kirsher <jeffrey.t.kirsher@intel.com>
      a211e013
    • A
      fm10k: Add support for debugfs · 7461fd91
      Alexander Duyck 提交于
      This patch adds limited debugfs support for the driver.  Most of the
      functionality needed for dumping registers is already provided via ethtool.
      The only thing we saw that we really neeed was the ability to dump the
      descriptor rings so as such this patch will add a fm10k directory containing a
      listing of directories each one with a unique PCI Bus, Device, and Function
      number.  Each of those BDF directories will have a list of q_vectors, and
      the q_vectors will contain a file for each of the Rx/Tx rings that are a part
      of the vector.  For example:
      
      # ls -RD /sys/kernel/debug/fm10k/
      /sys/kernel/debug/fm10k/:
      0000:01:00.0
      
      /sys/kernel/debug/fm10k/0000:01:00.0:
      q_vector.000  q_vector.001  q_vector.002  q_vector.003
      
      /sys/kernel/debug/fm10k/0000:01:00.0/q_vector.000:
      rx_ring.000  tx_ring.000
      
      /sys/kernel/debug/fm10k/0000:01:00.0/q_vector.001:
      rx_ring.001  tx_ring.001
      
      /sys/kernel/debug/fm10k/0000:01:00.0/q_vector.002:
      rx_ring.002  tx_ring.002
      
      /sys/kernel/debug/fm10k/0000:01:00.0/q_vector.003:
      rx_ring.003  tx_ring.003
      
      # cat /sys/kernel/debug/fm10k/0000:01:00.0/q_vector.000/rx_ring.000
      DES DATA       RSS        STATERR    LENGTH VLAN   DGLORT SGLORT TIMESTAMP
      ---------------------------------------------------------------------------
      000 0x00000000 0x00000000 0x00000003 0x002a 0x0000 0x0000 0x0000 0x13951807dc4fedf0
      001 0x00000000 0x00000000 0x00000003 0x002a 0x0000 0x0000 0x0000 0x1395180906c9f2c8
      002 0x3731c000 0x00000000 0x00000000 0x0000 0x0000 0x0000 0x0000 0x0000000000000000
      003 0x3731d000 0x00000000 0x00000000 0x0000 0x0000 0x0000 0x0000 0x0000000000000000
      004 0xaab3a000 0x00000000 0x00000000 0x0000 0x0000 0x0000 0x0000 0x0000000000000000
      ...
      
      # cat /sys/kernel/debug/fm10k/0000:01:00.0/q_vector.000/tx_ring.000
      DES BUFFER_ADDRESS     LENGTH VLAN   MSS    HDRLEN FLAGS
      ---------------------------------------------------------
      000 0x00000000aa8a1002 0x005a 0x0000 0x0000 0x0000 0xc0
      001 0x00000000aa8a2002 0x005a 0x0000 0x0000 0x0000 0xc0
      002 0x000000006bc13202 0x004e 0x0000 0x0000 0x0000 0xc0
      003 0x000000006bc13c02 0x002a 0x0000 0x0000 0x0000 0xe1
      004 0x000000006bc13602 0x0062 0x0000 0x0000 0x0000 0xc0
      Signed-off-by: NAlexander Duyck <alexander.h.duyck@intel.com>
      Signed-off-by: NJeff Kirsher <jeffrey.t.kirsher@intel.com>
      7461fd91
    • A
      fm10k: Add support for MACVLAN acceleration · 5cd5e2e9
      Alexander Duyck 提交于
      This patch adds support for L2 MACVLAN by making use of the fact that the
      RRC provides a unique tag per filter called a Global Resource Tag, or GLORT.
      In the case of this offload what I have done is assigned a linear block of
      these so that each GLORT represents one of the MACVLAN netdevs.  By doing
      this I can share the Rx queues and Tx queues for all of the MACVLAN netdevs
      while allowing them to be demuxed in the Rx cleanup path.
      Signed-off-by: NAlexander Duyck <alexander.h.duyck@intel.com>
      Signed-off-by: NJeff Kirsher <jeffrey.t.kirsher@intel.com>
      5cd5e2e9
    • A
      fm10k: Add support for netdev offloads · 76a540d4
      Alexander Duyck 提交于
      This patch adds support for basic offloads including TSO, Tx checksum, Rx
      checksum, Rx hash, and the same features applied to VXLAN/NVGRE tunnels.
      Signed-off-by: NAlexander Duyck <alexander.h.duyck@intel.com>
      Signed-off-by: NJeff Kirsher <jeffrey.t.kirsher@intel.com>
      76a540d4