1. 04 6月, 2019 1 次提交
  2. 05 5月, 2019 3 次提交
  3. 17 4月, 2019 2 次提交
  4. 24 3月, 2019 2 次提交
  5. 10 3月, 2019 1 次提交
    • M
      bnxt_en: Drop oversize TX packets to prevent errors. · 1713c8e1
      Michael Chan 提交于
      [ Upstream commit 2b3c6885386020b1b9d92d45e8349637e27d1f66 ]
      
      There have been reports of oversize UDP packets being sent to the
      driver to be transmitted, causing error conditions.  The issue is
      likely caused by the dst of the SKB switching between 'lo' with
      64K MTU and the hardware device with a smaller MTU.  Patches are
      being proposed by Mahesh Bandewar <maheshb@google.com> to fix the
      issue.
      
      In the meantime, add a quick length check in the driver to prevent
      the error.  The driver uses the TX packet size as index to look up an
      array to setup the TX BD.  The array is large enough to support all MTU
      sizes supported by the driver.  The oversize TX packet causes the
      driver to index beyond the array and put garbage values into the
      TX BD.  Add a simple check to prevent this.
      Signed-off-by: NMichael Chan <michael.chan@broadcom.com>
      Signed-off-by: NDavid S. Miller <davem@davemloft.net>
      Signed-off-by: NGreg Kroah-Hartman <gregkh@linuxfoundation.org>
      1713c8e1
  6. 13 2月, 2019 1 次提交
    • M
      bnxt_en: Disable MSIX before re-reserving NQs/CMPL rings. · 4cd197bf
      Michael Chan 提交于
      [ Upstream commit 36d65be9a88052cdfc8524eb591baf0e6c878408 ]
      
      When bringing up a device, the code checks to see if the number of
      MSIX has changed.  pci_disable_msix() should be called first before
      changing the number of reserved NQs/CMPL rings.  This ensures that
      the MSIX vectors associated with the NQs/CMPL rings are still
      properly mapped when pci_disable_msix() masks the vectors.
      
      This patch will prevent errors when RDMA support is added for the new
      57500 chips.  When the RDMA driver shuts down, the number of NQs is
      decreased and we must use the new sequence to prevent MSIX errors.
      Signed-off-by: NMichael Chan <michael.chan@broadcom.com>
      Signed-off-by: NDavid S. Miller <davem@davemloft.net>
      Signed-off-by: NSasha Levin <sashal@kernel.org>
      4cd197bf
  7. 05 10月, 2018 3 次提交
  8. 27 9月, 2018 1 次提交
    • M
      bnxt_en: Fix TX timeout during netpoll. · 73f21c65
      Michael Chan 提交于
      The current netpoll implementation in the bnxt_en driver has problems
      that may miss TX completion events.  bnxt_poll_work() in effect is
      only handling at most 1 TX packet before exiting.  In addition,
      there may be in flight TX completions that ->poll() may miss even
      after we fix bnxt_poll_work() to handle all visible TX completions.
      netpoll may not call ->poll() again and HW may not generate IRQ
      because the driver does not ARM the IRQ when the budget (0 for netpoll)
      is reached.
      
      We fix it by handling all TX completions and to always ARM the IRQ
      when we exit ->poll() with 0 budget.
      
      Also, the logic to ACK the completion ring in case it is almost filled
      with TX completions need to be adjusted to take care of the 0 budget
      case, as discussed with Eric Dumazet <edumazet@google.com>
      Reported-by: NSong Liu <songliubraving@fb.com>
      Reviewed-by: NSong Liu <songliubraving@fb.com>
      Tested-by: NSong Liu <songliubraving@fb.com>
      Signed-off-by: NMichael Chan <michael.chan@broadcom.com>
      Signed-off-by: NDavid S. Miller <davem@davemloft.net>
      73f21c65
  9. 24 9月, 2018 1 次提交
    • E
      bnxt: remove ndo_poll_controller · 58e0e22b
      Eric Dumazet 提交于
      As diagnosed by Song Liu, ndo_poll_controller() can
      be very dangerous on loaded hosts, since the cpu
      calling ndo_poll_controller() might steal all NAPI
      contexts (for all RX/TX queues of the NIC). This capture
      can last for unlimited amount of time, since one
      cpu is generally not able to drain all the queues under load.
      
      bnxt uses NAPI for TX completions, so we better let core
      networking stack call the napi->poll() to avoid the capture.
      Signed-off-by: NEric Dumazet <edumazet@google.com>
      Cc: Michael Chan <michael.chan@broadcom.com>
      Signed-off-by: NDavid S. Miller <davem@davemloft.net>
      58e0e22b
  10. 17 9月, 2018 1 次提交
  11. 04 9月, 2018 3 次提交
  12. 08 8月, 2018 1 次提交
  13. 06 8月, 2018 9 次提交
  14. 17 7月, 2018 1 次提交
  15. 10 7月, 2018 5 次提交
  16. 30 6月, 2018 1 次提交
  17. 26 6月, 2018 1 次提交
  18. 26 5月, 2018 1 次提交
    • B
      bnxt_en: Report PCIe link properties with pcie_print_link_status() · af125b75
      Bjorn Helgaas 提交于
      Previously the driver used pcie_get_minimum_link() to warn when the NIC
      is in a slot that can't supply as much bandwidth as the NIC could use.
      
      pcie_get_minimum_link() can be misleading because it finds the slowest link
      and the narrowest link (which may be different links) without considering
      the total bandwidth of each link.  For a path with a 16 GT/s x1 link and a
      2.5 GT/s x16 link, it returns 2.5 GT/s x1, which corresponds to 250 MB/s of
      bandwidth, not the true available bandwidth of about 1969 MB/s for a
      16 GT/s x1 link.
      
      Use pcie_print_link_status() to report PCIe link speed and possible
      limitations instead of implementing this in the driver itself.  This finds
      the slowest link in the path to the device by computing the total bandwidth
      of each link and compares that with the capabilities of the device.
      
      The dmesg change is:
      
        - PCIe: Speed %s Width x%d
        + %u.%03u Gb/s available PCIe bandwidth (%s x%d link)
      Signed-off-by: NBjorn Helgaas <bhelgaas@google.com>
      af125b75
  19. 08 5月, 2018 2 次提交