1. 24 5月, 2009 1 次提交
    • M
      [SCSI] iscsi: pass ep connect shost · 10eb0f01
      Mike Christie 提交于
      When we create the tcp/ip connection by calling ep_connect, we currently
      just go by the routing table info.
      
      I think there are two problems with this.
      
      1. Some drivers do not have access to a routing table. Some drivers like
      qla4xxx do not even know about other ports.
      
      2. If you have two initiator ports on the same subnet, the user may have
      set things up so that session1 was supposed to be run through port1. and
      session2 was supposed to be run through port2. It looks like we could
      end with both sessions going through one of the ports.
      
      Fixes for cxgb3i from Karen Xie.
      Signed-off-by: NMike Christie <michaelc@cs.wisc.edu>
      Signed-off-by: NJames Bottomley <James.Bottomley@HansenPartnership.com>
      10eb0f01
  2. 27 4月, 2009 1 次提交
  3. 21 4月, 2009 1 次提交
  4. 01 4月, 2009 1 次提交
  5. 22 3月, 2009 1 次提交
  6. 14 3月, 2009 5 次提交
  7. 28 2月, 2009 1 次提交
  8. 18 2月, 2009 1 次提交
  9. 22 1月, 2009 1 次提交
  10. 17 1月, 2009 1 次提交
  11. 15 1月, 2009 2 次提交
  12. 13 1月, 2009 2 次提交
    • R
      IB/iser: Add dependency on INFINIBAND_ADDR_TRANS · f5eb3b76
      Randy Dunlap 提交于
      Fix ib_iser build to depend on INFINIBAND_ADDR_TRANS; if INET=y but
      IPV6=n, then the RDMA CM is not built but INFINIBAND_ISER can be
      enabled, leading to:
      
          ERROR: "rdma_destroy_id" [drivers/infiniband/ulp/iser/ib_iser.ko] undefined!
          ERROR: "rdma_connect" [drivers/infiniband/ulp/iser/ib_iser.ko] undefined!
          ERROR: "rdma_destroy_qp" [drivers/infiniband/ulp/iser/ib_iser.ko] undefined!
          ERROR: "rdma_create_id" [drivers/infiniband/ulp/iser/ib_iser.ko] undefined!
          ERROR: "rdma_create_qp" [drivers/infiniband/ulp/iser/ib_iser.ko] undefined!
          ERROR: "rdma_resolve_route" [drivers/infiniband/ulp/iser/ib_iser.ko] undefined!
          ERROR: "rdma_disconnect" [drivers/infiniband/ulp/iser/ib_iser.ko] undefined!
          ERROR: "rdma_resolve_addr" [drivers/infiniband/ulp/iser/ib_iser.ko] undefined!
      Signed-off-by: NRandy Dunlap <randy.dunlap@oracle.com>
      f5eb3b76
    • Y
      IPoIB: Do not join broadcast group if interface is brought down · 50df48f5
      Yossi Etigin 提交于
      Because the ipoib_workqueue is not flushed when ipoib interface is
      brought down, ipoib_mcast_join() may trigger a join to the broadcast
      group after priv->broadcast was set to NULL (during cleanup).  This
      will cause the system to be a member of the broadcast group when
      interface is down.  As a side effect, this breaks the optimization of
      setting the Q_key only when joining the broadcast group.
      Signed-off-by: NYossi Etigin <yosefe@voltaire.com>
      Signed-off-by: NRoland Dreier <rolandd@cisco.com>
      50df48f5
  13. 10 1月, 2009 1 次提交
    • Y
      IPoIB: Fix loss of connectivity after bonding failover on both sides · a50df398
      Yossi Etigin 提交于
      Fix bonding failover in the case both peers failover and the
      gratuitous ARP is lost.  In that case, the sender side will create an
      ipoib_neigh and issue a path request with the old GID first.  When
      skb->dst->neighbour->ha changes due to ARP refresh, this ipoib_neigh
      will not be added to the path->list of the path of the new GID,
      because the ipoib_neigh already exists.  It will not have an AH
      either, because of sender-side failover.  Therefore, it will not get
      an AH when the path is resolved.
      
      The solution here is to compare GIDs in ipoib_start_xmit() even if
      neigh->ah is invalid.  Comparing with an uninitialized value of
      neigh->dgid should be fine, since a spurious match is harmless (and
      astronomically unlikely too).
      Signed-off-by: NMoni Shoua <monis@voltaire.com>
      Signed-off-by: NYossi Etigin <yosefe@voltaire.com>
      Signed-off-by: NRoland Dreier <rolandd@cisco.com>
      a50df398
  14. 07 1月, 2009 1 次提交
  15. 30 12月, 2008 2 次提交
  16. 23 12月, 2008 1 次提交
  17. 22 12月, 2008 1 次提交
    • D
      IB/iser: Avoid recv buffer exhaustion caused by unexpected PDUs · bba7ebba
      David Disseldorp 提交于
      iSCSI/iSER targets may send PDUs without a prior request from the
      initiator.  RFC 5046 refers to these PDUs as "unexpected".  NOP-In PDUs
      with itt=RESERVED and Asynchronous Message PDUs occupy this category.
      
      The amount of active "unexpected" PDU's an iSER target may have at any
      time is governed by the MaxOutstandingUnexpectedPDUs key, which is not
      yet supported.
      
      Currently when an iSER target sends an "unexpected" PDU, the
      initiators recv buffer consumed by the PDU is not replaced.  If over
      initial_post_recv_bufs_num "unexpected" PDUs are received then the
      receive queue will run out of receive work requests entirely.
      
      This patch ensures recv buffers consumed by "unexpected" PDUs are
      replaced in the next iser_post_receive_control() call.
      Signed-off-by: NDavid Disseldorp <ddiss@sgi.com>
      Signed-off-by: NKen Sandars <ksandars@sgi.com>
      Acked-by: NOr Gerlitz <ogerlitz@voltaire.com>
      Signed-off-by: NRoland Dreier <rolandd@cisco.com>
      bba7ebba
  18. 13 11月, 2008 3 次提交
  19. 31 10月, 2008 1 次提交
  20. 30 10月, 2008 1 次提交
  21. 29 10月, 2008 3 次提交
  22. 23 10月, 2008 2 次提交
  23. 14 10月, 2008 1 次提交
  24. 13 10月, 2008 2 次提交
  25. 11 10月, 2008 1 次提交
  26. 01 10月, 2008 1 次提交
    • R
      IPoIB: Use netif_tx_lock() and get rid of private tx_lock, LLTX · 943c246e
      Roland Dreier 提交于
      Currently, IPoIB is an LLTX driver that uses its own IRQ-disabling
      tx_lock.  Not only do we want to get rid of LLTX, this actually causes
      problems because of the skb_orphan() done with this tx_lock held: some
      skb destructors expect to be run with interrupts enabled.
      
      The simplest fix for this is to get rid of the driver-private tx_lock
      and stop using LLTX.  We kill off priv->tx_lock and use
      netif_tx_lock[_bh]() instead; the patch to do this is a tiny bit
      tricky because we need to update places that take priv->lock inside
      the tx_lock to disable IRQs, rather than relying on tx_lock having
      already disabled IRQs.
      
      Also, there are a couple of places where we need to disable BHs to
      make sure we have a consistent context to call netif_tx_lock() (since
      we no longer can use _irqsave() variants), and we also have to change
      ipoib_send_comp_handler() to call drain_tx_cq() through a timer rather
      than directly, because ipoib_send_comp_handler() runs in interrupt
      context and drain_tx_cq() must run in BH context so it can call
      netif_tx_lock().
      Signed-off-by: NRoland Dreier <rolandd@cisco.com>
      943c246e
  27. 26 9月, 2008 1 次提交