1. 18 10月, 2018 2 次提交
    • L
      nvmet: Optionally use PCI P2P memory · c6925093
      Logan Gunthorpe 提交于
      Create a configfs attribute in each nvme-fabrics namespace to enable P2P
      memory use.  The attribute may be enabled (with a boolean) or a specific
      P2P device may be given (with the device's PCI name).
      
      When enabled, the namespace will ensure the underlying block device
      supports P2P and is compatible with any specified P2P device.  If no device
      was specified it will ensure there is compatible P2P memory somewhere in
      the system.  Enabling a namespace with P2P memory will fail with EINVAL
      (and an appropriate dmesg error) if any of these conditions are not met.
      
      Once a controller is set up on a specific port, the P2P device to use for
      each namespace will be found and stored in a radix tree by namespace ID.
      When memory is allocated for a request, the tree is used to look up the P2P
      device to allocate memory against.  If no device is in the tree (because no
      appropriate device was found), or if allocation of P2P memory fails, fall
      back to using regular memory.
      Signed-off-by: NStephen Bates <sbates@raithlin.com>
      Signed-off-by: NSteve Wise <swise@opengridcomputing.com>
      [hch: partial rewrite of the initial code]
      Signed-off-by: NChristoph Hellwig <hch@lst.de>
      Signed-off-by: NLogan Gunthorpe <logang@deltatee.com>
      Signed-off-by: NBjorn Helgaas <bhelgaas@google.com>
      c6925093
    • L
      nvmet: Introduce helper functions to allocate and free request SGLs · 5b2322e4
      Logan Gunthorpe 提交于
      Add helpers to allocate and free the SGL in a struct nvmet_req:
      
        int nvmet_req_alloc_sgl(struct nvmet_req *req)
        void nvmet_req_free_sgl(struct nvmet_req *req)
      
      This will be expanded in a future patch to implement peer-to-peer memory
      DMAs and should be common with all target drivers.
      
      The new helpers are used in nvmet-rdma.  Seeing we use req.transfer_len as
      the length of the SGL it is set earlier and cleared on any error.  It also
      seems to be unnecessary to accumulate the length as the map_sgl functions
      should only ever be called once per request.
      Signed-off-by: NLogan Gunthorpe <logang@deltatee.com>
      Signed-off-by: NBjorn Helgaas <bhelgaas@google.com>
      Reviewed-by: NChristoph Hellwig <hch@lst.de>
      Acked-by: NSagi Grimberg <sagi@grimberg.me>
      5b2322e4
  2. 06 9月, 2018 1 次提交
    • S
      nvmet-rdma: fix possible bogus dereference under heavy load · 8407879c
      Sagi Grimberg 提交于
      Currently we always repost the recv buffer before we send a response
      capsule back to the host. Since ordering is not guaranteed for send
      and recv completions, it is posible that we will receive a new request
      from the host before we got a send completion for the response capsule.
      
      Today, we pre-allocate 2x rsps the length of the queue, but in reality,
      under heavy load there is nothing that is really preventing the gap to
      expand until we exhaust all our rsps.
      
      To fix this, if we don't have any pre-allocated rsps left, we dynamically
      allocate a rsp and make sure to free it when we are done. If under memory
      pressure we fail to allocate a rsp, we silently drop the command and
      wait for the host to retry.
      Reported-by: NSteve Wise <swise@opengridcomputing.com>
      Tested-by: NSteve Wise <swise@opengridcomputing.com>
      Signed-off-by: NSagi Grimberg <sagi@grimberg.me>
      [hch: dropped a superflous assignment]
      Signed-off-by: NChristoph Hellwig <hch@lst.de>
      8407879c
  3. 25 7月, 2018 1 次提交
  4. 23 7月, 2018 3 次提交
  5. 19 6月, 2018 1 次提交
  6. 26 3月, 2018 5 次提交
  7. 08 1月, 2018 2 次提交
  8. 07 1月, 2018 1 次提交
  9. 11 11月, 2017 2 次提交
  10. 18 8月, 2017 1 次提交
  11. 28 6月, 2017 2 次提交
  12. 21 5月, 2017 1 次提交
  13. 04 4月, 2017 3 次提交
  14. 17 3月, 2017 1 次提交
  15. 23 2月, 2017 2 次提交
  16. 26 1月, 2017 1 次提交
  17. 15 12月, 2016 1 次提交
  18. 06 12月, 2016 2 次提交
  19. 14 11月, 2016 3 次提交
  20. 24 9月, 2016 1 次提交
  21. 18 8月, 2016 1 次提交
  22. 16 8月, 2016 1 次提交
  23. 04 8月, 2016 2 次提交
    • S
      nvmet-rdma: Don't use the inline buffer in order to avoid allocation for small reads · 40e64e07
      Sagi Grimberg 提交于
      Under extreme conditions this might cause data corruptions. By doing that
      we we repost the buffer and then post this buffer for the device to send.
      If we happen to use shared receive queues the device might write to the
      buffer before it sends it (there is no ordering between send and recv
      queues). Without SRQs we probably won't get that if the host doesn't
      mis-behave and send more than we allowed it, but relying on that is not
      really a good idea.
      Signed-off-by: NSagi Grimberg <sagi@grimberg.me>
      Reviewed-by: NChristoph Hellwig <hch@lst.de>
      40e64e07
    • S
      nvmet-rdma: Correctly handle RDMA device hot removal · d8f7750a
      Sagi Grimberg 提交于
      When configuring a device attached listener, we may
      see device removal events. In this case we return a
      non-zero return code from the cm event handler which
      implicitly destroys the cm_id. It is possible that in
      the future the user will remove this listener and by
      that trigger a second call to rdma_destroy_id on an
      already destroyed cm_id -> BUG.
      
      In addition, when a queue bound (active session) cm_id
      generates a DEVICE_REMOVAL event we must guarantee all
      resources are cleaned up by the time we return from the
      event handler.
      
      Introduce nvmet_rdma_device_removal which addresses
      (or at least attempts to) both scenarios.
      Signed-off-by: NSagi Grimberg <sagi@grimberg.me>
      Reviewed-by: NChristoph Hellwig <hch@lst.de>
      d8f7750a