1. 06 7月, 2016 3 次提交
  2. 12 6月, 2016 1 次提交
  3. 08 6月, 2016 2 次提交
  4. 18 5月, 2016 6 次提交
  5. 04 5月, 2016 1 次提交
    • K
      NVMe: Fix reset/remove race · 87c32077
      Keith Busch 提交于
      This fixes a scenario where device is present and being reset, but a
      request to unbind the driver occurs.
      
      A previous patch series addressing a device failure removal scenario
      flushed reset_work after controller disable to unblock reset_work waiting
      on a completion that wouldn't occur. This isn't safe as-is. The broken
      scenario can potentially be induced with:
      
        modprobe nvme && modprobe -r nvme
      
      To fix, the reset work is flushed immediately after setting the controller
      removing flag, and any subsequent reset will not proceed with controller
      initialization if the flag is set.
      
      The controller status must be polled while active, so the watchdog timer
      is also left active until the controller is disabled to cleanup requests
      that may be stuck during namespace removal.
      
      [Fixes: ff23a2a1]
      Signed-off-by: NKeith Busch <keith.busch@intel.com>
      Reviewed-by: NChristoph Hellwig <hch@lst.de>
      Reviewed-by: NJohannes Thumshirn <jthumshirn@suse.de>
      Reviewed-by: NSagi Grimberg <sagi@grimberg.me>
      Signed-off-by: NJens Axboe <axboe@fb.com>
      87c32077
  6. 02 5月, 2016 7 次提交
  7. 15 4月, 2016 1 次提交
    • K
      NVMe: Always use MSI/MSI-x interrupts · a5229050
      Keith Busch 提交于
      Multiple users have reported device initialization failure due the driver
      not receiving legacy PCI interrupts. This is not unique to any particular
      controller, but has been observed on multiple platforms.
      
      There have been no issues reported or observed when with message signaled
      interrupts, so this patch attempts to use MSI-x during initialization,
      falling back to MSI. If that fails, legacy would become the default.
      
      The setup_io_queues error handling had to change as a result: the admin
      queue's msix_entry used to be initialized to the legacy IRQ. The case
      where nr_io_queues is 0 would fail request_irq when setting up the admin
      queue's interrupt since re-enabling MSI-x fails with 0 vectors, leaving
      the admin queue's msix_entry invalid. Instead, return success immediately.
      Reported-by: NTim Muhlemmer <muhlemmer@gmail.com>
      Reported-by: NJon Derrick <jonathan.derrick@intel.com>
      Signed-off-by: NKeith Busch <keith.busch@intel.com>
      Signed-off-by: NJens Axboe <axboe@fb.com>
      a5229050
  8. 13 4月, 2016 9 次提交
    • G
      nvme: Avoid reset work on watchdog timer function during error recovery · c875a709
      Guilherme G. Piccoli 提交于
      This patch adds a check on nvme_watchdog_timer() function to avoid the
      call to reset_work() when an error recovery process is ongoing on
      controller. The check is made by looking at pci_channel_offline()
      result.
      
      If we don't check for this on nvme_watchdog_timer(), error recovery
      mechanism can't recover well, because reset_work() won't be able to
      do its job (since we're in the middle of an error) and so the
      controller is removed from the system before error recovery mechanism
      can perform slot reset (which would allow the adapter to recover).
      
      In this patch we also have split the huge condition expression on
      nvme_watchdog_timer() by introducing an auxiliary function to help
      make the code more readable.
      Reviewed-by: NKeith Busch <keith.busch@intel.com>
      Reviewed-by: NJohannes Thumshirn <jthumshirn@suse.de>
      Signed-off-by: NGuilherme G. Piccoli <gpiccoli@linux.vnet.ibm.com>
      Signed-off-by: NJens Axboe <axboe@fb.com>
      c875a709
    • J
      NVMe: silence warning about unused 'dev' · 7e197930
      Jens Axboe 提交于
      Depending on options, we might not be using dev in nvme_cancel_io():
      
      drivers/nvme/host/pci.c: In function ‘nvme_cancel_io’:
      drivers/nvme/host/pci.c:970:19: warning: unused variable ‘dev’ [-Wunused-variable]
        struct nvme_dev *dev = data;
                         ^
      
      So get rid of it, and just cast for the dev_dbg_ratelimited() call.
      
      Fixes: 82b4552b ("nvme: Use blk-mq helper for IO termination")
      Signed-off-by: NJens Axboe <axboe@fb.com>
      7e197930
    • S
      nvme: Use blk-mq helper for IO termination · 82b4552b
      Sagi Grimberg 提交于
      blk-mq offers a tagset iterator so let's use that
      instead of using nvme_clear_queues.
      
      Note, we changed nvme_queue_cancel_ios name to nvme_cancel_io
      as there is no concept of a queue now in this function (we
      also lost the print).
      Signed-off-by: NSagi Grimberg <sagig@mellanox.com>
      Reviewed-by: NChristoph Hellwig <hch@lst.de>
      Acked-by: NKeith Busch <keith.busch@intel.com>
      Reviewed-by: NJohannes Thumshirn <jthumshirn@suse.de>
      Signed-off-by: NJens Axboe <axboe@fb.com>
      82b4552b
    • K
      NVMe: Skip async events for degraded controllers · 21f033f7
      Keith Busch 提交于
      If the controller is degraded, the driver should stay out of the way so
      the user can recover the drive. This patch skips driver initiated async
      event requests when the drive is in this state.
      Signed-off-by: NKeith Busch <keith.busch@intel.com>
      Reviewed-by: NSagi Grimberg <sagi@grimberg.me>
      Reviewed-by: NChristoph Hellwig <hch@lst.de>
      Signed-off-by: NJens Axboe <axboe@fb.com>
      21f033f7
    • M
      nvme: add helper nvme_setup_cmd() · 8093f7ca
      Ming Lin 提交于
      This moves nvme_setup_{flush,discard,rw} calls into a common
      nvme_setup_cmd() helper. So we can eventually hide all the command
      setup in the core module and don't even need to update the fabrics
      drivers for any specific command type.
      Signed-off-by: NMing Lin <ming.l@ssi.samsung.com>
      Reviewed-by: NChristoph Hellwig <hch@lst.de>
      Signed-off-by: NJens Axboe <axboe@fb.com>
      8093f7ca
    • M
      nvme: rewrite discard support · 03b5929e
      Ming Lin 提交于
      This rewrites nvme_setup_discard() with blk_add_request_payload().
      It allocates only the necessary amount(16 bytes) for the payload.
      Signed-off-by: NMing Lin <ming.l@ssi.samsung.com>
      Reviewed-by: NChristoph Hellwig <hch@lst.de>
      Signed-off-by: NJens Axboe <axboe@fb.com>
      03b5929e
    • M
      nvme: add helper nvme_map_len() · 58b45602
      Ming Lin 提交于
      The helper returns the number of bytes that need to be mapped
      using PRPs/SGL entries.
      Signed-off-by: NMing Lin <ming.l@ssi.samsung.com>
      Reviewed-by: NChristoph Hellwig <hch@lst.de>
      Signed-off-by: NJens Axboe <axboe@fb.com>
      58b45602
    • M
      nvme: add missing lock nesting notation · 2e39e0f6
      Ming Lin 提交于
      When unloading driver, nvme_disable_io_queues() calls nvme_delete_queue()
      that sends nvme_admin_delete_cq command to admin sq. So when the command
      completed, the lock acquired by nvme_irq() actually belongs to admin queue.
      
      While the lock that nvme_del_cq_end() trying to acquire belongs to io queue.
      So it will not deadlock.
      
      This patch adds lock nesting notation to fix following report.
      
      [  109.840952] =============================================
      [  109.846379] [ INFO: possible recursive locking detected ]
      [  109.851806] 4.5.0+ #180 Tainted: G            E
      [  109.856533] ---------------------------------------------
      [  109.861958] swapper/0/0 is trying to acquire lock:
      [  109.866771]  (&(&nvmeq->q_lock)->rlock){-.....}, at: [<ffffffffc0820bc6>] nvme_del_cq_end+0x26/0x70 [nvme]
      [  109.876535]
      [  109.876535] but task is already holding lock:
      [  109.882398]  (&(&nvmeq->q_lock)->rlock){-.....}, at: [<ffffffffc0820c2b>] nvme_irq+0x1b/0x50 [nvme]
      [  109.891547]
      [  109.891547] other info that might help us debug this:
      [  109.898107]  Possible unsafe locking scenario:
      [  109.898107]
      [  109.904056]        CPU0
      [  109.906515]        ----
      [  109.908974]   lock(&(&nvmeq->q_lock)->rlock);
      [  109.913381]   lock(&(&nvmeq->q_lock)->rlock);
      [  109.917787]
      [  109.917787]  *** DEADLOCK ***
      [  109.917787]
      [  109.923738]  May be due to missing lock nesting notation
      [  109.923738]
      [  109.930558] 1 lock held by swapper/0/0:
      [  109.934413]  #0:  (&(&nvmeq->q_lock)->rlock){-.....}, at: [<ffffffffc0820c2b>] nvme_irq+0x1b/0x50 [nvme]
      [  109.944010]
      [  109.944010] stack backtrace:
      [  109.948389] CPU: 0 PID: 0 Comm: swapper/0 Tainted: G            E   4.5.0+ #180
      [  109.955734] Hardware name: Dell Inc. OptiPlex 7010/0YXT71, BIOS A15 08/12/2013
      [  109.962989]  0000000000000000 ffff88011e203c38 ffffffff81383d9c ffffffff81c13540
      [  109.970478]  ffffffff826711d0 ffff88011e203ce8 ffffffff810bb429 0000000000000046
      [  109.977964]  0000000000000046 0000000000000000 0000000000b2e597 ffffffff81f4cb00
      [  109.985453] Call Trace:
      [  109.987911]  <IRQ>  [<ffffffff81383d9c>] dump_stack+0x85/0xc9
      [  109.993711]  [<ffffffff810bb429>] __lock_acquire+0x19b9/0x1c60
      [  109.999575]  [<ffffffff810b6d1d>] ? trace_hardirqs_off+0xd/0x10
      [  110.005524]  [<ffffffff810b386d>] ? complete+0x3d/0x50
      [  110.010688]  [<ffffffff810bb760>] lock_acquire+0x90/0xf0
      [  110.016029]  [<ffffffffc0820bc6>] ? nvme_del_cq_end+0x26/0x70 [nvme]
      [  110.022418]  [<ffffffff81772afb>] _raw_spin_lock_irqsave+0x4b/0x60
      [  110.028632]  [<ffffffffc0820bc6>] ? nvme_del_cq_end+0x26/0x70 [nvme]
      [  110.035019]  [<ffffffffc0820bc6>] nvme_del_cq_end+0x26/0x70 [nvme]
      [  110.041232]  [<ffffffff8135b485>] blk_mq_end_request+0x35/0x60
      [  110.047095]  [<ffffffffc0821ad8>] nvme_complete_rq+0x68/0x190 [nvme]
      [  110.053481]  [<ffffffff8135b53f>] __blk_mq_complete_request+0x8f/0x130
      [  110.060043]  [<ffffffff8135b611>] blk_mq_complete_request+0x31/0x40
      [  110.066343]  [<ffffffffc08209e3>] __nvme_process_cq+0x83/0x240 [nvme]
      [  110.072818]  [<ffffffffc0820c35>] nvme_irq+0x25/0x50 [nvme]
      [  110.078419]  [<ffffffff810cdb66>] handle_irq_event_percpu+0x36/0x110
      [  110.084804]  [<ffffffff810cdc77>] handle_irq_event+0x37/0x60
      [  110.090491]  [<ffffffff810d0ea3>] handle_edge_irq+0x93/0x150
      [  110.096180]  [<ffffffff81012306>] handle_irq+0xa6/0x130
      [  110.101431]  [<ffffffff81011abe>] do_IRQ+0x5e/0x120
      [  110.106333]  [<ffffffff8177384c>] common_interrupt+0x8c/0x8c
      Reviewed-by: NJohannes Thumshirn <jthumshirn@suse.de>
      Signed-off-by: NMing Lin <ming.l@ssi.samsung.com>
      Reviewed-by: NChristoph Hellwig <hch@lst.de>
      Reviewed-by: NSagi Grimberg <sagi@grimberg.me>
      Signed-off-by: NJens Axboe <axboe@fb.com>
      2e39e0f6
    • K
      NVMe: Always use MSI/MSI-x interrupts · 788e15ab
      Keith Busch 提交于
      Multiple users have reported device initialization failure due the driver
      not receiving legacy PCI interrupts. This is not unique to any particular
      controller, but has been observed on multiple platforms.
      
      There have been no issues reported or observed when with message signaled
      interrupts, so this patch attempts to use MSI-x during initialization,
      falling back to MSI. If that fails, legacy would become the default.
      
      The setup_io_queues error handling had to change as a result: the admin
      queue's msix_entry used to be initialized to the legacy IRQ. The case
      where nr_io_queues is 0 would fail request_irq when setting up the admin
      queue's interrupt since re-enabling MSI-x fails with 0 vectors, leaving
      the admin queue's msix_entry invalid. Instead, return success immediately.
      Reported-by: NTim Muhlemmer <muhlemmer@gmail.com>
      Reported-by: NJon Derrick <jonathan.derrick@intel.com>
      Signed-off-by: NKeith Busch <keith.busch@intel.com>
      Signed-off-by: NJens Axboe <axboe@fb.com>
      788e15ab
  9. 12 4月, 2016 1 次提交
    • K
      NVMe: Fix reset/remove race · 9bf2b972
      Keith Busch 提交于
      This fixes a scenario where device is present and being reset, but a
      request to unbind the driver occurs.
      
      A previous patch series addressing a device failure removal scenario
      flushed reset_work after controller disable to unblock reset_work waiting
      on a completion that wouldn't occur. This isn't safe as-is. The broken
      scenario can potentially be induced with:
      
        modprobe nvme && modprobe -r nvme
      
      To fix, the reset work is flushed immediately after setting the controller
      removing flag, and any subsequent reset will not proceed with controller
      initialization if the flag is set.
      
      The controller status must be polled while active, so the watchdog timer
      is also left active until the controller is disabled to cleanup requests
      that may be stuck during namespace removal.
      
      [Fixes: ff23a2a1]
      Signed-off-by: NKeith Busch <keith.busch@intel.com>
      Reviewed-by: NChristoph Hellwig <hch@lst.de>
      Reviewed-by: NJohannes Thumshirn <jthumshirn@suse.de>
      Reviewed-by: NSagi Grimberg <sagi@grimberg.me>
      Signed-off-by: NJens Axboe <axboe@fb.com>
      9bf2b972
  10. 23 3月, 2016 1 次提交
  11. 09 3月, 2016 1 次提交
  12. 08 3月, 2016 1 次提交
  13. 04 3月, 2016 4 次提交
  14. 29 2月, 2016 2 次提交