1. 31 3月, 2022 1 次提交
    • L
      drbd: fix potential silent data corruption · f4329d1f
      Lars Ellenberg 提交于
      Scenario:
      ---------
      
      bio chain generated by blk_queue_split().
      Some split bio fails and propagates its error status to the "parent" bio.
      But then the (last part of the) parent bio itself completes without error.
      
      We would clobber the already recorded error status with BLK_STS_OK,
      causing silent data corruption.
      
      Reproducer:
      -----------
      
      How to trigger this in the real world within seconds:
      
      DRBD on top of degraded parity raid,
      small stripe_cache_size, large read_ahead setting.
      Drop page cache (sysctl vm.drop_caches=1, fadvise "DONTNEED",
      umount and mount again, "reboot").
      
      Cause significant read ahead.
      
      Large read ahead request is split by blk_queue_split().
      Parts of the read ahead that are already in the stripe cache,
      or find an available stripe cache to use, can be serviced.
      Parts of the read ahead that would need "too much work",
      would need to wait for a "stripe_head" to become available,
      are rejected immediately.
      
      For larger read ahead requests that are split in many pieces, it is very
      likely that some "splits" will be serviced, but then the stripe cache is
      exhausted/busy, and the remaining ones will be rejected.
      Signed-off-by: NLars Ellenberg <lars.ellenberg@linbit.com>
      Signed-off-by: NChristoph Böhmwalder <christoph.boehmwalder@linbit.com>
      Cc: <stable@vger.kernel.org> # 4.13.x
      Link: https://lore.kernel.org/r/20220330185551.3553196-1-christoph.boehmwalder@linbit.comSigned-off-by: NJens Axboe <axboe@kernel.dk>
      f4329d1f
  2. 30 3月, 2022 1 次提交
  3. 29 3月, 2022 6 次提交
    • J
      Merge tag 'nvme-5.18-2022-03-29' of git://git.infradead.org/nvme into for-5.18/drivers · 1e06b3e7
      Jens Axboe 提交于
      Pull NVMe fixes from Christoph:
      
      "- fix multipath hang when disk goes live over reconnect (Anton Eidelman)
       - fix RCU hole that allowed for endless looping in multipath round robin
         (Chris Leech)
       - remove redundant assignment after left shift (Colin Ian King)
       - add quirks for Samsung X5 SSDs (Monish Kumar R)
       - fix the read-only state for zoned namespaces with unsupposed features
         (Pankaj Raghav)
       - use a private workqueue instead of the system workqueue in nvmet
         (Sagi Grimberg)
       - allow duplicate NSIDs for private namespaces (Sungup Moon)
       - expose use_threaded_interrupts read-only in sysfs (Xin Hao)"
      
      * tag 'nvme-5.18-2022-03-29' of git://git.infradead.org/nvme:
        nvme-multipath: fix hang when disk goes live over reconnect
        nvme: fix RCU hole that allowed for endless looping in multipath round robin
        nvme: allow duplicate NSIDs for private namespaces
        nvmet: remove redundant assignment after left shift
        nvmet: use a private workqueue instead of the system workqueue
        nvme-pci: add quirks for Samsung X5 SSDs
        nvme-pci: expose use_threaded_interrupts read-only in sysfs
        nvme: fix the read-only state for zoned namespaces with unsupposed features
      1e06b3e7
    • A
      nvme-multipath: fix hang when disk goes live over reconnect · a4a6f3c8
      Anton Eidelman 提交于
      nvme_mpath_init_identify() invoked from nvme_init_identify() fetches a
      fresh ANA log from the ctrl.  This is essential to have an up to date
      path states for both existing namespaces and for those scan_work may
      discover once the ctrl is up.
      
      This happens in the following cases:
        1) A new ctrl is being connected.
        2) An existing ctrl is successfully reconnected.
        3) An existing ctrl is being reset.
      
      While in (1) ctrl->namespaces is empty, (2 & 3) may have namespaces, and
      nvme_read_ana_log() may call nvme_update_ns_ana_state().
      
      This result in a hang when the ANA state of an existing namespace changes
      and makes the disk live: nvme_mpath_set_live() issues IO to the namespace
      through the ctrl, which does NOT have IO queues yet.
      
      See sample hang below.
      
      Solution:
      - nvme_update_ns_ana_state() to call set_live only if ctrl is live
      - nvme_read_ana_log() call from nvme_mpath_init_identify()
        therefore only fetches and parses the ANA log;
        any erros in this process will fail the ctrl setup as appropriate;
      - a separate function nvme_mpath_update()
        is called in nvme_start_ctrl();
        this parses the ANA log without fetching it.
        At this point the ctrl is live,
        therefore, disks can be set live normally.
      
      Sample failure:
          nvme nvme0: starting error recovery
          nvme nvme0: Reconnecting in 10 seconds...
          block nvme0n6: no usable path - requeuing I/O
          INFO: task kworker/u8:3:312 blocked for more than 122 seconds.
                Tainted: G            E     5.14.5-1.el7.elrepo.x86_64 #1
          Workqueue: nvme-wq nvme_tcp_reconnect_ctrl_work [nvme_tcp]
          Call Trace:
           __schedule+0x2a2/0x7e0
           schedule+0x4e/0xb0
           io_schedule+0x16/0x40
           wait_on_page_bit_common+0x15c/0x3e0
           do_read_cache_page+0x1e0/0x410
           read_cache_page+0x12/0x20
           read_part_sector+0x46/0x100
           read_lba+0x121/0x240
           efi_partition+0x1d2/0x6a0
           bdev_disk_changed.part.0+0x1df/0x430
           bdev_disk_changed+0x18/0x20
           blkdev_get_whole+0x77/0xe0
           blkdev_get_by_dev+0xd2/0x3a0
           __device_add_disk+0x1ed/0x310
           device_add_disk+0x13/0x20
           nvme_mpath_set_live+0x138/0x1b0 [nvme_core]
           nvme_update_ns_ana_state+0x2b/0x30 [nvme_core]
           nvme_update_ana_state+0xca/0xe0 [nvme_core]
           nvme_parse_ana_log+0xac/0x170 [nvme_core]
           nvme_read_ana_log+0x7d/0xe0 [nvme_core]
           nvme_mpath_init_identify+0x105/0x150 [nvme_core]
           nvme_init_identify+0x2df/0x4d0 [nvme_core]
           nvme_init_ctrl_finish+0x8d/0x3b0 [nvme_core]
           nvme_tcp_setup_ctrl+0x337/0x390 [nvme_tcp]
           nvme_tcp_reconnect_ctrl_work+0x24/0x40 [nvme_tcp]
           process_one_work+0x1bd/0x360
           worker_thread+0x50/0x3d0
      Signed-off-by: NAnton Eidelman <anton@lightbitslabs.com>
      Reviewed-by: NSagi Grimberg <sagi@grimberg.me>
      Signed-off-by: NChristoph Hellwig <hch@lst.de>
      a4a6f3c8
    • C
      nvme: fix RCU hole that allowed for endless looping in multipath round robin · d6d67427
      Chris Leech 提交于
      Make nvme_ns_remove match the assumptions elsewhere.
      
      1) !NVME_NS_READY needs to be srcu synchronized to make sure nothing is
         running in __nvme_find_path or nvme_round_robin_path that will
         re-assign this ns to current_path.
      
      2) Any matching current_path entries need to be cleared before removing
         from the siblings list, to prevent calling nvme_round_robin_path with
         an "old" ns that's off list.
      
      3) Finally the list_del_rcu can happen, and then synchronize again
         before releasing any reference counts.
      Signed-off-by: NChristoph Hellwig <hch@lst.de>
      d6d67427
    • S
      nvme: allow duplicate NSIDs for private namespaces · 5974ea7c
      Sungup Moon 提交于
      A NVMe subsystem with multiple controller can have private namespaces
      that use the same NSID under some conditions:
      
       "If Namespace Management, ANA Reporting, or NVM Sets are supported, the
        NSIDs shall be unique within the NVM subsystem. If the Namespace
        Management, ANA Reporting, and NVM Sets are not supported, then NSIDs:
         a) for shared namespace shall be unique; and
         b) for private namespace are not required to be unique."
      
      Reference: Section 6.1.6 NSID and Namespace Usage; NVM Express 1.4c spec.
      
      Make sure this specific setup is supported in Linux.
      
      Fixes: 9ad1927a ("nvme: always search for namespace head")
      Signed-off-by: NSungup Moon <sungup.moon@samsung.com>
      [hch: refactored and fixed the controller vs subsystem based naming
            conflict]
      Signed-off-by: NChristoph Hellwig <hch@lst.de>
      Reviewed-by: NSagi Grimberg <sagi@grimberg.me>
      5974ea7c
    • C
      nvmet: remove redundant assignment after left shift · 63bc732c
      Colin Ian King 提交于
      The left shift is followed by a re-assignment back to cc_css, the
      assignment is redundant.  Fix this by replacing the "<<=" operator with
      "<<" instead.
      
      This cleans up the clang scan build warning:
      
      drivers/nvme/target/core.c:1124:10: warning: Although the value stored to 'cc_css' is used in the enclosing expression, the value is never actually read from 'cc_css' [deadcode.DeadStores]
      Signed-off-by: NColin Ian King <colin.i.king@gmail.com>
      Reviewed-by: NKeith Busch <kbusch@kernel.org>
      Reviewed-by: NSagi Grimberg <sagi@grimberg.me>
      Reviewed-by: NChaitanya Kulkarni <kch@nvidia.com>
      Signed-off-by: NChristoph Hellwig <hch@lst.de>
      63bc732c
    • S
      nvmet: use a private workqueue instead of the system workqueue · 8832cf92
      Sagi Grimberg 提交于
      Any attempt to flush kernel-global WQs has possibility of deadlock
      so we should simply stop using them, instead introduce nvmet_wq
      which is the generic nvmet workqueue for work elements that
      don't explicitly require a dedicated workqueue (by the mere fact
      that they are using the system_wq).
      
      Changes were done using the following replaces:
      
       - s/schedule_work(/queue_work(nvmet_wq, /g
       - s/schedule_delayed_work(/queue_delayed_work(nvmet_wq, /g
       - s/flush_scheduled_work()/flush_workqueue(nvmet_wq)/g
      Reported-by: NTetsuo Handa <penguin-kernel@I-love.SAKURA.ne.jp>
      Signed-off-by: NSagi Grimberg <sagi@grimberg.me>
      Reviewed-by: NChaitanya Kulkarni <kch@nvidia.com>
      Signed-off-by: NChristoph Hellwig <hch@lst.de>
      8832cf92
  4. 23 3月, 2022 3 次提交
  5. 21 3月, 2022 3 次提交
  6. 18 3月, 2022 2 次提交
    • J
      Merge tag 'nvme-5.18-2022-03-17' of git://git.infradead.org/nvme into for-5.18/drivers · ae53aea6
      Jens Axboe 提交于
      Pull NVMe updates from Christoph:
      
      "Second round of nvme updates for Linux 5.18
      
       - add lockdep annotations for in-kernel sockets (Chris Leech)
       - use vmalloc for ANA log buffer (Hannes Reinecke)
       - kerneldoc fixes (Chaitanya Kulkarni)
       - cleanups (Guoqing Jiang, Chaitanya Kulkarni, me)
       - warn about shared namespaces without multipathing (me)"
      
      * tag 'nvme-5.18-2022-03-17' of git://git.infradead.org/nvme:
        nvme: warn about shared namespaces without CONFIG_NVME_MULTIPATH
        nvme: remove nvme_alloc_request and nvme_alloc_request_qid
        nvme: cleanup how disk->disk_name is assigned
        nvmet: move the call to nvmet_ns_changed out of nvmet_ns_revalidate
        nvmet: use snprintf() with PAGE_SIZE in configfs
        nvmet: don't fold lines
        nvmet-rdma: fix kernel-doc warning for nvmet_rdma_device_removal
        nvmet-fc: fix kernel-doc warning for nvmet_fc_unregister_targetport
        nvmet-fc: fix kernel-doc warning for nvmet_fc_register_targetport
        nvme-tcp: lockdep: annotate in-kernel sockets
        nvme-tcp: don't fold the line
        nvme-tcp: don't initialize ret variable
        nvme-multipath: call bio_io_error in nvme_ns_head_submit_bio
        nvme-multipath: use vmalloc for ANA log buffer
      ae53aea6
    • R
      virtio_blk: eliminate anonymous module_init & module_exit · bcfe9b6c
      Randy Dunlap 提交于
      Eliminate anonymous module_init() and module_exit(), which can lead to
      confusion or ambiguity when reading System.map, crashes/oops/bugs,
      or an initcall_debug log.
      
      Give each of these init and exit functions unique driver-specific
      names to eliminate the anonymous names.
      
      Example 1: (System.map)
       ffffffff832fc78c t init
       ffffffff832fc79e t init
       ffffffff832fc8f8 t init
      
      Example 2: (initcall_debug log)
       calling  init+0x0/0x12 @ 1
       initcall init+0x0/0x12 returned 0 after 15 usecs
       calling  init+0x0/0x60 @ 1
       initcall init+0x0/0x60 returned 0 after 2 usecs
       calling  init+0x0/0x9a @ 1
       initcall init+0x0/0x9a returned 0 after 74 usecs
      
      Fixes: e467cde2 ("Block driver using virtio.")
      Signed-off-by: NRandy Dunlap <rdunlap@infradead.org>
      Cc: "Michael S. Tsirkin" <mst@redhat.com>
      Cc: Jason Wang <jasowang@redhat.com>
      Cc: Paolo Bonzini <pbonzini@redhat.com>
      Cc: Stefan Hajnoczi <stefanha@redhat.com>
      Cc: virtualization@lists.linux-foundation.org
      Cc: Jens Axboe <axboe@kernel.dk>
      Cc: linux-block@vger.kernel.org
      Reviewed-by: NStefan Hajnoczi <stefanha@redhat.com>
      Acked-by: NJason Wang <jasowang@redhat.com>
      Reviewed-by: NIra Weiny <ira.weiny@intel.com>
      Link: https://lore.kernel.org/r/20220316192010.19001-2-rdunlap@infradead.orgSigned-off-by: NJens Axboe <axboe@kernel.dk>
      bcfe9b6c
  7. 16 3月, 2022 3 次提交
  8. 15 3月, 2022 1 次提交
  9. 14 3月, 2022 10 次提交
  10. 11 3月, 2022 2 次提交
  11. 09 3月, 2022 8 次提交