1. 22 6月, 2021 4 次提交
    • P
      block, bfq: boost throughput by extending queue-merging times · 7812472f
      Pietro Pedroni 提交于
      One of the methods with which bfq boosts throughput is by merging queues.
      One of the merging variants in bfq is the stable merge.
      This mechanism is activated between two queues only if they are created
      within a certain maximum time T1 from each other.
      Merging can happen soon or be delayed. In the second case, before
      merging, bfq needs to evaluate a throughput-boost parameter that
      indicates whether the queue generates a high throughput is served alone.
      Merging occurs when this throughput-boost is not high enough.
      In particular, this parameter is evaluated and late merging may occur
      only after at least a time T2 from the creation of the queue.
      
      Currently T1 and T2 are set to 180ms and 200ms, respectively.
      In this way the merging mechanism rarely occurs because time is not
      enough. This results in a noticeable lowering of the overall throughput
      with some workloads (see the example below).
      
      This commit introduces two constants bfq_activation_stable_merging and
      bfq_late_stable_merging in order to increase the duration of T1 and T2.
      Both the stable merging activation time and the late merging
      time are set to 600ms. This value has been experimentally evaluated
      using sqlite benchmark in the Phoronix Test Suite on a HDD.
      The duration of the benchmark before this fix was 111.02s, while now
      it has reached 97.02s, a better result than that of all the other
      schedulers.
      Signed-off-by: NPietro Pedroni <pedroni.pietro.96@gmail.com>
      Signed-off-by: NPaolo Valente <paolo.valente@linaro.org>
      Link: https://lore.kernel.org/r/20210619140948.98712-5-paolo.valente@linaro.orgSigned-off-by: NJens Axboe <axboe@kernel.dk>
      7812472f
    • P
      block, bfq: consider also creation time in delayed stable merge · d4f49983
      Paolo Valente 提交于
      Since commit 430a67f9 ("block, bfq: merge bursts of newly-created
      queues"), BFQ may schedule a merge between a newly created sync
      bfq_queue and the last sync bfq_queue created. Such a merging is not
      performed immediately, because BFQ needs first to find out whether the
      newly created queue actually reaches a higher throughput if not merged
      at all (and in that case BFQ will not perform any stable merging). To
      check that, a little time must be waited after the creation of the new
      queue, so that some I/O can flow in the queue, and statistics on such
      I/O can be computed.
      
      Yet, to evaluate the above waiting time, the last split time is
      considered as start time, instead of the creation time of the
      queue. This is a mistake, because considering the split time is
      correct only in the following scenario.
      
      The queue undergoes a non-stable merges on the arrival of its very
      first I/O request, due to close I/O with some other queue. While the
      queue is merged for close I/O, stable merging is not considered. Yet
      the queue may then happen to be split, if the close I/O finishes (or
      happens to be a false positive). From this time on, the queue can
      again be considered for stable merging. But, again, a little time must
      elapse, to let some new I/O flow in the queue and to get updated
      statistics. To wait for this time, the split time is to be taken into
      account.
      
      Yet, if the queue does not undergo a non-stable merge on the arrival
      of its very first request, then BFQ immediately checks whether the
      stable merge is to be performed. It happens because the split time for
      a queue is initialized to minus infinity when the queue is created.
      
      This commit fixes this mistake by adding the missing condition. Now
      the check for delayed stable-merge is performed after a little time is
      elapsed not only from the last queue split time, but also from the
      creation time of the queue.
      
      Fixes: 430a67f9 ("block, bfq: merge bursts of newly-created queues")
      Signed-off-by: NPaolo Valente <paolo.valente@linaro.org>
      Link: https://lore.kernel.org/r/20210619140948.98712-4-paolo.valente@linaro.orgSigned-off-by: NJens Axboe <axboe@kernel.dk>
      d4f49983
    • L
      block, bfq: fix delayed stable merge check · e03f2ab7
      Luca Mariotti 提交于
      When attempting to schedule a merge of a given bfq_queue with the currently
      in-service bfq_queue or with a cooperating bfq_queue among the scheduled
      bfq_queues, delayed stable merge is checked for rotational or non-queueing
      devs. For this stable merge to be performed, some conditions must be met.
      If the current bfq_queue underwent some split from some merged bfq_queue,
      one of these conditions is that two hundred milliseconds must elapse from
      split, otherwise this condition is always met.
      
      Unfortunately, by mistake, time_is_after_jiffies() was written instead of
      time_is_before_jiffies() for this check, verifying that less than two
      hundred milliseconds have elapsed instead of verifying that at least two
      hundred milliseconds have elapsed.
      
      Fix this issue by replacing time_is_after_jiffies() with
      time_is_before_jiffies().
      Signed-off-by: NLuca Mariotti <mariottiluca1@hotmail.it>
      Signed-off-by: NPaolo Valente <paolo.valente@unimore.it>
      Signed-off-by: NPietro Pedroni <pedroni.pietro.96@gmail.com>
      Link: https://lore.kernel.org/r/20210619140948.98712-3-paolo.valente@linaro.orgSigned-off-by: NJens Axboe <axboe@kernel.dk>
      e03f2ab7
    • P
      block, bfq: let also stably merged queues enjoy weight raising · 511a2699
      Paolo Valente 提交于
      Merged bfq_queues are kept out of weight-raising (low-latency)
      mechanisms. The reason is that these queues are usually created for
      non-interactive and non-soft-real-time tasks. Yet this is not the case
      for stably-merged queues. These queues are merged just because they
      are created shortly after each other. So they may easily serve the I/O
      of an interactive or soft-real time application, if the application
      happens to spawn multiple processes.
      
      To address this issue, this commits lets also stably-merged queued
      enjoy weight raising.
      Signed-off-by: NPaolo Valente <paolo.valente@linaro.org>
      Link: https://lore.kernel.org/r/20210619140948.98712-2-paolo.valente@linaro.orgSigned-off-by: NJens Axboe <axboe@kernel.dk>
      511a2699
  2. 12 5月, 2021 1 次提交
    • P
      block, bfq: avoid circular stable merges · 7ea96eef
      Paolo Valente 提交于
      BFQ may merge a new bfq_queue, stably, with the last bfq_queue
      created. In particular, BFQ first waits a little bit for some I/O to
      flow inside the new queue, say Q2, if this is needed to understand
      whether it is better or worse to merge Q2 with the last queue created,
      say Q1. This delayed stable merge is performed by assigning
      bic->stable_merge_bfqq = Q1, for the bic associated with Q1.
      
      Yet, while waiting for some I/O to flow in Q2, a non-stable queue
      merge of Q2 with Q1 may happen, causing the bic previously associated
      with Q2 to be associated with exactly Q1 (bic->bfqq = Q1). After that,
      Q2 and Q1 may happen to be split, and, in the split, Q1 may happen to
      be recycled as a non-shared bfq_queue. In that case, Q1 may then
      happen to undergo a stable merge with the bfq_queue pointed by
      bic->stable_merge_bfqq. Yet bic->stable_merge_bfqq still points to
      Q1. So Q1 would be merged with itself.
      
      This commit fixes this error by intercepting this situation, and
      canceling the schedule of the stable merge.
      
      Fixes: 430a67f9 ("block, bfq: merge bursts of newly-created queues")
      Signed-off-by: NPietro Pedroni <pedroni.pietro.96@gmail.com>
      Signed-off-by: NPaolo Valente <paolo.valente@linaro.org>
      Link: https://lore.kernel.org/r/20210512094352.85545-2-paolo.valente@linaro.orgSigned-off-by: NJens Axboe <axboe@kernel.dk>
      7ea96eef
  3. 11 5月, 2021 1 次提交
    • O
      kyber: fix out of bounds access when preempted · efed9a33
      Omar Sandoval 提交于
      __blk_mq_sched_bio_merge() gets the ctx and hctx for the current CPU and
      passes the hctx to ->bio_merge(). kyber_bio_merge() then gets the ctx
      for the current CPU again and uses that to get the corresponding Kyber
      context in the passed hctx. However, the thread may be preempted between
      the two calls to blk_mq_get_ctx(), and the ctx returned the second time
      may no longer correspond to the passed hctx. This "works" accidentally
      most of the time, but it can cause us to read garbage if the second ctx
      came from an hctx with more ctx's than the first one (i.e., if
      ctx->index_hw[hctx->type] > hctx->nr_ctx).
      
      This manifested as this UBSAN array index out of bounds error reported
      by Jakub:
      
      UBSAN: array-index-out-of-bounds in ../kernel/locking/qspinlock.c:130:9
      index 13106 is out of range for type 'long unsigned int [128]'
      Call Trace:
       dump_stack+0xa4/0xe5
       ubsan_epilogue+0x5/0x40
       __ubsan_handle_out_of_bounds.cold.13+0x2a/0x34
       queued_spin_lock_slowpath+0x476/0x480
       do_raw_spin_lock+0x1c2/0x1d0
       kyber_bio_merge+0x112/0x180
       blk_mq_submit_bio+0x1f5/0x1100
       submit_bio_noacct+0x7b0/0x870
       submit_bio+0xc2/0x3a0
       btrfs_map_bio+0x4f0/0x9d0
       btrfs_submit_data_bio+0x24e/0x310
       submit_one_bio+0x7f/0xb0
       submit_extent_page+0xc4/0x440
       __extent_writepage_io+0x2b8/0x5e0
       __extent_writepage+0x28d/0x6e0
       extent_write_cache_pages+0x4d7/0x7a0
       extent_writepages+0xa2/0x110
       do_writepages+0x8f/0x180
       __writeback_single_inode+0x99/0x7f0
       writeback_sb_inodes+0x34e/0x790
       __writeback_inodes_wb+0x9e/0x120
       wb_writeback+0x4d2/0x660
       wb_workfn+0x64d/0xa10
       process_one_work+0x53a/0xa80
       worker_thread+0x69/0x5b0
       kthread+0x20b/0x240
       ret_from_fork+0x1f/0x30
      
      Only Kyber uses the hctx, so fix it by passing the request_queue to
      ->bio_merge() instead. BFQ and mq-deadline just use that, and Kyber can
      map the queues itself to avoid the mismatch.
      
      Fixes: a6088845 ("block: kyber: make kyber more friendly with merging")
      Reported-by: NJakub Kicinski <kuba@kernel.org>
      Signed-off-by: NOmar Sandoval <osandov@fb.com>
      Link: https://lore.kernel.org/r/c7598605401a48d5cfeadebb678abd10af22b83f.1620691329.git.osandov@fb.comSigned-off-by: NJens Axboe <axboe@kernel.dk>
      efed9a33
  4. 16 4月, 2021 1 次提交
    • L
      bfq/mq-deadline: remove redundant check for passthrough request · 7687b38a
      Lin Feng 提交于
      Since commit 01e99aec 'blk-mq: insert passthrough request into
      hctx->dispatch directly', passthrough request should not appear in
      IO-scheduler any more, so blk_rq_is_passthrough checking in addon IO
      schedulers is redundant.
      
      (Notes: this patch passes generic IO load test with hdds under SAS
      controller and hdds under AHCI controller but obviously not covers all.
      Not sure if passthrough request can still escape into IO scheduler from
      blk_mq_sched_insert_requests, which is used by blk_mq_flush_plug_list and
      has lots of indirect callers.)
      Signed-off-by: NLin Feng <linf@wangsu.com>
      Reviewed-by: NMing Lei <ming.lei@redhat.com>
      Signed-off-by: NJens Axboe <axboe@kernel.dk>
      7687b38a
  5. 26 3月, 2021 6 次提交
  6. 03 3月, 2021 1 次提交
  7. 22 2月, 2021 1 次提交
  8. 03 2月, 2021 1 次提交
    • L
      bfq-iosched: Revert "bfq: Fix computation of shallow depth" · 388c705b
      Lin Feng 提交于
      This reverts commit 6d4d2735.
      
      bfq.limit_depth passes word_depths[] as shallow_depth down to sbitmap core
      sbitmap_get_shallow, which uses just the number to limit the scan depth of
      each bitmap word, formula:
      scan_percentage_for_each_word = shallow_depth / (1 << sbimap->shift) * 100%
      
      That means the comments's percentiles 50%, 75%, 18%, 37% of bfq are correct.
      But after commit patch 'bfq: Fix computation of shallow depth', we use
      sbitmap.depth instead, as a example in following case:
      
      sbitmap.depth = 256, map_nr = 4, shift = 6; sbitmap_word.depth = 64.
      The resulsts of computed bfqd->word_depths[] are {128, 192, 48, 96}, and
      three of the numbers exceed core dirver's 'sbitmap_word.depth=64' limit
      nothing.
      Signed-off-by: NLin Feng <linf@wangsu.com>
      Reviewed-by: NJan Kara <jack@suse.cz>
      Signed-off-by: NJens Axboe <axboe@kernel.dk>
      388c705b
  9. 28 1月, 2021 3 次提交
    • J
      bfq: Use only idle IO periods for think time calculations · 7684fbde
      Jan Kara 提交于
      Currently whenever bfq queue has a request queued we add now -
      last_completion_time to the think time statistics. This is however
      misleading in case the process is able to submit several requests in
      parallel because e.g. if the queue has request completed at time T0 and
      then queues new requests at times T1, T2, then we will add T1-T0 and
      T2-T0 to think time statistics which just doesn't make any sence (the
      queue's think time is penalized by the queue being able to submit more
      IO). So add to think time statistics only time intervals when the queue
      had no IO pending.
      Signed-off-by: NJan Kara <jack@suse.cz>
      Acked-by: NPaolo Valente <paolo.valente@linaro.org>
      [axboe: fix whitespace on empty line]
      Signed-off-by: NJens Axboe <axboe@kernel.dk>
      7684fbde
    • J
      bfq: Use 'ttime' local variable · 28c6def0
      Jan Kara 提交于
      Use local variable 'ttime' instead of dereferencing bfqq.
      Signed-off-by: NJan Kara <jack@suse.cz>
      Acked-by: NPaolo Valente <paolo.valente@linaro.org>
      Signed-off-by: NJens Axboe <axboe@kernel.dk>
      28c6def0
    • J
      bfq: Avoid false bfq queue merging · 41e76c85
      Jan Kara 提交于
      bfq_setup_cooperator() uses bfqd->in_serv_last_pos so detect whether it
      makes sense to merge current bfq queue with the in-service queue.
      However if the in-service queue is freshly scheduled and didn't dispatch
      any requests yet, bfqd->in_serv_last_pos is stale and contains value
      from the previously scheduled bfq queue which can thus result in a bogus
      decision that the two queues should be merged. This bug can be observed
      for example with the following fio jobfile:
      
      [global]
      direct=0
      ioengine=sync
      invalidate=1
      size=1g
      rw=read
      
      [reader]
      numjobs=4
      directory=/mnt
      
      where the 4 processes will end up in the one shared bfq queue although
      they do IO to physically very distant files (for some reason I was able to
      observe this only with slice_idle=1ms setting).
      
      Fix the problem by invalidating bfqd->in_serv_last_pos when switching
      in-service queue.
      
      Fixes: 058fdecc ("block, bfq: fix in-service-queue check for queue merging")
      CC: stable@vger.kernel.org
      Signed-off-by: NJan Kara <jack@suse.cz>
      Acked-by: NPaolo Valente <paolo.valente@linaro.org>
      Signed-off-by: NJens Axboe <axboe@kernel.dk>
      41e76c85
  10. 26 1月, 2021 7 次提交
  11. 25 1月, 2021 7 次提交
    • J
      Revert "blk-mq, elevator: Count requests per hctx to improve performance" · 5ac83c64
      Jan Kara 提交于
      This reverts commit b445547e.
      
      Since both mq-deadline and BFQ completely ignore hctx they are passed to
      their dispatch function and dispatch whatever request they deem fit
      checking whether any request for a particular hctx is queued is just
      pointless since we'll very likely get a request from a different hctx
      anyway. In the following commit we'll deal with lock contention in these
      IO schedulers in presence of multiple HW queues in a different way.
      Signed-off-by: NJan Kara <jack@suse.cz>
      Reviewed-by: NMing Lei <ming.lei@redhat.com>
      Signed-off-by: NJens Axboe <axboe@kernel.dk>
      5ac83c64
    • P
      block, bfq: do not expire a queue when it is the only busy one · 2391d13e
      Paolo Valente 提交于
      This commits preserves I/O-dispatch plugging for a special symmetric
      case that may suddenly turn into asymmetric: the case where only one
      bfq_queue, say bfqq, is busy. In this case, not expiring bfqq does not
      cause any harm to any other queues in terms of service guarantees. In
      contrast, it avoids the following unlucky sequence of events: (1) bfqq
      is expired, (2) a new queue with a lower weight than bfqq becomes busy
      (or more queues), (3) the new queue is served until a new request
      arrives for bfqq, (4) when bfqq is finally served, there are so many
      requests of the new queue in the drive that the pending requests for
      bfqq take a lot of time to be served. In particular, event (2) may
      case even already dispatched requests of bfqq to be delayed, inside
      the drive. So, to avoid this series of events, the scenario is
      preventively declared as asymmetric also if bfqq is the only busy
      queues. By doing so, I/O-dispatch plugging is performed for bfqq.
      Tested-by: NJan Kara <jack@suse.cz>
      Signed-off-by: NPaolo Valente <paolo.valente@linaro.org>
      Signed-off-by: NJens Axboe <axboe@kernel.dk>
      2391d13e
    • P
      block, bfq: avoid spurious switches to soft_rt of interactive queues · 3c337690
      Paolo Valente 提交于
      BFQ tags some bfq_queues as interactive or soft_rt if it deems that
      these bfq_queues contain the I/O of, respectively, interactive or soft
      real-time applications. BFQ privileges both these special types of
      bfq_queues over normal bfq_queues. To privilege a bfq_queue, BFQ
      mainly raises the weight of the bfq_queue. In particular, soft_rt
      bfq_queues get a higher weight than interactive bfq_queues.
      
      A bfq_queue may turn from interactive to soft_rt. And this leads to a
      tricky issue. Soft real-time applications usually start with an
      I/O-bound, interactive phase, in which they load themselves into main
      memory. BFQ correctly detects this phase, and keeps the bfq_queues
      associated with the application in interactive mode for a
      while. Problems arise when the I/O pattern of the application finally
      switches to soft real-time. One of the conditions for a bfq_queue to
      be deemed as soft_rt is that the bfq_queue does not consume too much
      bandwidth. But the bfq_queues associated with a soft real-time
      application consume as much bandwidth as they can in the loading phase
      of the application. So, after the application becomes truly soft
      real-time, a lot of time should pass before the average bandwidth
      consumed by its bfq_queues finally drops to a value acceptable for
      soft_rt bfq_queues. As a consequence, there might be a time gap during
      which the application is not privileged at all, because its bfq_queues
      are not interactive any longer, but cannot be deemed as soft_rt yet.
      
      To avoid this problem, BFQ pretends that an interactive bfq_queue
      consumes zero bandwidth, and allows an interactive bfq_queue to switch
      to soft_rt. Yet, this fake zero-bandwidth consumption easily causes
      the bfq_queue to often switch to soft_rt deceptively, during its
      loading phase. As in soft_rt mode, the bfq_queue gets its bandwidth
      correctly computed, and therefore soon switches back to
      interactive. Then it switches again to soft_rt, and so on. These
      spurious fluctuations usually cause losses of throughput, because they
      deceive BFQ's mechanisms for boosting throughput (injection,
      I/O-plugging avoidance, ...).
      
      This commit addresses this issue as follows:
      1) It does compute actual bandwidth consumption also for interactive
         bfq_queues. This avoids the above false positives.
      2) When a bfq_queue switches from interactive to normal mode, the
         consumed bandwidth is reset (forgotten). This allows the
         bfq_queue to enjoy soft_rt very quickly. In particular, two
         alternatives are possible in this switch:
          - the bfq_queue still has backlog, and therefore there is a budget
            already scheduled to serve the bfq_queue; in this case, the
            scheduling of the current budget of the bfq_queue is not
            hindered, because only the scheduling of the next budget will
            be affected by the weight drop. After that, if the bfq_queue is
            actually in a soft_rt phase, and becomes empty during the
            service of its current budget, which is the natural behavior of
            a soft_rt bfq_queue, then the bfq_queue will be considered as
            soft_rt when its next I/O arrives. If, in contrast, the
            bfq_queue remains constantly non-empty, then its next budget
            will be scheduled with a low weight, which is the natural
            treatment for an I/O-bound (non soft_rt) bfq_queue.
          - the bfq_queue is empty; in this case, the bfq_queue may be
            considered unjustly soft_rt when its new I/O arrives. Yet
            the problem is now much smaller than before, because it is
            unlikely that more than one spurious fluctuation occurs.
      Tested-by: NJan Kara <jack@suse.cz>
      Signed-off-by: NPaolo Valente <paolo.valente@linaro.org>
      Signed-off-by: NJens Axboe <axboe@kernel.dk>
      3c337690
    • P
      block, bfq: do not raise non-default weights · 91b896f6
      Paolo Valente 提交于
      BFQ heuristics try to detect interactive I/O, and raise the weight of
      the queues containing such an I/O. Yet, if also the user changes the
      weight of a queue (i.e., the user changes the ioprio of the process
      associated with that queue), then it is most likely better to prevent
      BFQ heuristics from silently changing the same weight.
      Tested-by: NJan Kara <jack@suse.cz>
      Signed-off-by: NPaolo Valente <paolo.valente@linaro.org>
      Signed-off-by: NJens Axboe <axboe@kernel.dk>
      91b896f6
    • P
      block, bfq: increase time window for waker detection · ab1fb47e
      Paolo Valente 提交于
      Tests on slower machines showed current window to be way too
      small. This commit increases it.
      Tested-by: NJan Kara <jack@suse.cz>
      Signed-off-by: NPaolo Valente <paolo.valente@linaro.org>
      Signed-off-by: NJens Axboe <axboe@kernel.dk>
      ab1fb47e
    • J
      block, bfq: set next_rq to waker_bfqq->next_rq in waker injection · d4fc3640
      Jia Cheng Hu 提交于
      Since commit c5089591c3ba ("block, bfq: detect wakers and
      unconditionally inject their I/O"), when the in-service bfq_queue, say
      Q, is temporarily empty, BFQ checks whether there are I/O requests to
      inject (also) from the waker bfq_queue for Q. To this goal, the value
      pointed by bfqq->waker_bfqq->next_rq must be controlled. However, the
      current implementation mistakenly looks at bfqq->next_rq, which
      instead points to the next request of the currently served queue.
      
      This mistake evidently causes losses of throughput in scenarios with
      waker bfq_queues.
      
      This commit corrects this mistake.
      
      Fixes: c5089591c3ba ("block, bfq: detect wakers and unconditionally inject their I/O")
      Signed-off-by: NJia Cheng Hu <jia.jiachenghu@gmail.com>
      Signed-off-by: NJan Kara <jack@suse.cz>
      Signed-off-by: NPaolo Valente <paolo.valente@linaro.org>
      Signed-off-by: NJens Axboe <axboe@kernel.dk>
      d4fc3640
    • P
      block, bfq: use half slice_idle as a threshold to check short ttime · b5f74eca
      Paolo Valente 提交于
      The value of the I/O plugging (idling) timeout is used also as the
      think-time threshold to decide whether a process has a short think
      time.  In this respect, a good value of this timeout for rotational
      drives is un the order of several ms. Yet, this is often too long a
      time interval to be effective as a think-time threshold. This commit
      mitigates this problem (by a lot, according to tests), by halving the
      threshold.
      Tested-by: NJan Kara <jack@suse.cz>
      Signed-off-by: NPaolo Valente <paolo.valente@linaro.org>
      Signed-off-by: NJens Axboe <axboe@kernel.dk>
      b5f74eca
  12. 06 1月, 2021 1 次提交
    • J
      bfq: Fix computation of shallow depth · 6d4d2735
      Jan Kara 提交于
      BFQ computes number of tags it allows to be allocated for each request type
      based on tag bitmap. However it uses 1 << bitmap.shift as number of
      available tags which is wrong. 'shift' is just an internal bitmap value
      containing logarithm of how many bits bitmap uses in each bitmap word.
      Thus number of tags allowed for some request types can be far to low.
      Use proper bitmap.depth which has the number of tags instead.
      Signed-off-by: NJan Kara <jack@suse.cz>
      Signed-off-by: NJens Axboe <axboe@kernel.dk>
      6d4d2735
  13. 09 9月, 2020 1 次提交
    • O
      block: only call sched requeue_request() for scheduled requests · e8a8a185
      Omar Sandoval 提交于
      Yang Yang reported the following crash caused by requeueing a flush
      request in Kyber:
      
        [    2.517297] Unable to handle kernel paging request at virtual address ffffffd8071c0b00
        ...
        [    2.517468] pc : clear_bit+0x18/0x2c
        [    2.517502] lr : sbitmap_queue_clear+0x40/0x228
        [    2.517503] sp : ffffff800832bc60 pstate : 00c00145
        ...
        [    2.517599] Process ksoftirqd/5 (pid: 51, stack limit = 0xffffff8008328000)
        [    2.517602] Call trace:
        [    2.517606]  clear_bit+0x18/0x2c
        [    2.517619]  kyber_finish_request+0x74/0x80
        [    2.517627]  blk_mq_requeue_request+0x3c/0xc0
        [    2.517637]  __scsi_queue_insert+0x11c/0x148
        [    2.517640]  scsi_softirq_done+0x114/0x130
        [    2.517643]  blk_done_softirq+0x7c/0xb0
        [    2.517651]  __do_softirq+0x208/0x3bc
        [    2.517657]  run_ksoftirqd+0x34/0x60
        [    2.517663]  smpboot_thread_fn+0x1c4/0x2c0
        [    2.517667]  kthread+0x110/0x120
        [    2.517669]  ret_from_fork+0x10/0x18
      
      This happens because Kyber doesn't track flush requests, so
      kyber_finish_request() reads a garbage domain token. Only call the
      scheduler's requeue_request() hook if RQF_ELVPRIV is set (like we do for
      the finish_request() hook in blk_mq_free_request()). Now that we're
      handling it in blk-mq, also remove the check from BFQ.
      Reported-by: NYang Yang <yang.yang@vivo.com>
      Signed-off-by: NOmar Sandoval <osandov@fb.com>
      Signed-off-by: NJens Axboe <axboe@kernel.dk>
      e8a8a185
  14. 04 9月, 2020 2 次提交
  15. 24 8月, 2020 1 次提交
  16. 01 8月, 2020 1 次提交
  17. 30 5月, 2020 1 次提交