1. 13 Feb, 2024 10 commits
  2. 12 Feb, 2024 8 commits
  3. 09 Feb, 2024 2 commits
  4. 08 Feb, 2024 13 commits
  5. 06 Feb, 2024 1 commit
  6. 05 Feb, 2024 6 commits
    • Tang Yizhou's avatar
      blk-throttle: Eliminate redundant checks for data direction · 3bca7640
      Tang Yizhou authored
      After calling throtl_peek_queued(), the data direction can be determined so
      there is no need to call bio_data_dir() to check the direction again.
      Signed-off-by: default avatarTang Yizhou <yizhou.tang@shopee.com>
      Reviewed-by: default avatarChristoph Hellwig <hch@lst.de>
      Link: https://lore.kernel.org/r/20240123081248.3752878-1-yizhou.tang@shopee.comSigned-off-by: default avatarJens Axboe <axboe@kernel.dk>
      3bca7640
    • Jens Axboe's avatar
      block: update cached timestamp post schedule/preemption · 06b23f92
      Jens Axboe authored
      Mark the task as having a cached timestamp when set assign it, so we
      can efficiently check if it needs updating post being scheduled back in.
      This covers both the actual schedule out case, which would've flushed
      the plug, and the preemption case which doesn't touch the plugged
      requests (for many reasons, one of them being then we'd need to have
      preemption disabled around plug state manipulation).
      Reviewed-by: default avatarJohannes Thumshirn <johannes.thumshirn@wdc.com>
      Signed-off-by: default avatarJens Axboe <axboe@kernel.dk>
      06b23f92
    • Jens Axboe's avatar
      block: cache current nsec time in struct blk_plug · da4c8c3d
      Jens Axboe authored
      Querying the current time is the most costly thing we do in the block
      layer per IO, and depending on kernel config settings, we may do it
      many times per IO.
      
      None of the callers actually need nsec granularity. Take advantage of
      that by caching the current time in the plug, with the assumption here
      being that any time checking will be temporally close enough that the
      slight loss of precision doesn't matter.
      
      If the block plug gets flushed, eg on preempt or schedule out, then
      we invalidate the cached clock.
      
      On a basic peak IOPS test case with iostats enabled, this changes
      the performance from:
      
      IOPS=108.41M, BW=52.93GiB/s, IOS/call=31/31
      IOPS=108.43M, BW=52.94GiB/s, IOS/call=32/32
      IOPS=108.29M, BW=52.88GiB/s, IOS/call=31/32
      IOPS=108.35M, BW=52.91GiB/s, IOS/call=32/32
      IOPS=108.42M, BW=52.94GiB/s, IOS/call=31/31
      IOPS=108.40M, BW=52.93GiB/s, IOS/call=32/32
      IOPS=108.31M, BW=52.89GiB/s, IOS/call=32/31
      
      to
      
      IOPS=118.79M, BW=58.00GiB/s, IOS/call=31/32
      IOPS=118.62M, BW=57.92GiB/s, IOS/call=31/31
      IOPS=118.80M, BW=58.01GiB/s, IOS/call=32/31
      IOPS=118.78M, BW=58.00GiB/s, IOS/call=32/32
      IOPS=118.69M, BW=57.95GiB/s, IOS/call=32/31
      IOPS=118.62M, BW=57.92GiB/s, IOS/call=32/31
      IOPS=118.63M, BW=57.92GiB/s, IOS/call=31/32
      
      which is more than a 9% improvement in performance. Looking at perf diff,
      we can see a huge reduction in time overhead:
      
          10.55%     -9.88%  [kernel.vmlinux]  [k] read_tsc
           1.31%     -1.22%  [kernel.vmlinux]  [k] ktime_get
      
      Note that since this relies on blk_plug for the caching, it's only
      applicable to the issue side. But this is where most of the time calls
      happen anyway. On the completion side, cached time stamping is done with
      struct io_comp patch, as long as the driver supports it.
      
      It's also worth noting that the above testing doesn't enable any of the
      higher cost CPU items on the block layer side, like wbt, cgroups,
      iocost, etc, which all would add additional time querying and hence
      overhead. IOW, results would likely look even better in comparison with
      those enabled, as distros would do.
      Reviewed-by: default avatarJohannes Thumshirn <johannes.thumshirn@wdc.com>
      Signed-off-by: default avatarJens Axboe <axboe@kernel.dk>
      da4c8c3d
    • Jens Axboe's avatar
      block: add blk_time_get_ns() and blk_time_get() helpers · 08420cf7
      Jens Axboe authored
      Convert any user of ktime_get_ns() to use blk_time_get_ns(), and
      ktime_get() to blk_time_get(), so we have a unified API for querying the
      current time in nanoseconds or as ktime.
      
      No functional changes intended, this patch just wraps ktime_get_ns()
      and ktime_get() with a block helper.
      Reviewed-by: default avatarJohannes Thumshirn <johannes.thumshirn@wdc.com>
      Signed-off-by: default avatarJens Axboe <axboe@kernel.dk>
      08420cf7
    • Jens Axboe's avatar
      block: move cgroup time handling code into blk.h · c4e47bbb
      Jens Axboe authored
      In preparation for moving time keeping into blk.h, move the cgroup
      related code for timestamps in here too. This will help avoid a circular
      dependency, and also moves it into a more appropriate header as this one
      is private to the block layer code.
      
      Leave struct bio_issue in blk_types.h as it's a proper time definition.
      Signed-off-by: default avatarJens Axboe <axboe@kernel.dk>
      c4e47bbb
    • Christoph Hellwig's avatar
      blk-mq: special case cached requests less · 72e84e90
      Christoph Hellwig authored
      Share the main merge / split / integrity preparation code between the
      cached request vs newly allocated request cases, and add comments
      explaining the cached request handling.
      Signed-off-by: default avatarChristoph Hellwig <hch@lst.de>
      Reviewed-by: default avatarDamien Le Moal <dlemoal@kernel.org>
      Tested-by: default avatarDamien Le Moal <dlemoal@kernel.org>
      Link: https://lore.kernel.org/r/20240124092658.2258309-4-hch@lst.deSigned-off-by: default avatarJens Axboe <axboe@kernel.dk>
      72e84e90