RE: scsi-mq
From: Elliott, Robert (Server Storage)
Date: Fri Jun 20 2014 - 20:55:44 EST
> -----Original Message-----
> From: Bart Van Assche [mailto:bvanassche@xxxxxxx]
> Sent: Wednesday, 18 June, 2014 2:09 AM
> To: Jens Axboe; Christoph Hellwig; James Bottomley
> Cc: Elliott, Robert (Server Storage); linux-scsi@xxxxxxxxxxxxxxx; linux-
> kernel@xxxxxxxxxxxxxxx
> Subject: Re: scsi-mq
>
...
> Hello Jens,
>
> Fio reports the same queue depth for use_blk_mq=Y (mq below) and
> use_blk_mq=N (sq below), namely ">=64". However, the number of context
> switches differs significantly for the random read-write tests.
>
...
> It seems like with the traditional SCSI mid-layer and block core (sq)
> that the number of context switches does not depend too much on the
> number of I/O operations but that for the multi-queue SCSI core there
> are a little bit more than two context switches per I/O in the
> particular test I ran. The "randrw" script I used for this test takes
> SCSI LUNs as arguments (/dev/sdX) and starts the fio tool as follows:
Some of those context switches might be from scsi_end_request(),
which always schedules the scsi_requeue_run_queue() function via the
requeue_work workqueue for scsi-mq. That causes lots of context
switches from a busy application thread (e.g., fio) to a
kworker thread.
As shown by ftrace:
fio-19340 [005] dNh. 12067.908444: scsi_io_completion <-scsi_finish_command
fio-19340 [005] dNh. 12067.908444: scsi_end_request <-scsi_io_completion
fio-19340 [005] dNh. 12067.908444: blk_update_request <-scsi_end_request
fio-19340 [005] dNh. 12067.908445: blk_account_io_completion <-blk_update_request
fio-19340 [005] dNh. 12067.908445: scsi_mq_free_sgtables <-scsi_end_request
fio-19340 [005] dNh. 12067.908445: scsi_free_sgtable <-scsi_mq_free_sgtables
fio-19340 [005] dNh. 12067.908445: blk_account_io_done <-__blk_mq_end_io
fio-19340 [005] dNh. 12067.908445: blk_mq_free_request <-__blk_mq_end_io
fio-19340 [005] dNh. 12067.908446: blk_mq_map_queue <-blk_mq_free_request
fio-19340 [005] dNh. 12067.908446: blk_mq_put_tag <-__blk_mq_free_request
fio-19340 [005] .N.. 12067.908446: blkdev_direct_IO <-generic_file_direct_write
kworker/5:1H-3207 [005] .... 12067.908448: scsi_requeue_run_queue <-process_one_work
kworker/5:1H-3207 [005] .... 12067.908448: scsi_run_queue <-scsi_requeue_run_queue
kworker/5:1H-3207 [005] .... 12067.908448: blk_mq_start_stopped_hw_queues <-scsi_run_queue
fio-19340 [005] .... 12067.908449: blk_start_plug <-do_blockdev_direct_IO
fio-19340 [005] .... 12067.908449: blkdev_get_block <-do_direct_IO
fio-19340 [005] .... 12067.908450: blk_throtl_bio <-generic_make_request_checks
fio-19340 [005] .... 12067.908450: blk_sq_make_request <-generic_make_request
fio-19340 [005] .... 12067.908450: blk_queue_bounce <-blk_sq_make_request
fio-19340 [005] .... 12067.908450: blk_mq_map_request <-blk_sq_make_request
fio-19340 [005] .... 12067.908451: blk_mq_queue_enter <-blk_mq_map_request
fio-19340 [005] .... 12067.908451: blk_mq_map_queue <-blk_mq_map_request
fio-19340 [005] .... 12067.908451: blk_mq_get_tag <-__blk_mq_alloc_request
fio-19340 [005] .... 12067.908451: blk_mq_bio_to_request <-blk_sq_make_request
fio-19340 [005] .... 12067.908451: blk_rq_bio_prep <-init_request_from_bio
fio-19340 [005] .... 12067.908451: blk_recount_segments <-bio_phys_segments
fio-19340 [005] .... 12067.908452: blk_account_io_start <-blk_mq_bio_to_request
fio-19340 [005] .... 12067.908452: blk_mq_hctx_mark_pending <-__blk_mq_insert_request
fio-19340 [005] .... 12067.908452: blk_mq_run_hw_queue <-blk_sq_make_request
fio-19340 [005] .... 12067.908452: blk_mq_start_request <-__blk_mq_run_hw_queue
In one snapshot just tracing scsi_end_request() and
scsi_request_run_queue(), 30K scsi_end_request() calls yielded
20k scsi_request_run_queue() calls.
In this case, blk_mq_start_stopped_hw_queues() doesn't end up
doing anything since there aren't any stopped queues to restart
(blk_mq_run_hw_queue() gets called a bit later during routine
fio work); the context switch turned out to be a waste of time.
If it did find a stopped queue, then it would call
blk_mq_run_hw_queue() itself.
---
Rob Elliott HP Server Storage
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/