summaryrefslogtreecommitdiffstats
path: root/include/net
diff options
context:
space:
mode:
authorXiaoguang Wang <xiaoguang.wang@linux.alibaba.com>2020-06-10 19:41:19 +0800
committerJens Axboe <axboe@kernel.dk>2020-06-10 17:58:46 -0600
commit7cdaf587de7c6f494b8433fded19f7728e70e1ef (patch)
tree14b4686517cdc4f0a7fa4ef2faf6b4658b198572 /include/net
parentc5b856255cbc3b664d686a83fa9397a835e063de (diff)
downloadlinux-7cdaf587de7c6f494b8433fded19f7728e70e1ef.tar.gz
linux-7cdaf587de7c6f494b8433fded19f7728e70e1ef.tar.bz2
linux-7cdaf587de7c6f494b8433fded19f7728e70e1ef.zip
io_uring: avoid whole io_wq_work copy for requests completed inline
If requests can be submitted and completed inline, we don't need to initialize whole io_wq_work in io_init_req(), which is an expensive operation, add a new 'REQ_F_WORK_INITIALIZED' to determine whether io_wq_work is initialized and add a helper io_req_init_async(), users must call io_req_init_async() for the first time touching any members of io_wq_work. I use /dev/nullb0 to evaluate performance improvement in my physical machine: modprobe null_blk nr_devices=1 completion_nsec=0 sudo taskset -c 60 fio -name=fiotest -filename=/dev/nullb0 -iodepth=128 -thread -rw=read -ioengine=io_uring -direct=1 -bs=4k -size=100G -numjobs=1 -time_based -runtime=120 before this patch: Run status group 0 (all jobs): READ: bw=724MiB/s (759MB/s), 724MiB/s-724MiB/s (759MB/s-759MB/s), io=84.8GiB (91.1GB), run=120001-120001msec With this patch: Run status group 0 (all jobs): READ: bw=761MiB/s (798MB/s), 761MiB/s-761MiB/s (798MB/s-798MB/s), io=89.2GiB (95.8GB), run=120001-120001msec About 5% improvement. Signed-off-by: Xiaoguang Wang <xiaoguang.wang@linux.alibaba.com> Signed-off-by: Jens Axboe <axboe@kernel.dk>
Diffstat (limited to 'include/net')
0 files changed, 0 insertions, 0 deletions