All of lore.kernel.org
 help / color / mirror / Atom feed
* CFQ is worse than other IO schedulers in some cases
@ 2009-02-18  6:00 Shan Wei
  2009-02-18  8:05 ` Mike Galbraith
  2009-02-18 11:37 ` Jens Axboe
  0 siblings, 2 replies; 12+ messages in thread
From: Shan Wei @ 2009-02-18  6:00 UTC (permalink / raw)
  To: jens.axboe; +Cc: linux-kernel

I found that CFQ's performance is worse than other IO scheduer in some cases
I confirmed its phenomenon when I executed dump command and sysbench on 2.6.28.


In dump(version:dump-0.4b41-2.fc6), I confirmed 
the speed under CFQ is slower than other IO schedulers.


The Test Result(dump):
   UNIT:Mb/sec
    _______________________
    |   IO       |        | 
    | scheduler  |  Speed |
    +------------|--------|
    |cfq         | 24.310 |  
    |noop        | 36.885 |  
    |anticipatory| 34.956 |  
    |deadline    | 36.758 |  
    +----------------------


Steps to reproduce(dump):
  #dump -0uf /dev/null /dev/sda6

  #df -h /dev/sda6
   Filesystem            Size  Used Avail Use% Mounted on
   /dev/sda6              19G   10G  7.6G  57% /mnt



In sysbench(version:sysbench-0.4.10), I confirmed followings.
  - CFQ's performance is worse than other IO schedulers when only multiple
    threads test.
    (There is no difference under single thread test.)
  - It is worse than other IO scheduler when
    I used read mode. (No regression in write mode).
  - There is no difference among other IO schedulers. (e.g noop deadline)


The Test Result(sysbench):
   UNIT:Mb/sec
    __________________________________________________
    |   IO       |      thread  number               |  
    | scheduler  |-----------------------------------|
    |            |  1   |  3    |  5   |   7  |   9  |
    +------------|------|-------|------|------|------|
    |cfq         | 77.8 |  32.4 | 43.3 | 55.8 | 58.5 | 
    |noop        | 78.2 |  79.0 | 78.2 | 77.2 | 77.0 |
    |anticipatory| 78.2 |  78.6 | 78.4 | 77.8 | 78.1 |
    |deadline    | 76.9 |  78.4 | 77.0 | 78.4 | 77.9 |
    +------------------------------------------------+


Steps to reproduce(sysbench):

  (1)#echo cfq > /sys/block/sda/queue/scheduler 

  (2)#sysbench --test=fileio --num-threads=1 --file-total-size=10G --file-test-mode=seqrd prepare

  (3)#sysbench --test=fileio --num-threads=1 --file-total-size=10G --file-test-mode=seqrd run
      [snip]
      Operations performed:  655360 Read, 0 Write, 0 Other = 655360 Total
      Read 10Gb  Written 0b  Total transferred 10Gb  (77.835Mb/sec)
      4981.44 Requests/sec executed                   ~~~~~~~~~~~
  (4)#sysbench --test=fileio --num-threads=1 --file-total-size=10G --file-test-mode=seqrd cleanup

  (5)#sysbench --test=fileio --num-threads=5 --file-total-size=10G --file-test-mode=seqrd prepare
  (6)#sysbench --test=fileio --num-threads=5 --file-total-size=10G --file-test-mode=seqrd run
      [snip]
      Operations performed:  655360 Read, 0 Write, 0 Other = 655360 Total
      Read 10Gb  Written 0b  Total transferred 10Gb  (43.396Mb/sec)
      2777.35 Requests/sec executed                   ~~~~~~~~~~~~
  (7)#sysbench --test=fileio --num-threads=5 --file-total-size=10G --file-test-mode=seqrd cleanup

when doing step 2 or 5, sysbench creats 128 files, and 80M each one. 
when doing step 4 or 7, sysbench deletes the files. 
when doing step 3 or 6, thread reads these files continuously and 
reads file-block-size(default:16Kbyte) at once, just like :

       t_0   t_0   t_0   t_0   t_0   t_0   t_0
        ^     ^     ^     ^     ^     ^     ^
     ---|-----|-----|-----|-----|-----|-----|--------
file | 16k | 16k | 16k | 16k | 16k | 16k | 16k | ... 
     ------------------------------------------------ 
                  (num-threads=1)

(t_0 stand for the first thread) 

       t_0   t_1   t_2   t_3   t_4   t_0   t_1
        ^     ^     ^     ^     ^     ^     ^
     ---|-----|-----|-----|-----|-----|-----|--------
file | 16k | 16k | 16k | 16k | 16k | 16k | 16k | ... 
     ------------------------------------------------ 
                  (num-threads=5)

(the executed threads are decide by the thread scheduler)


The Hardware Infos:
Arch    :x86_64
CPU     :4cpu; GenuineIntel 3325.087 MHz
MEMORY  :4044128kB

---- 
Shan Wei


^ permalink raw reply	[flat|nested] 12+ messages in thread

end of thread, other threads:[~2009-03-09 12:38 UTC | newest]

Thread overview: 12+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2009-02-18  6:00 CFQ is worse than other IO schedulers in some cases Shan Wei
2009-02-18  8:05 ` Mike Galbraith
2009-02-18 10:15   ` Shan Wei
2009-02-18 11:35     ` Mike Galbraith
2009-03-09  5:24   ` Shan Wei
2009-03-09  7:43     ` Jens Axboe
2009-03-09 12:02       ` Shan Wei
2009-03-09 12:14         ` Jens Axboe
2009-03-09 12:31           ` Shan Wei
2009-02-18 11:37 ` Jens Axboe
2009-02-19  9:28   ` Shan Wei
2009-02-19 15:26     ` Jeff Moyer

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.