[Stgt-devel] Performance of SCST versus STGT
Robin Humble
robin.humble+stgt
Thu Jan 24 12:32:15 CET 2008
On Thu, Jan 24, 2008 at 11:36:45AM +0100, Bart Van Assche wrote:
>On Jan 24, 2008 8:06 AM, Robin Humble <robin.humble+stgt at anu.edu.au> wrote:
>> On Tue, Jan 22, 2008 at 01:32:08PM +0100, Bart Van Assche wrote:
>> >.............................................................................................
>> >. . STGT read SCST read . STGT read SCST read .
>> >. . performance performance . performance performance .
>> >. . (0.5K, MB/s) (0.5K, MB/s) . (1 MB >MB/s) (1 MB, MB/s) .
>> >.............................................................................................
>> >. Ethernet (1 Gb/s network) . 77 78 . 77 89 .
>> >. IPoIB (8 Gb/s network) . 163 185 . 201 239 .
>> >. iSER (8 Gb/s network) . 250 N/A . 360 N/A .
>> >. SRP (8 Gb/s network) . N/A 421 . N/A 683 .
>> >............................................................................................
>>
>> how are write speeds with SCST SRP?
>> for some kernels and tests tgt writes at >2x the read speed.
>>
>> also I see much higher speeds that what you report in my DDR 4x IB tgt
>> testing... which could be taken as inferring that tgt is scaling quite
>> nicely on the faster fabric?
>> ib_write_bw of 1473 MB/s
>> ib_read_bw of 1378 MB/s
>>
>> iSER to 7G ramfs, x86_64, centos4.6, 2.6.22 kernels, git tgtd,
>> initiator end booted with mem=512M, target with 8G ram
>>
>> direct i/o dd
>> write/read 800/751 MB/s
>> dd if=/dev/zero of=/dev/sdc bs=1M count=5000 oflag=direct
>> dd of=/dev/null if=/dev/sdc bs=1M count=5000 iflag=direct
>>
>> buffered i/o dd
>> write/read 1109/350 MB/s
>> dd if=/dev/zero of=/dev/sdc bs=1M count=5000
>> dd of=/dev/null if=/dev/sdc bs=1M count=5000
>>
>> buffered i/o lmdd
>> write/read 682/438 MB/s
>> lmdd if=internal of=/dev/sdc bs=1M count=5000
>> lmdd of=internal if=/dev/sdc bs=1M count=5000
>The tests I performed were read performance tests with dd and with
>buffered I/O. For this test you obtained 350 MB/s with STGT on a DDR
... and 1.1GB/s writes :)
presumably because buffer aggregation works well.
>4x InfiniBand network, while I obtained 360 MB/s on a SDR 4x
>InfiniBand network. I don't think that we can call this "scaling up"
>...
the direct i/o read speed being twice the buffered i/o speed would seem
to imply that Linux's page cache is being slow and confused with this
particular set of kernel + OS + OFED versions.
I doubt that this result actually says that much about tgt really.
>Regarding write performance: the write tests were performed with a
>real target (three disks in RAID-0, write bandwidth about 100 MB/s). I
I'd be interested to see ramdisk writes.
cheers,
robin
More information about the stgt
mailing list