[dm-devel] Poor iSCSI performance
John A. Sullivan III
jsullivan at opensourcedevel.com
Mon Mar 16 09:58:57 UTC 2009
On Mon, 2009-03-16 at 05:57 -0400, John A. Sullivan III wrote:
> Hello, all. We've been struggling to tweak performance between on Linux
> iSCSI initiators (open-iscsi) and our opensolar iSCSI targets (Nexenta).
> On top of our generally poor performance (max 7000 IOPS per GbE NIC), we
> are seeing abysmal performance when we try to compensate by using either
> dm-multipath or dmadm to use multiple iSCSI LUNs.
>
> We have been testing using an eight processor Linux server with 6 GbE
> network interfaces speaking to a Nexenta based Z200 storage system from
> Pogo Linux with 10 GbE ports. I will attach a text file with some
> results using disktest.
>
> In summary, if we ran four completely independent tests against four
> separate targets on four separate NICs, we achieved an aggregate 24940
> IOPS with 512 byte blocks and 6713 IOPS with 64KB blocks.
>
> However, we would prefer to treat the storage as a single disk and so
> attempted to use software RAID, i.e., we created four LUNs, presented
> them as four separate disks and then used software RAID0 to stripe
> across all four targets. We expected slightly less than the performance
> cited above. Instead, we received 4450 IOPS for 512 and for 64KB.
>
> We then took a different approach and created one big LUN with eight
> paths to the target using dm-multipath multibus with round-robin
> scheduling and rr_min_io=100. Our numbers were 4350 IOPS for 512 and
> 1450 IOPS with 64KB.
>
> We then suspected it might an issue of the number of threads rather than
> just the number of disks, i.e., the four independent disk test was using
> four separate processes. So we ran four separate, concurrent tests
> against the RAID0 array and the multipath setup.
>
> RAID0 increased to 11720 IOPS for 512 and 3188 IOPS for 64 KB - still a
> far cry from 24900 and and 6713. dm-multipath numbers were 10140 IOPS
> for 512 and 2563 IOPS for 64KB. Moreover, the CPU utilization was
> brutal.
>
> /etc/multipath.conf:
> blacklist {
> # devnode "*"
> # sdb
> wwid SATA_ST3250310NS_9SF0L234
> #sda
> wwid SATA_ST3250310NS_9SF0LVSR
> # The above does not seem to be working thus we will do
> devnode "^sd[ab]$"
> # This is usually a bad idea as the device names can change
> # However, since we add our iSCSI devices long after boot, I
> think we are safe
> }
> defaults {
> udev_dir /dev
> polling_interval 5
> selector "round-robin 0"
> path_grouping_policy multibus
> getuid_callout "/sbin/scsi_id -g -u -s /block/%n"
> prio_callout /bin/true
> path_checker readsector0
> rr_min_io 100
> max_fds 8192
> rr_weight priorities
> failback immediate
> no_path_retry fail
> # user_friendly_names yes
> }
> multipaths {
> multipath {
> wwid
> 3600144f0e2824900000049b98e2b0001
> alias isda
> }
> multipath {
> wwid
> 3600144f0e2824900000049b062950002
> alias isdplain
> }
> multipath {
> wwid
> 3600144f0e2824900000049b9bb350001
> alias isdb
> }
> multipath {
> wwid
> 3600144f0e2824900000049b9bb350002
> alias isdc
> }
> multipath {
> wwid
> 3600144f0e2824900000049b9bb360003
> alias isdd
> }
> multipath {
> wwid
> 3600144f0e2824900000049b7878a0006
> alias isdtest
> }
> }
> devices {
> device {
> vendor "NEXENTA"
> product "COMSTAR"
> # vendor "SUN"
> # product "SOLARIS"
> getuid_callout "/sbin/scsi_id -g -u -s /block/%
> n"
> features "0"
> hardware_handler "0"
> # path_grouping_policy failover
> rr_weight uniform
> # rr_min_io 1000
> path_checker readsector0
> }
> }
>
> What would account for such miserable performance? How can we improve
> it? We do not want to proliferate disks just to increase aggregate
> performance. Thanks - John
Oops! Forgot to attach the file - John
--
John A. Sullivan III
Open Source Development Corporation
+1 207-985-7880
jsullivan at opensourcedevel.com
http://www.spiritualoutreach.com
Making Christianity intelligible to secular society
-------------- next part --------------
Disktest notes
./disktest -B512 -h1 -ID -pL -K100 -PT -T300 -r /dev/raw/raw1
Eight paths to one target - round robin multipath:
512 blocks - 4350 IOPS:
Tasks: 212 total, 1 running, 211 sleeping, 0 stopped, 0 zombie
Cpu0 : 0.0%us, 0.0%sy, 0.0%ni,100.0%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu1 : 0.0%us, 1.5%sy, 0.0%ni, 0.0%id, 96.9%wa, 0.2%hi, 1.4%si, 0.0%st
Cpu2 : 0.0%us, 1.0%sy, 0.0%ni, 0.4%id, 96.8%wa, 0.4%hi, 1.4%si, 0.0%st
Cpu3 : 0.0%us, 0.1%sy, 0.0%ni, 99.7%id, 0.0%wa, 0.1%hi, 0.1%si, 0.0%st
Cpu4 : 0.1%us, 1.8%sy, 0.0%ni, 0.0%id, 94.9%wa, 0.3%hi, 2.9%si, 0.0%st
Cpu5 : 0.2%us, 1.5%sy, 0.0%ni, 0.0%id, 96.2%wa, 0.3%hi, 1.8%si, 0.0%st
Cpu6 : 0.0%us, 0.1%sy, 0.0%ni, 99.8%id, 0.0%wa, 0.1%hi, 0.0%si, 0.0%st
Cpu7 : 0.0%us, 0.0%sy, 0.0%ni,100.0%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
64K blocks - 2000 - 2200 IOPS: (1450 with multipath - I forgot to start multipathd the first time)
top - 21:59:45 up 55 min, 2 users, load average: 89.52, 44.36, 23.60
Tasks: 212 total, 1 running, 211 sleeping, 0 stopped, 0 zombie
Cpu0 : 0.0%us, 0.0%sy, 0.0%ni, 99.7%id, 0.3%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu1 : 0.0%us, 0.2%sy, 0.0%ni, 85.7%id, 14.0%wa, 0.0%hi, 0.1%si, 0.0%st
Cpu2 : 0.1%us, 2.1%sy, 0.0%ni, 0.0%id, 91.3%wa, 1.2%hi, 5.3%si, 0.0%st
Cpu3 : 0.0%us, 0.0%sy, 0.0%ni, 98.6%id, 1.4%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu4 : 0.2%us, 1.8%sy, 0.0%ni, 0.0%id, 89.2%wa, 0.4%hi, 8.4%si, 0.0%st
Cpu5 : 0.2%us, 2.3%sy, 0.0%ni, 0.0%id, 93.3%wa, 0.2%hi, 4.0%si, 0.0%st
Cpu6 : 0.0%us, 1.7%sy, 0.0%ni, 0.0%id, 94.9%wa, 0.2%hi, 3.2%si, 0.0%st
Cpu7 : 0.0%us, 0.0%sy, 0.0%ni, 89.2%id, 10.8%wa, 0.0%hi, 0.0%si, 0.0%st
RAID0 across four targets
512 blocks - 4450 IOPS
top - 22:20:36 up 1:16, 2 users, load average: 81.25, 34.24, 24.49
Tasks: 288 total, 1 running, 287 sleeping, 0 stopped, 0 zombie
Cpu0 : 0.0%us, 0.0%sy, 0.0%ni,100.0%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu1 : 0.0%us, 1.2%sy, 0.0%ni, 0.0%id, 96.4%wa, 0.3%hi, 2.1%si, 0.0%st
Cpu2 : 0.0%us, 0.3%sy, 0.0%ni, 99.5%id, 0.0%wa, 0.0%hi, 0.2%si, 0.0%st
Cpu3 : 0.1%us, 3.3%sy, 0.0%ni, 0.0%id, 92.1%wa, 0.5%hi, 4.0%si, 0.0%st
Cpu4 : 0.1%us, 1.3%sy, 0.0%ni, 0.0%id, 96.1%wa, 0.7%hi, 1.8%si, 0.0%st
Cpu5 : 0.0%us, 0.0%sy, 0.0%ni, 99.6%id, 0.0%wa, 0.1%hi, 0.3%si, 0.0%st
Cpu6 : 0.0%us, 0.0%sy, 0.0%ni, 99.9%id, 0.0%wa, 0.1%hi, 0.0%si, 0.0%st
Cpu7 : 0.0%us, 0.0%sy, 0.0%ni,100.0%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
Mem: 33020800k total, 497424k used, 32523376k free, 26504k buffers
64K blocks - 1280 IOPS
top - 22:23:26 up 1:19, 2 users, load average: 97.37, 60.75, 36.23
Tasks: 288 total, 1 running, 287 sleeping, 0 stopped, 0 zombie
Cpu0 : 1.5%us, 0.5%sy, 0.0%ni, 85.3%id, 12.5%wa, 0.0%hi, 0.2%si, 0.0%st
Cpu1 : 0.1%us, 2.6%sy, 0.0%ni, 0.0%id, 92.2%wa, 0.5%hi, 4.6%si, 0.0%st
Cpu2 : 0.0%us, 0.1%sy, 0.0%ni, 88.6%id, 9.9%wa, 0.4%hi, 1.0%si, 0.0%st
Cpu3 : 0.0%us, 5.1%sy, 0.0%ni, 0.0%id, 87.2%wa, 0.5%hi, 7.2%si, 0.0%st
Cpu4 : 1.7%us, 1.0%sy, 0.0%ni, 13.1%id, 78.4%wa, 1.2%hi, 4.6%si, 0.0%st
Cpu5 : 0.0%us, 0.0%sy, 0.0%ni,100.0%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu6 : 0.0%us, 0.1%sy, 0.0%ni, 99.8%id, 0.0%wa, 0.1%hi, 0.0%si, 0.0%st
Cpu7 : 0.0%us, 0.0%sy, 0.0%ni,100.0%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
Four separate disks as four separate PVs placed in one VG and one LV
512 blocks - 5750 - 5800 IOPS
top - 22:44:12 up 13 min, 2 users, load average: 89.54, 36.43, 13.60
Tasks: 241 total, 1 running, 240 sleeping, 0 stopped, 0 zombie
Cpu0 : 0.0%us, 0.0%sy, 0.0%ni,100.0%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu1 : 0.1%us, 3.7%sy, 0.0%ni, 0.0%id, 90.1%wa, 1.6%hi, 4.5%si, 0.0%st
Cpu2 : 0.0%us, 0.0%sy, 0.0%ni,100.0%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu3 : 0.0%us, 1.0%sy, 0.0%ni, 98.3%id, 0.0%wa, 0.0%hi, 0.7%si, 0.0%st
Cpu4 : 0.2%us, 2.2%sy, 0.0%ni, 0.0%id, 93.6%wa, 0.6%hi, 3.4%si, 0.0%st
Cpu5 : 0.1%us, 2.4%sy, 0.0%ni, 0.0%id, 94.2%wa, 0.5%hi, 2.8%si, 0.0%st
Cpu6 : 0.1%us, 0.1%sy, 0.0%ni, 99.7%id, 0.0%wa, 0.1%hi, 0.0%si, 0.0%st
Cpu7 : 0.0%us, 0.0%sy, 0.0%ni,100.0%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
64K blocks - 1300 IOPS
top - 22:46:03 up 15 min, 2 users, load average: 94.37, 53.52, 22.27
Tasks: 241 total, 1 running, 240 sleeping, 0 stopped, 0 zombie
Cpu0 : 0.6%us, 0.1%sy, 0.0%ni, 99.3%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu1 : 0.0%us, 1.8%sy, 0.0%ni, 24.5%id, 61.0%wa, 2.7%hi, 10.0%si, 0.0%st
Cpu2 : 0.0%us, 0.0%sy, 0.0%ni,100.0%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu3 : 0.0%us, 0.0%sy, 0.0%ni,100.0%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu4 : 0.5%us, 5.9%sy, 0.0%ni, 0.0%id, 86.9%wa, 0.5%hi, 6.2%si, 0.0%st
Cpu5 : 0.1%us, 6.9%sy, 0.0%ni, 0.0%id, 85.7%wa, 0.8%hi, 6.5%si, 0.0%st
Cpu6 : 0.1%us, 0.1%sy, 0.0%ni, 99.8%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu7 : 0.0%us, 0.0%sy, 0.0%ni,100.0%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
Mem: 33020800k total, 400224k used, 32620576k free, 20896k buffers
Four separate disks - aggregate of four separate parallel tests
512 blocks - 24940 IOPS
top - 22:58:33 up 7 min, 5 users, load average: 329.69, 117.85, 42.52
Tasks: 249 total, 1 running, 248 sleeping, 0 stopped, 0 zombie
Cpu0 : 0.7%us, 0.1%sy, 0.0%ni, 99.2%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu1 : 1.9%us, 16.0%sy, 0.0%ni, 0.0%id, 51.5%wa, 8.3%hi, 22.3%si, 0.0%st
Cpu2 : 0.0%us, 0.0%sy, 0.0%ni,100.0%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu3 : 0.0%us, 2.7%sy, 0.0%ni, 21.4%id, 74.6%wa, 0.0%hi, 1.3%si, 0.0%st
Cpu4 : 0.7%us, 5.5%sy, 0.0%ni, 0.0%id, 82.5%wa, 1.4%hi, 9.9%si, 0.0%st
Cpu5 : 0.5%us, 6.0%sy, 0.0%ni, 0.0%id, 82.0%wa, 1.6%hi, 9.9%si, 0.0%st
Cpu6 : 0.0%us, 0.1%sy, 0.0%ni, 99.8%id, 0.0%wa, 0.1%hi, 0.0%si, 0.0%st
Cpu7 : 0.0%us, 5.2%sy, 0.0%ni, 20.2%id, 71.5%wa, 0.0%hi, 3.1%si, 0.0%st
64K blocks - 6713 IOPS
top - 23:01:53 up 10 min, 5 users, load average: 363.88, 219.53, 96.34
Tasks: 249 total, 1 running, 248 sleeping, 0 stopped, 0 zombie
Cpu0 : 0.9%us, 0.1%sy, 0.0%ni, 99.0%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu1 : 0.5%us, 5.1%sy, 0.0%ni, 0.0%id, 73.8%wa, 4.3%hi, 16.3%si, 0.0%st
Cpu2 : 0.0%us, 0.6%sy, 0.0%ni, 41.1%id, 58.0%wa, 0.0%hi, 0.3%si, 0.0%st
Cpu3 : 0.0%us, 0.1%sy, 0.0%ni, 99.9%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu4 : 0.4%us, 3.0%sy, 0.0%ni, 0.0%id, 89.4%wa, 0.6%hi, 6.6%si, 0.0%st
Cpu5 : 0.2%us, 2.5%sy, 0.0%ni, 0.0%id, 89.8%wa, 0.6%hi, 6.9%si, 0.0%st
Cpu6 : 0.1%us, 0.0%sy, 0.0%ni, 99.9%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu7 : 0.0%us, 1.2%sy, 0.0%ni, 40.9%id, 57.3%wa, 0.0%hi, 0.6%si, 0.0%st
3.4 Gbps however the switches do not seem to reflect this throughput:
switchdc1_01> sho int po
Status and Counters - Port Utilization
Rx Tx
Port Mode | --------------------------- | ---------------------------
| Kbits/sec Pkts/sec Util | Kbits/sec Pkts/sec Util
--------- -------- + ---------- ---------- ----- + ---------- ---------- -----
1 100FDx | 488 5 00.48 | 512 6 00.51
2-Trk24 1000FDx | 5000 0 00.50 | 1624 0 00.16
3 1000FDx | 4896 3 00.48 | 4888 0 00.48
4 1000FDx | 5000 5 00.50 | 5000 4 00.50
5 1000FDx | 0 0 0 | 0 0 0
6 1000FDx | 4896 1 00.48 | 5000 2 00.50
7 1000FDx | 5000 3 00.50 | 5000 0 00.50
8 1000FDx | 344904 6788 34.49 | 8616 4253 00.86
9 1000FDx | 0 0 0 | 5000 0 00.50
10 1000FDx | 4776 2 00.47 | 4888 2 00.48
11 100FDx | 0 0 0 | 496 0 00.49
12 1000FDx | 5000 3 00.50 | 5000 0 00.50
13 1000FDx | 5000 3 00.50 | 5008 8 00.50
14 1000FDx | 4856 1 00.48 | 5000 2 00.50
15 1000FDx | 0 0 0 | 0 0 0
16 1000FDx | 120 0 00.01 | 5000 0 00.50
17 1000FDx | 0 0 0 | 0 0 0
18 1000FDx | 8456 4224 00.84 | 344320 6757 34.43
19 1000FDx | 0 0 0 | 0 0 0
20 1000FDx | 1320 0 00.13 | 5000 0 00.50
21 100FDx | 0 0 0 | 496 0 00.49
22 1000FDx | 4864 1 00.48 | 5000 2 00.50
23-Trk24 1000FDx | 4848 8 00.48 | 2496 0 00.24
24-Trk24 1000FDx | 4984 0 00.49 | 4992 3 00.49
switchdc1_02> sho int po
Status and Counters - Port Utilization
Rx Tx
Port Mode | --------------------------- | ---------------------------
| Kbits/sec Pkts/sec Util | Kbits/sec Pkts/sec Util
--------- -------- + ---------- ---------- ----- + ---------- ---------- -----
1 1000FDx | 0 0 0 | 0 0 0
2-Trk24 1000FDx | 1672 0 00.16 | 5000 0 00.50
3 1000FDx | 4992 8 00.49 | 5000 8 00.50
4 1000FDx | 8400 4186 00.84 | 341664 6698 34.16
5 1000FDx | 0 0 0 | 0 0 0
6 1000FDx | 14160 10807 01.41 | 855656 17153 85.56
7 1000FDx | 0 0 0 | 0 0 0
8 1000FDx | 120 0 00.01 | 5000 0 00.50
9 1000FDx | 0 0 0 | 0 0 0
10 1000FDx | 160 0 00.01 | 5000 0 00.50
11 1000FDx | 0 0 0 | 0 0 0
12 1000FDx | 5000 3 00.50 | 5000 0 00.50
13 1000FDx | 5000 7 00.50 | 5000 3 00.50
14 1000FDx | 4856 1 00.48 | 5000 2 00.50
15 1000FDx | 0 0 0 | 5000 0 00.50
16 1000FDx | 855864 17156 85.58 | 14240 10806 01.42
17 1000FDx | 5000 3 00.50 | 5000 0 00.50
18 1000FDx | 4824 1 00.48 | 5000 2 00.50
19 1000FDx | 0 0 0 | 0 0 0
20 1000FDx | 342296 6734 34.22 | 8592 4219 00.85
21 100FDx | 0 0 0 | 496 0 00.49
22 1000FDx | 5008 6 00.50 | 5000 0 00.50
23-Trk24 1000FDx | 2408 0 00.24 | 4848 8 00.48
24-Trk24 1000FDx | 4992 4 00.49 | 4992 0 00.49
Does bonding targets (LVM, RAID) or multipathing dramatically reduce throughput or is it a consequence of running four separate tests? Let's run four separate tests agains RAID0 and against one target with eight paths in round robin.
RAID0 - four concurrent tests
512 blocks - 2930 IOPS = 11720 IOPS
top - 23:19:33 up 28 min, 5 users, load average: 381.54, 241.74, 193.55
Tasks: 249 total, 1 running, 248 sleeping, 0 stopped, 0 zombie
Cpu0 : 0.0%us, 0.0%sy, 0.0%ni,100.0%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu1 : 0.2%us, 1.8%sy, 0.0%ni, 0.0%id, 94.9%wa, 0.9%hi, 2.2%si, 0.0%st
Cpu2 : 0.0%us, 0.4%sy, 0.0%ni, 99.4%id, 0.0%wa, 0.0%hi, 0.2%si, 0.0%st
Cpu3 : 0.0%us, 0.0%sy, 0.0%ni,100.0%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu4 : 0.1%us, 1.0%sy, 0.0%ni, 0.0%id, 96.8%wa, 0.3%hi, 1.8%si, 0.0%st
Cpu5 : 0.0%us, 1.6%sy, 0.0%ni, 0.0%id, 96.3%wa, 0.3%hi, 1.8%si, 0.0%st
Cpu6 : 0.0%us, 0.1%sy, 0.0%ni, 99.8%id, 0.1%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu7 : 0.0%us, 0.0%sy, 0.0%ni,100.0%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
64K blocks - 797 IOPS * 4 = 3188 IOPS
top - 23:22:23 up 31 min, 5 users, load average: 365.45, 277.87, 214.36
Tasks: 249 total, 1 running, 248 sleeping, 0 stopped, 0 zombie
Cpu0 : 0.5%us, 0.0%sy, 0.0%ni, 99.5%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu1 : 0.0%us, 0.9%sy, 0.0%ni, 0.0%id, 94.0%wa, 0.9%hi, 4.2%si, 0.0%st
Cpu2 : 0.0%us, 0.0%sy, 0.0%ni,100.0%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu3 : 0.0%us, 0.0%sy, 0.0%ni,100.0%id, 0.0%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu4 : 0.6%us, 3.2%sy, 0.0%ni, 0.0%id, 93.8%wa, 0.1%hi, 2.3%si, 0.0%st
Cpu5 : 0.0%us, 2.4%sy, 0.0%ni, 0.0%id, 95.1%wa, 0.2%hi, 2.3%si, 0.0%st
Cpu6 : 0.0%us, 0.1%sy, 0.0%ni, 99.7%id, 0.1%wa, 0.1%hi, 0.0%si, 0.0%st
Cpu7 : 0.0%us, 0.1%sy, 0.0%ni, 76.6%id, 23.3%wa, 0.0%hi, 0.0%si, 0.0%st
Multipath round robin - four concurrent tests
512 blocks - 10140 IOPS
top - 23:31:43 up 5 min, 5 users, load average: 285.93, 89.10, 31.16
Tasks: 210 total, 1 running, 209 sleeping, 0 stopped, 0 zombie
Cpu0 : 0.0%us, 0.1%sy, 0.0%ni, 73.5%id, 26.4%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu1 : 0.1%us, 1.3%sy, 0.0%ni, 0.0%id, 98.2%wa, 0.0%hi, 0.4%si, 0.0%st
Cpu2 : 0.4%us, 4.2%sy, 0.0%ni, 0.0%id, 88.5%wa, 0.7%hi, 6.2%si, 0.0%st
Cpu3 : 0.1%us, 1.7%sy, 0.0%ni, 0.0%id, 97.4%wa, 0.0%hi, 0.8%si, 0.0%st
Cpu4 : 0.3%us, 3.7%sy, 0.0%ni, 0.0%id, 88.1%wa, 0.8%hi, 7.1%si, 0.0%st
Cpu5 : 0.1%us, 1.7%sy, 0.0%ni, 0.0%id, 97.2%wa, 0.0%hi, 1.0%si, 0.0%st
Cpu6 : 0.1%us, 1.5%sy, 0.0%ni, 0.0%id, 97.9%wa, 0.0%hi, 0.5%si, 0.0%st
Cpu7 : 0.0%us, 0.2%sy, 0.0%ni, 28.8%id, 71.0%wa, 0.0%hi, 0.0%si, 0.0%st
64K blocks - 2563 IOPS
top - 23:38:23 up 11 min, 5 users, load average: 388.21, 292.94, 147.88
Tasks: 210 total, 1 running, 209 sleeping, 0 stopped, 0 zombie
Cpu0 : 0.1%us, 0.5%sy, 0.0%ni, 0.0%id, 99.3%wa, 0.0%hi, 0.1%si, 0.0%st
Cpu1 : 0.1%us, 0.5%sy, 0.0%ni, 0.0%id, 99.3%wa, 0.0%hi, 0.1%si, 0.0%st
Cpu2 : 0.1%us, 1.0%sy, 0.0%ni, 0.0%id, 95.5%wa, 0.1%hi, 3.3%si, 0.0%st
Cpu3 : 0.1%us, 0.6%sy, 0.0%ni, 0.0%id, 99.3%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu4 : 0.0%us, 1.1%sy, 0.0%ni, 0.0%id, 94.7%wa, 0.3%hi, 3.9%si, 0.0%st
Cpu5 : 0.1%us, 0.6%sy, 0.0%ni, 0.0%id, 99.1%wa, 0.0%hi, 0.2%si, 0.0%st
Cpu6 : 0.0%us, 0.4%sy, 0.0%ni, 0.0%id, 99.6%wa, 0.0%hi, 0.0%si, 0.0%st
Cpu7 : 0.1%us, 0.4%sy, 0.0%ni, 0.0%id, 99.5%wa, 0.0%hi, 0.0%si, 0.0%st
More information about the dm-devel
mailing list