[dm-devel] Poor iSCSI performance

John A. Sullivan III jsullivan at opensourcedevel.com
Mon Mar 16 09:58:57 UTC 2009


On Mon, 2009-03-16 at 05:57 -0400, John A. Sullivan III wrote:
> Hello, all.  We've been struggling to tweak performance between on Linux
> iSCSI initiators (open-iscsi) and our opensolar iSCSI targets (Nexenta).
> On top of our generally poor performance (max 7000 IOPS per GbE NIC), we
> are seeing abysmal performance when we try to compensate by using either
> dm-multipath or dmadm to use multiple iSCSI LUNs.
> 
> We have been testing using an eight processor Linux server with 6 GbE
> network interfaces speaking to a Nexenta based Z200 storage system from
> Pogo Linux with 10 GbE ports.  I will attach a text file with some
> results using disktest.
> 
> In summary, if we ran four completely independent tests against four
> separate targets on four separate NICs, we achieved an aggregate 24940
> IOPS with 512 byte blocks and 6713 IOPS with 64KB blocks.
> 
> However, we would prefer to treat the storage as a single disk and so
> attempted to use software RAID, i.e., we created four LUNs, presented
> them as four separate disks and then used software RAID0 to stripe
> across all four targets.  We expected slightly less than the performance
> cited above.  Instead, we received 4450 IOPS for 512 and for 64KB.
> 
> We then took a different approach and created one big LUN with eight
> paths to the target using dm-multipath multibus with round-robin
> scheduling and rr_min_io=100.  Our numbers were 4350 IOPS for 512 and
> 1450 IOPS with 64KB.
> 
> We then suspected it might an issue of the number of threads rather than
> just the number of disks, i.e., the four independent disk test was using
> four separate processes.  So we ran four separate, concurrent tests
> against the RAID0 array and the multipath setup.
> 
> RAID0 increased to 11720 IOPS for 512 and 3188 IOPS for 64 KB - still a
> far cry from 24900 and and 6713.  dm-multipath numbers were 10140 IOPS
> for 512 and 2563 IOPS for 64KB.  Moreover, the CPU utilization was
> brutal.
> 
> /etc/multipath.conf:
> blacklist {
> #        devnode "*"
>         # sdb
>         wwid SATA_ST3250310NS_9SF0L234
>         #sda
>         wwid SATA_ST3250310NS_9SF0LVSR
>         # The above does not seem to be working thus we will do
>         devnode "^sd[ab]$"
>         # This is usually a bad idea as the device names can change
>         # However, since we add our iSCSI devices long after boot, I
> think we are safe
> }
> defaults {
>         udev_dir                /dev
>         polling_interval        5
>         selector                "round-robin 0"
>         path_grouping_policy    multibus
>         getuid_callout          "/sbin/scsi_id -g -u -s /block/%n"
>         prio_callout            /bin/true
>         path_checker            readsector0
>         rr_min_io               100
>         max_fds                 8192
>         rr_weight               priorities
>         failback                immediate
>         no_path_retry           fail
> #       user_friendly_names     yes
> }
> multipaths {
>         multipath {
>                 wwid
> 3600144f0e2824900000049b98e2b0001
>                 alias                   isda
>         }
>         multipath {
>                 wwid
> 3600144f0e2824900000049b062950002
>                 alias                   isdplain
>         }
>         multipath {
>                 wwid
> 3600144f0e2824900000049b9bb350001
>                 alias                   isdb
>         }
>         multipath {
>                 wwid
> 3600144f0e2824900000049b9bb350002
>                 alias                   isdc
>         }
>         multipath {
>                 wwid
> 3600144f0e2824900000049b9bb360003
>                 alias                   isdd
>         }
>         multipath {
>                 wwid
> 3600144f0e2824900000049b7878a0006
>                 alias                   isdtest
>         }
> }
> devices {
>        device {
>                vendor                  "NEXENTA"
>                product                 "COMSTAR"
> #               vendor                  "SUN"
> #               product                 "SOLARIS"
>                getuid_callout          "/sbin/scsi_id -g -u -s /block/%
> n"
>                features                "0"
>                hardware_handler        "0"
> #               path_grouping_policy    failover
>                rr_weight               uniform
> #               rr_min_io               1000
>                path_checker            readsector0
>        }
> }
> 
> What would account for such miserable performance? How can we improve
> it? We do not want to proliferate disks just to increase aggregate
> performance.  Thanks - John
Oops! Forgot to attach the file - John

-- 
John A. Sullivan III
Open Source Development Corporation
+1 207-985-7880
jsullivan at opensourcedevel.com

http://www.spiritualoutreach.com
Making Christianity intelligible to secular society
-------------- next part --------------
Disktest notes

./disktest -B512 -h1 -ID -pL -K100 -PT -T300 -r /dev/raw/raw1 

Eight paths to one target - round robin multipath:
512 blocks - 4350 IOPS:

Tasks: 212 total,   1 running, 211 sleeping,   0 stopped,   0 zombie
Cpu0  :  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu1  :  0.0%us,  1.5%sy,  0.0%ni,  0.0%id, 96.9%wa,  0.2%hi,  1.4%si,  0.0%st
Cpu2  :  0.0%us,  1.0%sy,  0.0%ni,  0.4%id, 96.8%wa,  0.4%hi,  1.4%si,  0.0%st
Cpu3  :  0.0%us,  0.1%sy,  0.0%ni, 99.7%id,  0.0%wa,  0.1%hi,  0.1%si,  0.0%st
Cpu4  :  0.1%us,  1.8%sy,  0.0%ni,  0.0%id, 94.9%wa,  0.3%hi,  2.9%si,  0.0%st
Cpu5  :  0.2%us,  1.5%sy,  0.0%ni,  0.0%id, 96.2%wa,  0.3%hi,  1.8%si,  0.0%st
Cpu6  :  0.0%us,  0.1%sy,  0.0%ni, 99.8%id,  0.0%wa,  0.1%hi,  0.0%si,  0.0%st
Cpu7  :  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st


64K blocks - 2000 - 2200 IOPS: (1450 with multipath - I forgot to start multipathd the first time)

top - 21:59:45 up 55 min,  2 users,  load average: 89.52, 44.36, 23.60
Tasks: 212 total,   1 running, 211 sleeping,   0 stopped,   0 zombie
Cpu0  :  0.0%us,  0.0%sy,  0.0%ni, 99.7%id,  0.3%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu1  :  0.0%us,  0.2%sy,  0.0%ni, 85.7%id, 14.0%wa,  0.0%hi,  0.1%si,  0.0%st
Cpu2  :  0.1%us,  2.1%sy,  0.0%ni,  0.0%id, 91.3%wa,  1.2%hi,  5.3%si,  0.0%st
Cpu3  :  0.0%us,  0.0%sy,  0.0%ni, 98.6%id,  1.4%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu4  :  0.2%us,  1.8%sy,  0.0%ni,  0.0%id, 89.2%wa,  0.4%hi,  8.4%si,  0.0%st
Cpu5  :  0.2%us,  2.3%sy,  0.0%ni,  0.0%id, 93.3%wa,  0.2%hi,  4.0%si,  0.0%st
Cpu6  :  0.0%us,  1.7%sy,  0.0%ni,  0.0%id, 94.9%wa,  0.2%hi,  3.2%si,  0.0%st
Cpu7  :  0.0%us,  0.0%sy,  0.0%ni, 89.2%id, 10.8%wa,  0.0%hi,  0.0%si,  0.0%st

RAID0 across four targets
512 blocks - 4450 IOPS
top - 22:20:36 up  1:16,  2 users,  load average: 81.25, 34.24, 24.49
Tasks: 288 total,   1 running, 287 sleeping,   0 stopped,   0 zombie
Cpu0  :  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu1  :  0.0%us,  1.2%sy,  0.0%ni,  0.0%id, 96.4%wa,  0.3%hi,  2.1%si,  0.0%st
Cpu2  :  0.0%us,  0.3%sy,  0.0%ni, 99.5%id,  0.0%wa,  0.0%hi,  0.2%si,  0.0%st
Cpu3  :  0.1%us,  3.3%sy,  0.0%ni,  0.0%id, 92.1%wa,  0.5%hi,  4.0%si,  0.0%st
Cpu4  :  0.1%us,  1.3%sy,  0.0%ni,  0.0%id, 96.1%wa,  0.7%hi,  1.8%si,  0.0%st
Cpu5  :  0.0%us,  0.0%sy,  0.0%ni, 99.6%id,  0.0%wa,  0.1%hi,  0.3%si,  0.0%st
Cpu6  :  0.0%us,  0.0%sy,  0.0%ni, 99.9%id,  0.0%wa,  0.1%hi,  0.0%si,  0.0%st
Cpu7  :  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st
Mem:  33020800k total,   497424k used, 32523376k free,    26504k buffers

64K blocks - 1280 IOPS

top - 22:23:26 up  1:19,  2 users,  load average: 97.37, 60.75, 36.23
Tasks: 288 total,   1 running, 287 sleeping,   0 stopped,   0 zombie
Cpu0  :  1.5%us,  0.5%sy,  0.0%ni, 85.3%id, 12.5%wa,  0.0%hi,  0.2%si,  0.0%st
Cpu1  :  0.1%us,  2.6%sy,  0.0%ni,  0.0%id, 92.2%wa,  0.5%hi,  4.6%si,  0.0%st
Cpu2  :  0.0%us,  0.1%sy,  0.0%ni, 88.6%id,  9.9%wa,  0.4%hi,  1.0%si,  0.0%st
Cpu3  :  0.0%us,  5.1%sy,  0.0%ni,  0.0%id, 87.2%wa,  0.5%hi,  7.2%si,  0.0%st
Cpu4  :  1.7%us,  1.0%sy,  0.0%ni, 13.1%id, 78.4%wa,  1.2%hi,  4.6%si,  0.0%st
Cpu5  :  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu6  :  0.0%us,  0.1%sy,  0.0%ni, 99.8%id,  0.0%wa,  0.1%hi,  0.0%si,  0.0%st
Cpu7  :  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st

Four separate disks as four separate PVs placed in one VG and one LV
512 blocks - 5750 - 5800 IOPS
top - 22:44:12 up 13 min,  2 users,  load average: 89.54, 36.43, 13.60
Tasks: 241 total,   1 running, 240 sleeping,   0 stopped,   0 zombie
Cpu0  :  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu1  :  0.1%us,  3.7%sy,  0.0%ni,  0.0%id, 90.1%wa,  1.6%hi,  4.5%si,  0.0%st
Cpu2  :  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu3  :  0.0%us,  1.0%sy,  0.0%ni, 98.3%id,  0.0%wa,  0.0%hi,  0.7%si,  0.0%st
Cpu4  :  0.2%us,  2.2%sy,  0.0%ni,  0.0%id, 93.6%wa,  0.6%hi,  3.4%si,  0.0%st
Cpu5  :  0.1%us,  2.4%sy,  0.0%ni,  0.0%id, 94.2%wa,  0.5%hi,  2.8%si,  0.0%st
Cpu6  :  0.1%us,  0.1%sy,  0.0%ni, 99.7%id,  0.0%wa,  0.1%hi,  0.0%si,  0.0%st
Cpu7  :  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st

64K blocks - 1300 IOPS
top - 22:46:03 up 15 min,  2 users,  load average: 94.37, 53.52, 22.27
Tasks: 241 total,   1 running, 240 sleeping,   0 stopped,   0 zombie
Cpu0  :  0.6%us,  0.1%sy,  0.0%ni, 99.3%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu1  :  0.0%us,  1.8%sy,  0.0%ni, 24.5%id, 61.0%wa,  2.7%hi, 10.0%si,  0.0%st
Cpu2  :  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu3  :  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu4  :  0.5%us,  5.9%sy,  0.0%ni,  0.0%id, 86.9%wa,  0.5%hi,  6.2%si,  0.0%st
Cpu5  :  0.1%us,  6.9%sy,  0.0%ni,  0.0%id, 85.7%wa,  0.8%hi,  6.5%si,  0.0%st
Cpu6  :  0.1%us,  0.1%sy,  0.0%ni, 99.8%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu7  :  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st
Mem:  33020800k total,   400224k used, 32620576k free,    20896k buffers

Four separate disks - aggregate of four separate parallel tests
512 blocks - 24940 IOPS
top - 22:58:33 up 7 min,  5 users,  load average: 329.69, 117.85, 42.52
Tasks: 249 total,   1 running, 248 sleeping,   0 stopped,   0 zombie
Cpu0  :  0.7%us,  0.1%sy,  0.0%ni, 99.2%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu1  :  1.9%us, 16.0%sy,  0.0%ni,  0.0%id, 51.5%wa,  8.3%hi, 22.3%si,  0.0%st
Cpu2  :  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu3  :  0.0%us,  2.7%sy,  0.0%ni, 21.4%id, 74.6%wa,  0.0%hi,  1.3%si,  0.0%st
Cpu4  :  0.7%us,  5.5%sy,  0.0%ni,  0.0%id, 82.5%wa,  1.4%hi,  9.9%si,  0.0%st
Cpu5  :  0.5%us,  6.0%sy,  0.0%ni,  0.0%id, 82.0%wa,  1.6%hi,  9.9%si,  0.0%st
Cpu6  :  0.0%us,  0.1%sy,  0.0%ni, 99.8%id,  0.0%wa,  0.1%hi,  0.0%si,  0.0%st
Cpu7  :  0.0%us,  5.2%sy,  0.0%ni, 20.2%id, 71.5%wa,  0.0%hi,  3.1%si,  0.0%st

64K blocks - 6713 IOPS
top - 23:01:53 up 10 min,  5 users,  load average: 363.88, 219.53, 96.34
Tasks: 249 total,   1 running, 248 sleeping,   0 stopped,   0 zombie
Cpu0  :  0.9%us,  0.1%sy,  0.0%ni, 99.0%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu1  :  0.5%us,  5.1%sy,  0.0%ni,  0.0%id, 73.8%wa,  4.3%hi, 16.3%si,  0.0%st
Cpu2  :  0.0%us,  0.6%sy,  0.0%ni, 41.1%id, 58.0%wa,  0.0%hi,  0.3%si,  0.0%st
Cpu3  :  0.0%us,  0.1%sy,  0.0%ni, 99.9%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu4  :  0.4%us,  3.0%sy,  0.0%ni,  0.0%id, 89.4%wa,  0.6%hi,  6.6%si,  0.0%st
Cpu5  :  0.2%us,  2.5%sy,  0.0%ni,  0.0%id, 89.8%wa,  0.6%hi,  6.9%si,  0.0%st
Cpu6  :  0.1%us,  0.0%sy,  0.0%ni, 99.9%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu7  :  0.0%us,  1.2%sy,  0.0%ni, 40.9%id, 57.3%wa,  0.0%hi,  0.6%si,  0.0%st

3.4 Gbps however the switches do not seem to reflect this throughput:

switchdc1_01> sho int po
 Status and Counters - Port Utilization

                                 Rx                           Tx
 Port      Mode     | --------------------------- | ---------------------------
                    | Kbits/sec   Pkts/sec  Util  | Kbits/sec  Pkts/sec   Util
 --------- -------- + ---------- ---------- ----- + ---------- ---------- -----
 1         100FDx   | 488        5          00.48 | 512        6          00.51
 2-Trk24   1000FDx  | 5000       0          00.50 | 1624       0          00.16
 3         1000FDx  | 4896       3          00.48 | 4888       0          00.48
 4         1000FDx  | 5000       5          00.50 | 5000       4          00.50
 5         1000FDx  | 0          0          0     | 0          0          0
 6         1000FDx  | 4896       1          00.48 | 5000       2          00.50
 7         1000FDx  | 5000       3          00.50 | 5000       0          00.50
 8         1000FDx  | 344904     6788       34.49 | 8616       4253       00.86
 9         1000FDx  | 0          0          0     | 5000       0          00.50
 10        1000FDx  | 4776       2          00.47 | 4888       2          00.48
 11        100FDx   | 0          0          0     | 496        0          00.49
 12        1000FDx  | 5000       3          00.50 | 5000       0          00.50
 13        1000FDx  | 5000       3          00.50 | 5008       8          00.50
 14        1000FDx  | 4856       1          00.48 | 5000       2          00.50
 15        1000FDx  | 0          0          0     | 0          0          0
 16        1000FDx  | 120        0          00.01 | 5000       0          00.50
 17        1000FDx  | 0          0          0     | 0          0          0
 18        1000FDx  | 8456       4224       00.84 | 344320     6757       34.43
 19        1000FDx  | 0          0          0     | 0          0          0
 20        1000FDx  | 1320       0          00.13 | 5000       0          00.50
 21        100FDx   | 0          0          0     | 496        0          00.49
 22        1000FDx  | 4864       1          00.48 | 5000       2          00.50
 23-Trk24  1000FDx  | 4848       8          00.48 | 2496       0          00.24
 24-Trk24  1000FDx  | 4984       0          00.49 | 4992       3          00.49

switchdc1_02> sho int po
 Status and Counters - Port Utilization

                                 Rx                           Tx
 Port      Mode     | --------------------------- | ---------------------------
                    | Kbits/sec   Pkts/sec  Util  | Kbits/sec  Pkts/sec   Util
 --------- -------- + ---------- ---------- ----- + ---------- ---------- -----
 1         1000FDx  | 0          0          0     | 0          0          0
 2-Trk24   1000FDx  | 1672       0          00.16 | 5000       0          00.50
 3         1000FDx  | 4992       8          00.49 | 5000       8          00.50
 4         1000FDx  | 8400       4186       00.84 | 341664     6698       34.16
 5         1000FDx  | 0          0          0     | 0          0          0
 6         1000FDx  | 14160      10807      01.41 | 855656     17153      85.56
 7         1000FDx  | 0          0          0     | 0          0          0
 8         1000FDx  | 120        0          00.01 | 5000       0          00.50
 9         1000FDx  | 0          0          0     | 0          0          0
 10        1000FDx  | 160        0          00.01 | 5000       0          00.50
 11        1000FDx  | 0          0          0     | 0          0          0
 12        1000FDx  | 5000       3          00.50 | 5000       0          00.50
 13        1000FDx  | 5000       7          00.50 | 5000       3          00.50
 14        1000FDx  | 4856       1          00.48 | 5000       2          00.50
 15        1000FDx  | 0          0          0     | 5000       0          00.50
 16        1000FDx  | 855864     17156      85.58 | 14240      10806      01.42
 17        1000FDx  | 5000       3          00.50 | 5000       0          00.50
 18        1000FDx  | 4824       1          00.48 | 5000       2          00.50
 19        1000FDx  | 0          0          0     | 0          0          0
 20        1000FDx  | 342296     6734       34.22 | 8592       4219       00.85
 21        100FDx   | 0          0          0     | 496        0          00.49
 22        1000FDx  | 5008       6          00.50 | 5000       0          00.50
 23-Trk24  1000FDx  | 2408       0          00.24 | 4848       8          00.48
 24-Trk24  1000FDx  | 4992       4          00.49 | 4992       0          00.49

Does bonding targets (LVM, RAID) or multipathing dramatically reduce throughput or is it a consequence of running four separate tests? Let's run four separate tests agains RAID0 and against one target with eight paths in round robin.

RAID0 - four concurrent tests
512 blocks - 2930 IOPS = 11720 IOPS
top - 23:19:33 up 28 min,  5 users,  load average: 381.54, 241.74, 193.55
Tasks: 249 total,   1 running, 248 sleeping,   0 stopped,   0 zombie
Cpu0  :  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu1  :  0.2%us,  1.8%sy,  0.0%ni,  0.0%id, 94.9%wa,  0.9%hi,  2.2%si,  0.0%st
Cpu2  :  0.0%us,  0.4%sy,  0.0%ni, 99.4%id,  0.0%wa,  0.0%hi,  0.2%si,  0.0%st
Cpu3  :  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu4  :  0.1%us,  1.0%sy,  0.0%ni,  0.0%id, 96.8%wa,  0.3%hi,  1.8%si,  0.0%st
Cpu5  :  0.0%us,  1.6%sy,  0.0%ni,  0.0%id, 96.3%wa,  0.3%hi,  1.8%si,  0.0%st
Cpu6  :  0.0%us,  0.1%sy,  0.0%ni, 99.8%id,  0.1%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu7  :  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st

64K blocks - 797 IOPS * 4 = 3188 IOPS
top - 23:22:23 up 31 min,  5 users,  load average: 365.45, 277.87, 214.36
Tasks: 249 total,   1 running, 248 sleeping,   0 stopped,   0 zombie
Cpu0  :  0.5%us,  0.0%sy,  0.0%ni, 99.5%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu1  :  0.0%us,  0.9%sy,  0.0%ni,  0.0%id, 94.0%wa,  0.9%hi,  4.2%si,  0.0%st
Cpu2  :  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu3  :  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu4  :  0.6%us,  3.2%sy,  0.0%ni,  0.0%id, 93.8%wa,  0.1%hi,  2.3%si,  0.0%st
Cpu5  :  0.0%us,  2.4%sy,  0.0%ni,  0.0%id, 95.1%wa,  0.2%hi,  2.3%si,  0.0%st
Cpu6  :  0.0%us,  0.1%sy,  0.0%ni, 99.7%id,  0.1%wa,  0.1%hi,  0.0%si,  0.0%st
Cpu7  :  0.0%us,  0.1%sy,  0.0%ni, 76.6%id, 23.3%wa,  0.0%hi,  0.0%si,  0.0%st

Multipath round robin - four concurrent tests
512 blocks - 10140 IOPS
top - 23:31:43 up 5 min,  5 users,  load average: 285.93, 89.10, 31.16
Tasks: 210 total,   1 running, 209 sleeping,   0 stopped,   0 zombie
Cpu0  :  0.0%us,  0.1%sy,  0.0%ni, 73.5%id, 26.4%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu1  :  0.1%us,  1.3%sy,  0.0%ni,  0.0%id, 98.2%wa,  0.0%hi,  0.4%si,  0.0%st
Cpu2  :  0.4%us,  4.2%sy,  0.0%ni,  0.0%id, 88.5%wa,  0.7%hi,  6.2%si,  0.0%st
Cpu3  :  0.1%us,  1.7%sy,  0.0%ni,  0.0%id, 97.4%wa,  0.0%hi,  0.8%si,  0.0%st
Cpu4  :  0.3%us,  3.7%sy,  0.0%ni,  0.0%id, 88.1%wa,  0.8%hi,  7.1%si,  0.0%st
Cpu5  :  0.1%us,  1.7%sy,  0.0%ni,  0.0%id, 97.2%wa,  0.0%hi,  1.0%si,  0.0%st
Cpu6  :  0.1%us,  1.5%sy,  0.0%ni,  0.0%id, 97.9%wa,  0.0%hi,  0.5%si,  0.0%st
Cpu7  :  0.0%us,  0.2%sy,  0.0%ni, 28.8%id, 71.0%wa,  0.0%hi,  0.0%si,  0.0%st

64K blocks - 2563 IOPS
top - 23:38:23 up 11 min,  5 users,  load average: 388.21, 292.94, 147.88
Tasks: 210 total,   1 running, 209 sleeping,   0 stopped,   0 zombie
Cpu0  :  0.1%us,  0.5%sy,  0.0%ni,  0.0%id, 99.3%wa,  0.0%hi,  0.1%si,  0.0%st
Cpu1  :  0.1%us,  0.5%sy,  0.0%ni,  0.0%id, 99.3%wa,  0.0%hi,  0.1%si,  0.0%st
Cpu2  :  0.1%us,  1.0%sy,  0.0%ni,  0.0%id, 95.5%wa,  0.1%hi,  3.3%si,  0.0%st
Cpu3  :  0.1%us,  0.6%sy,  0.0%ni,  0.0%id, 99.3%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu4  :  0.0%us,  1.1%sy,  0.0%ni,  0.0%id, 94.7%wa,  0.3%hi,  3.9%si,  0.0%st
Cpu5  :  0.1%us,  0.6%sy,  0.0%ni,  0.0%id, 99.1%wa,  0.0%hi,  0.2%si,  0.0%st
Cpu6  :  0.0%us,  0.4%sy,  0.0%ni,  0.0%id, 99.6%wa,  0.0%hi,  0.0%si,  0.0%st
Cpu7  :  0.1%us,  0.4%sy,  0.0%ni,  0.0%id, 99.5%wa,  0.0%hi,  0.0%si,  0.0%st


More information about the dm-devel mailing list