[wellylug] High load averages but no apparent cause

David Harrison david.harrison at stress-free.co.nz
Tue Mar 23 08:50:43 NZDT 2010


Thanks Daniel that switch for vmstat is very handy and I'd completely missed
the io wait value in top.
Googling based on your comments also brought up this page which is very
handy:
http://strugglers.net/wiki/Linux_performance_tuning

The problem is certainly is looking like an intermittent I/O issue.


Has anyone experience with the performance boost of a dedicated PCI-X SATA
controller for software RAID?

The server in question is a bog-standard HP ML110.
It isn't up to their needs, but it was a recent purchase by the previous IT
guys, so I'm afraid it is staying.

For practical reasons I want to keep the software RAID-5 (3x1TB drives), but
would putting some (or all) of these disks onto a dedicated
controller alleviate the I/O issue?

i.e. Is it worth recommending a $400 PCI-X SATA controller for the box, or
is that money better left on the table for a new server (ML310/330) in
twelve months time?
(My concern being that the card goes in and the problem remains the same.)


David



On Mon, Mar 22, 2010 at 10:51 PM, Daniel Reurich <daniel at centurion.net.nz>wrote:

> The top is interesting as it shows an iowait of 87%
>
> for current state of vmstat you need to run it like this:
> # vmstat 1 10
>
> The first result is an average since the last reboot, and the following
> ones are the state every second for 10 samples.  (See the man page).
>
> My initial thoughts are that your raid array it may be caused be the
> periodic raid scrub (runs a raid check over the entire array), so i'd
> check that via either "cat /proc/mdstat" or "mdadm -E /dev/md*"
>
> Also check that you don't have more than 1 disk at a time doing smart
> checks, as that really kills server performance.
>
> Daniel Reurich
>
>
> On Mon, 2010-03-22 at 17:12 +1300, David Harrison wrote:
> > I thought it was I/O bound too as it is running a software RAID5
> > array.
> > I would like it to be better, but the client can't afford hardware
> > upgrades right now.
>
> >
> > Does 6% IO wait time (from vmstat) constitute really bad disk
> > performance?
> > I've got systems with much higher wait times that have far lower
> > loads.
> >
> >
> > Here's the header output from top (the two active processes are
> > registering 1% cpu load each):
> >
> > top - 17:05:54 up 2 days, 20:25,  1 user,  load average: 2.74, 1.17,
> > 0.74
> > Tasks:  71 total,   2 running,  69 sleeping,   0 stopped,   0 zombie
> > Cpu(s):  0.0%us,  0.1%sy,  0.0%ni, 14.3%id, 85.4%wa,  0.0%hi,  0.1%
> > si,  0.0%st
> > Mem:   2074112k total,  2020768k used,    53344k free,    13128k
> > buffers
> > Swap:  3903608k total,      828k used,  3902780k free,  1770188k
> > cached
> >
> >
> >
> >
> > Here's the vmstat output:
> > procs -----------memory---------- ---swap-- -----io---- -system--
> > ----cpu----
> >  r  b   swpd   free   buff  cache   si   so    bi    bo   in   cs us
> > sy id wa
> >  0  0    828  53416  13228 1770304    0    0    12    22    0   24  0
> >  4 90  6
> >
> >
> > And finally iostat
> > avg-cpu:  %user   %nice %system %iowait  %steal   %idle
> >                   0.28     0.00        4.47      5.73      0.00
> >  89.52
> >
> >
> > Device:            tps   Blk_read/s   Blk_wrtn/s   Blk_read   Blk_wrtn
> > sda               4.12        22.55        56.06    5558094   13814570
> > sdb               3.79        22.91        56.63    5646538   13955010
> > sdc               4.06        23.30        56.80    5742544   13997936
> > md0               0.00         0.02         0.00       3756         10
> > md1               0.00         0.01         0.01       1592       2176
> > md2               1.26         2.60         9.28     640370    2287456
> > md3               9.40        46.87        58.95   11551194   14527312
> >
> >
> >
> >
> > On Mon, Mar 22, 2010 at 5:02 PM, Daniel Reurich
> > <daniel at centurion.net.nz> wrote:
> >
> >         On Mon, 2010-03-22 at 16:49 +1300, David Harrison wrote:
> >         > Hi,
> >         > Has anyone experienced high-load averages but haven't been
> >         able to see
> >         > processes that are causing it?
> >         >
> >         >
> >         > I've got an Ubuntu Server 9.10 instance who's load average
> >         ranges
> >         > between 1.0 and 3.0 for most of the day, yet tools like top
> >         and iostat
> >         > don't reveal any issues.
> >         > i.e. The load averages can be up around 1.5 whilst the
> >         maximum process
> >         > viewed in top is sitting at 5% of the CPU.
> >         >
> >         >
> >         > Anyone know of any other good tools for identifying the
> >         cause of
> >         > server load if the obvious ones fail?
> >         >
> >
> >         What's the wait state like (in top it's the %wa value).
> >
> >         Chances are that you have some serious I/O blocking going on
> >         which could
> >         be a slow or failing harddisk or something like that.
> >
> >
> >
> >         --
> >         Daniel Reurich.
> >
> >         Centurion Computer Technology (2005) Ltd
> >         Mobile 021 797 722
> >
> >
> >
> >
> >
> >         --
> >         Wellington Linux Users Group Mailing List:
> >         wellylug at lists.wellylug.org.nz
> >         To Leave:
> >          http://lists.wellylug.org.nz/mailman/listinfo/wellylug
> >
> >
> >
>
>
> --
> Daniel Reurich.
>
> Centurion Computer Technology (2005) Ltd
> Mobile 021 797 722
>
>
>
>
> --
> Wellington Linux Users Group Mailing List: wellylug at lists.wellylug.org.nz
> To Leave:  http://lists.wellylug.org.nz/mailman/listinfo/wellylug
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://lists.wellylug.org.nz/pipermail/wellylug/attachments/20100323/302bf597/attachment-0001.htm 


More information about the wellylug mailing list