High device bandwidth utilization
Hi All,
Please have a look on the below outputs of a same command in two different servers with same configuration.
iostat -x sda
Linux 2.6.32-220.el6.x86_64 (hostname 1) 04/01/2014 x86_64 (8 CPU)
avg-cpu: %user %nice %system %iowait %steal %idle
0.07 0.00 0.07 0.80 0.00 99.06
Device: rrqm/s wrqm/s r/s w/s rsec/s wsec/s avgrq-sz avgqu-sz await svctm %util
sda 0.94 0.43 3.84 0.41 109.91 6.75 27.43 0.21 48.63 13.08 5.56
iostat -x sda
Linux 2.6.32-220.el6.x86_64 (hostname 2) 04/01/2014 x86_64 (8 CPU)
avg-cpu: %user %nice %system %iowait %steal %idle
1.31 0.00 0.24 0.05 0.00 98.39
Device: rrqm/s wrqm/s r/s w/s rsec/s wsec/s avgrq-sz avgqu-sz await svctm %util
sda 0.11 1.89 0.03 1.71 2.74 28.78 18.14 0.01 8.29 3.65 0.63
I need to know why these bandwidth utilization values differ for both the servers though these are with a same configuration.
Please Suggest.
Regards,
Rahul
Responses
Rahul,
I am not sure what your specific concern is, but the method you are using is giving you numbers since the server was booted (default from iostat). From the man page:
The first report generated by the iostat command provides statistics concerning the time since the system was booted.
It's likely one of the servers has had more load at one point in time which has increased the overall average.
Use the following command to get output every 2 seconds which will better represent the workload at the time of testing:
iostat -x sda 2
The trailing '2' is the interval at which iostat will poll / display io information.
Hi Rahul - if I understand your question correctly, I believe you are wondering why 2 identical servers have different values for that particular query.
The OS runs a number of tools in the background:
updatedb (for locate commmand)
rhn-profile-sync (if your host is registered it will inventory installed packages)
etc...
You likely have looked at the 2 hosts while different things are going on.
I recommend that you look in to the command iotop.
I thought you were specifically wondering about only the disk statistics.
Performance Analysis (and tuning) is certainly challenging (the intro course is a week ;-)
I would try looking for information on the following:
iotop
sar -b
vmstat
iostat
I am also working through some performance issues myself right now. Unfortunately (for me) it involves 3 tiers of servers, multiple firewalls and load balancers.
It will likely take a bit of work to narrow it down, but first thing I would do is run 'sar' on both (no switches) and let us know what the '%iowait' figure shows for both servers (note: these are 10 minute averages). This will tell you if the server is spending time waiting on disk IO.
As mentioned by James as well, the 'sar -b' will give some statistics on performance of the disks over time too which may provide some more insight.
Also, can you provide the output of 'free' to show if the server is utilising swap?
Are these servers using SAN storage or local disk?
If it's SAN, is it FC or ISCSI?
If it's ISCSI are you using a dedicated network, or shared?
What is the performance issue you are seeing? slow network copy? slow filesystem copy? poor application performance? or are you just comparing the numbers between servers and have concerns based on the differences?
Welcome! Check out the Getting Started with Red Hat page for quick tours and guides for common tasks.
