Problems with h2

h2 is the machine which was up first and thus users like it and feel homely there; h1 came later and takes a bit of the load, but the main load is still on h2 and it's puzzling why it's soo slow. Hardware wise the machine is not exceptionally large, but should be

Symptoms

h2 is one of two Condor submit nodes, the load is shared with HA set-up with h1. Right now h1 is running the negotiator, thus matchmaking is done on h1, while most jobs are submitted on h2.

Current stats:
% uptime
 08:25:03 up 13 days, 11:57, 15 users,  load average: 19.70, 37.23, 39.23

% iostat -x 15
Linux 2.6.27.7-storage (h2)     02/09/09

avg-cpu:  %user   %nice %system %iowait  %steal   %idle
           8.53    0.01   42.11   11.06    0.00   38.28

Device:         rrqm/s   wrqm/s   r/s   w/s   rsec/s   wsec/s avgrq-sz avgqu-sz   await  svctm  %util
sda               0.33    13.36  9.47 24.88  1720.23  1627.08    97.45     0.36   10.40   1.03   3.54
sda1              0.21     0.14  0.03  0.01     1.93     1.20    82.32     0.00   14.04   8.92   0.03
sda2              0.00     0.00  0.00  0.00     0.00     0.00     6.26     0.00    4.56   4.26   0.00
sda3              0.01     3.28  0.08  0.65     1.63    31.41    45.47     0.00    6.76   4.69   0.34
sda4              0.00     0.00  0.00  0.00     0.00     0.00     2.00     0.00    0.00   0.00   0.00
sda5              0.02     3.44  0.21  2.00     9.28    43.51    23.90     0.03   12.41   2.29   0.51
sda6              0.00     3.47  0.01  1.16     0.18    37.09    31.77     0.01    7.46   3.19   0.37
sda7              0.08     2.59  0.58  0.63    10.65    25.79    29.90     0.01    5.00   3.13   0.38
sda8              0.00     0.00  0.00  0.00     0.00     0.00    19.98     0.00   30.19  30.00   0.00
sda9              0.01     0.44  8.56 20.42  1696.56  1488.08   109.88     0.31   10.68   1.13   3.28

avg-cpu:  %user   %nice %system %iowait  %steal   %idle
          17.92    0.00   38.20   25.97    0.00   17.92

Device:         rrqm/s   wrqm/s   r/s   w/s   rsec/s   wsec/s avgrq-sz avgqu-sz   await  svctm  %util
sda               0.00    19.97  0.00 50.13     0.00  6186.68   123.41     0.30    5.90   2.15  10.79
sda1              0.00     0.00  0.00  0.00     0.00     0.00     0.00     0.00    0.00   0.00   0.00
sda2              0.00     0.00  0.00  0.00     0.00     0.00     0.00     0.00    0.00   0.00   0.00
sda3              0.00     4.26  0.00  0.60     0.00    38.88    64.89     0.03   55.56  55.56   3.33
sda4              0.00     0.00  0.00  0.00     0.00     0.00     0.00     0.00    0.00   0.00   0.00
sda5              0.00     7.06  0.00  3.60     0.00    85.22    23.70     0.00    0.74   0.37   0.13
sda6              0.00     5.73  0.00  0.73     0.00    51.66    70.55     0.00    0.00   0.00   0.00
sda7              0.00     1.53  0.00  0.40     0.00    15.45    38.67     0.00    0.00   0.00   0.00
sda8              0.00     0.00  0.00  0.00     0.00     0.00     0.00     0.00    0.00   0.00   0.00
sda9              0.00     1.40  0.00 44.81     0.00  5995.47   133.81     0.26    5.79   2.42  10.85

% netstat -tn | awk '{print $6}' | sort | uniq -c
      1
     47 CLOSE_WAIT
   8920 ESTABLISHED
     14 FIN_WAIT2
      1 Foreign
      4 SYN_SENT
      7 TIME_WAIT

% nfsstat -o all
Server packet stats:
packets    udp        tcp        tcpconn
1          1          0          0

Server rpc stats:
calls      badcalls   badauth    badclnt    xdrcall
1          0          0          0          0

Server reply cache:
hits       misses     nocache
0          0          1

Server file handle cache:
lookup     anon       ncachedir  ncachedir  stale
0          0          0          0          0

Server nfs v3:
null         getattr      setattr      lookup       access       readlink
1       100% 0         0% 0         0% 0         0% 0         0% 0         0%
read         write        create       mkdir        symlink      mknod
0         0% 0         0% 0         0% 0         0% 0         0% 0         0%
remove       rmdir        rename       link         readdir      readdirplus
0         0% 0         0% 0         0% 0         0% 0         0% 0         0%
fsstat       fsinfo       pathconf     commit
0         0% 0         0% 0         0% 0         0%

Client packet stats:
packets    udp        tcp        tcpconn
0          0          0          0

Client rpc stats:
calls      retrans    authrefrsh
418351344   405        13

Client nfs v3:
null         getattr      setattr      lookup       access       readlink
0         0% 39698419  9% 1506477   0% 14595868  3% 30741922  7% 40807     0%
read         write        create       mkdir        symlink      mknod
281981726 67% 26535885  6% 1506773   0% 45541     0% 41580     0% 0         0%
remove       rmdir        rename       link         readdir      readdirplus
580526    0% 5751      0% 68646     0% 821       0% 57560     0% 1261537   0%
fsstat       fsinfo       pathconf     commit
366871    0% 34259     0% 17129     0% 19340899  4%

Client nfs v4:
null         read         write        commit       open         open_conf
0         0% 0         0% 0         0% 0         0% 0         0% 0         0%
open_noat    open_dgrd    close        setattr      fsinfo       renew
0         0% 0         0% 0         0% 0         0% 0         0% 0         0%
setclntid    confirm      lock         lockt        locku        access
0         0% 0         0% 0         0% 0         0% 0         0% 0         0%
getattr      lookup       lookup_root  remove       rename       link
0         0% 0         0% 0         0% 0         0% 0         0% 0         0%
symlink      create       pathconf     statfs       readlink     readdir
0         0% 0         0% 0         0% 0         0% 0         0% 0         0%
server_caps  delegreturn
0         0% 0         0%

Problems

This box seems to be really slow doing open/close operations, see a open/close test page. Sample runs of timeout -1 60 strace -tt -c -p PID would reveal that open or stat would take a long time to respond (of the order of 200-500ms) while the corresponding NFS server was more or less idle.

Solution

One item is to move the log files of certain job types off NFS since these are often opened/closed thus this should already help a bit. However it's not yet clear to me why this machine is so much slower as compared to h1 under somewhat comparable loads.

-- CarstenAulbert - 09 Feb 2009

This topic: ATLAS > H2Problems
Topic revision: 12 Feb 2009, CarstenAulbert
This site is powered by FoswikiCopyright © by the contributing authors. All material on this collaboration platform is the property of the contributing authors.
Ideas, requests, problems regarding Foswiki? Send feedback