Instances on integration project suffers from a 3 minutes initial boot delay because there is no NFS server available. The shared NFS has been disabled on integration T90610#1344487
It only happens on first boot though, The code from our firstboot.sh files has:
# Sleep until the nfs volumes we need are available. # Worst case, just time out after 3 minutes. tries=18 for i in `seq 1 ${tries}`; do prod_domain=`echo $domain | sed 's/wmflabs/wmnet/'` nfs_server="labstore.svc.${prod_domain}" echo $(showmount -e ${nfs_server} | egrep ^/exp/project/${project}\\s), | fgrep -q $ip, if [ $? -eq 0 ]; then break fi sleep 10 done
A couple console logs from T102108: integration-t102108-jessie-new2 and integration-t102108-trusty-new2 show:
+ tries=18 ++ seq 1 18 + for i in '`seq 1 ${tries}`' ++ sed s/wmflabs/wmnet/ ++ echo eqiad.wmflabs + prod_domain=eqiad.wmnet + nfs_server=labstore.svc.eqiad.wmnet + fgrep -q 10.68.17.6, ++ egrep '^/exp/project/integration\s' ++ showmount -e labstore.svc.eqiad.wmnet + echo /exp/project/integration 10.68.18.59,10.68.18.38,10.68.18.34,10.68.18.30,10.68.18.29,10.68.18.28,10.68.18.24,10.68.18.2,10.68.17.70,10.68.17.244,10.68.17.209,10.68.17.184,10.68.17.180,10.68.17.174,10.68.17.136,10.68.16.8,10.68.16.72,10.68.16.68,10.68.16.59,10.68.16.53,10.68.16.42,10.68.16.4,10.68.16.255,10.68.16.227,10.68.16.200, + '[' 1 -eq 0 ']' + sleep 10 ... + echo 'Warning: Timed out trying to detect NFS mounts.'
And indeed newly created instances are not exported since Shared NFS is disabled.
Maybe the information that there is no NFS share to expect can be retrieved by firstboot.sh? I have no idea how it is stored, I am guessing LDAP but as I understand it we want to move out of LDAP.