Description of problem: [root@localhost ~]# df -h Filesystem Size Used Avail Use% Mounted on /dev/vda3 16G 3.8G 11G 26% / tmpfs 1004M 0 1004M 0% /dev/shm /dev/vda1 194M 26M 158M 15% /boot /dev/loop0 1.9G 1.2G 586M 68% /mnt/test /dev/vdb1 98G 33M 98G 1% /export [root@localhost ~]# mount /dev/vda3 on / type ext4 (rw) proc on /proc type proc (rw) sysfs on /sys type sysfs (rw) devpts on /dev/pts type devpts (rw,gid=5,mode=620) tmpfs on /dev/shm type tmpfs (rw,rootcontext="system_u:object_r:tmpfs_t:s0") /dev/vda1 on /boot type ext3 (rw) none on /proc/sys/fs/binfmt_misc type binfmt_misc (rw) sunrpc on /var/lib/nfs/rpc_pipefs type rpc_pipefs (rw) /dev/loop0 on /mnt/test type ext4 (rw,user_xattr) /dev/vdb1 on /export type xfs (rw) ########################################################################### [root@localhost ~]# gluster volume info dist-rep Volume Name: dist-rep Type: Distributed-Replicate Volume ID: 3fffafbc-484c-4092-83ff-2365a68e9521 Status: Started Number of Bricks: 2 x 2 = 4 Transport-type: tcp Bricks: Brick1: 172.17.251.90:/export/dr Brick2: 172.17.251.92:/export/drr Brick3: 172.17.251.91:/export/ddr Brick4: 172.17.251.93:/export/ddrr [root@localhost ~]# [root@localhost ~]# [root@localhost ~]# [root@localhost ~]# gluster volume status Status of volume: dist-rep Gluster process Port Online Pid ------------------------------------------------------------------------------ Brick 172.17.251.90:/export/dr 24009 Y 22627 Brick 172.17.251.92:/export/drr 24009 Y 30745 Brick 172.17.251.91:/export/ddr 24009 Y 10915 Brick 172.17.251.93:/export/ddrr 24009 Y 31409 NFS Server on localhost 38467 Y 22633 Self-heal Daemon on localhost N/A Y 22638 NFS Server on 172.17.251.91 38467 Y 10920 Self-heal Daemon on 172.17.251.91 N/A Y 10927 NFS Server on 172.17.251.93 38467 Y 31415 Self-heal Daemon on 172.17.251.93 N/A Y 31420 NFS Server on 172.17.251.92 38467 Y 30751 Self-heal Daemon on 172.17.251.92 N/A Y 30756 [root@localhost ~]# rpcinfo -p program vers proto port service 100000 4 tcp 111 portmapper 100000 3 tcp 111 portmapper 100000 2 tcp 111 portmapper 100000 4 udp 111 portmapper 100000 3 udp 111 portmapper 100000 2 udp 111 portmapper 100005 3 tcp 38465 mountd 100005 1 tcp 38466 mountd 100003 3 tcp 38467 nfs 100024 1 udp 48503 status 100024 1 tcp 48767 status 100021 4 tcp 38468 nlockmgr 100021 1 udp 766 nlockmgr 100021 1 tcp 768 nlockmgr [root@localhost ~]# showmount -e 0 Export list for 0: /test * [root@localhost ~]# Version-Release number of selected component (if applicable): 3.3.0qa42 How reproducible: found once Steps to Reproduce: 1. create a loop device with ext4 as filesystem 2. create a volume over this device. 3. install 3.3.0qa42 4. create a new volume, after installing xfs on the storage available. 5. showmount -e 0 6. stop the glusterd on all nodes. 7. clear out data from /var/lib/glusterd 8. restart the glusterd 9. showmount -e 0 Actual results:- step 5 and 9 does not display the newly created volume in the list Expected results: the volume should be displyed in the list for showmount Additional info: even restart of the volume fails to display it.
killing the nfs process using kill <pid> and restarting the volume displays the correct info for showmount, whereas on the other node of the cluster the information is still stale. This is still a bug as the volume start also restarts the nfs server and information updation should be taken at that time itself.
*** Bug 823763 has been marked as a duplicate of this bug. ***
Talked to Saurabh. A stale nfs process is alive whose pid file has been deleted (/var/lib/glusterd was cleaned up) and hence "glusterd stop" command never kills this nfs process. I have requested Saurabh to reproduce this issue, in case its not reproducible we can close this bug.