Hide Forgot
Yes. In my replicate setup too quota shows different sizes on different backends [root@centos-rahul-2 ~]# getfattr -d -m . -e hex /export/mirror/ getfattr: Removing leading '/' from absolute path names # file: export/mirror trusted.gfid=0x00000000000000000000000000000001 trusted.glusterfs.676687e0-4f58-4343-b4d8-61309744ae9a.xtime=0x4da1a1730000b11a trusted.glusterfs.quota.dirty=0x3000 trusted.glusterfs.quota.size=0x0000000184cdf7b4 trusted.glusterfs.test=0x776f726b696e6700 [root@centos-rahul-3 ~]# getfattr -d -m . -e hex /export/mirror/ getfattr: Removing leading '/' from absolute path names # file: export/mirror trusted.gfid=0x00000000000000000000000000000001 trusted.glusterfs.676687e0-4f58-4343-b4d8-61309744ae9a.xtime=0x4da2385000004abf trusted.glusterfs.quota.dirty=0x3000 trusted.glusterfs.quota.size=0x000000012b9f2fce trusted.glusterfs.test=0x776f726b696e6700
root@booradley:/home/raghu/work/git/glusterfs# ./extras/quota-metadata-cleanup.sh /home/export/[56] && getfattr -e hex /home/export/[56]// -n trusted.glusterfs.quota.size && find /mnt/glusterfs | xargs touch && ./extras/quota-metadata-cleanup.sh /home/export/[56] && find /mnt/raghu | xargs touch && getfattr -e hex /home/export/[56]// -n trusted.glusterfs.quota.size /home/export/5/: trusted.glusterfs.quota.size: No such attribute /home/export/6/: trusted.glusterfs.quota.size: No such attribute getfattr: Removing leading '/' from absolute path names # file: home/export/5/ trusted.glusterfs.quota.size=0x0000000005180dfa # file: home/export/6/ trusted.glusterfs.quota.size=0x0000000005686567 However the disk-usage is same: root@booradley:/home/raghu/work/git/glusterfs# du -hs /home/export/5 396M /home/export/5 root@booradley:/home/raghu/work/git/glusterfs# du -hs /home/export/6 396M /home/export/6 Also note that the size as indicated by xattrs (around 84MB) is not the same as shown by du. regards, Raghavendra.
Tried the same test with tag v3.2.0qa12, and everything seems to be working fine. Can some one in QA too confirm this is fixed.
Tested on top of fixes to bug 764429 and everything is working fine. Hence marking this bug as resolved.