Bug 1005575
| Summary: | Dist-geo-rep : gluster volume geo <master_vol> <slave_ip>::<slave_vol> config throws error 'Staging failed..command failed' after adding brick to master volume(before reconfiguration of session) | ||
|---|---|---|---|
| Product: | [Red Hat Storage] Red Hat Gluster Storage | Reporter: | Rachana Patel <racpatel> |
| Component: | geo-replication | Assignee: | Bug Updates Notification Mailing List <rhs-bugs> |
| Status: | CLOSED EOL | QA Contact: | storage-qa-internal <storage-qa-internal> |
| Severity: | medium | Docs Contact: | |
| Priority: | medium | ||
| Version: | 2.1 | CC: | avishwan, chrisw, csaba, mzywusko, nsathyan, rhs-bugs, vagarwal |
| Target Milestone: | --- | ||
| Target Release: | --- | ||
| Hardware: | x86_64 | ||
| OS: | Linux | ||
| Whiteboard: | config | ||
| Fixed In Version: | Doc Type: | Bug Fix | |
| Doc Text: |
Cause: This behaviour is expected and happens because glusterd today is cluster aware and not volume aware. Section 11.4 of the admin guide specifically asks the admin to perform a series of steps when a new brick on a new node is added in the cluster. By choice, we do not document the exact error messages that the command will output if the admin fails to perform these steps, as it will make the admin guide more complicated, and the error messages themselves are self explanatory. Hence the above mentioned behaviour is not an issue.
Consequence: This behaviour will also be seen, if a new node is added to the cluster(even if a new brick is not added) and the steps mentioned in 11.4 are not followed. This is because of the way glusterd functions today(not volume aware). This is normal glusterd behaviour and is same as every other gluster command which performs the same set of operations on all the nodes in the cluster(irrespective of whether the node is a part of the volume on which the operations are being performed or not).
Fix: We can update 11.4 section of the admin guide, so that it talks about adding new node in the cluster, along with adding new bricks in the volume.
Result:
|
Story Points: | --- |
| Clone Of: | Environment: | ||
| Last Closed: | 2015-11-25 08:47:48 UTC | Type: | Bug |
| Regression: | --- | Mount Type: | --- |
| Documentation: | --- | CRM: | |
| Verified Versions: | Category: | --- | |
| oVirt Team: | --- | RHEL 7.3 requirements from Atomic Host: | |
| Cloudforms Team: | --- | Target Upstream Version: | |
| Embargoed: | |||
|
Description
Rachana Patel
2013-09-08 15:22:58 UTC
log snippet:- [2013-09-08 12:14:27.905956] W [glusterd-geo-rep.c:1404:glusterd_op_gsync_args_get] 0-: master not found [2013-09-08 12:14:28.026453] E [glusterd-geo-rep.c:1745:glusterd_mountbroker_check] (-->/usr/lib64/libglusterfs.so.0(dict_foreach+0x45) [0x7f3a1310b4e5] (-->/usr/lib64/glusterfs/3.4.0.32rhs/xlator/mgmt/glusterd.so(+0x81dde) [0x7f3a0f6dedde] (-->/usr/lib64/glusterfs/3.4. 0.32rhs/xlator/mgmt/glusterd.so(+0x7aa43) [0x7f3a0f6d7a43]))) 0-: Assertion failed: op_errstr [2013-09-08 12:14:28.507895] E [glusterd-geo-rep.c:1745:glusterd_mountbroker_check] (-->/usr/lib64/libglusterfs.so.0(dict_foreach+0x45) [0x7f3a1310b4e5] (-->/usr/lib64/glusterfs/3.4.0.32rhs/xlator/mgmt/glusterd.so(+0x81dde) [0x7f3a0f6dedde] (-->/usr/lib64/glusterfs/3.4. 0.32rhs/xlator/mgmt/glusterd.so(+0x7aa43) [0x7f3a0f6d7a43]))) 0-: Assertion failed: op_errstr [2013-09-08 12:15:35.460649] I [glusterd-geo-rep.c:1573:glusterd_get_statefile_name] 0-: Using passed config template(/var/lib/glusterd /geo-replication/m_master1_rhsauto031.lab.eng.blr.redhat.com_slave1/gsyncd.conf). [2013-09-08 12:15:35.784245] E [glusterd-syncop.c:102:gd_collate_errors] 0-: Staging failed on 10.70.35.26. Error: Geo-replication session between m_master1 and rhsauto031.lab.eng.blr.redhat.com::slave1 does not exist. [2013-09-08 12:15:57.088611] I [glusterd-geo-rep.c:1573:glusterd_get_statefile_name] 0-: Using passed config template(/var/lib/glusterd/geo-replication/m_master1_rhsauto031.lab.eng.blr.redhat.com_slave1/gsyncd.conf). [2013-09-08 12:15:57.393328] E [glusterd-syncop.c:102:gd_collate_errors] 0-: Staging failed on 10.70.35.26. Error: Geo-replication session between m_master1 and rhsauto031.lab.eng.blr.redhat.com::slave1 does not exist. [2013-09-08 12:16:58.551798] I [glusterd-geo-rep.c:1573:glusterd_get_statefile_name] 0-: Using passed config template(/var/lib/glusterd/geo-replication/m_master1_rhsauto031.lab.eng.blr.redhat.com_slave1/gsyncd.conf). [2013-09-08 12:16:58.878982] E [glusterd-syncop.c:102:gd_collate_errors] 0-: Staging failed on 10.70.35.26. Error: Geo-replication session between m_master1 and rhsauto031.lab.eng.blr.redhat.com::slave1 does not exist. [2013-09-08 12:18:00.354277] I [glusterd-geo-rep.c:1573:glusterd_get_statefile_name] 0-: Using passed config template(/var/lib/glusterd/geo-replication/m_master1_rhsauto031.lab.eng.blr.redhat.com_slave1/gsyncd.conf). [2013-09-08 12:18:40.690326] I [glusterd-geo-rep.c:1573:glusterd_get_statefile_name] 0-: Using passed config template(/var/lib/glusterd/geo-replication/m_master1_rhsauto031.lab.eng.blr.redhat.com_slave1/gsyncd.conf). [2013-09-08 12:18:40.864229] I [glusterd-geo-rep.c:1991:glusterd_op_stage_gsync_create] 0-: Session between m_master1 and rhsauto031.lab.eng.blr.redhat.com::slave1 is already created. Force creating again. [2013-09-08 12:18:41.278845] I [run.c:190:runner_log] 0-management: Ran script: /var/lib/glusterd/hooks/1/gsync-create/post/S56glusterd-geo-rep-create-post.sh --volname=m_master1 This argument will stop the hooks script [2013-09-08 12:19:29.759510] I [glusterd-geo-rep.c:1573:glusterd_get_statefile_name] 0-: Using passed config template(/var/lib/glusterd/geo-replication/m_rep_master1_rhsauto031.lab.eng.blr.redhat.com_slave2/gsyncd.conf). [2013-09-08 12:19:29.862407] I [glusterd-geo-rep.c:1991:glusterd_op_stage_gsync_create] 0-: Session between m_rep_master1 and rhsauto031.lab.eng.blr.redhat.com::slave2 is already created. Force creating again. Targeting for 3.0.0 (Denali) release. Dev ack to 3.0 RHS BZs Closing this bug since RHGS 2.1 release reached EOL. Required bugs are cloned to RHGS 3.1. Please re-open this issue if found again. Closing this bug since RHGS 2.1 release reached EOL. Required bugs are cloned to RHGS 3.1. Please re-open this issue if found again. |