RHEL Engineering is moving the tracking of its product development work on RHEL 6 through RHEL 9 to Red Hat Jira (issues.redhat.com). If you're a Red Hat customer, please continue to file support cases via the Red Hat customer portal. If you're not, please head to the "RHEL project" in Red Hat Jira and file new tickets here. Individual Bugzilla bugs in the statuses "NEW", "ASSIGNED", and "POST" are being migrated throughout September 2023. Bugs of Red Hat partners with an assigned Engineering Partner Manager (EPM) are migrated in late September as per pre-agreed dates. Bugs against components "kernel", "kernel-rt", and "kpatch" are only migrated if still in "NEW" or "ASSIGNED". If you cannot log in to RH Jira, please consult article #7032570. That failing, please send an e-mail to the RH Jira admins at rh-issues@redhat.com to troubleshoot your issue as a user management inquiry. The email creates a ServiceNow ticket with Red Hat. Individual Bugzilla bugs that are migrated will be moved to status "CLOSED", resolution "MIGRATED", and set with "MigratedToJIRA" in "Keywords". The link to the successor Jira issue will be found under "Links", have a little "two-footprint" icon next to it, and direct you to the "RHEL project" in Red Hat Jira (issue links are of type "https://issues.redhat.com/browse/RHEL-XXXX", where "X" is a digit). This same link will be available in a blue banner at the top of the page informing you that that bug has been migrated.
Bug 845525 - beta2 - VSDM is not reporting the BONDING_OPTS for bonds
Summary: beta2 - VSDM is not reporting the BONDING_OPTS for bonds
Keywords:
Status: CLOSED ERRATA
Alias: None
Product: Red Hat Enterprise Linux 6
Classification: Red Hat
Component: vdsm
Version: 6.4
Hardware: x86_64
OS: Linux
unspecified
high
Target Milestone: beta
: 6.4
Assignee: Dan Kenigsberg
QA Contact: Martin Pavlik
URL:
Whiteboard: network
: 845804 (view as bug list)
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2012-08-03 10:13 UTC by Martin Pavlik
Modified: 2022-07-09 05:37 UTC (History)
16 users (show)

Fixed In Version: v4.9.6-27.0
Doc Type: Bug Fix
Doc Text:
Previously, when you added a bridged logical network to a cluster and you created a bond in mode=1, and you attached the bridged logical network to the bond, Red Hat Enterprise Virtualization Manager displayed the wrong bond mode. This was due to VDSM failing to report the BONDING_OPTS for the bonds, which caused the backend to be unable to resolve the bond mode. When this happened, the default behavior was to set the bond mode to "custom", which was incorrect. Now, an update to VDSM makes sure that the bond mode is properly set.
Clone Of:
Environment:
Last Closed: 2012-12-04 19:04:30 UTC
Target Upstream Version:
Embargoed:


Attachments (Terms of Use)
logs (413.42 KB, application/x-gzip)
2012-08-03 10:13 UTC, Martin Pavlik
no flags Details
screenshot 1 (191.56 KB, image/png)
2012-08-03 10:13 UTC, Martin Pavlik
no flags Details


Links
System ID Private Priority Status Summary Last Updated
Red Hat Product Errata RHSA-2012:1508 0 normal SHIPPED_LIVE Important: rhev-3.1.0 vdsm security, bug fix, and enhancement update 2012-12-04 23:48:05 UTC

Description Martin Pavlik 2012-08-03 10:13:04 UTC
Created attachment 602101 [details]
logs

Description of problem:
if bond (tested with mode=1) is created via setup networks, bond gets created properly on host, but rhevm is diplaying mode custom:

Version-Release number of selected component (if applicable):
Red Hat Enterprise Virtualization Manager Version: '3.1.0-10.el6ev' 

How reproducible:
100%

Steps to Reproduce:
1. Add bridged logical network to cluster
2. Host -> your host -> network interfaces -> setup host networks -> create bond (mode=1) and attach network from step 1 to bond -> click OK
3. Host -> your host -> network interfaces -> setup host networks -> click pencil icon on bond (see screenshot1)
  
Actual results:
wrong bonf mode is diplayed

Expected results:
correct bond mode should be displayed

Additional info:

paramters are send correctly

MainProcess|Thread-5765::DEBUG::2012-08-03 11:58:41,945::configNetwork::1189::setupNetworks::(setupNetworks) Setting up network according to configuration: networks:{'NET1': {'bonding': 'bond1', 'STP': 'no', 'bridged': 'true'}}, bondings:{'bond1': {'nics': ['p1p1', 'p1p2'], 'options': 'mode=1 miimon=100'}}, options:{'connectivityCheck': 'true', 'connectivityTimeout': 120}


bond gets created properly

[root@dell-r210ii-07 ~]# cat /etc/sysconfig/network-scripts/ifcfg-bond1 
DEVICE=bond1
ONBOOT=yes
BONDING_OPTS='mode=1 miimon=100'
BRIDGE=NET1
NM_CONTROLLED=no
STP=no
[root@dell-r210ii-07 ~]# cat /proc/net/bonding/bond0
Ethernet Channel Bonding Driver: v3.6.0 (September 26, 2009)

Bonding Mode: load balancing (round-robin)
MII Status: down
MII Polling Interval (ms): 0
Up Delay (ms): 0
Down Delay (ms): 0
[root@dell-r210ii-07 ~]# cat /proc/net/bonding/bond1
Ethernet Channel Bonding Driver: v3.6.0 (September 26, 2009)

Bonding Mode: fault-tolerance (active-backup)
Primary Slave: None
Currently Active Slave: p1p1
MII Status: up
MII Polling Interval (ms): 100
Up Delay (ms): 0
Down Delay (ms): 0

Slave Interface: p1p1
MII Status: up
Speed: 1000 Mbps
Duplex: full
Link Failure Count: 0
Permanent HW addr: 90:e2:ba:04:29:88
Slave queue ID: 0

Slave Interface: p1p2
MII Status: up
Speed: 1000 Mbps
Duplex: full
Link Failure Count: 0
Permanent HW addr: 90:e2:ba:04:29:89
Slave queue ID: 0

Comment 1 Martin Pavlik 2012-08-03 10:13:49 UTC
Created attachment 602102 [details]
screenshot 1

Comment 3 Simon Grinberg 2012-08-06 08:14:06 UTC
*** Bug 845804 has been marked as a duplicate of this bug. ***

Comment 6 Moti Asayag 2012-08-06 09:21:00 UTC
Martin, can you add the output of 'vdsClient -s 0 getVdsCaps' to verify how VDSM report for that bond ?

Comment 7 Martin Pavlik 2012-08-06 10:08:58 UTC
vdsm-4.9.6-26.0.el6_3.x86_64

[root@dell-r210ii-07 ~]# vdsClient -s 0 getVdsCaps
	HBAInventory = {'iSCSI': [{'InitiatorName': 'iqn.1994-05.com.redhat:2bf48d38b7da'}], 'FC': []}
	ISCSIInitiatorName = iqn.1994-05.com.redhat:2bf48d38b7da
	bondings = {'bond4': {'addr': '', 'cfg': {'SLAVE': 'yes', 'ONBOOT': 'yes', 'NM_CONTROLLED': 'no', 'STP': 'no', 'HWADDR': '90:E2:BA:04:29:89', 'MASTER': 'bond4', 'DEVICE': 'p1p2'}, 'mtu': '1500', 'netmask': '', 'slaves': ['p1p1', 'p1p2'], 'hwaddr': '90:E2:BA:04:29:88'}, 'bond0': {'addr': '', 'cfg': {'SLAVE': 'yes', 'ONBOOT': 'yes', 'NM_CONTROLLED': 'no', 'STP': 'no', 'HWADDR': '90:E2:BA:04:29:89', 'MASTER': 'bond4', 'DEVICE': 'p1p2'}, 'mtu': '1500', 'netmask': '', 'slaves': [], 'hwaddr': '00:00:00:00:00:00'}, 'bond1': {'addr': '', 'cfg': {'SLAVE': 'yes', 'ONBOOT': 'yes', 'NM_CONTROLLED': 'no', 'STP': 'no', 'HWADDR': '90:E2:BA:04:29:89', 'MASTER': 'bond4', 'DEVICE': 'p1p2'}, 'mtu': '1500', 'netmask': '', 'slaves': [], 'hwaddr': '00:00:00:00:00:00'}, 'bond2': {'addr': '', 'cfg': {'SLAVE': 'yes', 'ONBOOT': 'yes', 'NM_CONTROLLED': 'no', 'STP': 'no', 'HWADDR': '90:E2:BA:04:29:89', 'MASTER': 'bond4', 'DEVICE': 'p1p2'}, 'mtu': '1500', 'netmask': '', 'slaves': [], 'hwaddr': '00:00:00:00:00:00'}, 'bond3': {'addr': '', 'cfg': {'SLAVE': 'yes', 'ONBOOT': 'yes', 'NM_CONTROLLED': 'no', 'STP': 'no', 'HWADDR': '90:E2:BA:04:29:89', 'MASTER': 'bond4', 'DEVICE': 'p1p2'}, 'mtu': '1500', 'netmask': '', 'slaves': [], 'hwaddr': '00:00:00:00:00:00'}}
	bridges = {'rhevm': {'addr': '10.34.66.71', 'cfg': {'DELAY': '0', 'NM_CONTROLLED': 'no', 'BOOTPROTO': 'dhcp', 'DEVICE': 'rhevm', 'TYPE': 'Bridge', 'ONBOOT': 'yes'}, 'mtu': '1500', 'netmask': '255.255.255.0', 'stp': 'off', 'ports': ['em1']}, 'NET1': {'addr': '', 'cfg': {'DELAY': '0', 'NM_CONTROLLED': 'no', 'STP': 'no', 'DEVICE': 'NET1', 'TYPE': 'Bridge', 'ONBOOT': 'yes'}, 'mtu': '1500', 'netmask': '', 'stp': 'off', 'ports': ['bond4']}}
	clusterLevels = ['3.0', '3.1']
	cpuCores = 4
	cpuFlags = fpu,vme,de,pse,tsc,msr,pae,mce,cx8,apic,mtrr,pge,mca,cmov,pat,pse36,clflush,dts,acpi,mmx,fxsr,sse,sse2,ss,ht,tm,pbe,syscall,nx,rdtscp,lm,constant_tsc,arch_perfmon,pebs,bts,rep_good,xtopology,nonstop_tsc,aperfmperf,pni,pclmulqdq,dtes64,monitor,ds_cpl,vmx,smx,est,tm2,ssse3,cx16,xtpr,pdcm,sse4_1,sse4_2,x2apic,popcnt,aes,xsave,avx,lahf_lm,ida,arat,epb,xsaveopt,pln,pts,dts,tpr_shadow,vnmi,flexpriority,ept,vpid,model_Nehalem,model_Conroe,model_Penryn,model_Westmere,model_SandyBridge
	cpuModel = Intel(R) Xeon(R) CPU E31240 @ 3.30GHz
	cpuSockets = 1
	cpuSpeed = 3292.421
	emulatedMachines = ['rhel6.3.0', 'pc', 'rhel6.2.0', 'rhel6.1.0', 'rhel6.0.0', 'rhel5.5.0', 'rhel5.4.4', 'rhel5.4.0']
	guestOverhead = 65
	hooks = {}
	kvmEnabled = true
	lastClient = 10.34.63.65
	lastClientIface = rhevm
	management_ip = 
	memSize = 7862
	netConfigDirty = True
	networks = {'rhevm': {'iface': 'rhevm', 'addr': '10.34.66.71', 'cfg': {'DELAY': '0', 'NM_CONTROLLED': 'no', 'BOOTPROTO': 'dhcp', 'DEVICE': 'rhevm', 'TYPE': 'Bridge', 'ONBOOT': 'yes'}, 'mtu': '1500', 'netmask': '255.255.255.0', 'stp': 'off', 'bridged': True, 'gateway': '10.34.66.254', 'ports': ['em1']}, 'NET1': {'iface': 'NET1', 'addr': '', 'cfg': {'DELAY': '0', 'NM_CONTROLLED': 'no', 'STP': 'no', 'DEVICE': 'NET1', 'TYPE': 'Bridge', 'ONBOOT': 'yes'}, 'mtu': '1500', 'netmask': '', 'stp': 'off', 'bridged': True, 'gateway': '0.0.0.0', 'ports': ['bond4']}}
	nics = {'p1p1': {'permhwaddr': '90:E2:BA:04:29:88', 'addr': '', 'cfg': {'SLAVE': 'yes', 'ONBOOT': 'yes', 'NM_CONTROLLED': 'no', 'STP': 'no', 'HWADDR': '90:E2:BA:04:29:88', 'MASTER': 'bond4', 'DEVICE': 'p1p1'}, 'mtu': '1500', 'netmask': '', 'hwaddr': '90:E2:BA:04:29:88', 'speed': 1000}, 'p1p2': {'permhwaddr': '90:E2:BA:04:29:89', 'addr': '', 'cfg': {'SLAVE': 'yes', 'ONBOOT': 'yes', 'NM_CONTROLLED': 'no', 'STP': 'no', 'HWADDR': '90:E2:BA:04:29:89', 'MASTER': 'bond4', 'DEVICE': 'p1p2'}, 'mtu': '1500', 'netmask': '', 'hwaddr': '90:E2:BA:04:29:88', 'speed': 1000}, 'em1': {'addr': '', 'cfg': {'BRIDGE': 'rhevm', 'NM_CONTROLLED': 'no', 'DEVICE': 'em1', 'BOOTPROTO': 'dhcp', 'HWADDR': 'D0:67:E5:F0:82:44', 'ONBOOT': 'yes'}, 'mtu': '1500', 'netmask': '', 'hwaddr': 'D0:67:E5:F0:82:44', 'speed': 1000}, 'em2': {'addr': '', 'cfg': {}, 'mtu': '1500', 'netmask': '', 'hwaddr': 'D0:67:E5:F0:82:45', 'speed': 0}}
	operatingSystem = {'release': '6.3.0.3.el6', 'version': '6Server', 'name': 'RHEL'}
	packages2 = {'kernel': {'release': '279.el6.x86_64', 'buildtime': 1339604676.0, 'version': '2.6.32'}, 'spice-server': {'release': '10.el6', 'buildtime': '1337611492', 'version': '0.10.1'}, 'vdsm': {'release': '26.0.el6_3', 'buildtime': '1343838445', 'version': '4.9.6'}, 'qemu-kvm': {'release': '2.298.el6_3', 'buildtime': '1343061272', 'version': '0.12.1.2'}, 'libvirt': {'release': '21.el6_3.3', 'buildtime': '1341975208', 'version': '0.9.10'}, 'qemu-img': {'release': '2.298.el6_3', 'buildtime': '1343061272', 'version': '0.12.1.2'}}
	reservedMem = 321
	software_revision = 26.0
	software_version = 4.9
	supportedProtocols = ['2.2', '2.3']
	supportedRHEVMs = ['3.0', '3.1']
	uuid = 4C4C4544-0037-5410-8033-C3C04F39354A_90:E2:BA:04:29:88
	version_name = Snow Man
	vlans = {}
	vmTypes = ['kvm']

Comment 8 Moti Asayag 2012-08-06 10:26:59 UTC
VSDM is not reporting the BONDING_OPTS for the bonds, therefore the backend is not capable to resolve the bond mode and sets its mode to 'custom' with no value.

Assigning the bug to VDSM component.

Comment 11 Dan Kenigsberg 2012-08-06 13:51:44 UTC
argh, broken merger fixed in

http://gerrit.usersys.redhat.com/1381

Comment 15 errata-xmlrpc 2012-12-04 19:04:30 UTC
Since the problem described in this bug report should be
resolved in a recent advisory, it has been closed with a
resolution of ERRATA.

For information on the advisory, and where to find the updated
files, follow the link below.

If the solution does not work for you, open a new bug report.

http://rhn.redhat.com/errata/RHSA-2012-1508.html


Note You need to log in before you can comment on or make changes to this bug.