Bug 974917 - Can't launch VM with 2048 GB memory but 2047 GB is OK.
Summary: Can't launch VM with 2048 GB memory but 2047 GB is OK.
Keywords:
Status: CLOSED CURRENTRELEASE
Alias: None
Product: Red Hat Enterprise Virtualization Manager
Classification: Red Hat
Component: ovirt-engine
Version: 3.2.0
Hardware: x86_64
OS: Linux
unspecified
unspecified
Target Milestone: ---
: 3.3.0
Assignee: Martin Sivák
QA Contact: Tareq Alayan
URL:
Whiteboard: sla
: 975945 (view as bug list)
Depends On:
Blocks: 985973
TreeView+ depends on / blocked
 
Reported: 2013-06-17 06:42 UTC by Hajime Taira
Modified: 2016-02-10 20:17 UTC (History)
13 users (show)

Fixed In Version:
Doc Type: Bug Fix
Doc Text:
Clone Of:
: 985973 (view as bug list)
Environment:
Last Closed: 2014-01-21 22:14:20 UTC
oVirt Team: SLA
Target Upstream Version:
Embargoed:


Attachments (Terms of Use)
packed log file about engine.log and vdsm.log (54.16 KB, application/x-gzip-compressed)
2013-06-21 02:33 UTC, Hajime Taira
no flags Details
Update ovirt mempolicy script (411 bytes, text/plain)
2013-06-24 02:39 UTC, Hajime Taira
no flags Details
vdsm.log - succeeded to run VM with 2TB memory (24.95 KB, text/plain)
2013-09-02 03:05 UTC, Hajime Taira
no flags Details


Links
System ID Private Priority Status Summary Last Updated
oVirt gerrit 17086 0 None None None Never

Description Hajime Taira 2013-06-17 06:42:25 UTC
Description of problem:
I cloud not launch VM with 2048 GB memory but 2047 GB is OK.

It seems that value overflowed in communication of XML-RPC.
RHEV engine logged following messages in /var/log/ovirt-engine/engine.log

2013-06-17 15:36:43,587 INFO  [org.ovirt.engine.core.vdsbroker.CreateVmVDSCommand] (ajp-/127.0.0.1:8702-8) [1275c72b] START, CreateVmVDSCommand(HostName = cpc-dl980g7.hp.cpc, HostId = baec873f-dc2f-44a2-ac3c-021100675f8b, vmId=84178847-54ac-4ea0-adc3-65ec16fe70a0, vm=VM [RHEL64-2TB]), log id: 473f9162
2013-06-17 15:36:43,627 INFO  [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand] (ajp-/127.0.0.1:8702-8) [1275c72b] START, CreateVDSCommand(HostName = cpc-dl980g7.hp.cpc, HostId = baec873f-dc2f-44a2-ac3c-021100675f8b, vmId=84178847-54ac-4ea0-adc3-65ec16fe70a0, vm=VM [RHEL64-2TB]), log id: 285aecb1
2013-06-17 15:36:44,236 ERROR [org.ovirt.engine.core.vdsbroker.VDSCommandBase] (ajp-/127.0.0.1:8702-8) [1275c72b] Command CreateVDS execution failed. Exception: VDSNetworkException: org.apache.xmlrpc.XmlRpcException: <type 'exceptions.OverflowError'>:int exceeds XML-RPC limits
2013-06-17 15:36:44,236 INFO  [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand] (ajp-/127.0.0.1:8702-8) [1275c72b] FINISH, CreateVDSCommand, log id: 285aecb1
2013-06-17 15:36:44,236 ERROR [org.ovirt.engine.core.vdsbroker.CreateVmVDSCommand] (ajp-/127.0.0.1:8702-8) [1275c72b] VDS::create Failed creating vm RHEL64-2TB in vds = baec873f-dc2f-44a2-ac3c-021100675f8b : cpc-dl980g7.hp.cpc error = org.ovirt.engine.core.vdsbroker.vdsbroker.VDSNetworkException: org.apache.xmlrpc.XmlRpcException: <type 'exceptions.OverflowError'>:int exceeds XML-RPC limits

Please see attached file for detail.

Version-Release number of selected component (if applicable):
rhevm-3.2.0-11.30.el6ev.noarch

How reproducible:
Yes.

Steps to Reproduce:
1. define new VM with 2TB (2048 GB) memory.
2. launch the VM.

Comment 1 Hajime Taira 2013-06-18 00:46:32 UTC
If we can't fix this issue before rhev-3.3. Please cap limitation of memory size by rhevm-web admin and user portal for work around on rhev-3.2.z.

Comment 2 Dan Kenigsberg 2013-06-19 09:05:02 UTC
Would you attach the vmCreate call (and return) from vdsm.log? I would like to understand where we report/expect memory size in KiB, and overflow the 31 bit limitation of xmlrpc integers.

(we have a worse issue with setting/reporting balloon size, where we use integer with bytes as units: balloon can grow no bigger than 2GiB.)

Comment 3 Doron Fediuck 2013-06-19 12:28:38 UTC
(In reply to Dan Kenigsberg from comment #2)
> Would you attach the vmCreate call (and return) from vdsm.log? I would like
> to understand where we report/expect memory size in KiB, and overflow the 31
> bit limitation of xmlrpc integers.
> 
> (we have a worse issue with setting/reporting balloon size, where we use
> integer with bytes as units: balloon can grow no bigger than 2GiB.)

Thanks Dan.
Since the balloon wasn't in use so far, we can handle it without backwards compatibility issues.

Comment 4 Hajime Taira 2013-06-21 02:33:55 UTC
Created attachment 763621 [details]
packed log file about engine.log and vdsm.log

Hi, Dan

Please confirm attached log files.

At first, I set memory size 2097151 MB RHEL64-2TB(84178847-54ac-4ea0-adc3-65ec16fe70a0).
When RHEL64-2TB is boot successful.

At second, I change memory size 2097152 MB to RHEL64-2TB(84178847-54ac-4ea0-adc3-65ec16fe70a0).
When RHEL64-2TB is boot fail. But qemu-kvm process is running on RHEV-H.

Finally, I killed qemu-kvm process manually.

Comment 5 Dan Kenigsberg 2013-06-23 09:41:16 UTC
Oh it seems that the balloon issue *is* our issue: vmCreate succeeds, however following calls to getVmStats returns a balloon size that is too big to digest

Thread-195703::DEBUG::2013-06-21 01:23:01,049::BindingXMLRPC::920::vds::(wrapper) return vmGetStats with {'status': {'message': 'Done', 'code': 0}, 'statsList': [{'status': 'Powering up', 'username': 'Unknown', 'memUsage': '0', 'acpiEnable': 'true', 'pid': '23104', 'displayIp': '0', 'displayPort': u'5901', 'session': 'Unknown', 'displaySecurePort': '-1', 'timeOffset': '-2', 'hash': '8092988585526967060', 'balloonInfo': {'balloon_max': 2147483648, 'balloon_cur': 2147483648}, 'pauseCode': 'NOERR', 'clientIp': '', 'kvmEnable': 'true', 'network': {u'vnet1': {'macAddr': '00:1a:4a:a8:7a:09', 'rxDropped': '0', 'rxErrors': '0', 'txDropped': '0', 'txRate': '0.0', 'rxRate': '0.0', 'txErrors': '0', 'state': 'unknown', 'speed': '1000', 'name': u'vnet1'}}, 'vmId': '84178847-54ac-4ea0-adc3-65ec16fe70a0', 'displayType': 'vnc', 'cpuUser': '42.34', 'disks': {u'vda': {'truesize': '107374182400', 'apparentsize': '107374182400', 'imageID': 'dc55cddb-84cb-4c0f-9095-ca244df6ddde'}, u'hdc': {'truesize': '0', 'apparentsize': '0'}}, 'monitorResponse': '0', 'statsAge': '1.06', 'elapsedTime': '55', 'vmType': 'kvm', 'cpuSys': '57.51', 'appsList': [], 'guestIPs': ''}]}

Comment 6 Doron Fediuck 2013-06-23 12:14:28 UTC
A quick work around would be to remove the balloon,
which is accessible via the REST API:
<memory_policy>
 <guaranteed>1610612736</guaranteed>
 <ballooning>false</ballooning>
</memory_policy>

Simply set it to false as in this sample.

Comment 7 Hajime Taira 2013-06-24 02:39:11 UTC
Created attachment 764401 [details]
Update ovirt mempolicy script

I updated mempolicy of VM (RHEL64-2TB) when I can launch it.

Please confirm following log.

2013-06-24 19:58:15,138 INFO  [org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand] (ajp-/127.0.0.1:8702-11) START, IsVmDuringInitiati
ngVDSCommand( vmId = 84178847-54ac-4ea0-adc3-65ec16fe70a0), log id: 34619876
2013-06-24 19:58:15,138 INFO  [org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand] (ajp-/127.0.0.1:8702-11) FINISH, IsVmDuringInitiat
ingVDSCommand, return: false, log id: 34619876
2013-06-24 19:58:15,180 INFO  [org.ovirt.engine.core.bll.RunVmCommand] (pool-4-thread-50) [3811d911] Running command: RunVmCommand internal: fals
e. Entities affected :  ID: 84178847-54ac-4ea0-adc3-65ec16fe70a0 Type: VM
2013-06-24 19:58:15,247 INFO  [org.ovirt.engine.core.vdsbroker.irsbroker.IsoPrefixVDSCommand] (pool-4-thread-50) [3811d911] START, IsoPrefixVDSCo
mmand( storagePoolId = 5849b030-626e-47cb-ad90-3ce782d831b3, ignoreFailoverLimit = false, compatabilityVersion = null), log id: 3fb22615
2013-06-24 19:58:15,247 INFO  [org.ovirt.engine.core.vdsbroker.irsbroker.IsoPrefixVDSCommand] (pool-4-thread-50) [3811d911] FINISH, IsoPrefixVDSC
ommand, return: /rhev/data-center/mnt/rhevm32.hp.cpc:_srv_nfs_iso/4921cd96-781a-43fa-a279-4c945904312e/images/11111111-1111-1111-1111-11111111111
1, log id: 3fb22615
2013-06-24 19:58:15,269 INFO  [org.ovirt.engine.core.vdsbroker.CreateVmVDSCommand] (pool-4-thread-50) [3811d911] START, CreateVmVDSCommand(HostNa
me = cpc-dl980g7.hp.cpc, HostId = baec873f-dc2f-44a2-ac3c-021100675f8b, vmId=84178847-54ac-4ea0-adc3-65ec16fe70a0, vm=VM [RHEL64-2TB]), log id: 4
8e24d5e
2013-06-24 19:58:15,288 INFO  [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand] (pool-4-thread-50) [3811d911] START, CreateVDSCommand(
HostName = cpc-dl980g7.hp.cpc, HostId = baec873f-dc2f-44a2-ac3c-021100675f8b, vmId=84178847-54ac-4ea0-adc3-65ec16fe70a0, vm=VM [RHEL64-2TB]), log
 id: 7900ac9a
2013-06-24 19:58:15,357 INFO  [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand] (pool-4-thread-50) [3811d911] org.ovirt.engine.core.vd
sbroker.vdsbroker.CreateVDSCommand spiceSslCipherSuite=DEFAULT,memSize=2097152,kvmEnable=true,smp=40,vmType=kvm,emulatedMachine=rhel6.4.0,keyboar
dLayout=en-us,pitReinjection=false,nice=0,display=vnc,smartcardEnable=false,tabletEnable=true,smpCoresPerSocket=4,spiceSecureChannels=smain,sinpu
ts,scursor,splayback,srecord,sdisplay,susbredir,ssmartcard,timeOffset=-2,transparentHugePages=true,vmId=84178847-54ac-4ea0-adc3-65ec16fe70a0,devi
ces=[Ljava.util.Map;@4ae74e50,acpiEnable=true,vmName=RHEL64-2TB,cpuType=Westmere,custom={device_1ecc0c93-4bba-4676-8e1a-785e87bfbd4b=VmDevice {vm
Id=84178847-54ac-4ea0-adc3-65ec16fe70a0, deviceId=1ecc0c93-4bba-4676-8e1a-785e87bfbd4b, device=ide, type=controller, bootOrder=0, specParams={}, 
address={bus=0x00, domain=0x0000, type=pci, slot=0x01, function=0x1}, managed=false, plugged=true, readOnly=false, deviceAlias=ide0}, device_1ecc
0c93-4bba-4676-8e1a-785e87bfbd4bdevice_cc47705b-e514-4232-b120-457062072439=VmDevice {vmId=84178847-54ac-4ea0-adc3-65ec16fe70a0, deviceId=cc47705
b-e514-4232-b120-457062072439, device=virtio-serial, type=controller, bootOrder=0, specParams={}, address={bus=0x00, domain=0x0000, type=pci, slo
t=0x04, function=0x0}, managed=false, plugged=true, readOnly=false, deviceAlias=virtio-serial0}, device_1ecc0c93-4bba-4676-8e1a-785e87bfbd4bdevic
e_cc47705b-e514-4232-b120-457062072439device_f4912582-8e6a-4381-84d7-9bbb0bf9842cdevice_14b9ef9d-c95b-463f-822b-717c44e6ff5b=VmDevice {vmId=84178
847-54ac-4ea0-adc3-65ec16fe70a0, deviceId=14b9ef9d-c95b-463f-822b-717c44e6ff5b, device=unix, type=channel, bootOrder=0, specParams={}, address={p
ort=2, bus=0, controller=0, type=virtio-serial}, managed=false, plugged=true, readOnly=false, deviceAlias=channel1}, device_1ecc0c93-4bba-4676-8e
1a-785e87bfbd4bdevice_cc47705b-e514-4232-b120-457062072439device_f4912582-8e6a-4381-84d7-9bbb0bf9842c=VmDevice {vmId=84178847-54ac-4ea0-adc3-65ec
16fe70a0, deviceId=f4912582-8e6a-4381-84d7-9bbb0bf9842c, device=unix, type=channel, bootOrder=0, specParams={}, address={port=1, bus=0, controlle
r=0, type=virtio-serial}, managed=false, plugged=true, readOnly=false, deviceAlias=channel0}}
2013-06-24 19:58:15,357 INFO  [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand] (pool-4-thread-50) [3811d911] FINISH, CreateVDSCommand
, log id: 7900ac9a
2013-06-24 19:58:15,362 INFO  [org.ovirt.engine.core.vdsbroker.CreateVmVDSCommand] (pool-4-thread-50) [3811d911] IncreasePendingVms::CreateVmIncr
easing vds cpc-dl980g7.hp.cpc pending vcpu count, now 40. Vm: RHEL64-2TB
2013-06-24 19:58:15,365 INFO  [org.ovirt.engine.core.vdsbroker.CreateVmVDSCommand] (pool-4-thread-50) [3811d911] FINISH, CreateVmVDSCommand, retu
rn: WaitForLaunch, log id: 48e24d5e
2013-06-24 19:58:15,365 INFO  [org.ovirt.engine.core.bll.RunVmCommand] (pool-4-thread-50) [3811d911] Lock freed to object EngineLock [exclusiveLo
cks= key: 84178847-54ac-4ea0-adc3-65ec16fe70a0 value: VM
, sharedLocks= ]
2013-06-24 19:58:15,366 WARN  [org.ovirt.engine.core.compat.backendcompat.PropertyInfo] (pool-4-thread-50) Unable to get value of property: glust
erVolume for class org.ovirt.engine.core.bll.RunVmCommand
2013-06-24 19:58:15,366 WARN  [org.ovirt.engine.core.compat.backendcompat.PropertyInfo] (pool-4-thread-50) Unable to get value of property: vds f
or class org.ovirt.engine.core.bll.RunVmCommand
2013-06-24 19:58:35,398 INFO  [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] (QuartzScheduler_Worker-28) [73388428] VM RHEL64-2TB 8417884
7-54ac-4ea0-adc3-65ec16fe70a0 moved from WaitForLaunch --> PoweringUp
2013-06-24 19:58:35,398 INFO  [org.ovirt.engine.core.vdsbroker.vdsbroker.FullListVdsCommand] (QuartzScheduler_Worker-28) [73388428] START, FullLi
stVdsCommand( HostId = baec873f-dc2f-44a2-ac3c-021100675f8b, vds=null, vmIds=[84178847-54ac-4ea0-adc3-65ec16fe70a0]), log id: ca57c25
2013-06-24 19:58:35,428 INFO  [org.ovirt.engine.core.vdsbroker.vdsbroker.FullListVdsCommand] (QuartzScheduler_Worker-28) [73388428] FINISH, FullL
istVdsCommand, return: [Lorg.ovirt.engine.core.vdsbroker.xmlrpc.XmlRpcStruct;@4479679f, log id: ca57c25
2013-06-24 19:58:35,430 INFO  [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] (QuartzScheduler_Worker-28) [73388428] Received a memballoon
 Device without an address when processing VM 84178847-54ac-4ea0-adc3-65ec16fe70a0 devices, skipping device: {specParams={model=none}, device=mem
balloon, type=balloon}
2013-06-24 19:58:35,461 INFO  [org.ovirt.engine.core.vdsbroker.UpdateVdsDynamicDataVDSCommand] (QuartzScheduler_Worker-28) [73388428] START, Upda
teVdsDynamicDataVDSCommand(HostName = cpc-dl980g7.hp.cpc, HostId = baec873f-dc2f-44a2-ac3c-021100675f8b, vdsDynamic=org.ovirt.engine.core.common.
businessentities.VdsDynamic@99368aca), log id: 3a1e30d
2013-06-24 19:58:35,462 INFO  [org.ovirt.engine.core.vdsbroker.UpdateVdsDynamicDataVDSCommand] (QuartzScheduler_Worker-28) [73388428] FINISH, UpdateVdsDynamicDataVDSCommand, log id: 3a1e30d
2013-06-24 19:59:15,760 INFO  [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] (QuartzScheduler_Worker-56) VM RHEL64-2TB 84178847-54ac-4ea0-adc3-65ec16fe70a0 moved from PoweringUp --> Up

Comment 9 Doron Fediuck 2013-07-01 15:12:59 UTC
*** Bug 975945 has been marked as a duplicate of this bug. ***

Comment 13 Hajime Taira 2013-09-02 03:05:46 UTC
Created attachment 792704 [details]
vdsm.log - succeeded to run VM with 2TB memory

I tested it again with vdsm 4.10.2-25.0.el6ev.
It succeeded to run VM with 2TB memory on HP ProLiant DL980 G7 (4TB Phys Memory).
Please confirm log file.

Comment 14 Hajime Taira 2013-09-02 05:17:45 UTC
And I tested currently released version vdsm-4.10.2-24.1.el6ev in same PoC environment.
It's works fine.

Comment 15 Doron Fediuck 2013-09-08 23:45:43 UTC
Moving to verified per comment 14.

Comment 16 Itamar Heim 2014-01-21 22:14:20 UTC
Closing - RHEV 3.3 Released

Comment 17 Itamar Heim 2014-01-21 22:22:04 UTC
Closing - RHEV 3.3 Released


Note You need to log in before you can comment on or make changes to this bug.