← Back to team overview

openstack team mailing list archive

Re: [OpenStack][Nova] Live Migration Error

 

That did! Thanks :)

Do you by change have any pointer on getting the live-migration to work
without running libvirt under root?

On Tue, Jul 10, 2012 at 10:55 AM, Sébastien Han <han.sebastien@xxxxxxxxx>wrote:

> Change the vncserver_listen to 0.0.0.0 and re-try the live-migration, you
> should get better results :)
>
>
>
> On Tue, Jul 10, 2012 at 11:52 AM, Leander Bessa Beernaert <
> leanderbb@xxxxxxxxx> wrote:
>
>> If i don't run libvirt with root, it can't write to the shared folder.
>> It's the only way i've been able to get this to work. :S
>>
>> Below follows the configuration of one of the compute nodes. 10.0.1.1 is
>> the controller and 10.0.1.2 is the compute node.
>>
>> novncproxy_base_url=http://10.0.1.1:6080/vnc_auto.html
>>> xvpvncproxy_base_url=http://10.0.1.1:6081/console
>>> vncserver_proxyclient_address=10.0.1.2
>>> vncserver_listen=10.0.1.2
>>
>>
>> On Tue, Jul 10, 2012 at 10:49 AM, Sébastien Han <han.sebastien@xxxxxxxxx>wrote:
>>
>>> I forgot to ask, did you enable the vnc console?
>>>
>>> If so, with which parameters?
>>>
>>>
>>> On Tue, Jul 10, 2012 at 11:48 AM, Sébastien Han <han.sebastien@xxxxxxxxx
>>> > wrote:
>>>
>>>> Hi!
>>>>
>>>> Usually you get:
>>>>
>>>> 2012-07-09 13:58:27.179+0000: 10227: warning : qemuDomainObjTaint:1134
>>>>> : Domain id=2 name='instance-00000002'
>>>>> uuid=57aca8a6-d062-4a08-8d87-e4d11d259ac7 is tainted: high-privileges
>>>>
>>>>
>>>> when you change permission in libvirt (root I presumed) which is not
>>>> necessary.
>>>>
>>>> 2012-07-10 09:26:02.244+0000: 10229: error : virExecWithHook:328 :
>>>>> Cannot find 'pm-is-supported' in path: No such file or directory
>>>>
>>>>
>>>> This error is harmless and can be easily solved by installing the
>>>> following package:
>>>>
>>>> sudo apt-get install pm-utils -y
>>>>
>>>>
>>>> Do you have something in the nova-scheduler logs?
>>>>
>>>> Cheers!
>>>>
>>>> On Tue, Jul 10, 2012 at 11:29 AM, Leander Bessa Beernaert <
>>>> leanderbb@xxxxxxxxx> wrote:
>>>>
>>>>> Well i've checked the libvirt logs on both nodes and i found these two
>>>>> lines:
>>>>>
>>>>> 2012-07-09 13:58:27.179+0000: 10227: warning : qemuDomainObjTaint:1134
>>>>>> : Domain id=2 name='instance-00000002'
>>>>>> uuid=57aca8a6-d062-4a08-8d87-e4d11d259ac7 is tainted: high-privileges
>>>>>> 2012-07-09 13:58:27.736+0000: 10226: error : qemuMonitorIORead:513 :
>>>>>> Unable to read from monitor: Connection reset by peer
>>>>>
>>>>>
>>>>> The log is alos filled with the message below, it gets repeated over
>>>>> and over and over.
>>>>>
>>>>> 2012-07-10 09:26:02.244+0000: 10229: error : virExecWithHook:328 :
>>>>>> Cannot find 'pm-is-supported' in path: No such file or directory
>>>>>> 2012-07-10 09:26:02.244+0000: 10229: warning : qemuCapsInit:856 :
>>>>>> Failed to get host power management capabilities
>>>>>
>>>>>
>>>>> On Tue, Jul 10, 2012 at 8:16 AM, Razique Mahroua <
>>>>> razique.mahroua@xxxxxxxxx> wrote:
>>>>>
>>>>>> Hi Leander,
>>>>>> try to check libvirtd.log files,
>>>>>> is the instance still running on the first node while you are
>>>>>> lauching the migration process ?
>>>>>>
>>>>>> Razique
>>>>>>  *Nuage & Co - Razique Mahroua** *
>>>>>> razique.mahroua@xxxxxxxxx
>>>>>>
>>>>>>
>>>>>> Le 9 juil. 2012 à 16:09, Leander Bessa Beernaert a écrit :
>>>>>>
>>>>>> Ok, so i've updated to the test packages from
>>>>>>
>>>>>> The migration still fails, but i see no errors in the logs. I'm
>>>>>> trying to migrate a VM with the m1.tiny flavor from one machine to another.
>>>>>> Their hardware are identical and they have more than enough resources to
>>>>>> support the m1.tiny flavor:
>>>>>>
>>>>>> cloud35         (total)                         4    3867     186
>>>>>>> cloud35         (used_now)                      0     312       5
>>>>>>> cloud35         (used_max)                      0       0       0
>>>>>>
>>>>>>
>>>>>> These are the logs from the origin compute node:
>>>>>> http://paste.openstack.org/show/19319/  and  the destination compute
>>>>>> node: http://paste.openstack.org/show/19318/ . The scheduler's log
>>>>>> has no visible errors or stack traces.
>>>>>>
>>>>>> I'm still using nfsv4.
>>>>>>
>>>>>> Any ideas?
>>>>>>
>>>>>>
>>>>>> On Fri, Jul 6, 2012 at 7:57 PM, Leander Bessa Beernaert <
>>>>>> leanderbb@xxxxxxxxx> wrote:
>>>>>>
>>>>>>> Thanks for the tip, it's a better than nothing :)
>>>>>>>
>>>>>>> Regards,
>>>>>>> Leander
>>>>>>>
>>>>>>> On Fri, Jul 6, 2012 at 6:32 PM, Mandar Vaze / मंदार वझे <
>>>>>>> mandarvaze@xxxxxxxxx> wrote:
>>>>>>>
>>>>>>>> Not sure if you are able to debug this, but a while ago there was a
>>>>>>>> bug where instance.id was passed where instance.uuid was expected.
>>>>>>>> This used to cause some problem.
>>>>>>>> It looks like you are using distribution package rather than
>>>>>>>> devstack installation, so it is likely that the issue is now fixed. Can you
>>>>>>>> try latest packages (and/or try devstack if you can)
>>>>>>>>
>>>>>>>> I wish I could help more.
>>>>>>>>
>>>>>>>> -Mandar
>>>>>>>>
>>>>>>>>
>>>>>>>> On Fri, Jul 6, 2012 at 3:26 PM, Leander Bessa Beernaert <
>>>>>>>> leanderbb@xxxxxxxxx> wrote:
>>>>>>>>
>>>>>>>>> Hello,
>>>>>>>>>
>>>>>>>>> I've recently setup a system to test out the live migration
>>>>>>>>> feature. So far i've been able to launch the instances with the shared nfs
>>>>>>>>> folder. However, when i run the live-migration command i encounter this
>>>>>>>>> error in the destination compute node:
>>>>>>>>>
>>>>>>>>> 2012-07-05 09:33:48 ERROR nova.manager [-] Error during
>>>>>>>>>> ComputeManager.update_available_resource: Domain not found: no domain with
>>>>>>>>>> matching id 2
>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager Traceback (most recent
>>>>>>>>>> call last):
>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager   File
>>>>>>>>>> "/usr/lib/python2.7/dist-packages/nova/manager.py", line 155, in
>>>>>>>>>> periodic_tasks
>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager     task(self, context)
>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager   File
>>>>>>>>>> "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 2409, in
>>>>>>>>>> update_available_resource
>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager
>>>>>>>>>> self.driver.update_available_resource(context, self.host)
>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager   File
>>>>>>>>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/connection.py", line
>>>>>>>>>> 1936, in update_available_resource
>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager     'vcpus_used':
>>>>>>>>>> self.get_vcpu_used(),
>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager   File
>>>>>>>>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/connection.py", line
>>>>>>>>>> 1743, in get_vcpu_used
>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager     dom =
>>>>>>>>>> self._conn.lookupByID(dom_id)
>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager   File
>>>>>>>>>> "/usr/lib/python2.7/dist-packages/libvirt.py", line 2363, in lookupByID
>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager     if ret is None:raise
>>>>>>>>>> libvirtError('virDomainLookupByID() failed', conn=self)
>>>>>>>>>> 2012-07-05 09:33:48 TRACE nova.manager libvirtError: Domain not
>>>>>>>>>> found: no domain with matching id 2
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Any ideas on how to solve this?
>>>>>>>>>
>>>>>>>>> Regards,
>>>>>>>>> Leander
>>>>>>>>>
>>>>>>>>> _______________________________________________
>>>>>>>>> Mailing list: https://launchpad.net/~openstack
>>>>>>>>> Post to     : openstack@xxxxxxxxxxxxxxxxxxx
>>>>>>>>> Unsubscribe : https://launchpad.net/~openstack
>>>>>>>>> More help   : https://help.launchpad.net/ListHelp
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>> _______________________________________________
>>>>>> Mailing list: https://launchpad.net/~openstack
>>>>>> Post to     : openstack@xxxxxxxxxxxxxxxxxxx
>>>>>> Unsubscribe : https://launchpad.net/~openstack
>>>>>> More help   : https://help.launchpad.net/ListHelp
>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> Mailing list: https://launchpad.net/~openstack
>>>>> Post to     : openstack@xxxxxxxxxxxxxxxxxxx
>>>>> Unsubscribe : https://launchpad.net/~openstack
>>>>> More help   : https://help.launchpad.net/ListHelp
>>>>>
>>>>>
>>>>
>>>
>>
>

Follow ups

References