← Back to team overview

yahoo-eng-team team mailing list archive

[Bug 1157725] Re: Snapshot stuck in 'queued' for VM after Volume detach

 

[Expired for OpenStack Compute (nova) because there has been no activity
for 60 days.]

** Changed in: nova
       Status: Incomplete => Expired

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1157725

Title:
  Snapshot stuck in 'queued' for VM after Volume detach

Status in OpenStack Compute (Nova):
  Expired

Bug description:
  Problem statement:
  A Vm can't have a snapshot taken successfully once it has had a volume attached and detached.

  Current versions - 
  Current version of OS -  Linux 3.2.0-39-generic #62-Ubuntu SMP Thu Feb 28 00:28:53 UTC 2013 x86_64 x86_64 x86_64 GNU/Linux
  Current version of Nova-compute - 2012.2.1+stable-20121212-a99a802e-0ubuntu1.2~cloud0
  Current version of CEPH - 0.56.3-42-ga30903c-1precise
  Consistency of error - 100%
  No. of Openstack instances tested on - 2

  Steps to reproduce:
  Create a VM via the horizon dashboard
  Create a volume with cinder - any size (10gb used for testing)
  Attach the volume to the Created VM
  format/use the volume and then unmount
  detach the volume from the created VM
  click on create snapshot , name and click on create snapshot
  The Snapshot will remain in the queued state.
  Vm reboot actions fail to start the VM - it is shutdown and offline

  The Vm is shutdown and cant be started. If you attempt to start the VM via virsh you get this error message
  virsh start instance-0000001a
  error: Failed to start domain instance-0000001a
  error: internal error process exited while connecting to monitor: char device redirected to /dev/pts/5
  Unknown savevm section or instance '0000:00:06.0/virtio-blk' 0
  load of migration failed

  the only way to restore the Vm to service is to dump the saved state from virsh as follows:
  virsh managedsave-remove NameOfDomain

  The Vm will the able to be restarted

  workarounds identified:
  Not a real workaround but brings the Vm back online:
  ID the host running the VM, log onto that host
  virsh managedsave-remove <NameOfDomain>  to dump the failed suspend state data
  virsh start <NameOfDomain> OR nova reboot <Name of VM>
  Vm should then restart

  error messages produced:

  Nova-compute.log

  2013-03-20 13:22:29 AUDIT nova.compute.manager [req-4f9d1cc7-c476-4add-9cc2-e58801d752c1 e560e0362abe492db88c655dc043ce4f cbdfbd6c58d444fcb858e77fc7135f04] [instance: 7a0df355-6094-4d2d-8ca9-f2453222
  76e5] instance snapshotting
  2013-03-20 13:23:06 1939 ERROR nova.openstack.common.rpc.amqp [-] Exception during message handling
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp Traceback (most recent call last):
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp   File "/usr/lib/python2.7/dist-packages/nova/openstack/common/rpc/amqp.py", line 276, in _process_data
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp     rval = self.proxy.dispatch(ctxt, version, method, **args)
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp   File "/usr/lib/python2.7/dist-packages/nova/openstack/common/rpc/dispatcher.py", line 145, in dispatch
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp     return getattr(proxyobj, method)(ctxt, **kwargs)
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp   File "/usr/lib/python2.7/dist-packages/nova/exception.py", line 117, in wrapped
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp     temp_level, payload)
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp   File "/usr/lib/python2.7/contextlib.py", line 24, in __exit__
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp     self.gen.next()
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp   File "/usr/lib/python2.7/dist-packages/nova/exception.py", line 92, in wrapped
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp     return f(*args, **kw)
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp   File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 175, in decorated_function
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp     pass
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp   File "/usr/lib/python2.7/contextlib.py", line 24, in __exit__
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp     self.gen.next()
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp   File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 161, in decorated_function
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp     return function(self, context, *args, **kwargs)
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp   File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 196, in decorated_function
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp     kwargs['instance']['uuid'], e, sys.exc_info())
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp   File "/usr/lib/python2.7/contextlib.py", line 24, in __exit__
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp     self.gen.next()
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp   File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 190, in decorated_function
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp     return function(self, context, *args, **kwargs)
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp   File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 1174, in snapshot_instance
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp     self.driver.snapshot(context, instance, image_id)
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp   File "/usr/lib/python2.7/dist-packages/nova/exception.py", line 117, in wrapped
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp     temp_level, payload)
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp   File "/usr/lib/python2.7/contextlib.py", line 24, in __exit__
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp     self.gen.next()
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp   File "/usr/lib/python2.7/dist-packages/nova/exception.py", line 92, in wrapped
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp     return f(*args, **kw)
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp   File "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/driver.py", line 875, in snapshot
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp     self._create_domain(domain=virt_dom)
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp   File "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/driver.py", line 1910, in _create_domain
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp     domain.createWithFlags(launch_flags)
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp   File "/usr/lib/python2.7/dist-packages/eventlet/tpool.py", line 187, in doit
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp     result = proxy_call(self._autowrap, f, *args, **kwargs)
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp   File "/usr/lib/python2.7/dist-packages/eventlet/tpool.py", line 147, in proxy_call
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp     rv = execute(f,*args,**kwargs)
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp   File "/usr/lib/python2.7/dist-packages/eventlet/tpool.py", line 76, in tworker
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp     rv = meth(*args,**kwargs)
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp   File "/usr/lib/python2.7/dist-packages/libvirt.py", line 650, in createWithFlags
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp     if ret == -1: raise libvirtError ('virDomainCreateWithFlags() failed', dom=self)
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp libvirtError: internal error process exited while connecting to monitor: char device redirected to /dev/pts/5
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp Unknown savevm section or instance '0000:00:06.0/virtio-blk' 0
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp load of migration failed
  2013-03-20 13:23:06 1939 TRACE nova.openstack.common.rpc.amqp 2013-03-20 13:23:03.154+0000: shutting down

  Libvirt instance log

  2013-03-20 13:23:00.091+0000: starting up
  LC_ALL=C PATH=/usr/local/sbin:/usr/local/bin:/usr/bin:/usr/sbin:/sbin:/bin QEMU_AUDIO_DRV=none /usr/bin/kvm -name instance-0000001a -S -M pc-1.0 -cpu core2duo,+lahf_lm,+rdtscp,+pdpe1gb,+popcnt,+sse4.2,+sse4.1,+dca,+pdcm,+xtpr,+cx16,+tm2,+est,+smx,+vmx,+ds_cpl,+dtes64,+pbe,+tm,+ht,+ss,+acpi,+ds -enable-kvm -m 2048 -smp 2,sockets=2,cores=1,threads=1 -uuid 7a0df355-6094-4d2d-8ca9-f245322276e5 -nodefconfig -nodefaults -chardev socket,id=charmonitor,path=/var/lib/libvirt/qemu/instance-0000001a.monitor,server,nowait -mon chardev=charmonitor,id=monitor,mode=control -rtc base=utc,driftfix=slew -no-kvm-pit-reinjection -no-shutdown -device piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -drive file=/var/lib/nova/instances/instance-0000001a/disk,if=none,id=drive-virtio-disk0,format=qcow2,cache=none -device virtio-blk-pci,scsi=off,bus=pci.0,addr=0x4,drive=drive-virtio-disk0,id=virtio-disk0,bootindex=1 -netdev tap,fd=28,id=hostnet0 -device virtio-net-pci,netdev=hostnet0,id=net0,mac=fa:16:3e:48:bc:cc,bus=pci.0,addr=0x3 -chardev file,id=charserial0,path=/var/lib/nova/instances/instance-0000001a/console.log -device isa-serial,chardev=charserial0,id=serial0 -chardev pty,id=charserial1 -device isa-serial,chardev=charserial1,id=serial1 -device usb-tablet,id=input0 -vnc 172.28.107.199:3 -k en-us -vga cirrus -incoming fd:25 -device virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x5
  char device redirected to /dev/pts/5
  Unknown savevm section or instance '0000:00:06.0/virtio-blk' 0
  load of migration failed
  2013-03-20 13:23:03.154+0000: shutting down

  
  Any help or guidance on how to resolve or fully work around this error would be really appreciated

  thanks

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1157725/+subscriptions