← Back to team overview

yahoo-eng-team team mailing list archive

[Bug 1669503] Re: [Libvirt] VM does not boot in Ocata (libvirt.xml missing, wrong owner for console.log)

 

I have found a solution. Adding the following configuration to my nova-
compute.cong solved the issue:

[libvirt]
hw_machine_type = "x86_64=pc-i440fx-xenial,i686=pc-i440fx-xenial"


** Changed in: nova
       Status: Confirmed => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1669503

Title:
  [Libvirt] VM does not boot in Ocata (libvirt.xml missing, wrong owner
  for console.log)

Status in OpenStack Compute (nova):
  Invalid

Bug description:
  I am setting up OpenStack Ocata as HA deployment on Ubuntu 16.04. and
  I am confronted with a heavy blocker. I have set up Keystone, Glance,
  Nova, Neutron and Horizon. I am able to create and spawn a VM, but it
  does not boot.

  The block device is created and available in my Ceph cluster. Libvirt gives me only:
  Mar  2 17:09:11 os-compute01 virtlogd[3218]: End of file while reading data: Input/output error

  The instances folder located under "/var/lib/nova/instances" contains
  a folder for the VM, but it contains only the empty file "console.log"
  and NO "libvirt.xml". Another strange thing is that "console.log" is
  owned by root:root. Under Newton the file is owned by libvirt-
  qemu:kvm.

  I have tried to correct the ownership manaully and copied the VM
  specification file from "/etc/libvirt/qemu/instance-*.xml" to
  "/var/lib/nova/instancs/[VM-UUID]/libvirt.xml". After rebooting the VM
  the console.log is owned by root:root again and the VM does not boot.

  There seems to be problem in the current Ocata release of Nova.

  Any hints how to workaround or fix the problem?

  This is my nova.conf on os-compute01:
  [DEFAULT]
  compute_monitors = cpu.virt_driver
  debug = true
  firewall_driver = nova.virt.firewall.NoopFirewallDriver
  host = os-compute01
  instance_usage_audit = true
  instance_usage_audit_period = hour
  notify_on_state_change = vm_and_task_state
  instances_path = $state_path/instances
  linuxnet_interface_driver = nova.network.linux_net.LinuxOVSInterfaceDriver
  log_dir = /var/log/nova
  memcached_servers = os-memcache:11211
  my_ip = 10.30.200.111
  notification_driver = messagingv2
  resume_guests_state_on_host_boot = true
  state_path = /var/lib/nova
  transport_url = rabbit://nova:SECRET@os-rabbit01:5672,nova:SECRET@os-rabbit02:5672/openstack
  use_neutron = true

  [api]
  auth_strategy = keystone

  [cache]
  enabled = true
  backend = oslo_cache.memcache_pool
  memcache_servers = os-memcache:11211

  [cinder]
  catalog_info = volumev2:cinderv2:internalURL

  [conductor]
  use_local = false

  [glance]
  api_servers = http://os-image:9292

  [keystone_authtoken]
  auth_type = password
  auth_uri = http://os-identity:5000
  auth_url = http://os-identity:35357
  memcached_servers = os-memcache:11211
  password = SECRET
  project_domain_name = default
  project_name = service
  service_token_roles_required = true
  user_domain_name = default
  username = nova

  [neutron]
  auth_type = password
  auth_uri = http://os-identity:5000
  auth_url = http://os-identity:35357
  password = SECRET
  project_domain_name = default
  project_name = service
  region_name = RegionOne
  url = http://os-network:9696
  user_domain_name = default
  username = neutron

  [oslo_concurrency]
  lock_path = /var/lock/nova

  [oslo_messaging_rabbit]
  amqp_durable_queues = true
  rabbit_ha_queues = true
  rabbit_retry_backoff = 2
  rabbit_retry_interval = 1

  [placement]
  auth_type = password
  auth_uri = http://os-identity:5000
  auth_url = http://os-identity:35357
  username = placement
  password = SECRET
  user_domain_name = default
  project_name = service
  project_domain_name = default
  os_interface = internal
  os_region_name = RegionOne

  [vnc]
  enabled = true
  novncproxy_base_url = https://os-cloud.materna.com:6080/vnc_auto.html
  vncserver_listen = 0.0.0.0
  vncserver_proxyclient_address = $my_ip

  and my nova-compute.conf:
  [DEFAULT]
  compute_driver = libvirt.LibvirtDriver

  [libvirt]
  cpu_mode = custom
  cpu_model = SandyBridge
  disk_cachemodes="network=writeback"
  hw_disk_discard = unmap
  images_rbd_ceph_conf = /etc/ceph/ceph.conf
  images_rbd_pool = vms
  images_type = rbd
  inject_key = false
  inject_partition = -2
  inject_password = false
  rbd_secret_uuid = SECRET
  rbd_user = cinder
  virt_type = kvm

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1669503/+subscriptions


References