← Back to team overview

yahoo-eng-team team mailing list archive

[Bug 1975637] [NEW] Ceph VM images leak on instance deletion if there are snapshots of that image

 

Public bug reported:

Description
===========


I'm using backy2 to back up instance images. For the sake of incremental backups, we keep a pending last-backed-up snapshot associated with each instance at all times.

When an instance is deleted, the rbd delete call fails silently and
leaves both the VM and snapshot behind forever.

Ideally two things would be different:

1) The logs would reflect this failure
2) A config option would allow me to demand that all associated snaps are purged on instance deletion.

I thought I had a weirdo edge use-case but I see at least one other user
encountering this same leakage, here:
https://heiterbiswolkig.blogs.nde.ag/2019/03/07/orphaned-instances-
part-2/

(I'm running version Wallaby but the code is the same on the current git
head)

Steps to reproduce
==================

* Create an instance backed with ceph/rbd
* Take a snapshot of that instance (in my case, I'm doing this out of band using rbd commands, not a nova API)
* Delete the instance
* Note that the volume is still present on the storage backend

   
Expected result

* Log messages should announce a failure to delete, or
* (optionally) server volume is actually deleted
   
Actual result
=============

* Logfile is silent about failure to delete
* server volume is leaked and lives on forever, invisible to Nova

Environment
===========

Openstack Wallaby installed from the debian BPO on Debian Bullseye

# dpkg --list | grep nova
ii  nova-common                          2:23.1.0-2~bpo11+1                        all          OpenStack Compute - common files
ii  nova-compute                         2:23.1.0-2~bpo11+1                        all          OpenStack Compute - compute node
ii  nova-compute-kvm                     2:23.1.0-2~bpo11+1                        all          OpenStack Compute - compute node (KVM)
ii  python3-nova                         2:23.1.0-2~bpo11+1                        all          OpenStack Compute - libraries
ii  python3-novaclient                   2:17.4.0-1~bpo11+1                        all          client library for OpenStack Compute API - 3.x

The same issue is present in the latest git code.

** Affects: nova
     Importance: Undecided
         Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1975637

Title:
  Ceph VM images leak on instance deletion if there are snapshots of
  that image

Status in OpenStack Compute (nova):
  New

Bug description:
  Description
  ===========

  
  I'm using backy2 to back up instance images. For the sake of incremental backups, we keep a pending last-backed-up snapshot associated with each instance at all times.

  When an instance is deleted, the rbd delete call fails silently and
  leaves both the VM and snapshot behind forever.

  Ideally two things would be different:

  1) The logs would reflect this failure
  2) A config option would allow me to demand that all associated snaps are purged on instance deletion.

  I thought I had a weirdo edge use-case but I see at least one other
  user encountering this same leakage, here:
  https://heiterbiswolkig.blogs.nde.ag/2019/03/07/orphaned-instances-
  part-2/

  (I'm running version Wallaby but the code is the same on the current
  git head)

  Steps to reproduce
  ==================

  * Create an instance backed with ceph/rbd
  * Take a snapshot of that instance (in my case, I'm doing this out of band using rbd commands, not a nova API)
  * Delete the instance
  * Note that the volume is still present on the storage backend

     
  Expected result

  * Log messages should announce a failure to delete, or
  * (optionally) server volume is actually deleted
     
  Actual result
  =============

  * Logfile is silent about failure to delete
  * server volume is leaked and lives on forever, invisible to Nova

  Environment
  ===========

  Openstack Wallaby installed from the debian BPO on Debian Bullseye

  # dpkg --list | grep nova
  ii  nova-common                          2:23.1.0-2~bpo11+1                        all          OpenStack Compute - common files
  ii  nova-compute                         2:23.1.0-2~bpo11+1                        all          OpenStack Compute - compute node
  ii  nova-compute-kvm                     2:23.1.0-2~bpo11+1                        all          OpenStack Compute - compute node (KVM)
  ii  python3-nova                         2:23.1.0-2~bpo11+1                        all          OpenStack Compute - libraries
  ii  python3-novaclient                   2:17.4.0-1~bpo11+1                        all          client library for OpenStack Compute API - 3.x

  The same issue is present in the latest git code.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1975637/+subscriptions