yahoo-eng-team team mailing list archive
-
yahoo-eng-team team
-
Mailing list archive
-
Message #88948
[Bug 1975637] [NEW] Ceph VM images leak on instance deletion if there are snapshots of that image
Public bug reported:
Description
===========
I'm using backy2 to back up instance images. For the sake of incremental backups, we keep a pending last-backed-up snapshot associated with each instance at all times.
When an instance is deleted, the rbd delete call fails silently and
leaves both the VM and snapshot behind forever.
Ideally two things would be different:
1) The logs would reflect this failure
2) A config option would allow me to demand that all associated snaps are purged on instance deletion.
I thought I had a weirdo edge use-case but I see at least one other user
encountering this same leakage, here:
https://heiterbiswolkig.blogs.nde.ag/2019/03/07/orphaned-instances-
part-2/
(I'm running version Wallaby but the code is the same on the current git
head)
Steps to reproduce
==================
* Create an instance backed with ceph/rbd
* Take a snapshot of that instance (in my case, I'm doing this out of band using rbd commands, not a nova API)
* Delete the instance
* Note that the volume is still present on the storage backend
Expected result
* Log messages should announce a failure to delete, or
* (optionally) server volume is actually deleted
Actual result
=============
* Logfile is silent about failure to delete
* server volume is leaked and lives on forever, invisible to Nova
Environment
===========
Openstack Wallaby installed from the debian BPO on Debian Bullseye
# dpkg --list | grep nova
ii nova-common 2:23.1.0-2~bpo11+1 all OpenStack Compute - common files
ii nova-compute 2:23.1.0-2~bpo11+1 all OpenStack Compute - compute node
ii nova-compute-kvm 2:23.1.0-2~bpo11+1 all OpenStack Compute - compute node (KVM)
ii python3-nova 2:23.1.0-2~bpo11+1 all OpenStack Compute - libraries
ii python3-novaclient 2:17.4.0-1~bpo11+1 all client library for OpenStack Compute API - 3.x
The same issue is present in the latest git code.
** Affects: nova
Importance: Undecided
Status: New
--
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1975637
Title:
Ceph VM images leak on instance deletion if there are snapshots of
that image
Status in OpenStack Compute (nova):
New
Bug description:
Description
===========
I'm using backy2 to back up instance images. For the sake of incremental backups, we keep a pending last-backed-up snapshot associated with each instance at all times.
When an instance is deleted, the rbd delete call fails silently and
leaves both the VM and snapshot behind forever.
Ideally two things would be different:
1) The logs would reflect this failure
2) A config option would allow me to demand that all associated snaps are purged on instance deletion.
I thought I had a weirdo edge use-case but I see at least one other
user encountering this same leakage, here:
https://heiterbiswolkig.blogs.nde.ag/2019/03/07/orphaned-instances-
part-2/
(I'm running version Wallaby but the code is the same on the current
git head)
Steps to reproduce
==================
* Create an instance backed with ceph/rbd
* Take a snapshot of that instance (in my case, I'm doing this out of band using rbd commands, not a nova API)
* Delete the instance
* Note that the volume is still present on the storage backend
Expected result
* Log messages should announce a failure to delete, or
* (optionally) server volume is actually deleted
Actual result
=============
* Logfile is silent about failure to delete
* server volume is leaked and lives on forever, invisible to Nova
Environment
===========
Openstack Wallaby installed from the debian BPO on Debian Bullseye
# dpkg --list | grep nova
ii nova-common 2:23.1.0-2~bpo11+1 all OpenStack Compute - common files
ii nova-compute 2:23.1.0-2~bpo11+1 all OpenStack Compute - compute node
ii nova-compute-kvm 2:23.1.0-2~bpo11+1 all OpenStack Compute - compute node (KVM)
ii python3-nova 2:23.1.0-2~bpo11+1 all OpenStack Compute - libraries
ii python3-novaclient 2:17.4.0-1~bpo11+1 all client library for OpenStack Compute API - 3.x
The same issue is present in the latest git code.
To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1975637/+subscriptions