Hello, I just release script for backing up QCOW2 datastores with live snapshot support.
Backup live VMs with snapshots (--quiesce --atomic)
Feature Request : VM Backup & Disk resize
Hi all, addon was updated with new features and fixes.
Actual version 1.2.0
Changes since first version:
- Live snapshotting added fallback in case that FS freeze fails
- Fix backup persistent used images if VM is not in ACTIVE state
- new option -k --insecure for use rsync with weakest but fastest SSH encryption
- other rsync options adjustments to improve speed
- new option -n --netcat for use netcat instead of rsync to transfer main image (*.snap dirs still use rsync)
- extended -i --image option to support multiple image IDs separated by comma
thanks a lot for this contribution. I believe this is a very important piece for the OpenNebula project.
I did some tests and found a problem with this addon in my setup.
In general I followed the standard installation instructions to install OpenNebula 5.4.10 on CentOS 7. The only difference is that I installed the Enterprise version of qemu-kvm (Package qemu-kvm-ev) in order to have access to some features not present in the standard package.
Besides this everything is according standard instructions.
The point is that in this setup libvirt is not enabled for tcp connections (I get a connection refused when your script tries to create the snapshot by running virsh). And to enable libvirt for tcp with proper authentication adds some complexity to my setup.
I understand that the virsh command is anyhow executed locally on the remote kvm node. Your are using
virsh -c qemu+tcp://localhost/system … for snapshot and blockcommit.
And I found out that
virsh -c qemu:///system … works in my setup without any other changes in my setup.
Do you see a possibility to change the virsh command in one-image-backup.js to virsh -c qemu:///system?
Thanks and Best Regards
EDIT: Of course, I can add it to config.
Ok, thanks. That link to the Tuning Guide makes it very clear, although I have some security related doubts with this setup (non-authenticated and non-encrypted). Network isolation should be bullet proof. Just my opinion…
Anyway I would appreciate it if you can add it to the config. Keep up the good work!
Hi all, I just released version 1.4.1.
Changes since version 1.2.0
- new config option libvirtUri to configure custom hypervisor connection URI
- new option -S --start-image <image_id> image id to start backup from
- rsync use inplace option because we use custom tmp files
- don’t update images backup info in opennebula when dry run is used
- removed netcat option from backup.sh script due to experimental nature of that feature
- backup main images to tmp file and after copy replace original one
thanks a lot for the quick update. I may have another feature request, if I may ask.
Similar to the -i option I would love to see a --datastore option to include all images from the datastores in a comma separated listed.
This is similiar to the way we handle “protected” and “unprotected” datastores in our vSphere environment, because we do not want to backup all images. The placement of vm-image decides if it is backuped or not.
Hello, I pushed updated version to develop branch. Please test it. If all goes ok, them I’ll publish new release.
Yesterday I tested the develop branch with the -a (datastores) option and your addon backuped all images from the specified datastores. Brilliant!
One last idea came to my mind to further enhance the addon:
Backup all datastores and images with a specific label (e.g. ‘protected’). This logic would be simple and similar to your ansible inventory script and would give users much more control over the backup in sunstone.
But for now I am already very happy with the datastores option. Thanks Kristian!
New release 1.6.1!
Changes since version 1.4.1:
- new option
-a --datastoreto backup from specific datastore(s)
- new option
-l --labelto backup image(s) and/or datastore(s) by specific label(s)
- code enhancements
I can confirm both options are working in our environment (also with nested labes like “images/protected”).
Hi @feldsam ,
I just want to ensure myself that I understand how your addon works. I wonder how your script will backup my VM images. I have image type DATABLOCK with “Persistent” setting set to “no”, this image is used by VMs as OS/DATA disk.
VM disk image looks like this :
qemu-img info 0
file format: qcow2
virtual size: 10G (10737418240 bytes)
disk size: 9.7G
backing file: /var/lib/one/datastores/103/cdfd110a37e569165315b96f913293a2
Format specific information:
lazy refcounts: false
refcount bits: 16
Can you explain please if VM will be snapshoted, than backed up and block commited, or not ?
Hello, in OpenNebula there are two types of images, as you already know:
You can deploy many VMs using same non-persistent image. On the background, oned creates new qcow2 image which have backing image of that non-persistent one. That new image is deployed to system datastore.
When you terminate VM instance, that image deployed to system datastore will be deleted. So from nature of this functionality, VMs deployed from non-persistent images have character of instances of base image and should not persist data and should not be backuped. This is ideal for applications, which can run several times at once for horizontal scaling. These instances usually connects to same shared database server and/or shared file server (nfs) to persist data. When VM instance need to persist data, them there should be attached persistent datablock or VM should be instatiated as persistent.
Persistent image can be attached to only one VM and changes made in VM instance persists after undeploys/terminates. This images are backuped and also, if attached to running VM, live snapshoted.
This is kind of pitty because I use these non-persistent images for all VMs. In my environment after undeploy action machine goes to shutdown and than undeployed state (it free up resources on node) but disks or machine image files stays on hypervisor datastore untouched, so it could be backed-up. (this behaviour is not the same as you described - dunno why) I use GlusterFS (via libgfapi) to access disk images. Termination state deletes VM and related files of course.
So to use your script I have to make all disks of all VMs persistent. (now images are presented as non-persistent images backed by image from image datastore). Do you think this is good approach or I should rewrite your script in the way it will backup also non-persistent images with same logic as persistent ?
Thanks for any suggestions in advance.
Hmm, also name “non-persistent” have character of “temporary” disk, where data doesn’t matter. Base images are of course backuped, but not non-persistent part of attached image.
I think, that every VM, which have data, that needs to be persistent, have to use persistent images. On instantiate tab, you can select “instantiate as persistent”, or you can already instatiated non-persistent VMs “save as” and make them persistent. When you need terminate VM, them you should delete also VM Template with All disks, so you free up space.
Of cource, I can add new option to backup also non persistent VMs, but from nature of non-persistent vs persistent feature, you should decide rework your setup.
I will do some research and let you know, because probably this is related to GlusterFS and OpenNebula driver for Gluster itself. This is probably why my non-persistent VMs and their images stayed untouched after undeploy operation.
Anyway thanks for help for now
EDIT: Also non-persistent images as I am using them saving space on storage because base system is not copied multiple times for every VM, it can also be “super cached” to some hot-tier cache because multiple VMs using same file to work (base image)
From the official documentation (https://docs.opennebula.org/5.4/operation/vm_management/vm_instances.html):
“Undeploy -> The VM is shut down. The VM disks are transfered to the system datastore. The VM can be resumed later.”
I understand that undeployed VMs still hold their “non-persistent” data. But as soon as you terminate a VM with non-persistent disks all changed data (to the base image) is lost.
Undeploy and Terminate are two separate actions. The transaction of the states is as follow
Undeploy: --> [epilog] --> [undeployed]
Terminate: --> [epilog] --> [done]