mirror of
https://github.com/TECHNOFAB11/zfs-localpv.git
synced 2025-12-12 14:30:12 +01:00
feat(zfspv): adding backup and restore support (#162)
This commit adds support for Backup and Restore controller, which will be watching for
the events. The velero plugin will create a Backup CR to create a backup
with the remote location information, the controller will send the data
to that remote location.
In the same way, the velero plugin will create a Restore CR to restore the
volume from the the remote location and the restore controller will restore
the data.
Steps to use velero plugin for ZFS-LocalPV are :
1. install velero
2. add openebs plugin
velero plugin add openebs/velero-plugin:latest
3. Create the volumesnapshot location :
for full backup :-
```yaml
apiVersion: velero.io/v1
kind: VolumeSnapshotLocation
metadata:
name: default
namespace: velero
spec:
provider: openebs.io/zfspv-blockstore
config:
bucket: velero
prefix: zfs
namespace: openebs
provider: aws
region: minio
s3ForcePathStyle: "true"
s3Url: http://minio.velero.svc:9000
```
for incremental backup :-
```yaml
apiVersion: velero.io/v1
kind: VolumeSnapshotLocation
metadata:
name: default
namespace: velero
spec:
provider: openebs.io/zfspv-blockstore
config:
bucket: velero
prefix: zfs
backup: incremental
namespace: openebs
provider: aws
region: minio
s3ForcePathStyle: "true"
s3Url: http://minio.velero.svc:9000
```
4. Create backup
velero backup create my-backup --snapshot-volumes --include-namespaces=velero-ns --volume-snapshot-locations=aws-cloud-default --storage-location=default
5. Create Schedule
velero create schedule newschedule --schedule="*/1 * * * *" --snapshot-volumes --include-namespaces=velero-ns --volume-snapshot-locations=aws-local-default --storage-location=default
6. Restore from backup
velero restore create --from-backup my-backup --restore-volumes=true --namespace-mappings velero-ns:ns1
Signed-off-by: Pawan <pawan@mayadata.io>
This commit is contained in:
parent
a5e645b43d
commit
e40026c98a
48 changed files with 5148 additions and 7 deletions
245
pkg/mgmt/restore/restore.go
Normal file
245
pkg/mgmt/restore/restore.go
Normal file
|
|
@ -0,0 +1,245 @@
|
|||
/*
|
||||
Copyright 2020 The OpenEBS Authors
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License.
|
||||
*/
|
||||
|
||||
package restore
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"time"
|
||||
|
||||
"k8s.io/klog"
|
||||
|
||||
apis "github.com/openebs/zfs-localpv/pkg/apis/openebs.io/zfs/v1"
|
||||
zfs "github.com/openebs/zfs-localpv/pkg/zfs"
|
||||
k8serror "k8s.io/apimachinery/pkg/api/errors"
|
||||
"k8s.io/apimachinery/pkg/util/runtime"
|
||||
"k8s.io/apimachinery/pkg/util/wait"
|
||||
"k8s.io/client-go/tools/cache"
|
||||
)
|
||||
|
||||
// isDeletionCandidate checks if a zfs backup is a deletion candidate.
|
||||
func (c *RstrController) isDeletionCandidate(rstr *apis.ZFSRestore) bool {
|
||||
return rstr.ObjectMeta.DeletionTimestamp != nil
|
||||
}
|
||||
|
||||
// syncHandler compares the actual state with the desired, and attempts to
|
||||
// converge the two.
|
||||
func (c *RstrController) syncHandler(key string) error {
|
||||
// Convert the namespace/name string into a distinct namespace and name
|
||||
namespace, name, err := cache.SplitMetaNamespaceKey(key)
|
||||
if err != nil {
|
||||
runtime.HandleError(fmt.Errorf("invalid resource key: %s", key))
|
||||
return nil
|
||||
}
|
||||
|
||||
// Get the rstr resource with this namespace/name
|
||||
rstr, err := c.rstrLister.ZFSRestores(namespace).Get(name)
|
||||
if k8serror.IsNotFound(err) {
|
||||
runtime.HandleError(fmt.Errorf("zfs restore '%s' has been deleted", key))
|
||||
return nil
|
||||
}
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
rstrCopy := rstr.DeepCopy()
|
||||
err = c.syncRestore(rstrCopy)
|
||||
return err
|
||||
}
|
||||
|
||||
// enqueueRestore takes a ZFSRestore resource and converts it into a namespace/name
|
||||
// string which is then put onto the work queue. This method should *not* be
|
||||
// passed resources of any type other than ZFSRestore.
|
||||
func (c *RstrController) enqueueRestore(obj interface{}) {
|
||||
var key string
|
||||
var err error
|
||||
if key, err = cache.MetaNamespaceKeyFunc(obj); err != nil {
|
||||
runtime.HandleError(err)
|
||||
return
|
||||
}
|
||||
c.workqueue.Add(key)
|
||||
}
|
||||
|
||||
// synRestore is the function which tries to converge to a desired state for the
|
||||
// ZFSRestore
|
||||
func (c *RstrController) syncRestore(rstr *apis.ZFSRestore) error {
|
||||
var err error = nil
|
||||
// ZFSRestore should not be deleted. Check if deletion timestamp is set
|
||||
if !c.isDeletionCandidate(rstr) {
|
||||
// if status is Init, then only do the restore
|
||||
if rstr.Status == apis.RSTZFSStatusInit {
|
||||
err = zfs.CreateRestore(rstr)
|
||||
if err == nil {
|
||||
klog.Infof("restore %s done %s", rstr.Name, rstr.Spec.VolumeName)
|
||||
err = zfs.UpdateRestoreInfo(rstr, apis.RSTZFSStatusDone)
|
||||
} else {
|
||||
klog.Errorf("restore %s failed %s err %v", rstr.Name, rstr.Spec.VolumeName, err)
|
||||
err = zfs.UpdateRestoreInfo(rstr, apis.RSTZFSStatusFailed)
|
||||
}
|
||||
}
|
||||
}
|
||||
return err
|
||||
}
|
||||
|
||||
// addRestore is the add event handler for ZFSRestore
|
||||
func (c *RstrController) addRestore(obj interface{}) {
|
||||
rstr, ok := obj.(*apis.ZFSRestore)
|
||||
if !ok {
|
||||
runtime.HandleError(fmt.Errorf("Couldn't get rstr object %#v", obj))
|
||||
return
|
||||
}
|
||||
|
||||
if zfs.NodeID != rstr.Spec.OwnerNodeID {
|
||||
return
|
||||
}
|
||||
klog.Infof("Got add event for Restore %s vol %s", rstr.Name, rstr.Spec.VolumeName)
|
||||
c.enqueueRestore(rstr)
|
||||
}
|
||||
|
||||
// updateRestore is the update event handler for ZFSRestore
|
||||
func (c *RstrController) updateRestore(oldObj, newObj interface{}) {
|
||||
|
||||
newRstr, ok := newObj.(*apis.ZFSRestore)
|
||||
if !ok {
|
||||
runtime.HandleError(fmt.Errorf("Couldn't get rstr object %#v", newRstr))
|
||||
return
|
||||
}
|
||||
|
||||
if zfs.NodeID != newRstr.Spec.OwnerNodeID {
|
||||
return
|
||||
}
|
||||
|
||||
if c.isDeletionCandidate(newRstr) {
|
||||
klog.Infof("Got update event for Restore %s vol %s", newRstr.Name, newRstr.Spec.VolumeName)
|
||||
c.enqueueRestore(newRstr)
|
||||
}
|
||||
}
|
||||
|
||||
// deleteRestore is the delete event handler for ZFSRestore
|
||||
func (c *RstrController) deleteRestore(obj interface{}) {
|
||||
rstr, ok := obj.(*apis.ZFSRestore)
|
||||
if !ok {
|
||||
tombstone, ok := obj.(cache.DeletedFinalStateUnknown)
|
||||
if !ok {
|
||||
runtime.HandleError(fmt.Errorf("Couldn't get object from tombstone %#v", obj))
|
||||
return
|
||||
}
|
||||
rstr, ok = tombstone.Obj.(*apis.ZFSRestore)
|
||||
if !ok {
|
||||
runtime.HandleError(fmt.Errorf("Tombstone contained object that is not a zfsbackup %#v", obj))
|
||||
return
|
||||
}
|
||||
}
|
||||
|
||||
if zfs.NodeID != rstr.Spec.OwnerNodeID {
|
||||
return
|
||||
}
|
||||
|
||||
klog.Infof("Got delete event for Restore %s", rstr.Spec.VolumeName)
|
||||
c.enqueueRestore(rstr)
|
||||
}
|
||||
|
||||
// Run will set up the event handlers for types we are interested in, as well
|
||||
// as syncing informer caches and starting workers. It will block until stopCh
|
||||
// is closed, at which point it will shutdown the workqueue and wait for
|
||||
// workers to finish processing their current work items.
|
||||
func (c *RstrController) Run(threadiness int, stopCh <-chan struct{}) error {
|
||||
defer runtime.HandleCrash()
|
||||
defer c.workqueue.ShutDown()
|
||||
|
||||
// Start the informer factories to begin populating the informer caches
|
||||
klog.Info("Starting Restore controller")
|
||||
|
||||
// Wait for the k8s caches to be synced before starting workers
|
||||
klog.Info("Waiting for informer caches to sync")
|
||||
if ok := cache.WaitForCacheSync(stopCh, c.rstrSynced); !ok {
|
||||
return fmt.Errorf("failed to wait for caches to sync")
|
||||
}
|
||||
klog.Info("Starting Restore workers")
|
||||
// Launch worker to process Restore resources
|
||||
// Threadiness will decide the number of workers you want to launch to process work items from queue
|
||||
for i := 0; i < threadiness; i++ {
|
||||
go wait.Until(c.runWorker, time.Second, stopCh)
|
||||
}
|
||||
|
||||
klog.Info("Started Restore workers")
|
||||
<-stopCh
|
||||
klog.Info("Shutting down Restore workers")
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
// runWorker is a long-running function that will continually call the
|
||||
// processNextWorkItem function in order to read and process a message on the
|
||||
// workqueue.
|
||||
func (c *RstrController) runWorker() {
|
||||
for c.processNextWorkItem() {
|
||||
}
|
||||
}
|
||||
|
||||
// processNextWorkItem will read a single work item off the workqueue and
|
||||
// attempt to process it, by calling the syncHandler.
|
||||
func (c *RstrController) processNextWorkItem() bool {
|
||||
obj, shutdown := c.workqueue.Get()
|
||||
|
||||
if shutdown {
|
||||
return false
|
||||
}
|
||||
|
||||
// We wrap this block in a func so we can defer c.workqueue.Done.
|
||||
err := func(obj interface{}) error {
|
||||
// We call Done here so the workqueue knows we have finished
|
||||
// processing this item. We also must remember to call Forget if we
|
||||
// do not want this work item being re-queued. For example, we do
|
||||
// not call Forget if a transient error occurs, instead the item is
|
||||
// put back on the workqueue and attempted again after a back-off
|
||||
// period.
|
||||
defer c.workqueue.Done(obj)
|
||||
var key string
|
||||
var ok bool
|
||||
// We expect strings to come off the workqueue. These are of the
|
||||
// form namespace/name. We do this as the delayed nature of the
|
||||
// workqueue means the items in the informer cache may actually be
|
||||
// more up to date that when the item was initially put onto the
|
||||
// workqueue.
|
||||
if key, ok = obj.(string); !ok {
|
||||
// As the item in the workqueue is actually invalid, we call
|
||||
// Forget here else we'd go into a loop of attempting to
|
||||
// process a work item that is invalid.
|
||||
c.workqueue.Forget(obj)
|
||||
runtime.HandleError(fmt.Errorf("expected string in workqueue but got %#v", obj))
|
||||
return nil
|
||||
}
|
||||
// Run the syncHandler, passing it the namespace/name string of the
|
||||
// Restore resource to be synced.
|
||||
if err := c.syncHandler(key); err != nil {
|
||||
// Put the item back on the workqueue to handle any transient errors.
|
||||
c.workqueue.AddRateLimited(key)
|
||||
return fmt.Errorf("error syncing '%s': %s, requeuing", key, err.Error())
|
||||
}
|
||||
// Finally, if no error occurs we Forget this item so it does not
|
||||
// get queued again until another change happens.
|
||||
c.workqueue.Forget(obj)
|
||||
klog.Infof("Successfully synced '%s'", key)
|
||||
return nil
|
||||
}(obj)
|
||||
|
||||
if err != nil {
|
||||
runtime.HandleError(err)
|
||||
return true
|
||||
}
|
||||
|
||||
return true
|
||||
}
|
||||
Loading…
Add table
Add a link
Reference in a new issue