Collectd Global CRD
[demo.git] / vnfs / DAaaS / microservices / collectd-operator / pkg / controller / collectdglobal / collectdglobal_controller.go
diff --git a/vnfs/DAaaS/microservices/collectd-operator/pkg/controller/collectdglobal/collectdglobal_controller.go b/vnfs/DAaaS/microservices/collectd-operator/pkg/controller/collectdglobal/collectdglobal_controller.go
new file mode 100644 (file)
index 0000000..701580b
--- /dev/null
@@ -0,0 +1,324 @@
+package collectdglobal
+
+import (
+       "context"
+       "fmt"
+       //"path/filepath"
+       "reflect"
+       "strings"
+
+       "github.com/go-logr/logr"
+       "github.com/operator-framework/operator-sdk/pkg/predicate"
+
+       onapv1alpha1 "demo/vnfs/DAaaS/microservices/collectd-operator/pkg/apis/onap/v1alpha1"
+       collectdutils "demo/vnfs/DAaaS/microservices/collectd-operator/pkg/controller/utils"
+
+       appsv1 "k8s.io/api/apps/v1"
+       corev1 "k8s.io/api/core/v1"
+       extensionsv1beta1 "k8s.io/api/extensions/v1beta1"
+       "k8s.io/apimachinery/pkg/api/errors"
+       "k8s.io/apimachinery/pkg/runtime"
+       "k8s.io/apimachinery/pkg/types"
+       "k8s.io/client-go/util/retry"
+       "sigs.k8s.io/controller-runtime/pkg/client"
+       "sigs.k8s.io/controller-runtime/pkg/controller"
+       "sigs.k8s.io/controller-runtime/pkg/handler"
+       "sigs.k8s.io/controller-runtime/pkg/manager"
+       "sigs.k8s.io/controller-runtime/pkg/reconcile"
+       logf "sigs.k8s.io/controller-runtime/pkg/runtime/log"
+       "sigs.k8s.io/controller-runtime/pkg/source"
+)
+
+var log = logf.Log.WithName("controller_collectdglobal")
+
+// Add creates a new CollectdGlobal Controller and adds it to the Manager. The Manager will set fields on the Controller
+// and Start it when the Manager is Started.
+func Add(mgr manager.Manager) error {
+       return add(mgr, newReconciler(mgr))
+}
+
+// newReconciler returns a new reconcile.Reconciler
+func newReconciler(mgr manager.Manager) reconcile.Reconciler {
+       return &ReconcileCollectdGlobal{client: mgr.GetClient(), scheme: mgr.GetScheme()}
+}
+
+// add adds a new Controller to mgr with r as the reconcile.Reconciler
+func add(mgr manager.Manager, r reconcile.Reconciler) error {
+       // Create a new controller
+       log.V(1).Info("Creating a new controller for CollectdGlobal")
+       c, err := controller.New("collectdglobal-controller", mgr, controller.Options{Reconciler: r})
+       if err != nil {
+               return err
+       }
+
+       // Watch for changes to primary resource CollectdGlobal
+       log.V(1).Info("Add watcher for primary resource CollectdGlobal")
+       err = c.Watch(&source.Kind{Type: &onapv1alpha1.CollectdGlobal{}}, &handler.EnqueueRequestForObject{}, predicate.GenerationChangedPredicate{})
+       if err != nil {
+               return err
+       }
+
+       log.V(1).Info("Add watcher for secondary resource Collectd Daemonset")
+       err = c.Watch(
+               &source.Kind{Type: &appsv1.DaemonSet{}},
+               &handler.EnqueueRequestsFromMapFunc{
+                       ToRequests: handler.ToRequestsFunc(func(a handler.MapObject) []reconcile.Request {
+                               labelSelector, err := collectdutils.GetWatchLabels()
+                               labels := strings.Split(labelSelector, "=")
+                               if err != nil {
+                                       log.Error(err, "Failed to get watch labels, continuing with default label")
+                               }
+                               rcp := r.(*ReconcileCollectdGlobal)
+                               // Select the Daemonset with labelSelector (Defautl  is app=collectd)
+                               if a.Meta.GetLabels()[labels[0]] == labels[1] {
+                                       var requests []reconcile.Request
+                                       cg, err := collectdutils.GetCollectdGlobal(rcp.client, a.Meta.GetNamespace())
+                                       if err != nil || cg == nil {
+                                               return nil
+                                       }
+                                       requests = append(requests, reconcile.Request{
+                                               NamespacedName: client.ObjectKey{Namespace: cg.Namespace, Name: cg.Name}})
+                                       return requests
+                               }
+                               return nil
+                       }),
+               })
+       if err != nil {
+               return err
+       }
+
+       return nil
+}
+
+// blank assignment to verify that ReconcileCollectdGlobal implements reconcile.Reconciler
+var _ reconcile.Reconciler = &ReconcileCollectdGlobal{}
+
+// ReconcileCollectdGlobal reconciles a CollectdGlobal object
+type ReconcileCollectdGlobal struct {
+       // This client, initialized using mgr.Client() above, is a split client
+       // that reads objects from the cache and writes to the apiserver
+       client client.Client
+       scheme *runtime.Scheme
+}
+
+// Reconcile reads that state of the cluster for a CollectdGlobal object and makes changes based on the state read
+// and what is in the CollectdGlobal.Spec
+// TODO(user): Modify this Reconcile function to implement your Controller logic.  This example creates
+// a Pod as an example
+// Note:
+// The Controller will requeue the Request to be processed again if the returned error is non-nil or
+// Result.Requeue is true, otherwise upon completion it will remove the work from the queue.
+func (r *ReconcileCollectdGlobal) Reconcile(request reconcile.Request) (reconcile.Result, error) {
+       reqLogger := log.WithValues("Request.Namespace", request.Namespace, "Request.Name", request.Name)
+       reqLogger.Info("Reconciling CollectdGlobal")
+
+       // Fetch the CollectdGlobal instance
+       instance := &onapv1alpha1.CollectdGlobal{}
+       err := r.client.Get(context.TODO(), request.NamespacedName, instance)
+       if err != nil {
+               if errors.IsNotFound(err) {
+                       // Request object not found, could have been deleted after reconcile request.
+                       // Owned objects are automatically garbage collected. For additional cleanup logic use finalizers.
+                       // Return and don't requeue
+                       reqLogger.V(1).Info("CollectdGlobal object Not found")
+                       return reconcile.Result{}, nil
+               }
+               // Error reading the object - requeue the request.
+               reqLogger.V(1).Info("Error reading the CollectdGlobal object, Requeuing")
+               return reconcile.Result{}, err
+       }
+
+       // Handle Delete CR for additional cleanup
+       isDelete, err := r.handleDelete(reqLogger, instance)
+       if isDelete {
+               return reconcile.Result{}, err
+       }
+
+       // Add finalizer for this CR
+       if !collectdutils.Contains(instance.GetFinalizers(), collectdutils.CollectdFinalizer) {
+               if err := r.addFinalizer(reqLogger, instance); err != nil {
+                       return reconcile.Result{}, err
+               }
+               return reconcile.Result{}, nil
+       }
+       // Handle the reconciliation for CollectdGlobal.
+       // At this stage the Status of the CollectdGlobal should NOT be ""
+       err = r.handleCollectdGlobal(reqLogger, instance, false)
+       return reconcile.Result{}, err
+}
+
+// handleCollectdGlobal regenerates the collectd conf on CR Create, Update, Delete events
+func (r *ReconcileCollectdGlobal) handleCollectdGlobal(reqLogger logr.Logger, cr *onapv1alpha1.CollectdGlobal, isDelete bool) error {
+
+       rmap, err := collectdutils.FindResourceMapForCR(r.client, reqLogger, cr.Namespace)
+       if err != nil {
+               reqLogger.Info(":::: Skip current reconcile:::: Resources not found. Cache might be stale. Requeue")
+               return err
+       }
+
+       cm := rmap.ConfigMap
+       reqLogger.V(1).Info("Found ResourceMap")
+       reqLogger.V(1).Info(":::: ConfigMap Info ::::", "ConfigMap.Namespace", cm.Namespace, "ConfigMap.Name", cm.Name)
+
+       collectdConf, err := collectdutils.RebuildCollectdConf(r.client, cr.Namespace, isDelete, "")
+       if err != nil {
+               reqLogger.Error(err, "Skip reconcile: Rebuild conf failed")
+               return err
+       }
+
+       cm.SetAnnotations(map[string]string{
+               "daaas-random": collectdutils.ComputeSHA256([]byte(collectdConf)),
+       })
+
+       cm.Data["collectd.conf"] = collectdConf
+       retryErr := retry.RetryOnConflict(retry.DefaultRetry, func() error {
+               // Update the ConfigMap with new Spec and reload DaemonSets
+               reqLogger.Info("Updating the ConfigMap", "ConfigMap.Namespace", cm.Namespace, "ConfigMap.Name", cm.Name)
+               log.V(1).Info("ConfigMap Data", "Map: ", cm.Data)
+               err = r.client.Update(context.TODO(), cm)
+               if err != nil {
+                       reqLogger.Error(err, "Update the ConfigMap failed", "ConfigMap.Namespace", cm.Namespace, "ConfigMap.Name", cm.Name)
+                       return err
+               }
+               return nil
+       })
+       if retryErr != nil {
+               panic(fmt.Errorf("Update failed: %v", retryErr))
+       }
+
+       retryErr = retry.RetryOnConflict(retry.DefaultRetry, func() error {
+               // Retrieve the latest version of Daemonset before attempting update
+               // RetryOnConflict uses exponential backoff to avoid exhausting the apiserver
+               // Select DaemonSets with label
+               dsList := &extensionsv1beta1.DaemonSetList{}
+               opts := &client.ListOptions{}
+               labelSelector, err := collectdutils.GetWatchLabels()
+               if err != nil {
+                       reqLogger.Error(err, "Failed to get watch labels, continuing with default label")
+               }
+               opts.SetLabelSelector(labelSelector)
+               opts.InNamespace(cr.Namespace)
+               err = r.client.List(context.TODO(), opts, dsList)
+               if err != nil {
+                       panic(fmt.Errorf("Failed to get latest version of DaemonSet: %v", err))
+               }
+
+               if dsList.Items == nil || len(dsList.Items) == 0 {
+                       return errors.NewNotFound(corev1.Resource("daemonset"), "DaemonSet")
+               }
+               ds := &dsList.Items[0]
+               //Restart Collectd Pods
+               reqLogger.Info("Reloading the Daemonset", "DaemonSet.Namespace", ds.Namespace, "DaemonSet.Name", ds.Name)
+               //Restart only if hash of conf has changed.
+               ds.Spec.Template.SetAnnotations(map[string]string{
+                       "daaas-random": collectdutils.ComputeSHA256([]byte(collectdConf)),
+               })
+               r.handleAdditonalConfigMap(reqLogger, cr, ds)
+               updateErr := r.client.Update(context.TODO(), ds)
+               return updateErr
+       })
+       if retryErr != nil {
+               panic(fmt.Errorf("Update failed: %v", retryErr))
+       }
+
+       err = r.updateStatus(cr)
+       if err != nil {
+               reqLogger.Error(err, "Unable to update status")
+               return err
+       }
+       // Reconcile success
+       reqLogger.Info("Reconcile success!!")
+       return nil
+}
+
+// Handle Delete CR event for additional cleanup
+func (r *ReconcileCollectdGlobal) handleDelete(reqLogger logr.Logger, cr *onapv1alpha1.CollectdGlobal) (bool, error) {
+       // Check if the CollectdGlobal instance is marked to be deleted, which is
+       // indicated by the deletion timestamp being set.
+       isMarkedToBeDeleted := cr.GetDeletionTimestamp() != nil
+       if isMarkedToBeDeleted {
+               // Update status to Deleting state
+               cr.Status.Status = onapv1alpha1.Deleting
+               cr.Status.CollectdAgents = nil
+               _ = r.client.Status().Update(context.TODO(), cr)
+
+               if collectdutils.Contains(cr.GetFinalizers(), collectdutils.CollectdFinalizer) {
+                       // Run finalization logic for CollectdFinalizer. If the
+                       // finalization logic fails, don't remove the finalizer so
+                       // that we can retry during the next reconciliation.
+                       if err := r.finalizeCollectdGlobal(reqLogger, cr); err != nil {
+                               return isMarkedToBeDeleted, err
+                       }
+
+                       // Remove CollectdFinalizer. Once all finalizers have been
+                       // removed, the object will be deleted.
+                       cr.SetFinalizers(collectdutils.Remove(cr.GetFinalizers(), collectdutils.CollectdFinalizer))
+                       err := r.client.Update(context.TODO(), cr)
+                       if err != nil {
+                               return isMarkedToBeDeleted, err
+                       }
+               }
+       }
+       return isMarkedToBeDeleted, nil
+}
+
+func (r *ReconcileCollectdGlobal) updateStatus(cr *onapv1alpha1.CollectdGlobal) error {
+       switch cr.Status.Status {
+       case onapv1alpha1.Initial:
+               cr.Status.Status = onapv1alpha1.Created
+       case onapv1alpha1.Created, onapv1alpha1.Enabled:
+               pods, err := collectdutils.GetPodList(r.client, cr.Namespace)
+               if err != nil {
+                       return err
+               }
+               if !reflect.DeepEqual(pods, cr.Status.CollectdAgents) {
+                       cr.Status.CollectdAgents = pods
+                       cr.Status.Status = onapv1alpha1.Enabled
+               }
+       case onapv1alpha1.Deleting, onapv1alpha1.Deprecated:
+               return nil
+       }
+       err := r.client.Status().Update(context.TODO(), cr)
+       return err
+}
+
+func (r *ReconcileCollectdGlobal) finalizeCollectdGlobal(reqLogger logr.Logger, cr *onapv1alpha1.CollectdGlobal) error {
+       // Cleanup by regenerating new collectd conf and rolling update of DaemonSet
+       if err := r.handleCollectdGlobal(reqLogger, cr, true); err != nil {
+               reqLogger.Error(err, "Finalize CollectdGlobal failed!!")
+               return err
+       }
+       reqLogger.Info("Successfully finalized CollectdGlobal!!")
+       return nil
+}
+
+func (r *ReconcileCollectdGlobal) addFinalizer(reqLogger logr.Logger, cr *onapv1alpha1.CollectdGlobal) error {
+       reqLogger.Info("Adding Finalizer for the CollectdGlobal")
+       cr.SetFinalizers(append(cr.GetFinalizers(), collectdutils.CollectdFinalizer))
+       // Update status from Initial to Created
+       // Since addFinalizer will be executed only once,
+       // the status will be changed from Initial state to Created
+       updateErr := r.updateStatus(cr)
+       if updateErr != nil {
+               reqLogger.Error(updateErr, "Failed to update status from Initial state")
+       }
+       // Update CR
+       err := r.client.Update(context.TODO(), cr)
+       if err != nil {
+               reqLogger.Error(err, "Failed to update CollectdGlobal with finalizer")
+               return err
+       }
+       return nil
+}
+
+func (r *ReconcileCollectdGlobal) handleAdditonalConfigMap(reqLogger logr.Logger, cr *onapv1alpha1.CollectdGlobal, ds *extensionsv1beta1.DaemonSet) error {
+       cm := &corev1.ConfigMap{}
+       key := types.NamespacedName{Namespace: cr.Namespace, Name: cr.Spec.ConfigMap}
+       err := r.client.Get(context.TODO(), key, cm)
+       if err != nil {
+               reqLogger.Info("Error getting TypesDB")
+               return nil
+       }
+       // TODO: Implement Types.DB mounting
+       return nil
+}