Refactor Distributed Analytics project structure
[demo.git] / vnfs / DAaaS / operator / charts / sparkoperator / templates / spark-operator-deployment.yaml
diff --git a/vnfs/DAaaS/operator/charts/sparkoperator/templates/spark-operator-deployment.yaml b/vnfs/DAaaS/operator/charts/sparkoperator/templates/spark-operator-deployment.yaml
deleted file mode 100755 (executable)
index fdfc51a..0000000
+++ /dev/null
@@ -1,79 +0,0 @@
-# If the admission webhook is enabled, then a post-install step is required
-# to generate and install the secret in the operator namespace.
-
-# In the post-install hook, the token corresponding to the operator service account
-# is used to authenticate with the Kubernetes API server to install the secret bundle.
-
-apiVersion: apps/v1
-kind: Deployment
-metadata:
-  name: {{ include "sparkoperator.fullname" . }}
-  labels:
-    app.kubernetes.io/name: {{ include "sparkoperator.name" . }}
-    helm.sh/chart: {{ include "sparkoperator.chart" . }}
-    app.kubernetes.io/instance: {{ .Release.Name }}
-    app.kubernetes.io/managed-by: {{ .Release.Service }}
-spec:
-  replicas: 1
-  selector:
-    matchLabels:
-      app.kubernetes.io/name: {{ include "sparkoperator.name" . }}
-      app.kubernetes.io/version: {{ .Values.operatorVersion }}
-  strategy:
-    type: Recreate
-  template:
-    metadata:
-      {{- if .Values.enableMetrics }}
-      annotations:
-        prometheus.io/scrape: "true"
-        prometheus.io/port: "{{ .Values.metricsPort }}"
-        prometheus.io/path: {{ .Values.metricsEndpoint }}
-      {{- end }}
-      labels:
-        app.kubernetes.io/name: {{ include "sparkoperator.name" . }}
-        app.kubernetes.io/version: {{ .Values.operatorVersion }}
-      initializers:
-        pending: []
-    spec:
-      serviceAccountName: {{ include "sparkoperator.serviceAccountName" . }}
-      {{- if .Values.enableWebhook }}
-      volumes:
-        - name: webhook-certs
-          secret:
-            secretName: spark-webhook-certs
-      {{- end }}
-      containers:
-      - name: sparkoperator
-        image: {{ .Values.operatorImageName }}:{{ .Values.operatorVersion }}
-        imagePullPolicy: {{ .Values.imagePullPolicy }}
-        {{- if .Values.enableWebhook }}
-        volumeMounts:
-          - name: webhook-certs
-            mountPath: /etc/webhook-certs
-        {{- end }}
-        {{- if .Values.enableMetrics }}
-        ports:
-          - containerPort: {{ .Values.metricsPort }}
-        {{ end }}
-        args:
-        - -v=2
-        - -namespace={{ .Values.sparkJobNamespace }}
-        - -ingress-url-format={{ .Values.ingressUrlFormat }}
-        - -install-crds={{ .Values.installCrds }}
-        - -controller-threads={{ .Values.controllerThreads }}
-        - -resync-interval={{ .Values.resyncInterval }}
-        - -logtostderr
-        {{- if .Values.enableMetrics }}
-        - -enable-metrics=true
-        - -metrics-labels=app_type
-        - -metrics-port={{ .Values.metricsPort }}
-        - -metrics-endpoint={{ .Values.metricsEndpoint }}
-        - -metrics-prefix={{ .Values.metricsPrefix }}
-        {{- end }}
-        {{- if .Values.enableWebhook }}
-        - -enable-webhook=true
-        - -webhook-svc-namespace={{ .Release.Namespace }}
-        - -webhook-port={{ .Values.webhookPort }}
-        - -webhook-svc-name={{ .Release.Name }}-webhook
-        - -webhook-config-name={{ include "sparkoperator.fullname" . }}-webhook-config
-        {{- end }}