Merge "Add FCAPS to oom"
[oom.git] / kubernetes / common / etcd / templates / statefulset.yaml
1 # Copyright © 2019 Intel Corporation Inc
2 #
3 # Licensed under the Apache License, Version 2.0 (the "License");
4 # you may not use this file except in compliance with the License.
5 # You may obtain a copy of the License at
6 #
7 #       http://www.apache.org/licenses/LICENSE-2.0
8 #
9 # Unless required by applicable law or agreed to in writing, software
10 # distributed under the License is distributed on an "AS IS" BASIS,
11 # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 # See the License for the specific language governing permissions and
13 # limitations under the License.
14
15 apiVersion: apps/v1beta1
16 kind: StatefulSet
17 metadata:
18   name: {{ include "common.servicename" .  }}
19   labels:
20     heritage: "{{ .Release.Service }}"
21     release: "{{ .Release.Name }}"
22     chart: "{{ .Chart.Name }}-{{ .Chart.Version }}"
23     app: {{ template "common.name" . }}
24 spec:
25   serviceName: {{ include "common.servicename" .  }}
26   replicas: {{ .Values.replicaCount }}
27   template:
28     metadata:
29       labels:
30         heritage: "{{ .Release.Service }}"
31         release: "{{ .Release.Name }}"
32         chart: "{{ .Chart.Name }}-{{ .Chart.Version }}"
33         app: {{ include "common.name" . }}
34     spec:
35 {{- if .Values.affinity }}
36       affinity:
37 {{ toYaml .Values.affinity | indent 8 }}
38 {{- end }}
39 {{- if .Values.nodeSelector }}
40       nodeSelector:
41 {{ toYaml .Values.nodeSelector | indent 8 }}
42 {{- end }}
43 {{- if .Values.tolerations }}
44       tolerations:
45 {{ toYaml .Values.tolerations | indent 8 }}
46 {{- end }}
47       containers:
48       - name: {{ include "common.servicename" .  }}
49         image: "{{ .Values.repository }}/{{ .Values.image }}"
50         imagePullPolicy: "{{ .Values.pullPolicy }}"
51         ports:
52         - containerPort: {{ .Values.service.peerInternalPort }}
53           name: {{ .Values.service.peerPortName }}
54         - containerPort: {{ .Values.service.clientInternalPort }}
55           name: {{ .Values.service.clientPortName }}
56         {{- if eq .Values.liveness.enabled true }}
57         livenessProbe:
58           exec:
59             command: ["/bin/sh", "-c", "etcdctl cluster-health | grep -w healthy" ]
60             initialDelaySeconds: {{ .Values.liveness.initialDelaySeconds }}
61             periodSeconds: {{ .Values.liveness.periodSeconds }}
62             timeoutSeconds: {{ .Values.liveness.timeoutSeconds }}
63           {{ end -}}
64         readinessProbe:
65           exec:
66             command: ["/bin/sh", "-c", "etcdctl cluster-health | grep -w healthy" ]
67             initialDelaySeconds: {{ .Values.readiness.initialDelaySeconds }}
68             periodSeconds: {{ .Values.readiness.periodSeconds }}
69         resources:
70 {{ include "common.resources" . | indent 10 }}
71         env:
72         - name: INITIAL_CLUSTER_SIZE
73           value: {{ .Values.replicaCount | quote }}
74         - name: SET_NAME
75           value: {{ include "common.servicename" . }}
76 {{- if .Values.extraEnv }}
77 {{ toYaml .Values.extraEnv | indent 8 }}
78 {{- end }}
79         lifecycle:
80           preStop:
81             exec:
82               command:
83                 - "/bin/sh"
84                 - "-ec"
85                 - |
86                   EPS=""
87                   for i in $(seq 0 $((${INITIAL_CLUSTER_SIZE} - 1))); do
88                       EPS="${EPS}${EPS:+,}http://${SET_NAME}-${i}.${SET_NAME}:2379"
89                   done
90
91                   HOSTNAME=$(hostname)
92
93                   member_hash() {
94                       etcdctl member list | grep http://${HOSTNAME}.${SET_NAME}:2380 | cut -d':' -f1 | cut -d'[' -f1
95                   }
96
97                   SET_ID=${HOSTNAME##*[^0-9]}
98
99                   if [ "${SET_ID}" -ge ${INITIAL_CLUSTER_SIZE} ]; then
100                       echo "Removing ${HOSTNAME} from etcd cluster"
101                       ETCDCTL_ENDPOINT=${EPS} etcdctl member remove $(member_hash)
102                       if [ $? -eq 0 ]; then
103                           # Remove everything otherwise the cluster will no longer scale-up
104                           rm -rf /var/run/etcd/*
105                       fi
106                   fi
107         command:
108           - "/bin/sh"
109           - "-ec"
110           - |
111             HOSTNAME=$(hostname)
112
113             # store member id into PVC for later member replacement
114             collect_member() {
115                 while ! etcdctl member list &>/dev/null; do sleep 1; done
116                 etcdctl member list | grep http://${HOSTNAME}.${SET_NAME}:2380 | cut -d':' -f1 | cut -d'[' -f1 > /var/run/etcd/member_id
117                 exit 0
118             }
119
120             eps() {
121                 EPS=""
122                 for i in $(seq 0 $((${INITIAL_CLUSTER_SIZE} - 1))); do
123                     EPS="${EPS}${EPS:+,}http://${SET_NAME}-${i}.${SET_NAME}:2379"
124                 done
125                 echo ${EPS}
126             }
127
128             member_hash() {
129                 etcdctl member list | grep http://${HOSTNAME}.${SET_NAME}:2380 | cut -d':' -f1 | cut -d'[' -f1
130             }
131
132             # we should wait for other pods to be up before trying to join
133             # otherwise we got "no such host" errors when trying to resolve other members
134             for i in $(seq 0 $((${INITIAL_CLUSTER_SIZE} - 1))); do
135                 while true; do
136                     echo "Waiting for ${SET_NAME}-${i}.${SET_NAME} to come up"
137                     ping -W 1 -c 1 ${SET_NAME}-${i}.${SET_NAME} > /dev/null && break
138                     sleep 1s
139                 done
140             done
141
142             # re-joining after failure?
143             if [ -e /var/run/etcd/default.etcd ]; then
144                 echo "Re-joining etcd member"
145                 member_id=$(cat /var/run/etcd/member_id)
146
147                 # re-join member
148                 ETCDCTL_ENDPOINT=$(eps) etcdctl member update ${member_id} http://${HOSTNAME}.${SET_NAME}:2380 | true
149                 exec etcd --name ${HOSTNAME} \
150                     --listen-peer-urls http://0.0.0.0:2380 \
151                     --listen-client-urls http://0.0.0.0:2379\
152                     --advertise-client-urls http://${HOSTNAME}.${SET_NAME}:2379 \
153                     --data-dir /var/run/etcd/default.etcd
154             fi
155
156             # etcd-SET_ID
157             SET_ID=${HOSTNAME##*[^0-9]}
158
159             # adding a new member to existing cluster (assuming all initial pods are available)
160             if [ "${SET_ID}" -ge ${INITIAL_CLUSTER_SIZE} ]; then
161                 export ETCDCTL_ENDPOINT=$(eps)
162
163                 # member already added?
164                 MEMBER_HASH=$(member_hash)
165                 if [ -n "${MEMBER_HASH}" ]; then
166                     # the member hash exists but for some reason etcd failed
167                     # as the datadir has not be created, we can remove the member
168                     # and retrieve new hash
169                     etcdctl member remove ${MEMBER_HASH}
170                 fi
171
172                 echo "Adding new member"
173                 etcdctl member add ${HOSTNAME} http://${HOSTNAME}.${SET_NAME}:2380 | grep "^ETCD_" > /var/run/etcd/new_member_envs
174
175                 if [ $? -ne 0 ]; then
176                     echo "Exiting"
177                     rm -f /var/run/etcd/new_member_envs
178                     exit 1
179                 fi
180
181                 cat /var/run/etcd/new_member_envs
182                 source /var/run/etcd/new_member_envs
183
184                 collect_member &
185
186                 exec etcd --name ${HOSTNAME} \
187                     --listen-peer-urls http://0.0.0.0:2380 \
188                     --listen-client-urls http://0.0.0.0:2379 \
189                     --advertise-client-urls http://${HOSTNAME}.${SET_NAME}:2379 \
190                     --data-dir /var/run/etcd/default.etcd \
191                     --initial-advertise-peer-urls http://${HOSTNAME}.${SET_NAME}:2380 \
192                     --initial-cluster ${ETCD_INITIAL_CLUSTER} \
193                     --initial-cluster-state ${ETCD_INITIAL_CLUSTER_STATE}
194             fi
195
196             PEERS=""
197             for i in $(seq 0 $((${INITIAL_CLUSTER_SIZE} - 1))); do
198                 PEERS="${PEERS}${PEERS:+,}${SET_NAME}-${i}=http://${SET_NAME}-${i}.${SET_NAME}:2380"
199             done
200
201             collect_member &
202
203             # join member
204             exec etcd --name ${HOSTNAME} \
205                 --initial-advertise-peer-urls http://${HOSTNAME}.${SET_NAME}:2380 \
206                 --listen-peer-urls http://0.0.0.0:2380 \
207                 --listen-client-urls http://0.0.0.0:2379 \
208                 --advertise-client-urls http://${HOSTNAME}.${SET_NAME}:2379 \
209                 --initial-cluster-token etcd-cluster-1 \
210                 --initial-cluster ${PEERS} \
211                 --initial-cluster-state new \
212                 --data-dir /var/run/etcd/default.etcd
213         volumeMounts:
214         - name: {{ include "common.servicename" . }}-datadir
215           mountPath: /var/run/etcd
216   {{- if .Values.persistence.enabled }}
217   volumeClaimTemplates:
218   - metadata:
219       name: {{ include "common.servicename" . }}-data
220     spec:
221       accessModes:
222         - "{{ .Values.persistence.accessMode }}"
223       resources:
224         requests:
225           # upstream recommended max is 700M
226           storage: "{{ .Values.persistence.storage }}"
227     {{- if .Values.persistence.storageClass }}
228     {{- if (eq "-" .Values.persistence.storageClass) }}
229       storageClassName: ""
230     {{- else }}
231       storageClassName: "{{ .Values.persistence.storageClass }}"
232     {{- end }}
233     {{- end }}
234   {{- else }}
235       volumes:
236       - name: {{ include "common.servicename" . }}-datadir
237       {{- if .Values.memoryMode }}
238         emptyDir:
239           medium: Memory
240       {{- else }}
241         emptyDir: {}
242       {{- end }}
243   {{- end }}
244