Inference package with Tensorflow Serving 45/83645/2
authorDileep Ranganathan <dileep.ranganathan@intel.com>
Thu, 28 Mar 2019 18:29:36 +0000 (11:29 -0700)
committerDileep Ranganathan <dileep.ranganathan@intel.com>
Thu, 28 Mar 2019 18:34:45 +0000 (11:34 -0700)
Initial inference package with TF Serving, Tensorboard.
Added Minio/S3 support.

Change-Id: If896a7c8fc150bf6a552bdcf7d1096d357520751
Issue-ID: ONAPARC-454
Signed-off-by: Dileep Ranganathan <dileep.ranganathan@intel.com>
13 files changed:
vnfs/DAaaS/inference-core/.helmignore [new file with mode: 0644]
vnfs/DAaaS/inference-core/Chart.yaml [new file with mode: 0644]
vnfs/DAaaS/inference-core/charts/tf-serving/.helmignore [new file with mode: 0644]
vnfs/DAaaS/inference-core/charts/tf-serving/Chart.yaml [new file with mode: 0644]
vnfs/DAaaS/inference-core/charts/tf-serving/templates/NOTES.txt [new file with mode: 0644]
vnfs/DAaaS/inference-core/charts/tf-serving/templates/_helpers.tpl [new file with mode: 0644]
vnfs/DAaaS/inference-core/charts/tf-serving/templates/deployment.yaml [new file with mode: 0644]
vnfs/DAaaS/inference-core/charts/tf-serving/templates/ingress.yaml [new file with mode: 0644]
vnfs/DAaaS/inference-core/charts/tf-serving/templates/secrets.yaml [new file with mode: 0644]
vnfs/DAaaS/inference-core/charts/tf-serving/templates/service.yaml [new file with mode: 0644]
vnfs/DAaaS/inference-core/charts/tf-serving/templates/serviceaccount.yaml [new file with mode: 0644]
vnfs/DAaaS/inference-core/charts/tf-serving/values.yaml [new file with mode: 0644]
vnfs/DAaaS/inference-core/values.yaml [new file with mode: 0644]

diff --git a/vnfs/DAaaS/inference-core/.helmignore b/vnfs/DAaaS/inference-core/.helmignore
new file mode 100644 (file)
index 0000000..50af031
--- /dev/null
@@ -0,0 +1,22 @@
+# Patterns to ignore when building packages.
+# This supports shell glob matching, relative path matching, and
+# negation (prefixed with !). Only one pattern per line.
+.DS_Store
+# Common VCS dirs
+.git/
+.gitignore
+.bzr/
+.bzrignore
+.hg/
+.hgignore
+.svn/
+# Common backup files
+*.swp
+*.bak
+*.tmp
+*~
+# Various IDEs
+.project
+.idea/
+*.tmproj
+.vscode/
diff --git a/vnfs/DAaaS/inference-core/Chart.yaml b/vnfs/DAaaS/inference-core/Chart.yaml
new file mode 100644 (file)
index 0000000..836b1a8
--- /dev/null
@@ -0,0 +1,5 @@
+apiVersion: v1
+appVersion: "1.0"
+description: A Helm chart for inference framework components
+name: inference-core
+version: 0.1.0
diff --git a/vnfs/DAaaS/inference-core/charts/tf-serving/.helmignore b/vnfs/DAaaS/inference-core/charts/tf-serving/.helmignore
new file mode 100644 (file)
index 0000000..50af031
--- /dev/null
@@ -0,0 +1,22 @@
+# Patterns to ignore when building packages.
+# This supports shell glob matching, relative path matching, and
+# negation (prefixed with !). Only one pattern per line.
+.DS_Store
+# Common VCS dirs
+.git/
+.gitignore
+.bzr/
+.bzrignore
+.hg/
+.hgignore
+.svn/
+# Common backup files
+*.swp
+*.bak
+*.tmp
+*~
+# Various IDEs
+.project
+.idea/
+*.tmproj
+.vscode/
diff --git a/vnfs/DAaaS/inference-core/charts/tf-serving/Chart.yaml b/vnfs/DAaaS/inference-core/charts/tf-serving/Chart.yaml
new file mode 100644 (file)
index 0000000..2db1a48
--- /dev/null
@@ -0,0 +1,5 @@
+apiVersion: v1
+appVersion: "1.0"
+description: Helm chart for Tensorflow serving model server
+name: tf-serving
+version: 0.1.0
diff --git a/vnfs/DAaaS/inference-core/charts/tf-serving/templates/NOTES.txt b/vnfs/DAaaS/inference-core/charts/tf-serving/templates/NOTES.txt
new file mode 100644 (file)
index 0000000..2dcf639
--- /dev/null
@@ -0,0 +1,20 @@
+1. Get the tensorflow serving URL by running these commands:
+{{- if .Values.ingress.enabled }}
+{{- range $host := .Values.ingress.hosts }}
+  {{- range $.Values.ingress.paths }}
+  http{{ if $.Values.ingress.tls }}s{{ end }}://{{ $host }}{{ . }}
+  {{- end }}
+{{- end }}
+{{- else if contains "NodePort" .Values.service.type }}
+  export NODE_PORT=$(kubectl get --namespace {{ .Release.Namespace }} -o jsonpath="{.spec.ports[0].nodePort}" services {{ include "tf-serving.fullname" . }})
+  export NODE_IP=$(kubectl get nodes --namespace {{ .Release.Namespace }} -o jsonpath="{.items[0].status.addresses[0].address}")
+  echo http://$NODE_IP:$NODE_PORT
+{{- else if contains "LoadBalancer" .Values.service.type }}
+     NOTE: It may take a few minutes for the LoadBalancer IP to be available.
+           You can watch the status of by running 'kubectl get svc -w {{ include "tf-serving.fullname" . }}'
+  export SERVICE_IP=$(kubectl get svc --namespace {{ .Release.Namespace }} {{ include "tf-serving.fullname" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')
+  echo http://$SERVICE_IP:{{ .Values.service.port }}
+{{- else if contains "ClusterIP" .Values.service.type }}
+  export SVC_NAME=$(kubectl get svc --namespace {{ .Release.Namespace }} -l "app={{ include "tf-serving.name" . }},modelName={{ .Values.modelName }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
+  kubectl port-forward svc/$SVC_NAME 8500
+{{- end }}
diff --git a/vnfs/DAaaS/inference-core/charts/tf-serving/templates/_helpers.tpl b/vnfs/DAaaS/inference-core/charts/tf-serving/templates/_helpers.tpl
new file mode 100644 (file)
index 0000000..ce3ce91
--- /dev/null
@@ -0,0 +1,41 @@
+{{/* vim: set filetype=mustache: */}}
+{{/*
+Expand the name of the chart.
+*/}}
+{{- define "tf-serving.name" -}}
+{{- default .Chart.Name .Values.nameOverride | trunc 63 | trimSuffix "-" -}}
+{{- end -}}
+
+{{/*
+Create a default fully qualified app name.
+We truncate at 63 chars because some Kubernetes name fields are limited to this (by the DNS naming spec).
+If release name contains chart name it will be used as a full name.
+*/}}
+{{- define "tf-serving.fullname" -}}
+{{- if .Values.fullnameOverride -}}
+{{- .Values.fullnameOverride | trunc 63 | trimSuffix "-" -}}
+{{- else -}}
+{{- $name := default .Chart.Name .Values.nameOverride -}}
+{{- if contains $name .Release.Name -}}
+{{- .Release.Name | trunc 63 | trimSuffix "-" -}}
+{{- else -}}
+{{- printf "%s-%s-%s" .Release.Name $name .Values.modelName | trunc 63 | trimSuffix "-" -}}
+{{- end -}}
+{{- end -}}
+{{- end -}}
+
+{{/*
+Create chart name and version as used by the chart label.
+*/}}
+{{- define "tf-serving.chart" -}}
+{{- printf "%s-%s" .Chart.Name .Chart.Version | replace "+" "_" | trunc 63 | trimSuffix "-" -}}
+{{- end -}}
+
+{{/* Create the name of tf-serving service account to use */}}
+{{- define "tf-serving.serviceAccountName" -}}
+{{- if and .Values.global.rbac .Values.serviceAccount.create -}}
+    {{ default (include "tf-serving.fullname" .) .Values.serviceAccount.name }}
+{{- else -}}
+    {{ default "default" .Values.serviceAccount.name }}
+{{- end -}}
+{{- end -}}
\ No newline at end of file
diff --git a/vnfs/DAaaS/inference-core/charts/tf-serving/templates/deployment.yaml b/vnfs/DAaaS/inference-core/charts/tf-serving/templates/deployment.yaml
new file mode 100644 (file)
index 0000000..0a909e9
--- /dev/null
@@ -0,0 +1,138 @@
+{{/*
+# Copyright 2019 Intel Corporation, Inc
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+#       http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+*/}}
+
+apiVersion: extensions/v1beta1
+kind: Deployment
+metadata:
+  name: {{ include "tf-serving.fullname" . }}
+  labels:
+    app: {{ include "tf-serving.name" . }}
+    chart: {{ .Chart.Name }}-{{ .Chart.Version | replace "+" "_" }}
+    release: {{ .Release.Name }}
+    heritage: {{ .Release.Service | quote }}
+spec:
+  replicas: 1
+  selector:
+    matchLabels:
+      app: {{ include "tf-serving.name" . }}
+      release: {{ .Release.Name }}
+      modelName: {{ .Values.modelName }}
+  template:
+    metadata:
+      labels:
+        app: {{ include "tf-serving.name" . }}
+        release: {{ .Release.Name }}
+        modelName: {{ .Values.modelName }}
+    spec:
+      serviceAccountName: {{ template "tf-serving.serviceAccountName" . }}
+      containers:
+        - name: tf-serving
+          image: "{{ .Values.image.tensorflowServing.repository }}:{{ .Values.image.tensorflowServing.tag }}"
+          imagePullPolicy: {{ .Values.image.pullPolicy }}
+          command:
+            - "/usr/bin/tensorflow_model_server"
+          args:
+            - "--port={{ .Values.service.tensorflowServingPort }}"
+            - "--model_name={{ .Values.modelName }}"
+            - "--model_base_path={{ .Values.modelBasePath }}"
+          ports:
+            - name: tf-serving
+              containerPort: {{ .Values.service.tensorflowServingPort }}
+              protocol: TCP
+          env:
+            - name: AWS_ACCESS_KEY_ID
+              valueFrom:
+                secretKeyRef:
+                  name: {{ if .Values.minio.existingSecret }}{{ .Values.minio.existingSecret }}{{ else }}{{ template "tf-serving.fullname" . }}{{ end }}
+                  key: accesskey
+            - name: AWS_SECRET_ACCESS_KEY
+              valueFrom:
+                secretKeyRef:
+                  name: {{ if .Values.minio.existingSecret }}{{ .Values.minio.existingSecret }}{{ else }}{{ template "tf-serving.fullname" . }}{{ end }}
+                  key: secretkey
+            {{- range $key, $val := .Values.minio.environment }}
+            - name: {{ $key }}
+              value: {{ $val | quote }}
+            {{- end}}
+          readinessProbe:
+            tcpSocket:
+              port: tf-serving
+            initialDelaySeconds: 15
+            timeoutSeconds: 1
+          resources:
+            {{- toYaml .Values.resources | nindent 12 }}
+          volumeMounts:
+            - mountPath: /models
+              name: models
+        - name: tensorboard
+          image: "{{ .Values.image.tensorboard.repository }}:{{ .Values.image.tensorboard.tag }}"
+          imagePullPolicy: {{ .Values.image.pullPolicy }}
+          command:
+            - tensorboard
+          args:
+             - --logdir
+            - {{ .Values.modelBasePath }}
+            - --host
+            - 0.0.0.0
+          ports:
+            - name: tensorboard
+              containerPort: {{ .Values.service.tensorboardPort }}
+              protocol: TCP
+          env:
+            - name: AWS_ACCESS_KEY_ID
+              valueFrom:
+                secretKeyRef:
+                  name: {{ if .Values.minio.existingSecret }}{{ .Values.minio.existingSecret }}{{ else }}{{ template "tf-serving.fullname" . }}{{ end }}
+                  key: accesskey
+            - name: AWS_SECRET_ACCESS_KEY
+              valueFrom:
+                secretKeyRef:
+                  name: {{ if .Values.minio.existingSecret }}{{ .Values.minio.existingSecret }}{{ else }}{{ template "tf-serving.fullname" . }}{{ end }}
+                  key: secretkey
+            {{- range $key, $val := .Values.minio.environment }}
+            - name: {{ $key }}
+              value: {{ $val | quote }}
+            {{- end}}
+          livenessProbe:
+            httpGet:
+              path: /
+              port: {{ .Values.service.tensorboardPort }}
+          readinessProbe:
+            httpGet:
+              path: /
+              port: {{ .Values.service.tensorboardPort }}
+          volumeMounts:
+            - mountPath: /output/training_logs
+              name: training-logs-volume
+          resources:
+            {{- toYaml .Values.resources | nindent 12 }}
+      volumes:
+        - name: models
+          emptyDir: {}
+        - name: training-logs-volume
+          emptyDir: {}
+      {{- with .Values.nodeSelector }}
+      nodeSelector:
+        {{- toYaml . | nindent 8 }}
+      {{- end }}
+    {{- with .Values.affinity }}
+      affinity:
+        {{- toYaml . | nindent 8 }}
+    {{- end }}
+    {{- with .Values.tolerations }}
+      tolerations:
+        {{- toYaml . | nindent 8 }}
+    {{- end }}
diff --git a/vnfs/DAaaS/inference-core/charts/tf-serving/templates/ingress.yaml b/vnfs/DAaaS/inference-core/charts/tf-serving/templates/ingress.yaml
new file mode 100644 (file)
index 0000000..b02fc8b
--- /dev/null
@@ -0,0 +1,55 @@
+{{/*
+# Copyright 2019 Intel Corporation, Inc
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+#       http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+*/}}
+
+{{- if .Values.ingress.enabled -}}
+{{- $fullName := include "tf-serving.fullname" . -}}
+{{- $ingressPaths := .Values.ingress.paths -}}
+apiVersion: extensions/v1beta1
+kind: Ingress
+metadata:
+  name: {{ $fullName }}
+  labels:
+    app: {{ include "tf-serving.name" . }}
+    release: {{ .Release.Name }}
+    modelName: {{ .Values.modelName }}
+  {{- with .Values.ingress.annotations }}
+  annotations:
+    {{- toYaml . | nindent 4 }}
+  {{- end }}
+spec:
+{{- if .Values.ingress.tls }}
+  tls:
+  {{- range .Values.ingress.tls }}
+    - hosts:
+      {{- range .hosts }}
+        - {{ . | quote }}
+      {{- end }}
+      secretName: {{ .secretName }}
+  {{- end }}
+{{- end }}
+  rules:
+  {{- range .Values.ingress.hosts }}
+    - host: {{ . | quote }}
+      http:
+        paths:
+       {{- range $ingressPaths }}
+          - path: {{ . }}
+            backend:
+              serviceName: {{ $fullName }}
+              servicePort: http
+       {{- end }}
+  {{- end }}
+{{- end }}
diff --git a/vnfs/DAaaS/inference-core/charts/tf-serving/templates/secrets.yaml b/vnfs/DAaaS/inference-core/charts/tf-serving/templates/secrets.yaml
new file mode 100644 (file)
index 0000000..f4b8fe8
--- /dev/null
@@ -0,0 +1,31 @@
+{{/*
+# Copyright 2019 Intel Corporation, Inc
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+#       http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+*/}}
+
+{{- if not .Values.minio.existingSecret }}
+apiVersion: v1
+kind: Secret
+metadata:
+  name: {{ template "tf-serving.fullname" . }}
+  labels:
+    app: {{ include "tf-serving.name" . }}
+    chart: {{ .Chart.Name }}-{{ .Chart.Version | replace "+" "_" }}
+    release: {{ .Release.Name }}
+    heritage: {{ .Release.Service | quote }}
+type: Opaque
+data:
+  accesskey: {{ .Values.minio.accessKey | b64enc }}
+  secretkey: {{ .Values.minio.secretKey | b64enc }}
+{{- end }}
diff --git a/vnfs/DAaaS/inference-core/charts/tf-serving/templates/service.yaml b/vnfs/DAaaS/inference-core/charts/tf-serving/templates/service.yaml
new file mode 100644 (file)
index 0000000..3ddcca6
--- /dev/null
@@ -0,0 +1,39 @@
+{{/*
+# Copyright 2019 Intel Corporation, Inc
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+#       http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+*/}}
+
+apiVersion: v1
+kind: Service
+metadata:
+  name: {{ include "tf-serving.fullname" . }}
+  labels:
+    app: {{ include "tf-serving.name" . }}
+    release: {{ .Release.Name }}
+    modelName: {{ .Values.modelName }}
+spec:
+  type: {{ .Values.service.type }}
+  ports:
+    - port: {{ .Values.service.tensorflowServingPort }}
+      targetPort: tf-serving
+      protocol: TCP
+      name: tf-serving
+    - port: {{ .Values.service.tensorboardPort }}
+      targetPort: tensorboard
+      protocol: TCP
+      name: tensorboard
+  selector:
+    app: {{ include "tf-serving.name" . }}
+    release: {{ .Release.Name }}
+    modelName: {{ .Values.modelName }}
diff --git a/vnfs/DAaaS/inference-core/charts/tf-serving/templates/serviceaccount.yaml b/vnfs/DAaaS/inference-core/charts/tf-serving/templates/serviceaccount.yaml
new file mode 100644 (file)
index 0000000..af4987d
--- /dev/null
@@ -0,0 +1,25 @@
+{{/*
+# Copyright 2019 Intel Corporation, Inc
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+#       http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+*/}}
+
+apiVersion: v1
+kind: ServiceAccount
+metadata:
+    name: {{ include "tf-serving.fullname" . }}
+    labels:
+       app: {{ include "tf-serving.name" . }}
+       chart: {{ .Chart.Name }}-{{ .Chart.Version | replace "+" "_" }}
+       release: {{ .Release.Name }}
+       heritage: {{ .Release.Service | quote }}
diff --git a/vnfs/DAaaS/inference-core/charts/tf-serving/values.yaml b/vnfs/DAaaS/inference-core/charts/tf-serving/values.yaml
new file mode 100644 (file)
index 0000000..ebf8868
--- /dev/null
@@ -0,0 +1,71 @@
+# Default values for tf-serving.
+# This is a YAML-formatted file.
+# Declare variables to be passed into your templates.
+
+replicaCount: 1
+
+image:
+  tensorflowServing:
+    repository: tensorflow/serving
+    tag: latest
+  tensorboard:
+    repository: tensorflow/tensorflow
+    tag: latest
+  pullPolicy: IfNotPresent
+
+nameOverride: ""
+fullnameOverride: ""
+
+## Model information for tf-serving
+modelName: "mnist"
+modelBasePath: "s3://models/mnist"
+
+## Model repository information (Minio)
+minio:
+  existingSecret: ""
+  accessKey: "onapdaas"
+  secretKey: "onapsecretdaas"
+  environment:
+    AWS_REGION: ""
+    S3_REGION: ""
+    S3_ENDPOINT: "minio.minio.svc.cluster.local:9000"
+    AWS_ENDPOINT_URL: "http://minio.minio.svc.cluster.local:9000"
+    S3_USE_HTTPS: 0
+    S3_VERIFY_SSL: 0
+
+
+## Service account for tf-serving to use.
+serviceAccount:
+  create: true
+  name: ""
+
+service:
+  type: ClusterIP
+  tensorflowServingPort: 8500
+  tensorboardPort: 6006
+
+ingress:
+  enabled: false
+  annotations: {}
+  paths: []
+  hosts:
+    - chart-example.local
+  tls: []
+
+resources: {}
+  # We usually recommend not to specify default resources and to leave this as a conscious
+  # choice for the user. This also increases chances charts run on environments with little
+  # resources, such as Minikube. If you do want to specify resources, uncomment the following
+  # lines, adjust them as necessary, and remove the curly braces after 'resources:'.
+  # limits:
+  #  cpu: 100m
+  #  memory: 128Mi
+  # requests:
+  #  cpu: 100m
+  #  memory: 128Mi
+
+nodeSelector: {}
+
+tolerations: []
+
+affinity: {}
diff --git a/vnfs/DAaaS/inference-core/values.yaml b/vnfs/DAaaS/inference-core/values.yaml
new file mode 100644 (file)
index 0000000..10714b8
--- /dev/null
@@ -0,0 +1,30 @@
+# Copyright © 2019 Intel Corporation
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+#       http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+#################################################################
+# Global configuration defaults.
+#################################################################
+global:
+  nodePortPrefix: 310
+  rbac: true
+  repository: nexus3.onap.org:10001
+  readinessRepository: oomk8s
+  readinessImage: readiness-check:2.0.0
+  loggingRepository: docker.elastic.co
+  loggingImage: beats/filebeat:5.5.0
+
+#################################################################
+# k8s Operator Day-0 configuration defaults.
+#################################################################
+