kubernetes/config/onap-parameters.yaml
.idea/*
+kubernetes/dist/*
+requirements.lock
+**/charts/*.tgz
relationships:
- type: cloudify.relationships.connected_to
target: k8s_master
-
- kube2msb:
- type: cloudify.nodes.Helm
- properties:
- args:
- - { get_input: namespace_perfix }
- - kube2msb
- relationships:
- - type: cloudify.relationships.connected_to
- target: k8s_master
- [kubectl, 1.8.6-0]
- [kubernetes-cni, 0.5.1-1]
- [nfs-utils]
+ - git
+ - wget
runcmd:
- [ setenforce, 0 ]
- [ sysctl , '--system' ]
User for connecting to agent VMs
default: centos
+ helm_version:
+ description: helm version to be installed
+ default: v2.7.0
+
dsl_definitions:
openstack_config: &openstack_config
- type: cloudify.relationships.contained_in
target: kubernetes_master_host
+ k8s_tools:
+ type: cloudify.nodes.SoftwareComponent
+ properties:
+ interfaces:
+ cloudify.interfaces.lifecycle:
+ start:
+ implementation: fabric.fabric_plugin.tasks.run_script
+ inputs:
+ script_path: scripts/k8s_tools.sh
+ process:
+ args: [{ get_input: helm_version }]
+ fabric_env:
+ host_string: { get_attribute: [ kubernetes_master_host, ip ] }
+ user: { get_input: agent_user }
+ key: { get_secret: agent_key_private }
+ relationships:
+ - type: cloudify.relationships.contained_in
+ target: kubernetes_master
+
kubernetes_master_host:
type: cloudify.openstack.nodes.Server
properties:
--- /dev/null
+#!/bin/sh
+
+# ============LICENSE_START==========================================
+# ===================================================================
+# Copyright (c) 2017 AT&T
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+#============LICENSE_END============================================
+
+# this script will install dashboard on k8s master.
+
+#install heapster
+git clone -b release-1.5 https://github.com/kubernetes/heapster.git
+
+kubectl create -f heapster/deploy/kube-config/influxdb/
+kubectl create -f heapster/deploy/kube-config/rbac/heapster-rbac.yaml
+
+#install dashboard
+kubectl apply -f https://raw.githubusercontent.com/kubernetes/dashboard/master/src/deploy/alternative/kubernetes-dashboard.yaml
+
+##Change spec.type from ClusterIP to NodePort and save.
+kubectl get svc kubernetes-dashboard --namespace=kube-system -o yaml | sed 's/type: ClusterIP/type: NodePort/' | kubectl replace -f -
+
+cat <<EOF >>dashboard-admin.yaml
+apiVersion: rbac.authorization.k8s.io/v1beta1
+kind: ClusterRoleBinding
+metadata:
+ name: kubernetes-dashboard
+ labels:
+ k8s-app: kubernetes-dashboard
+roleRef:
+ apiGroup: rbac.authorization.k8s.io
+ kind: ClusterRole
+ name: cluster-admin
+subjects:
+- kind: ServiceAccount
+ name: kubernetes-dashboard
+ namespace: kube-system
+EOF
+
+kubectl create -f dashboard-admin.yaml
+
+#install helm
+wget http://storage.googleapis.com/kubernetes-helm/helm-$1-linux-amd64.tar.gz
+tar -zxvf helm-$1-linux-amd64.tar.gz
+sudo mv linux-amd64/helm /usr/bin/helm
+
+kubectl -n kube-system create sa tiller
+kubectl create clusterrolebinding tiller --clusterrole cluster-admin --serviceaccount=kube-system:tiller
+helm init --service-account tiller
\ No newline at end of file
ONAP Operations Manager Project
###############################
+.. contents::
+ :depth: 3
+..
+
Introduction
============
Quick Start Guide
=================
-Once a kubernetes environment is available (check out `ONAP on Kubernetes <https://wiki.onap.org/display/DW/ONAP+on+Kubernetes>`__ if you're
-getting started) and the deployment artifacts have been customized for your location, ONAP is ready to be installed.Â
+Pre-requisites
+--------------
+
+* Your Kubernetes environment must be available. For more information see, `ONAP on Kubernetes <https://wiki.onap.org/display/DW/ONAP+on+Kubernetes>`__.
+* Deployment artifacts are customized for your location.
+
+**Step 1**
+
+Review and optionally change configuration parameters:
+
+Setup the `/oom/kubernetes/config/onap-parameters.yaml <https://gerrit.onap.org/r/gitweb?p=oom.git;a=blob;f=kubernetes/config/onap-parameters.yaml;h=7ddaf4d4c3dccf2fad515265f0da9c31ec0e64b1;hb=refs/heads/master>`__ file with key-value pairs specific to your OpenStack environment.
+
+OR
+
+There is a `sample <https://gerrit.onap.org/r/gitweb?p=oom.git;a=blob;f=kubernetes/config/onap-parameters-sample.yaml;h=3a74beddbbf7f9f9ec8e5a6abaecb7cb238bd519;hb=refs/heads/master>`__ that may help you out or even be usable directly if you don't intend to actually use OpenStack resources.
+
+**Step 2**
+
+In-order to be able to support multiple ONAP instances within a single kubernetes environment, a configuration set is required. To do this, execute the `createConfig.sh <https://gerrit.onap.org/r/gitweb?p=oom.git;a=blob;f=kubernetes/config/createConfig.sh;h=f226ccae47ca6de15c1da49be4b8b6de974895ed;hb=refs/heads/master>`__ script::
+
+ oom/kubernetes/config/createConfig.sh -n onap
+
+Where:
+
+* onap' refers to the name of the instance. This serves as the Namespace prefix for each deployed ONAP component (for example, onap-mso).
+
+**Step 3**
+
+The bash script `createAll.bash <https://gerrit.onap.org/r/gitweb?p=oom.git;a=blob;f=kubernetes/oneclick/createAll.bash;h=5e5f2dc76ea7739452e757282e750638b4e3e1de;hb=refs/heads/master>`__ is used to create an ONAP deployment with kubernetes. It has two primary functions:
+
+* Creating the namespaces used to encapsulate the ONAP components, and
+* Creating the services, pods and containers within each of these namespaces that provide the core functionality of ONAP.
+
+Before you execute the createAll.bash. script, pod config-init (`pod-config-init.yaml <https://gerrit.onap.org/r/gitweb?p=oom.git;a=blob;f=kubernetes/config/pod-config-init.yaml;h=b1285ce21d61815c082f6d6aa3c43d00561811c7;hb=refs/heads/master>`__) may need editing to match your environment and deployment into the default namespace.
+
+To deploy the containers and create your ONAP system, execute the following command::
+
+ oom/kubernetes/oneclick/createAll.bash -n onap
+
+Additional information on usage of createAll.bash
+-------------------------------------------------
+
+Namespaces provide isolation between ONAP components as ONAP release 1.0 contains duplicate application (for example, mariadb) and port usage.
-The first step is to setup
-the \ `/oom/kubernetes/config/onap-parameters.yaml <https://gerrit.onap.org/r/gitweb?p=oom.git;a=blob;f=kubernetes/config/onap-parameters.yaml;h=7ddaf4d4c3dccf2fad515265f0da9c31ec0e64b1;hb=refs/heads/master>`__
-file with key-value pairs specific to your OpenStack environment. Â There is a
-`sample <https://gerrit.onap.org/r/gitweb?p=oom.git;a=blob;f=kubernetes/config/onap-parameters-sample.yaml;h=3a74beddbbf7f9f9ec8e5a6abaecb7cb238bd519;hb=refs/heads/master>`__
-that may help you out or even be usable directly if you don't intend to actually use OpenStack resources.
+As such createAll.bash requires the user to enter a namespace prefix string that can be used to separate multiple deployments of onap. The result will be set of 10 namespaces (for example, onap-sdc, onap-aai, onap-mso, onap-message-router, onap-robot, onap-vid, onap-sdnc, onap-portal, onap-policy, onap-appc) being created within the kubernetes environment.
-In-order to be able to support multiple ONAP instances within a single kubernetes environment a configuration set is required.
- The `createConfig.sh <https://gerrit.onap.org/r/gitweb?p=oom.git;a=blob;f=kubernetes/config/createConfig.sh;h=f226ccae47ca6de15c1da49be4b8b6de974895ed;hb=refs/heads/master>`__ script
-is used to do this.::
+Deploying multiple ONAP instances within the same Kubernetes cluster
+--------------------------------------------------------------------
- > ./createConfig.sh -n onapTrial
+To deploy multiple ONAP instances, you must specify the number of Instances you would like to create in a Kubernetes cluster using createAllbash.
-The bash scriptÂ
-\ `createAll.bash <https://gerrit.onap.org/r/gitweb?p=oom.git;a=blob;f=kubernetes/oneclick/createAll.bash;h=5e5f2dc76ea7739452e757282e750638b4e3e1de;hb=refs/heads/master>`__Â is
-used to create an ONAP deployment with kubernetes. It has two primary
-functions:
+This is currently required due to the use of NodePort ranges. NodePorts allow external IP:Port access to containers that are running inside a Kubernetes cluster.
-- Creating the namespaces used to encapsulate the ONAP components, and
+To create multiple instances of an ONAP deployment in the cluster, use the following commands::
+
+ oom/kubernetes/config/createConfig.sh -n onap
+
+ oom/kubernetes/oneclick/createAll.bash -n onap -i 2
+
+Where:
-- Creating the services, pods and containers within each of these
- namespaces that provide the core functionality of ONAP.
+* 'onap' refers to the name of the instance.
+* ‘i 2’ refers to the number of instances of an ONAP deployment in the cluster.
-To deploy the containers and create your ONAP system enter::
+To delete a deployed instance
+-----------------------------
- > ./createAll.bash -n onapTrial
+To delete a deployed instance, use the following command::
+
+ oom/kubernetes/oneclick/deleteAll.bash -n onap
+
+**Note**: Deleting the runtime containers does not remove the configuration created in step 2.
-Namespaces provide isolation between ONAP components as ONAP release 1.0
-contains duplicate application (e.g. mariadb) and port usage. As
-such createAll.bash requires the user to enter a namespace prefix string
-that can be used to separate multiple deployments of onap. The result
-will be set of 10 namespaces (e.g. onapTrial-sdc, onapTrial-aai,
-onapTrial-mso, onapTrial-message-router, onapTrial-robot, onapTrial-vid,
-onapTrial-sdnc, onapTrial-portal, onapTrial-policy, onapTrial-appc)
-being created within the kubernetes environment. Â A prerequisite pod
-config-init (\ `pod-config-init.yaml <https://gerrit.onap.org/r/gitweb?p=oom.git;a=blob;f=kubernetes/config/pod-config-init.yaml;h=b1285ce21d61815c082f6d6aa3c43d00561811c7;hb=refs/heads/master>`__)
-may need editing to match your environment and deployment into the
-default namespace before running createAll.bash.
+For more information on OOM project documentation, refer to `Quick Start Guide on Wiki <https://wiki.onap.org/display/DW/ONAP+Operations+Manager+Project#ONAPOperationsManagerProject-QuickStartGuide>`__.
Demo Video
----------
PARENT_CHART := onap
COMMON_CHARTS_DIR := common
-
+SETUP_CHARTS_DIR := setup
# FIXME OOM-765
ROOT_DIR := $(shell dirname $(realpath $(lastword $(MAKEFILE_LIST))))
OUTPUT_DIR := $(ROOT_DIR)/dist
PACKAGE_DIR := $(OUTPUT_DIR)/packages
SECRET_DIR := $(OUTPUT_DIR)/secrets
-EXCLUDES := $(COMMON_CHARTS_DIR) config oneclick readiness test dist $(PARENT_CHART) dcae
-HELM_CHARTS := $(COMMON_CHARTS_DIR) $(filter-out $(EXCLUDES), $(patsubst %/.,%,$(wildcard */.))) $(PARENT_CHART)
+EXCLUDES := $(SETUP_CHARTS_DIR) $(COMMON_CHARTS_DIR) config oneclick readiness test dist $(PARENT_CHART) dcae
+HELM_CHARTS := $(filter-out $(EXCLUDES), $(patsubst %/.,%,$(wildcard */.))) $(PARENT_CHART)
.PHONY: $(EXCLUDES) $(HELM_CHARTS)
-all: $(HELM_CHARTS)
+all: $(COMMON_CHARTS_DIR) $(SETUP_CHARTS_DIR) $(HELM_CHARTS)
+
+common:
+ @echo "\n[$@]"
+ @make package-$@
+
+setup:
+ @echo "\n[$@]"
+ @make package-$@
$(HELM_CHARTS):
@echo "\n[$@]"
kind: ConfigMap
metadata:
name: aaf-data-configmap
- namespace: {{ .Values.nsPrefix }}-aaf
+ namespace: {{ .Values.nsPrefix }}
data:
{{ (.Files.Glob "resources/config/aaf-data/*").AsConfig | indent 2 }}
#{{ end }}
kind: Deployment
metadata:
name: aaf-cs
- namespace: "{{ .Values.nsPrefix }}-aaf"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
- replicas: 1
+ replicas: {{ .Values.aafcsReplicas }}
selector:
matchLabels:
app: aaf-cs
labels:
app: aaf
name: aaf
- namespace: "{{ .Values.nsPrefix }}-aaf"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
+ replicas: {{ .Values.aafReplicas }}
selector:
matchLabels:
app: aaf
kind: Secret
metadata:
name: aaf-cs-data-secret
- namespace: {{ .Values.nsPrefix }}-aaf
+ namespace: {{ .Values.nsPrefix }}
type: Opaque
data:
{{ (.Files.Glob "resources/config/aaf-cs-data/*").AsSecrets | indent 2 }}
kind: Service
metadata:
name: aaf-cs
- namespace: "{{ .Values.nsPrefix }}-aaf"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: aaf-cs
spec:
kind: Service
metadata:
name: aaf-authz-service
- namespace: "{{ .Values.nsPrefix }}-aaf"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: aaf-authz-service
spec:
nsPrefix: onap
pullPolicy: Always
nodePortPrefix: 302
+aafcsReplicas: 1
+aafReplicas: 1
image:
readiness: oomk8s/readiness-check:1.1.0
aafImage: nexus3.onap.org:10001/onap/aaf/authz-service
"PROJECT_HOME": "/opt/app/aai-traversal",
"LOGROOT": "/opt/aai/logroot",
"JAVA_HOME": "/usr/lib/jvm/java-8-openjdk-amd64",
- "AAI_SERVER_URL_BASE": "https://aai-service.onap-aai:8443/aai/",
- "AAI_SERVER_URL": "https://aai-service.onap-aai:8443/aai/v11/",
- "AAI_GLOBAL_CALLBACK_URL": "https://aai-service.onap-aai:8443/aai/",
+ "AAI_SERVER_URL_BASE": "https://aai-servicei.{{.Values.nsPrefix}}:8443/aai/",
+ "AAI_SERVER_URL": "https://aai-service.{{.Values.nsPrefix}}:8443/aai/v11/",
+ "AAI_GLOBAL_CALLBACK_URL": "https://aai-service.{{.Values.nsPrefix}}:8443/aai/",
"AAI_TRUSTSTORE_FILENAME": "aai_keystore",
"AAI_TRUSTSTORE_PASSWD_X": "OBF:1vn21ugu1saj1v9i1v941sar1ugw1vo0",
"AAI_KEYSTORE_FILENAME": "aai_keystore",
"AAI_KEYSTORE_PASSWD_X": "OBF:1vn21ugu1saj1v9i1v941sar1ugw1vo0",
- "APPLICATION_SERVERS": "aai-service.onap-aai",
+ "APPLICATION_SERVERS": "aai-service.{{.Values.nsPrefix}}",
"AAI_DMAAP_PROTOCOL": "http",
- "AAI_DMAAP_HOST_PORT": "dmaap.onap-message-router:3904",
+ "AAI_DMAAP_HOST_PORT": "dmaap.{{.Values.nsPrefix}}:3904",
"AAI_DMAAP_TOPIC_NAME": "AAI-EVENT",
"AAI_NOTIFICATION_EVENT_DEFAULT_EVENT_STATUS": "UNPROCESSED",
"AAI_NOTIFICATION_EVENT_DEFAULT_EVENT_TYPE": "AAI-EVENT",
"AAI_NOTIFICATION_CURRENT_VERSION": "v11",
"RESOURCE_VERSION_ENABLE_FLAG": "true",
"TXN_HBASE_TABLE_NAME": "aailogging.dev",
- "TXN_ZOOKEEPER_QUORUM": "hbase.onap-aai",
+ "TXN_ZOOKEEPER_QUORUM": "hbase.{{.Values.nsPrefix}}",
"TXN_ZOOKEEPER_PROPERTY_CLIENTPORT": "2181",
"TXN_HBASE_ZOOKEEPER_ZNODE_PARENT": "/hbase",
"AAI_WORKLOAD_PREFERRED_ROUTE_KEY": "MR1",
- "STORAGE_HOSTNAME": "hbase.onap-aai",
+ "STORAGE_HOSTNAME": "hbase.{{.Values.nsPrefix}}",
"STORAGE_HBASE_TABLE": "aaigraph.dev",
"STORAGE_HBASE_ZOOKEEPER_ZNODE_PARENT": "/hbase",
"DB_CACHE_CLEAN_WAIT": "20",
"PROJECT_HOME": "/opt/app/aai-resources",
"LOGROOT": "/opt/aai/logroot",
"JAVA_HOME": "/usr/lib/jvm/java-8-openjdk-amd64",
- "AAI_SERVER_URL_BASE": "https://aai-service.onap-aai:8443/aai/",
- "AAI_SERVER_URL": "https://aai-service.onap-aai:8443/aai/v11/",
- "AAI_GLOBAL_CALLBACK_URL": "https://aai-service.onap-aai:8443/aai/",
+ "AAI_SERVER_URL_BASE": "https://aai-service.{{.Values.nsPrefix}}:8443/aai/",
+ "AAI_SERVER_URL": "https://aai-service.{{.Values.nsPrefix}}:8443/aai/v11/",
+ "AAI_GLOBAL_CALLBACK_URL": "https://aai-service.{{.Values.nsPrefix}}:8443/aai/",
"AAI_TRUSTSTORE_FILENAME": "aai_keystore",
"AAI_TRUSTSTORE_PASSWD_X": "OBF:1vn21ugu1saj1v9i1v941sar1ugw1vo0",
"AAI_KEYSTORE_FILENAME": "aai_keystore",
"AAI_KEYSTORE_PASSWD_X": "OBF:1vn21ugu1saj1v9i1v941sar1ugw1vo0",
- "APPLICATION_SERVERS": "aai-service.onap-aai",
+ "APPLICATION_SERVERS": "aai-service.{{.Values.nsPrefix}}",
"AAI_DMAAP_PROTOCOL": "http",
- "AAI_DMAAP_HOST_PORT": "dmaap.onap-message-router:3904",
+ "AAI_DMAAP_HOST_PORT": "dmaap.{{.Values.nsPrefix}}:3904",
"AAI_DMAAP_TOPIC_NAME": "AAI-EVENT",
"AAI_NOTIFICATION_EVENT_DEFAULT_EVENT_STATUS": "UNPROCESSED",
"AAI_NOTIFICATION_EVENT_DEFAULT_EVENT_TYPE": "AAI-EVENT",
"AAI_NOTIFICATION_CURRENT_VERSION": "v11",
"RESOURCE_VERSION_ENABLE_FLAG": "true",
"TXN_HBASE_TABLE_NAME": "aailogging.dev",
- "TXN_ZOOKEEPER_QUORUM": "hbase.onap-aai",
+ "TXN_ZOOKEEPER_QUORUM": "hbase.{{.Values.nsPrefix}}",
"TXN_ZOOKEEPER_PROPERTY_CLIENTPORT": "2181",
"TXN_HBASE_ZOOKEEPER_ZNODE_PARENT": "/hbase",
"AAI_WORKLOAD_PREFERRED_ROUTE_KEY": "MR1",
- "STORAGE_HOSTNAME": "hbase.onap-aai",
+ "STORAGE_HOSTNAME": "hbase.{{.Values.nsPrefix}}",
"STORAGE_HBASE_TABLE": "aaigraph.dev",
"STORAGE_HBASE_ZOOKEEPER_ZNODE_PARENT": "/hbase",
"DB_CACHE_CLEAN_WAIT": "20",
<route xmlns="http://camel.apache.org/schema/spring" trace="true">
- <from uri="event-bus:mybus/?eventTopic=AAI-EVENT&groupName=datarouter&groupId=datarouter&url=http://dmaap.onap-message-router:3904"/>
+ <from uri="event-bus:mybus/?eventTopic=AAI-EVENT&groupName=datarouter&groupId=datarouter&url=http://dmaap.{{.Values.nsPrefix}}:3904"/>
<to uri="bean:entityEventPolicy?method=process"/>
</route>
balance roundrobin
http-request set-header X-Forwarded-Port %[src_port]
http-response set-header Strict-Transport-Security max-age=16000000;\ includeSubDomains;\ preload;
- server aai-resources.onap-aai aai-resources.onap-aai:8447 port 8447 ssl verify none
+ server aai-resources.{{.Values.nsPrefix}} aai-resources.{{.Values.nsPrefix}}:8447 port 8447 ssl verify none
#######################
# BACKEND 8446#########
balance roundrobin
http-request set-header X-Forwarded-Port %[src_port]
http-response set-header Strict-Transport-Security max-age=16000000;\ includeSubDomains;\ preload;
- server aai-traversal.onap-aai aai-traversal.onap-aai:8446 port 8446 ssl verify none
+ server aai-traversal.{{.Values.nsPrefix}} aai-traversal.{{.Values.nsPrefix}}:8446 port 8446 ssl verify none
listen IST_AAI_STATS
mode http
#List of logstash server ip addresses with port number.
#But, in our case, this will be the loadbalancer IP address.
#For the below property to work the loadbalancer or logstash should expose 5044 port to listen the filebeat events or port in the property should be changed appropriately.
- hosts: ["logstash.onap-log:5044"]
+ hosts: ["logstash.{{.Values.nsPrefix}}:5044"]
#If enable will do load balancing among availabe Logstash, automatically.
loadbalance: true
# Model Loader Distribution Client Configuration
ml.distribution.ACTIVE_SERVER_TLS_AUTH=false
-ml.distribution.ASDC_ADDRESS=sdc-be.onap-sdc:8443
+ml.distribution.ASDC_ADDRESS=sdc-be.{{.Values.nsPrefix}}:8443
ml.distribution.CONSUMER_GROUP=aai-ml-group
ml.distribution.CONSUMER_ID=aai-ml
ml.distribution.ENVIRONMENT_NAME=AUTO
ml.distribution.ARTIFACT_TYPES=MODEL_INVENTORY_PROFILE,MODEL_QUERY_SPEC,VNF_CATALOG
# Model Loader AAI REST Client Configuration
-ml.aai.BASE_URL=https://aai-service.onap-aai:8443
+ml.aai.BASE_URL=https://aai-service.{{.Values.nsPrefix}}:8443
ml.aai.MODEL_URL=/aai/v10/service-design-and-creation/models/model/
ml.aai.NAMED_QUERY_URL=/aai/v10/service-design-and-creation/named-queries/named-query/
ml.aai.VNF_IMAGE_URL=/aai/v8/service-design-and-creation/vnf-images
# ElasticSearch Configuration
es.cluster-name=ES_AAI
-es.ip-address=elasticsearch.onap-aai
+es.ip-address=aai-elasticsearch.{{.Values.nsPrefix}}
es.http-port=9200
#
# The ip address/hostname and port to the desired AAI instance
#
-aai.rest.host=aai-service.onap-aai
+aai.rest.host=aai-service.{{.Values.nsPrefix}}
aai.rest.port=8443
############################## REST ##############################
aai.taskProcessor.maxConcurrentWorkers=5
aai.taskProcessor.transactionRateControllerEnabled=false
aai.taskProcessor.numSamplesPerThreadForRunningAverage=100
-aai.taskProcessor.targetTPS=100
\ No newline at end of file
+aai.taskProcessor.targetTPS=100
# The ip address/hostname and port to the desired AAI instance
# For development it's recommended to use a local instance of ES
#
-elasticsearch.ipAddress=elasticsearch.onap-aai
+elasticsearch.ipAddress=aai-elasticsearch.{{.Values.nsPrefix}}
elasticsearch.httpPort=9200
elasticsearch.javaApiPort=8443
elasticsearch.taskProcessor.transactionRateControllerEnabled=false
elasticsearch.taskProcessor.numSamplesPerThreadForRunningAverage=100
elasticsearch.taskProcessor.targetTPS=100
-elasticsearch.clusterName=ES_AAI_LOCALHOST
\ No newline at end of file
+elasticsearch.clusterName=ES_AAI_LOCALHOST
# Instance of ECOMP Portal where the app has been on-boarded
# use insecure http for dev purposes to avoid self-signed certificate
-ecomp_rest_url = http://portalapps.onap-portal:8989/ONAPPORTAL/auxapi
+ecomp_rest_url = http://portalapps.{{.Values.nsPrefix}}:8989/ONAPPORTAL/auxapi
# Standard global logon page
-ecomp_redirect_url = http://portalapps.onap-portal:8989/ONAPPORTAL/login.htm
+ecomp_redirect_url = http://portalapps.{{.Values.nsPrefix}}:8989/ONAPPORTAL/login.htm
# Name of cookie to extract on login request
csp_cookie_name = EPService
#
# The ip address/hostname and port to the desired Search Data Service instance
#
-search-service.ipAddress=search-data-service.onap-aai
+search-service.ipAddress=search-data-service.{{.Values.nsPrefix}}
search-service.httpPort=9509
############################## Indexes ##############################
############################## Certs ##############################
search-service.ssl.cert-name=client-cert-onap.p12
search-service.ssl.keystore-password=OBF:1y0q1uvc1uum1uvg1pil1pjl1uuq1uvk1uuu1y10
-search-service.ssl.keystore=tomcat_keystore
\ No newline at end of file
+search-service.ssl.keystore=tomcat_keystore
kind: ConfigMap
metadata:
name: aai-deployment-configmap
- namespace: {{ .Values.nsPrefix }}-aai
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/haproxy/*").AsConfig . | indent 2 }}
#{{ end }}
kind: Deployment
metadata:
name: aai-service
- namespace: "{{ .Values.nsPrefix }}-aai"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.aaiServiceReplicas }}
selector:
kind: ConfigMap
metadata:
name: aai-filebeat-configmap
- namespace: {{ .Values.nsPrefix }}-aai
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/log/filebeat/*").AsConfig . | indent 2 }}
#{{ end }}
kind: Deployment
metadata:
name: aai-resources
- namespace: "{{ .Values.nsPrefix }}-aai"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.aaiResourceReplicas }}
selector:
labels:
app: aai-resources
name: aai-resources
+ annotations:
+ msb.onap.org/service-info: '[
+ {
+ "serviceName": "_aai-cloudInfrastructure",
+ "version": "v11",
+ "url": "/aai/v11/cloud-infrastructure",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v11/cloud-infrastructure"
+ },
+ {
+ "serviceName": "_aai-cloudInfrastructure",
+ "version": "v12",
+ "url": "/aai/v12/cloud-infrastructure",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v12/cloud-infrastructure"
+ },
+ {
+ "serviceName": "_aai-cloudInfrastructure",
+ "version": "v13",
+ "url": "/aai/v13/cloud-infrastructure",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v13/cloud-infrastructure"
+ },
+ {
+ "serviceName": "_aai-business",
+ "version": "v11",
+ "url": "/aai/v11/business",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v11/business"
+ },
+ {
+ "serviceName": "_aai-business",
+ "version": "v12",
+ "url": "/aai/v12/business",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v12/business"
+ },
+ {
+ "serviceName": "_aai-business",
+ "version": "v13",
+ "url": "/aai/v13/business",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v13/business"
+ },
+ {
+ "serviceName": "_aai-actions",
+ "version": "v11",
+ "url": "/aai/v11/actions",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v11/actions"
+ },
+ {
+ "serviceName": "_aai-actions",
+ "version": "v12",
+ "url": "/aai/v12/actions",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v12/actions"
+ },
+ {
+ "serviceName": "_aai-actions",
+ "version": "v13",
+ "url": "/aai/v13/actions",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v13/actions"
+ },
+ {
+ "serviceName": "_aai-service-design-and-creation",
+ "version": "v11",
+ "url": "/aai/v11/service-design-and-creation",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v11/service-design-and-creation"
+ },
+ {
+ "serviceName": "_aai-service-design-and-creation",
+ "version": "v12",
+ "url": "/aai/v12/service-design-and-creation",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v12/service-design-and-creation"
+ },
+ {
+ "serviceName": "_aai-service-design-and-creation",
+ "version": "v13",
+ "url": "/aai/v13/service-design-and-creation",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v13/service-design-and-creation"
+ },
+ {
+ "serviceName": "_aai-network",
+ "version": "v11",
+ "url": "/aai/v11/network",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v11/network"
+ },
+ {
+ "serviceName": "_aai-network",
+ "version": "v12",
+ "url": "/aai/v12/network",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v12/network"
+ },
+ {
+ "serviceName": "_aai-network",
+ "version": "v13",
+ "url": "/aai/v13/network",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v13/network"
+ },
+ {
+ "serviceName": "_aai-externalSystem",
+ "version": "v11",
+ "url": "/aai/v11/external-system",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v11/external-system"
+ },
+ {
+ "serviceName": "_aai-externalSystem",
+ "version": "v12",
+ "url": "/aai/v12/external-system",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v12/external-system"
+ },
+ {
+ "serviceName": "_aai-externalSystem",
+ "version": "v13",
+ "url": "/aai/v13/external-system",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v13/external-system"
+ },
+ {
+ "serviceName": "aai-cloudInfrastructure",
+ "version": "v11",
+ "url": "/aai/v11/cloud-infrastructure",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-cloudInfrastructure",
+ "version": "v12",
+ "url": "/aai/v12/cloud-infrastructure",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-cloudInfrastructure",
+ "version": "v13",
+ "url": "/aai/v13/cloud-infrastructure",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-business",
+ "version": "v11",
+ "url": "/aai/v11/business",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-business",
+ "version": "v12",
+ "url": "/aai/v12/business",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-business",
+ "version": "v13",
+ "url": "/aai/v13/business",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-actions",
+ "version": "v11",
+ "url": "/aai/v11/actions",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-actions",
+ "version": "v12",
+ "url": "/aai/v12/actions",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-actions",
+ "version": "v13",
+ "url": "/aai/v13/actions",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-service-design-and-creation",
+ "version": "v11",
+ "url": "/aai/v11/service-design-and-creation",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-service-design-and-creation",
+ "version": "v12",
+ "url": "/aai/v12/service-design-and-creation",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-service-design-and-creation",
+ "version": "v13",
+ "url": "/aai/v13/service-design-and-creation",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-network",
+ "version": "v11",
+ "url": "/aai/v11/network",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-network",
+ "version": "v12",
+ "url": "/aai/v12/network",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-network",
+ "version": "v13",
+ "url": "/aai/v13/network",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-externalSystem",
+ "version": "v11",
+ "url": "/aai/v11/external-system",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-externalSystem",
+ "version": "v12",
+ "url": "/aai/v12/external-system",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-externalSystem",
+ "version": "v13",
+ "url": "/aai/v13/external-system",
+ "protocol": "REST",
+ "port": "8447",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ }
+ ]'
spec:
initContainers:
- command:
kind: ConfigMap
metadata:
name: aai-chef-config-configmap
- namespace: {{ .Values.nsPrefix }}-aai
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/aai-data/chef-config/dev/.knife/solo.rb").AsConfig . | indent 2 }}
---
kind: ConfigMap
metadata:
name: aai-resources-environments-configmap
- namespace: {{ .Values.nsPrefix }}-aai
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/aai-data/environments/*").AsConfig . | indent 2 }}
---
kind: ConfigMap
metadata:
name: aai-resources-log-configmap
- namespace: {{ .Values.nsPrefix }}-aai
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/log/resources/logback.xml").AsConfig . | indent 2 }}
---
kind: ConfigMap
metadata:
name: aai-traversal-log-configmap
- namespace: {{ .Values.nsPrefix }}-aai
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/log/traversal/logback.xml").AsConfig . | indent 2 }}
#{{ end }}
kind: Deployment
metadata:
name: aai-traversal
- namespace: "{{ .Values.nsPrefix }}-aai"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.aaiTraversalReplicas }}
selector:
labels:
app: aai-traversal
name: aai-traversal
+ annotations:
+ msb.onap.org/service-info: '[
+ {
+ "serviceName": "_aai-generic-query",
+ "version": "v11",
+ "url": "/aai/v11/search/generic-query",
+ "protocol": "REST",
+ "port": "8446",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v11/search/generic-query"
+ },
+ {
+ "serviceName": "_aai-generic-query",
+ "version": "v12",
+ "url": "/aai/v12/search/generic-query",
+ "protocol": "REST",
+ "port": "8446",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v12/search/generic-query"
+ },
+ {
+ "serviceName": "_aai-generic-query",
+ "version": "v13",
+ "url": "/aai/v13/search/generic-query",
+ "protocol": "REST",
+ "port": "8446",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v13/search/generic-query"
+ },
+ {
+ "serviceName": "_aai-nodes-query",
+ "version": "v11",
+ "url": "/aai/v11/search/nodes-query",
+ "protocol": "REST",
+ "port": "8446",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v11/search/nodes-query"
+ },
+ {
+ "serviceName": "_aai-nodes-query",
+ "version": "v12",
+ "url": "/aai/v12/search/nodes-query",
+ "protocol": "REST",
+ "port": "8446",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v12/search/nodes-query"
+ },
+ {
+ "serviceName": "_aai-nodes-query",
+ "version": "v13",
+ "url": "/aai/v13/search/nodes-query",
+ "protocol": "REST",
+ "port": "8446",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v13/search/nodes-query"
+ },
+ {
+ "serviceName": "_aai-query",
+ "version": "v11",
+ "url": "/aai/v11/query",
+ "protocol": "REST",
+ "port": "8446",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v11/query"
+ },
+ {
+ "serviceName": "_aai-query",
+ "version": "v12",
+ "url": "/aai/v12/query",
+ "protocol": "REST",
+ "port": "8446",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v12/query"
+ },
+ {
+ "serviceName": "_aai-query",
+ "version": "v13",
+ "url": "/aai/v13/query",
+ "protocol": "REST",
+ "port": "8446",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/v13/query"
+ },
+ {
+ "serviceName": "_aai-named-query",
+ "url": "/aai/search",
+ "protocol": "REST",
+ "port": "8446",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1",
+ "path": "/aai/search"
+ },
+ {
+ "serviceName": "aai-generic-query",
+ "version": "v11",
+ "url": "/aai/v11/search/generic-query",
+ "protocol": "REST",
+ "port": "8446",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-generic-query",
+ "version": "v12",
+ "url": "/aai/v12/search/generic-query",
+ "protocol": "REST",
+ "port": "8446",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-generic-query",
+ "version": "v13",
+ "url": "/aai/v13/search/generic-query",
+ "protocol": "REST",
+ "port": "8446",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-nodes-query",
+ "version": "v11",
+ "url": "/aai/v11/search/nodes-query",
+ "protocol": "REST",
+ "port": "8446",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-nodes-query",
+ "version": "v12",
+ "url": "/aai/v12/search/nodes-query",
+ "protocol": "REST",
+ "port": "8446",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-nodes-query",
+ "version": "v13",
+ "url": "/aai/v13/search/nodes-query",
+ "protocol": "REST",
+ "port": "8446",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-query",
+ "version": "v11",
+ "url": "/aai/v11/query",
+ "protocol": "REST",
+ "port": "8446",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-query",
+ "version": "v12",
+ "url": "/aai/v12/query",
+ "protocol": "REST",
+ "port": "8446",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-query",
+ "version": "v13",
+ "url": "/aai/v13/query",
+ "protocol": "REST",
+ "port": "8446",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ },
+ {
+ "serviceName": "aai-named-query",
+ "url": "/aai/search",
+ "protocol": "REST",
+ "port": "8446",
+ "enable_ssl": true,
+ "lb_policy":"ip_hash",
+ "visualRange": "1"
+ }
+ ]'
spec:
initContainers:
- command:
- name: CHEF_GIT_URL
value: http://gerrit.onap.org/r/aai
- name: RESOURCES_HOSTNAME
- value: aai-resources.{{ .Values.nsPrefix }}-aai
+ value: aai-resources.{{ .Values.nsPrefix }}
volumeMounts:
- mountPath: /etc/localtime
name: localtime
kind: Service
metadata:
name: hbase
- namespace: "{{ .Values.nsPrefix }}-aai"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: hbase
spec:
kind: Service
metadata:
name: aai-service
- namespace: "{{ .Values.nsPrefix }}-aai"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: aai-service
- annotations:
- msb.onap.org/service-info: '[
- {
- "serviceName": "aai-cloudInfrastructure",
- "version": "v11",
- "url": "/aai/v11/cloud-infrastructure",
- "protocol": "REST",
- "port": "8443",
- "enable_ssl": true,
- "visualRange": "1"
- },
- {
- "serviceName": "aai-cloudInfrastructure-deprecated",
- "version": "v11",
- "url": "/aai/v11/cloud-infrastructure",
- "protocol": "REST",
- "port": "8443",
- "enable_ssl": true,
- "visualRange": "1",
- "path": "/aai/v11/cloud-infrastructure"
- },
- {
- "serviceName": "aai-business",
- "version": "v11",
- "url": "/aai/v11/business",
- "protocol": "REST",
- "port": "8443",
- "enable_ssl": true,
- "visualRange": "1"
- },
- {
- "serviceName": "aai-business-deprecated",
- "version": "v11",
- "url": "/aai/v11/business",
- "protocol": "REST",
- "port": "8443",
- "enable_ssl": true,
- "visualRange": "1",
- "path": "/aai/v11/business"
- },
- {
- "serviceName": "aai-search",
- "version": "v11",
- "url": "/aai/v11/search",
- "protocol": "REST",
- "port": "8443",
- "enable_ssl": true,
- "visualRange": "1"
- },
- {
- "serviceName": "aai-search-deprecated",
- "version": "v11",
- "url": "/aai/v11/search",
- "protocol": "REST",
- "port": "8443",
- "enable_ssl": true,
- "visualRange": "1",
- "path": "/aai/v11/search"
- },
- {
- "serviceName": "aai-actions",
- "version": "v11",
- "url": "/aai/v11/actions",
- "protocol": "REST",
- "port": "8443",
- "enable_ssl": true,
- "visualRange": "1"
- },
- {
- "serviceName": "aai-actions-deprecated",
- "version": "v11",
- "url": "/aai/v11/actions",
- "protocol": "REST",
- "port": "8443",
- "enable_ssl": true,
- "visualRange": "1",
- "path": "/aai/v11/actions"
- },
- {
- "serviceName": "aai-service-design-and-creation",
- "version": "v11",
- "url": "/aai/v11/service-design-and-creation",
- "protocol": "REST",
- "port": "8443",
- "enable_ssl": true,
- "visualRange": "1"
- },
- {
- "serviceName": "aai-service-design-and-creation-deprecated",
- "version": "v11",
- "url": "/aai/v11/service-design-and-creation",
- "protocol": "REST",
- "port": "8443",
- "enable_ssl": true,
- "visualRange": "1",
- "path": "/aai/v11/service-design-and-creation"
- },
- {
- "serviceName": "aai-network",
- "version": "v11",
- "url": "/aai/v11/network",
- "protocol": "REST",
- "port": "8443",
- "enable_ssl": true,
- "visualRange": "1"
- },
- {
- "serviceName": "aai-network-deprecated",
- "version": "v11",
- "url": "/aai/v11/network",
- "protocol": "REST",
- "port": "8443",
- "enable_ssl": true,
- "visualRange": "1",
- "path": "/aai/v11/network"
- },
- {
- "serviceName": "aai-externalSystem",
- "version": "v11",
- "url": "/aai/v11/external-system",
- "protocol": "REST",
- "port": "8443",
- "enable_ssl": true,
- "visualRange": "1"
- },
- {
- "serviceName": "aai-externalSystem-deprecated",
- "version": "v11",
- "url": "/aai/v11/external-system",
- "protocol": "REST",
- "port": "8443",
- "enable_ssl": true,
- "visualRange": "1",
- "path": "/aai/v11/external-system"
- }
- ]'
spec:
ports:
- name: "aai-service-port-8443"
kind: Service
metadata:
name: model-loader-service
- namespace: "{{ .Values.nsPrefix }}-aai"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: model-loader-service
spec:
kind: Service
metadata:
name: gremlin
- namespace: "{{ .Values.nsPrefix }}-aai"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: gremlin
spec:
apiVersion: v1
kind: Service
metadata:
- name: elasticsearch
- namespace: "{{ .Values.nsPrefix }}-aai"
+ name: aai-elasticsearch
+ namespace: "{{ .Values.nsPrefix }}"
labels:
- app: elasticsearch
+ app: aai-elasticsearch
spec:
ports:
- - name: "elasticsearch-port"
+ - name: "aai-elasticsearch-port"
port: 9200
selector:
- app: elasticsearch
+ app: aai-elasticsearch
clusterIP: None
#{{ end }}
#{{ if not .Values.disableAaiSearchDataService }}
kind: Service
metadata:
name: search-data-service
- namespace: "{{ .Values.nsPrefix }}-aai"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: search-data-service
spec:
kind: Service
metadata:
name: aai-traversal
- namespace: "{{ .Values.nsPrefix }}-aai"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: aai-traversal
spec:
kind: Service
metadata:
name: aai-resources
- namespace: "{{ .Values.nsPrefix }}-aai"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: aai-resources
spec:
kind: Service
metadata:
name: sparky-be
- namespace: "{{ .Values.nsPrefix }}-aai"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: sparky-be
spec:
selector:
app: sparky-be
clusterIP: None
-#{{ end }}
\ No newline at end of file
+#{{ end }}
apiVersion: v1
kind: ConfigMap
metadata:
- name: data-router-prop-configmap
- namespace: {{ .Values.nsPrefix }}-aai
+ name: aai-data-router-prop-configmap
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/data-router/appconfig/data-router.properties").AsConfig . | indent 2 }}
---
apiVersion: v1
kind: ConfigMap
metadata:
- name: data-router-model-v8-configmap
- namespace: {{ .Values.nsPrefix }}-aai
+ name: aai-data-router-model-v8-configmap
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/data-router/appconfig/model/aai_oxm_v8.xml").AsConfig . | indent 2 }}
---
apiVersion: v1
kind: ConfigMap
metadata:
- name: data-router-model-v9-configmap
- namespace: {{ .Values.nsPrefix }}-aai
+ name: aai-data-router-model-v9-configmap
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/data-router/appconfig/model/aai_oxm_v9.xml").AsConfig . | indent 2 }}
---
apiVersion: v1
kind: ConfigMap
metadata:
- name: data-router-model-v10-configmap
- namespace: {{ .Values.nsPrefix }}-aai
+ name: aai-data-router-model-v10-configmap
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/data-router/appconfig/model/aai_oxm_v10.xml").AsConfig . | indent 2 }}
---
apiVersion: v1
kind: ConfigMap
metadata:
- name: data-router-model-v11-configmap
- namespace: {{ .Values.nsPrefix }}-aai
+ name: aai-data-router-model-v11-configmap
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/data-router/appconfig/model/aai_oxm_v11.xml").AsConfig . | indent 2 }}
---
apiVersion: v1
kind: Secret
metadata:
- name: data-router-secret
- namespace: {{ .Values.nsPrefix }}-aai
+ name: aai-data-router-secret
+ namespace: {{ .Values.nsPrefix }}
type: Opaque
data:
{{ tpl (.Files.Glob "resources/config/data-router/appconfig/auth/*").AsSecrets . | indent 2 }}
apiVersion: v1
kind: ConfigMap
metadata:
- name: data-router-dynamic-configmap
- namespace: {{ .Values.nsPrefix }}-aai
+ name: aai-data-router-dynamic-configmap
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/data-router/dynamic/routes/entity-event.route").AsConfig . | indent 2 }}
{{ tpl (.Files.Glob "resources/config/data-router/dynamic/conf/entity-event-policy.xml").AsConfig . | indent 2 }}
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
- name: data-router
- namespace: "{{ .Values.nsPrefix }}-aai"
+ name: aai-data-router
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.dataRouterReplicas }}
selector:
matchLabels:
- app: data-router
+ app: aai-data-router
template:
metadata:
labels:
- app: data-router
- name: data-router
+ app: aai-data-router
+ name: aai-data-router
spec:
initContainers:
- command:
imagePullPolicy: {{ .Values.pullPolicy }}
name: init-sysctl
volumeMounts:
- - name: data-router-logs
+ - name: aai-data-router-logs
mountPath: /logroot/
containers:
- - name: data-router
+ - name: aai-data-router
image: "{{ .Values.image.dataRouterImage }}:{{ .Values.image.dataRouterVersion }}"
imagePullPolicy: {{ .Values.pullPolicy }}
env:
readOnly: true
- mountPath: /opt/app/data-router/config/model/aai_oxm_v8.xml
subPath: aai_oxm_v8.xml
- name: data-router-model-v8
+ name: aai-data-router-model-v8
- mountPath: /opt/app/data-router/config/model/aai_oxm_v9.xml
subPath: aai_oxm_v9.xml
- name: data-router-model-v9
+ name: aai-data-router-model-v9
- mountPath: /opt/app/data-router/config/model/aai_oxm_v10.xml
subPath: aai_oxm_v10.xml
- name: data-router-model-v10
+ name: aai-data-router-model-v10
- mountPath: /opt/app/data-router/config/model/aai_oxm_v11.xml
subPath: aai_oxm_v11.xml
- name: data-router-model-v11
+ name: aai-data-router-model-v11
- mountPath: /opt/app/data-router/config/auth
- name: data-router-auth
+ name: aai-data-router-auth
- mountPath: /opt/app/data-router/config/data-router.properties
- name: data-router-properties
+ name: aai-data-router-properties
subPath: data-router.properties
- mountPath: /opt/app/data-router/dynamic/routes/entity-event.route
subPath: entity-event.route
- name: data-router-dynamic-route
+ name: aai-data-router-dynamic-route
- mountPath: /opt/app/data-router/dynamic/conf/entity-event-policy.xml
subPath: entity-event-policy.xml
- name: data-router-dynamic-policy
+ name: aai-data-router-dynamic-policy
- mountPath: /logs/
- name: data-router-logs
+ name: aai-data-router-logs
ports:
- containerPort: 9502
readinessProbe:
- name: localtime
hostPath:
path: /etc/localtime
- - name: data-router-model-v8
+ - name: aai-data-router-model-v8
configMap:
- name: data-router-model-v8-configmap
- - name: data-router-model-v9
+ name: aai-data-router-model-v8-configmap
+ - name: aai-data-router-model-v9
configMap:
- name: data-router-model-v9-configmap
- - name: data-router-model-v10
+ name: aai-data-router-model-v9-configmap
+ - name: aai-data-router-model-v10
configMap:
- name: data-router-model-v10-configmap
- - name: data-router-model-v11
+ name: aai-data-router-model-v10-configmap
+ - name: aai-data-router-model-v11
configMap:
- name: data-router-model-v11-configmap
- - name: data-router-auth
+ name: aai-data-router-model-v11-configmap
+ - name: aai-data-router-auth
secret:
- secretName: data-router-secret
- - name: data-router-properties
+ secretName: aai-data-router-secret
+ - name: aai-data-router-properties
configMap:
- name: data-router-prop-configmap
- - name: data-router-dynamic-route
+ name: aai-data-router-prop-configmap
+ - name: aai-data-router-dynamic-route
configMap:
- name: data-router-dynamic-configmap
- - name: data-router-dynamic-policy
+ name: aai-data-router-dynamic-configmap
+ - name: aai-data-router-dynamic-policy
configMap:
- name: data-router-dynamic-configmap
- - name: data-router-logs
+ name: aai-data-router-dynamic-configmap
+ - name: aai-data-router-logs
hostPath:
path: "/dockerdata-nfs/{{ .Values.nsPrefix }}/aai/data-router/logs"
restartPolicy: Always
kind: ConfigMap
metadata:
name: aai-elasticsearch-configmap
- namespace: {{ .Values.nsPrefix }}-aai
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/elasticsearch/config/elasticsearch.yml").AsConfig . | indent 2 }}
#{{ end }}
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
- name: elasticsearch
- namespace: "{{ .Values.nsPrefix }}-aai"
+ name: aai-elasticsearch
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.elasticsearchReplicas }}
selector:
matchLabels:
- app: elasticsearch
+ app: aai-elasticsearch
template:
metadata:
labels:
- app: elasticsearch
- name: elasticsearch
+ app: aai-elasticsearch
+ name: aai-elasticsearch
spec:
initContainers:
- command:
volumeMounts:
- name: elasticsearch-data
mountPath: /logroot/
- hostname: elasticsearch
+ hostname: aai-elasticsearch
containers:
- - name: elasticsearch
+ - name: aai-elasticsearch
image: "{{ .Values.image.elasticsearchImage }}:{{ .Values.image.elasticsearchVersion }}"
imagePullPolicy: {{ .Values.pullPolicy }}
ports:
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
- name: hbase
- namespace: "{{ .Values.nsPrefix }}-aai"
+ name: aai-hbase
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.hbaseReplicas }}
selector:
metadata:
labels:
app: hbase
- name: hbase
+ name: aai-hbase
spec:
hostname: hbase
containers:
apiVersion: v1
kind: ConfigMap
metadata:
- name: model-loader-prop-configmap
- namespace: {{ .Values.nsPrefix }}-aai
+ name: aai-model-loader-prop-configmap
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/model-loader/appconfig/model-loader.properties").AsConfig . | indent 2 }}
---
apiVersion: v1
kind: Secret
metadata:
- name: model-loader-secret
- namespace: {{ .Values.nsPrefix }}-aai
+ name: aai-model-loader-secret
+ namespace: {{ .Values.nsPrefix }}
type: Opaque
data:
{{ tpl (.Files.Glob "resources/config/model-loader/appconfig/auth/*").AsSecrets . | indent 2 }}
kind: ConfigMap
metadata:
name: aai-model-loader-log-configmap
- namespace: {{ .Values.nsPrefix }}-aai
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/log/model-loader/logback.xml").AsConfig . | indent 2 }}
#{{ end }}
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
- name: model-loader-service
- namespace: "{{ .Values.nsPrefix }}-aai"
+ name: aai-model-loader-service
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.modelLoaderReplicas }}
selector:
metadata:
labels:
app: model-loader-service
- name: model-loader-service
+ name: aai-model-loader-service
spec:
containers:
- name: model-loader-service
path: /etc/localtime
- name: aai-model-loader-prop-config
configMap:
- name: model-loader-prop-configmap
+ name: aai-model-loader-prop-configmap
- name: aai-model-loader-auth-config
secret:
- secretName: model-loader-secret
+ secretName: aai-model-loader-secret
- name: filebeat-conf
configMap:
name: aai-filebeat-configmap
apiVersion: v1
kind: ConfigMap
metadata:
- name: search-data-configmap
- namespace: {{ .Values.nsPrefix }}-aai
+ name: aai-search-data-configmap
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/search-data-service/appconfig/*").AsConfig . | indent 2 }}
---
apiVersion: v1
kind: Secret
metadata:
- name: search-data-keystone-secret
- namespace: {{ .Values.nsPrefix }}-aai
+ name: aai-search-data-keystone-secret
+ namespace: {{ .Values.nsPrefix }}
type: Opaque
data:
{{ tpl (.Files.Glob "resources/config/search-data-service/appconfig/auth/tomcat_keystore").AsSecrets . | indent 2 }}
kind: ConfigMap
metadata:
name: aai-search-policy-configmap
- namespace: {{ .Values.nsPrefix }}-aai
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/search-data-service/appconfig/auth/search_policy.json").AsSecrets . | indent 2 }}
---
kind: ConfigMap
metadata:
name: aai-search-data-service-log-configmap
- namespace: {{ .Values.nsPrefix }}-aai
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/log/search-data-service/logback.xml").AsConfig . | indent 2 }}
#{{ end }}
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
- name: search-data-service
- namespace: "{{ .Values.nsPrefix }}-aai"
+ name: aai-search-data-service
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.searchDataServiceReplicas }}
selector:
metadata:
labels:
app: search-data-service
- name: search-data-service
+ name: aai-search-data-service
spec:
containers:
- name: search-data-service
path: /etc/localtime
- name: aai-search-data-service-config
configMap:
- name: search-data-configmap
+ name: aai-search-data-configmap
- name: aai-search-data-service-auth-config
secret:
- secretName: search-data-keystone-secret
+ secretName: aai-search-data-keystone-secret
- name: aai-search-data-search-policy-config
configMap:
name: aai-search-policy-configmap
apiVersion: v1
kind: ConfigMap
metadata:
- name: sparky-be-configmap
- namespace: {{ .Values.nsPrefix }}-aai
+ name: aai-sparky-be-configmap
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/sparky-be/appconfig/*").AsConfig . | indent 2 }}
---
apiVersion: v1
kind: ConfigMap
metadata:
- name: sparky-be-model-configmap
- namespace: {{ .Values.nsPrefix }}-aai
+ name: aai-sparky-be-model-configmap
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/sparky-be/appconfig/model/*").AsConfig . | indent 2 }}
---
apiVersion: v1
kind: ConfigMap
metadata:
- name: sparky-be-portal-configmap
- namespace: {{ .Values.nsPrefix }}-aai
+ name: aai-sparky-be-portal-configmap
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/sparky-be/appconfig/portal/*").AsConfig . | indent 2 }}
---
apiVersion: v1
kind: Secret
metadata:
- name: sparky-be-secret
- namespace: {{ .Values.nsPrefix }}-aai
+ name: aai-sparky-be-secret
+ namespace: {{ .Values.nsPrefix }}
type: Opaque
data:
{{ tpl (.Files.Glob "resources/config/sparky-be/appconfig/auth/*").AsSecrets . | indent 2 }}
kind: ConfigMap
metadata:
name: aai-sparky-be-log-configmap
- namespace: {{ .Values.nsPrefix }}-aai
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/log/sparky-be/logback.xml").AsConfig . | indent 2 }}
#{{ end }}
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
- name: sparky-be
- namespace: "{{ .Values.nsPrefix }}-aai"
+ name: aai-sparky-be
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.sparkyReplicas }}
selector:
metadata:
labels:
app: sparky-be
- name: sparky-be
+ name: aai-sparky-be
spec:
containers:
- name: sparky-be
path: /etc/localtime
- name: aai-sparky-be-config
configMap:
- name: sparky-be-configmap
+ name: aai-sparky-be-configmap
- name: aai-sparky-be-model-config
configMap:
- name: sparky-be-model-configmap
+ name: aai-sparky-be-model-configmap
- name: aai-sparky-be-portal-config
configMap:
- name: sparky-be-portal-configmap
+ name: aai-sparky-be-portal-configmap
- name: aai-sparky-be-auth-config
secret:
- secretName: sparky-be-secret
+ secretName: aai-sparky-be-secret
- name: filebeat-conf
configMap:
name: aai-filebeat-configmap
--- /dev/null
+#!/bin/bash
+
+###
+# ============LICENSE_START=======================================================
+# openECOMP : SDN-C
+# ================================================================================
+# Copyright (C) 2017 AT&T Intellectual Property. All rights
+# reserved.
+# ================================================================================
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+# ============LICENSE_END=========================================================
+###
+
+SDNC_HOME=${SDNC_HOME:-/opt/onap/sdnc}
+MYSQL_PASSWD=${MYSQL_PASSWD:-openECOMP1.0}
+
+SDNC_DB_USER=${SDNC_DB_USER:-sdnctl}
+SDNC_DB_PASSWD=${SDNC_DB_PASSWD:-gamma}
+SDNC_DB_DATABASE=${SDN_DB_DATABASE:-sdnctl}
+
+
+# Create tablespace and user account
+mysql -h appc-dbhost.{{.Values.nsPrefix}} -u root -p${MYSQL_PASSWD} mysql <<-END
+CREATE DATABASE ${SDNC_DB_DATABASE};
+CREATE USER '${SDNC_DB_USER}'@'localhost' IDENTIFIED BY '${SDNC_DB_PASSWD}';
+CREATE USER '${SDNC_DB_USER}'@'%' IDENTIFIED BY '${SDNC_DB_PASSWD}';
+GRANT ALL PRIVILEGES ON ${SDNC_DB_DATABASE}.* TO '${SDNC_DB_USER}'@'localhost' WITH GRANT OPTION;
+GRANT ALL PRIVILEGES ON ${SDNC_DB_DATABASE}.* TO '${SDNC_DB_USER}'@'%' WITH GRANT OPTION;
+commit;
+END
+
+if [ -f ${SDNC_HOME}/data/odlsli.dump ]
+then
+mysql -h appc-dbhost.{{.Values.nsPrefix}} -u root -p${MYSQL_PASSWD} sdnctl < ${SDNC_HOME}/data/odlsli.dump
+fi
+
--- /dev/null
+#!/bin/bash
+
+###
+# ============LICENSE_START=======================================================
+# openECOMP : SDN-C
+# ================================================================================
+# Copyright (C) 2017 AT&T Intellectual Property. All rights
+# reserved.
+# ================================================================================
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+# ============LICENSE_END=========================================================
+###
+
+
+# Install SDN-C platform components if not already installed and start container
+
+ODL_HOME=${ODL_HOME:-/opt/opendaylight/current}
+ODL_ADMIN_PASSWORD=${ODL_ADMIN_PASSWORD:-Kp8bJ4SXszM0WXlhak3eHlcse2gAw84vaoGGmJvUy2U}
+SDNC_HOME=${SDNC_HOME:-/opt/onap/sdnc}
+SLEEP_TIME=${SLEEP_TIME:-120}
+MYSQL_PASSWD=${MYSQL_PASSWD:-openECOMP1.0}
+
+#
+# Wait for database
+#
+echo "Waiting for mysql"
+until mysql -h appc-dbhost.{{.Values.nsPrefix}} -u root -p${MYSQL_PASSWD} mysql &> /dev/null
+do
+ printf "."
+ sleep 1
+done
+echo -e "\nmysql ready"
+
+if [ ! -f ${SDNC_HOME}/.installed ]
+then
+ echo "Installing SDN-C database"
+ ${SDNC_HOME}/bin/installSdncDb.sh
+ echo "Starting OpenDaylight"
+ ${ODL_HOME}/bin/start
+ echo "Waiting ${SLEEP_TIME} seconds for OpenDaylight to initialize"
+ sleep ${SLEEP_TIME}
+ echo "Installing SDN-C platform features"
+ ${SDNC_HOME}/bin/installFeatures.sh
+ if [ -x ${SDNC_HOME}/svclogic/bin/install.sh ]
+ then
+ echo "Installing directed graphs"
+ ${SDNC_HOME}/svclogic/bin/install.sh
+ fi
+
+
+ echo "Restarting OpenDaylight"
+ ${ODL_HOME}/bin/stop
+ echo "Installed at `date`" > ${SDNC_HOME}/.installed
+fi
+
+exec ${ODL_HOME}/bin/karaf
+
#
# Configuration file for A&AI Client
#
-org.onap.ccsdk.sli.adaptors.aai.uri=https://aai-service.onap-aai:8443
+org.onap.ccsdk.sli.adaptors.aai.uri=https://aai-service.{{.Values.nsPrefix}}:8443
# query
# UBB Notify
org.onap.ccsdk.sli.adaptors.aai.path.notify=/aai/v11/actions/notify
-org.onap.ccsdk.sli.adaptors.aai.notify.selflink.fqdn=https://aai-service.onap-aai:8443/restconf/config/L3SDN-API:services/layer3-service-list/{service-instance-id}
-org.onap.ccsdk.sli.adaptors.aai.notify.selflink.avpn=https://aai-service.onap-aai:8443/restconf/config/L3AVPN-EVC-API:services/service-list/{service-instance-id}/service-data/avpn-logicalchannel-information
+org.onap.ccsdk.sli.adaptors.aai.notify.selflink.fqdn=https://aai-service.{{.Values.nsPrefix}}:8443/restconf/config/L3SDN-API:services/layer3-service-list/{service-instance-id}
+org.onap.ccsdk.sli.adaptors.aai.notify.selflink.avpn=https://aai-service.{{.Values.nsPrefix}}:8443/restconf/config/L3AVPN-EVC-API:services/service-list/{service-instance-id}/service-data/avpn-logicalchannel-information
# Service
org.onap.ccsdk.sli.adaptors.aai.path.service=/aai/v11/service-design-and-creation/services/service/{service-id}
--- /dev/null
+###
+# ============LICENSE_START=======================================================
+# openECOMP : SDN-C
+# ================================================================================
+# Copyright (C) 2017 AT&T Intellectual Property. All rights
+# reserved.
+# ================================================================================
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+# ============LICENSE_END=========================================================
+###
+
+# dblib.properrties
+org.onap.ccsdk.sli.dbtype=jdbc
+
+org.onap.ccsdk.sli.jdbc.hosts=sdnctldb01
+org.onap.ccsdk.sli.jdbc.url=jdbc:mysql://appc-dbhost.{{.Values.nsPrefix}}:3306/sdnctl
+org.onap.ccsdk.sli.jdbc.driver=org.mariadb.jdbc.Driver
+org.onap.ccsdk.sli.jdbc.database=sdnctl
+org.onap.ccsdk.sli.jdbc.user=sdnctl
+org.onap.ccsdk.sli.jdbc.password=gamma
+org.onap.ccsdk.sli.jdbc.connection.name=sdnctldb01
+org.onap.ccsdk.sli.jdbc.connection.timeout=50
+org.onap.ccsdk.sli.jdbc.request.timeout=100
+org.onap.ccsdk.sli.jdbc.limit.init=10
+org.onap.ccsdk.sli.jdbc.limit.min=10
+org.onap.ccsdk.sli.jdbc.limit.max=20
+org.onap.dblib.connection.recovery=false
+
--- /dev/null
+###
+# ============LICENSE_START=======================================================
+# openECOMP : SDN-C
+# ================================================================================
+# Copyright (C) 2017 AT&T Intellectual Property. All rights
+# reserved.
+# ================================================================================
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+# ============LICENSE_END=========================================================
+###
+
+org.onap.ccsdk.sli.dbtype = dblib
+#Note : the next 4 fields are only used if org.onap.ccsdk.sli.dbtype = jdbc
+org.onap.ccsdk.sli.jdbc.url=jdbc:mysql://appc-dbhost.{{.Values.nsPrefix}}:3306/sdnctl
+org.onap.ccsdk.sli.jdbc.database=sdnctl
+org.onap.ccsdk.sli.jdbc.user=sdnctl
+org.onap.ccsdk.sli.jdbc.password=gamma
+
+org.xml.sax.driver=org.apache.xerces.parsers.SAXParser
+
--- /dev/null
+###
+# ============LICENSE_START=======================================================
+# openECOMP : SDN-C
+# ================================================================================
+# Copyright (C) 2017 AT&T Intellectual Property. All rights
+# reserved.
+# ================================================================================
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+# ============LICENSE_END=========================================================
+###
+
+MYSQL_USER=${MYSQL_USER:-sdnctl}
+MYSQL_PWD=${MYSQL_PWD:-gamma}
+MYSQL_DB=${MYSQL_DB:-sdnctl}
+MYSQL_HOST=${MYSQL_HOST:-appc-dbhost.{{.Values.nsPrefix}}}
+
+mysql --user=${MYSQL_USER} --password=${MYSQL_PWD} --host=${MYSQL_HOST} ${MYSQL_DB} <<-END
+SELECT module, rpc, version, mode from SVC_LOGIC where active='Y';
+END
+
--- /dev/null
+###
+# ============LICENSE_START=======================================================
+# openECOMP : SDN-C
+# ================================================================================
+# Copyright (C) 2017 AT&T Intellectual Property. All rights
+# reserved.
+# ================================================================================
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+# ============LICENSE_END=========================================================
+###
+
+org.onap.ccsdk.sli.dbtype = jdbc
+org.onap.ccsdk.sli.jdbc.url = jdbc:mysql://appc-sdnctldb01.{{.Values.nsPrefix}}:3306/sdnctl
+org.onap.ccsdk.sli.jdbc.database = sdnctl
+org.onap.ccsdk.sli.jdbc.user = sdnctl
+org.onap.ccsdk.sli.jdbc.password = gamma
+
--- /dev/null
+###
+# ============LICENSE_START=======================================================
+# APPC
+# ================================================================================
+# Copyright (C) 2017 AT&T Intellectual Property. All rights reserved.
+# ================================================================================
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+# ============LICENSE_END=========================================================
+###
+
+SDNC_HOME=${SDNC_HOME:-/opt/onap/sdnc}
+APPC_HOME=${APPC_HOME:-/opt/openecomp/appc}
+MYSQL_PASSWD=${MYSQL_PASSWD:-openECOMP1.0}
+
+APPC_DB_USER=${APPC_DB_USER:-appcctl}
+APPC_DB_PASSWD=${APPC_DB_PASSWD:-appcctl}
+APPC_DB_DATABASE=${SDN_DB_DATABASE:-appcctl}
+
+
+# Create tablespace and user account
+mysql -h appc-dbhost.{{.Values.nsPrefix}} -u root -p${MYSQL_PASSWD} mysql <<-END
+CREATE DATABASE ${APPC_DB_DATABASE};
+CREATE USER '${APPC_DB_USER}'@'localhost' IDENTIFIED BY '${APPC_DB_PASSWD}';
+CREATE USER '${APPC_DB_USER}'@'%' IDENTIFIED BY '${APPC_DB_PASSWD}';
+GRANT ALL PRIVILEGES ON ${APPC_DB_DATABASE}.* TO '${APPC_DB_USER}'@'localhost' WITH GRANT OPTION;
+GRANT ALL PRIVILEGES ON ${APPC_DB_DATABASE}.* TO '${APPC_DB_USER}'@'%' WITH GRANT OPTION;
+commit;
+END
+
+if [ -f ${APPC_HOME}/data/appcctl.dump ]
+then
+ mysql -h appc-dbhost.{{.Values.nsPrefix}} -u root -p${MYSQL_PASSWD} appcctl < ${APPC_HOME}/data/appcctl.dump
+fi
+
+if [ -f ${APPC_HOME}/data/sdnctl.dump ]
+then
+ mysql -h appc-dbhost.{{.Values.nsPrefix}} -u root -p${MYSQL_PASSWD} sdnctl < ${APPC_HOME}/data/sdnctl.dump
+fi
+
+if [ -f ${APPC_HOME}/data/sqlData.dump ]
+then
+ mysql -h appc-dbhost.{{.Values.nsPrefix}} -u root -p${MYSQL_PASSWD} sdnctl < ${APPC_HOME}/data/sqlData.dump
+fi
+
--- /dev/null
+#!/bin/bash
+
+###
+# ============LICENSE_START=======================================================
+# APPC
+# ================================================================================
+# Copyright (C) 2017 AT&T Intellectual Property. All rights reserved.
+# ================================================================================
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+# ============LICENSE_END=========================================================
+# ECOMP is a trademark and service mark of AT&T Intellectual Property.
+###
+
+#
+# This script takes care of installing the SDNC & APPC platform components
+# if not already installed, and starts the APPC Docker Container
+#
+
+ODL_HOME=${ODL_HOME:-/opt/opendaylight/current}
+ODL_ADMIN_PASSWORD=${ODL_ADMIN_PASSWORD:-Kp8bJ4SXszM0WXlhak3eHlcse2gAw84vaoGGmJvUy2U}
+SDNC_HOME=${SDNC_HOME:-/opt/onap/sdnc}
+APPC_HOME=${APPC_HOME:-/opt/openecomp/appc}
+SLEEP_TIME=${SLEEP_TIME:-120}
+MYSQL_PASSWD=${MYSQL_PASSWD:-openECOMP1.0}
+
+#
+# Adding the DMAAP_TOPIC_ENV variable into APPC-ASDC-LISTENER properties
+#
+DMAAP_TOPIC_ENV=${DMAAP_TOPIC_ENV}
+
+if [ -z "$DMAAP_TOPIC_ENV" ]
+ then
+ echo "DMAAP_TOPIC_ENV shell variable is empty. Adding default value OS-ETE-DFW"
+ DMAAP_TOPIC_ENV="OS-ETE-DFW"
+ else
+ echo "DMAAP_TOPIC_ENV shell variable exists and it's $DMAAP_TOPIC_ENV"
+fi
+
+echo "Adding a value to property appc.asdc.env in appc.properties for appc-asdc-listener feature"
+echo "" >> $APPC_HOME/data/properties/appc.properties
+echo "appc.asdc.env=$DMAAP_TOPIC_ENV" >> $APPC_HOME/data/properties/appc.properties
+echo "" >> $APPC_HOME/data/properties/appc.properties
+
+#
+# Wait for database to init properly
+#
+echo "Waiting for mysql"
+until mysql -h appc-dbhost.{{.Values.nsPrefix}} -u root -p${MYSQL_PASSWD} mysql &> /dev/null
+do
+ printf "."
+ sleep 1
+done
+echo -e "\nmysql ready"
+
+if [ ! -f ${SDNC_HOME}/.installed ]
+then
+ echo "Installing SDNC database"
+ ${SDNC_HOME}/bin/installSdncDb.sh
+ echo "Installing APPC database"
+ ${APPC_HOME}/bin/installAppcDb.sh
+ echo "Starting OpenDaylight"
+ ${ODL_HOME}/bin/start
+ echo "Waiting ${SLEEP_TIME} seconds for OpenDaylight to initialize"
+ sleep ${SLEEP_TIME}
+ echo "Inserting modified installFeatures.sh for sdnc"
+ rm ${SDNC_HOME}/bin/installFeatures.sh
+ cp ${APPC_HOME}/data/sdncInstallFeatures.sh ${SDNC_HOME}/bin/installFeatures.sh
+ echo "Installing SDNC platform features"
+ ${SDNC_HOME}/bin/installFeatures.sh
+ if [ -x ${SDNC_HOME}/svclogic/bin/install.sh ]
+ then
+ echo "Installing directed graphs"
+ ${SDNC_HOME}/svclogic/bin/install.sh
+ fi
+
+ echo "Installing APPC platform features"
+ ${APPC_HOME}/bin/installFeatures.sh
+ if [ -x ${APPC_HOME}/svclogic/bin/install.sh ]
+ then
+ echo "Installing APPC DGs using platform-logic"
+ ${APPC_HOME}/svclogic/bin/install.sh
+ fi
+
+ if [ -x ${APPC_HOME}/svclogic/bin/install-converted-dgs.sh ]
+ then
+ echo "Installing APPC JSON DGs converted to XML using dg-loader"
+ ${APPC_HOME}/svclogic/bin/install-converted-dgs.sh
+ fi
+
+ echo "Restarting OpenDaylight"
+ ${ODL_HOME}/bin/stop
+ echo "Installed at `date`" > ${SDNC_HOME}/.installed
+fi
+
+exec ${ODL_HOME}/bin/karaf
+
--- /dev/null
+###
+# ============LICENSE_START=======================================================
+# APPC
+# ================================================================================
+# Copyright (C) 2017 AT&T Intellectual Property. All rights reserved.
+# ================================================================================
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+# ============LICENSE_END=========================================================
+###
+
+#
+# Configuration file for A&AI Client
+#
+
+#
+# Certificate keystore and truststore
+#
+org.onap.ccsdk.sli.adaptors.aai.ssl.trust=/opt/openecomp/appc/data/stores/truststore.openecomp.client.jks
+org.onap.ccsdk.sli.adaptors.aai.ssl.trust.psswd=adminadmin
+org.onap.ccsdk.sli.adaptors.aai.host.certificate.ignore=true
+
+org.onap.ccsdk.sli.adaptors.aai.client.name=AAI
+org.onap.ccsdk.sli.adaptors.aai.client.psswd=AAI
+
+org.onap.ccsdk.sli.adaptors.aai.application=openECOMP
+#connection.timeout=1000
+#read.timeout=2000
+
+#
+# Configuration file for A&AI Client
+#
+org.onap.ccsdk.sli.adaptors.aai.uri=https://aai-service.{{.Values.nsPrefix}}:8443
+
+
+# query
+org.onap.ccsdk.sli.adaptors.aai.path.query=/aai/v11/search/sdn-zone-query
+org.onap.ccsdk.sli.adaptors.aai.query.nodes=/aai/v11/search/nodes-query?search-node-type={node-type}&filter={entity-identifier}:EQUALS:{entity-name}
+org.onap.ccsdk.sli.adaptors.aai.query.generic=/aai/v11/search/generic-query?key={identifier}:{value}&start-node-type={start-node-type}&include=complex&depth=3
+
+# named query
+org.onap.ccsdk.sli.adaptors.aai.query.named=/aai/search/named-query
+
+
+#update
+org.onap.ccsdk.sli.adaptors.aai.update=/aai/v11/actions/update
+
+# vce
+org.onap.ccsdk.sli.adaptors.aai.path.vce =/aai/v11/network/vces/vce/
+org.onap.ccsdk.sli.adaptors.aai.path.vces=/aai/v11/network/vces/
+
+# vpe
+org.onap.ccsdk.sli.adaptors.aai.path.vpe =/aai/v11/network/vpes/vpe/
+org.onap.ccsdk.sli.adaptors.aai.path.vpes=/aai/v11/network/vpes/
+
+# customer
+org.onap.ccsdk.sli.adaptors.aai.path.customer=/aai/v11/business/customers/customer/{customer-id}
+
+# service subscription
+org.onap.ccsdk.sli.adaptors.aai.path.service.subscription=/aai/v11/business/customers/customer/{global-customer-id}/service-subscriptions/service-subscription/{service-type}
+
+# service instance
+org.onap.ccsdk.sli.adaptors.aai.path.svcinst=/aai/v11/business/customers/customer/{customer-id}/service-subscriptions/service-subscription/{service-type}/service-instances
+org.onap.ccsdk.sli.adaptors.aai.path.svcinst.query=/aai/v11/search/generic-query?key=service-instance.service-instance-id:{svc-instance-id}&start-node-type=service-instance&include=service-instance
+org.onap.ccsdk.sli.adaptors.aai.path.service.instance=/aai/v11/business/customers/customer/{global-customer-id}/service-subscriptions/service-subscription/{service-type}/service-instances/service-instance/{service-instance-id}
+
+# complex
+org.onap.ccsdk.sli.adaptors.aai.path.complexes=/aai/v11/cloud-infrastructure/complexes
+org.onap.ccsdk.sli.adaptors.aai.path.complex=/aai/v11/cloud-infrastructure/complexes/complex/{physical-location-id}
+
+# tenant
+org.onap.ccsdk.sli.adaptors.aai.path.tenant=/aai/v11/cloud-infrastructure/cloud-regions/cloud-region/{cloud-owner}/{cloud-region-id}/tenants/tenant/{tenant-id}
+org.onap.ccsdk.sli.adaptors.aai.path.tenant.query=/aai/v11/cloud-infrastructure/cloud-regions/cloud-region/{cloud-owner}/{cloud-region-id}/tenants/tenant?tenant-name={tenant-name}
+
+# vservers
+org.onap.ccsdk.sli.adaptors.aai.path.vservers=/aai/v11/cloud-infrastructure/cloud-regions/cloud-region/{cloud-owner}/{cloud-region-id}/tenants/tenant/{tenant-id}/vservers/
+org.onap.ccsdk.sli.adaptors.aai.path.vserver=/aai/v11/cloud-infrastructure/cloud-regions/cloud-region/{cloud-owner}/{cloud-region-id}/tenants/tenant/{tenant-id}/vservers/vserver/{vserver-id}
+
+# vpls-pe
+org.onap.ccsdk.sli.adaptors.aai.path.vpls.pes=/aai/v11/network/vpls-pes/
+org.onap.ccsdk.sli.adaptors.aai.path.vpls.pe =/aai/v11/network/vpls-pes/vpls-pe/
+
+# ctag-pool
+org.onap.ccsdk.sli.adaptors.aai.path.ctag.pools=/aai/v11/cloud-infrastructure/complexes/complex/{physical-location-id}/ctag-pools
+org.onap.ccsdk.sli.adaptors.aai.path.ctag.pool=/aai/v11/cloud-infrastructure/complexes/complex/{physical-location-id}/ctag-pools/ctag-pool/{target-pe}/{availability-zone-name}
+
+#
+#-------------- 1510 ----------------------
+#
+
+# pservers
+org.onap.ccsdk.sli.adaptors.aai.path.pservers=/aai/v11/cloud-infrastructure/pservers
+org.onap.ccsdk.sli.adaptors.aai.path.pserver=/aai/v11/cloud-infrastructure/pservers/pserver/{hostname}
+
+# generic-vnf
+org.onap.ccsdk.sli.adaptors.aai.path.generic.vnfs=/aai/v11/network/generic-vnfs
+org.onap.ccsdk.sli.adaptors.aai.path.generic.vnf=/aai/v11/network/generic-vnfs/generic-vnf/{vnf-id}
+
+# dvs-switch
+org.onap.ccsdk.sli.adaptors.aai.path.dvsswitches=/aai/v11/cloud-infrastructure/cloud-regions/cloud-region/{cloud-owner}/{cloud-region-id}/dvs-switches
+org.onap.ccsdk.sli.adaptors.aai.path.dvsswitch=/aai/v11/cloud-infrastructure/cloud-regions/cloud-region/{cloud-owner}/{cloud-region-id}/dvs-switches/dvs-switch/{switch-name}
+
+# L3 Networks
+org.onap.ccsdk.sli.adaptors.aai.path.l3networks=/aai/v11/network/l3-networks
+org.onap.ccsdk.sli.adaptors.aai.path.l3network=/aai/v11/network/l3-networks/l3-network/{network-id}
+org.onap.ccsdk.sli.adaptors.aai.path.l3network.query.name=/aai/v11/network/l3-networks/l3-network?network-name={network-name}
+
+# P-Interfaces
+org.onap.ccsdk.sli.adaptors.aai.path.pserver.pinterfaces=/aai/v11/cloud-infrastructure/pservers/pserver/{hostname}/p-interfaces
+org.onap.ccsdk.sli.adaptors.aai.path.pserver.pinterface=/aai/v11/cloud-infrastructure/pservers/pserver/{hostname}/p-interfaces/p-interface/{interface-name}
+
+# Physical Link
+org.onap.ccsdk.sli.adaptors.aai.path.physical.links=/aai/v11/network/physical-links
+org.onap.ccsdk.sli.adaptors.aai.path.physical.link=/aai/v11/network/physical-links/physical-link/{link-name}
+
+# VPN Bindings
+org.onap.ccsdk.sli.adaptors.aai.path.vpn.bindings=/aai/v11/network/vpn-bindings/
+org.onap.ccsdk.sli.adaptors.aai.path.vpn.binding=/aai/v11/network/vpn-bindings/vpn-binding/{vpn-id}
+
+# VNF IMAGES
+org.onap.ccsdk.sli.adaptors.aai.path.vnf.images=/aai/v11/service-design-and-creation/vnf-images
+org.onap.ccsdk.sli.adaptors.aai.path.vnf.image=/aai/v11/service-design-and-creation/vnf-images/vnf-image/{att-uuid}
+org.onap.ccsdk.sli.adaptors.aai.path.vnf.image.query=/aai/v11/service-design-and-creation/vnf-images/vnf-image?application={application_model}&application-vendor={application_vendor}
+
+# UBB Notify
+org.onap.ccsdk.sli.adaptors.aai.path.notify=/aai/v11/actions/notify
+org.onap.ccsdk.sli.adaptors.aai.notify.selflink.fqdn=https://aai-service.{{.Values.nsPrefix}}:8443/restconf/config/L3SDN-API:services/layer3-service-list/{service-instance-id}
+org.onap.ccsdk.sli.adaptors.aai.notify.selflink.avpn=https://aai-service.{{.Values.nsPrefix}}:8443/restconf/config/L3AVPN-EVC-API:services/service-list/{service-instance-id}/service-data/avpn-logicalchannel-information
+
+# Service
+org.onap.ccsdk.sli.adaptors.aai.path.service=/aai/v11/service-design-and-creation/services/service/{service-id}
+org.onap.ccsdk.sli.adaptors.aai.path.services=/aai/v11/service-design-and-creation/services
+
+
+#
+#-------------- 1604 ----------------------
+#
+
+# VNFC
+org.onap.ccsdk.sli.adaptors.aai.path.vnfc=/aai/v11/network/vnfcs/vnfc/{vnfc-name}
+
+# class-of-service
+org.onap.ccsdk.sli.adaptors.aai.path.class.of.service=/aai/v11/network/site-pair-sets/site-pair-set/{site-pair-set-id}/routing-instances/routing-instance/{routing-instance-id}/site-pairs/site-pair/{site-pair-id}/classes-of-service/class-of-service/{cos-id}
+
+# site-pair
+org.onap.ccsdk.sli.adaptors.aai.path.site.pair=/aai/v11/network/site-pair-sets/site-pair-set/{site-pair-set-id}/routing-instances/routing-instance/{routing-instance-id}/site-pairs/site-pair/{site-pair-id}
+
+# routing-instance
+org.onap.ccsdk.sli.adaptors.aai.path.routing.instance=/aai/v11/network/site-pair-sets/site-pair-set/{site-pair-set-id}/routing-instances/routing-instance/{routing-instance-id}
+
+# site-pair-set
+org.onap.ccsdk.sli.adaptors.aai.path.site.pair.set=/aai/v11/network/site-pair-sets/site-pair-set/{site-pair-set-id}
+
+# license key resource
+org.onap.ccsdk.sli.adaptors.aai.path.license.acquire=/aai/v11/actions/assignment/license-management/assignment-group-uuid/{assignment-group-uuid}
+org.onap.ccsdk.sli.adaptors.aai.path.license=/aai/v11/license-management/license-key-resources/license-key-resource/{att-uuid}
+
+# logical-link
+org.onap.ccsdk.sli.adaptors.aai.path.logical.link =/aai/v11/network/logical-links/logical-link/{link-name}
+
+# virtual-data-center
+org.onap.ccsdk.sli.adaptors.aai.path.virtual.data.center=/aai/v11/cloud-infrastructure/virtual-data-centers/virtual-data-center/{vdc-id}
+
+# wan-connector
+org.onap.ccsdk.sli.adaptors.aai.path.wan.connector=/aai/v11/business/connectors/connector/{resource-instance-id}
+
+# l-interface
+org.onap.ccsdk.sli.adaptors.aai.path.lag.interface.l.interface=/aai/v11/cloud-infrastructure/pservers/pserver/{hostname}/lag-interfaces/lag-interface/{lag-interface.interface-name}/l-interfaces/l-interface/{interface-name}
+org.onap.ccsdk.sli.adaptors.aai.path.p.interface.l.interface=/aai/v11/cloud-infrastructure/pservers/pserver/{hostname}/p-interfaces/p-interface/{p-interface.interface-name}/l-interfaces/l-interface/{interface-name}
+
+# l-interface pnf
+org.onap.ccsdk.sli.adaptors.aai.path.lag.interface.l.interface.pnf=/aai/v11/network/pnfs/pnf/{pnf-name}/lag-interfaces/lag-interface/{lag-interface.interface-name}/l-interfaces/l-interface/{interface-name}
+org.onap.ccsdk.sli.adaptors.aai.path.p.interface.l.interface.pnf=/aai/v11/network/pnfs/pnf/{pnf-name}/p-interfaces/p-interface/{p-interface.interface-name}/l-interfaces/l-interface/{interface-name}
+
+# subinterface
+org.onap.ccsdk.sli.adaptors.aai.path.pnf.lag.interface.subinterface=/aai/v11/network/pnfs/pnf/{pnf-name}/lag-interfaces/lag-interface/{lag-interface.interface-name}/l-interfaces/l-interface/{interface-name}
+org.onap.ccsdk.sli.adaptors.aai.path.pnf.p.interface.l.interface=/aai/v11/network/pnfs/pnf/{pnf-name}/p-interfaces/p-interface/{p-interface.interface-name}/l-interfaces/l-interface/{interface-name}
+
+# vlans
+org.onap.ccsdk.sli.adaptors.aai.path.vlan=/aai/v11/cloud-infrastructure/cloud-regions/cloud-region/{cloud-owner}/{cloud-region-id}/tenants/tenant/{tenant-id}/vservers/vserver/{vserver-id}/l-interfaces/l-interface/{interface-name}/vlans/vlan/{vlan-interface}
+org.onap.ccsdk.sli.adaptors.aai.path.generic.vnf.vlan=/aai/v11/network/generic-vnfs/generic-vnf/{vnf-id}/l-interfaces/l-interface/{interface-name}/vlans/vlan/{vlan-interface}
+
+# l3-interface-ipv4-address-list
+org.onap.ccsdk.sli.adaptors.aai.path.l3.interface.ipv4.address.list=/aai/v11/network/generic-vnfs/generic-vnf/{vnf-id}/l-interfaces/l-interface/{interface-name}/l3-interface-ipv4-address-list/{l3-interface-ipv4-address}
+org.onap.ccsdk.sli.adaptors.aai.path.vlan.l3.interface.ipv4.address.list=/aai/v11/network/generic-vnfs/generic-vnf/{vnf-id}/l-interfaces/l-interface/{interface-name}/vlans/vlan/{vlan-interface}/l3-interface-ipv4-address-list/{l3-interface-ipv4-address}
+
+# l3-interface-ipv6-address-list
+org.onap.ccsdk.sli.adaptors.aai.path.l3.interface.ipv6.address.list=/aai/v11/network/generic-vnfs/generic-vnf/{vnf-id}/l-interfaces/l-interface/{interface-name}/l3-interface-ipv6-address-list/{l3-interface-ipv6-address}
+org.onap.ccsdk.sli.adaptors.aai.path.vlan.l3.interface.ipv6.address.list=/aai/v11/network/generic-vnfs/generic-vnf/{vnf-id}/l-interfaces/l-interface/{interface-name}/vlans/vlan/{vlan-interface}/l3-interface-ipv6-address-list/{l3-interface-ipv6-address}
+
+# ipsec-configuration
+org.onap.ccsdk.sli.adaptors.aai.path.ipsec.configuration=/aai/v11/network/ipsec-configurations/ipsec-configuration/{ipsec-configuration-id}
+
+# vig server
+org.onap.ccsdk.sli.adaptors.aai.path.vig.server=/aai/v11/network/ipsec-configurations/ipsec-configuration/{ipsec-configuration-id}/vig-servers/vig-server/{vig-address-type}
+
+# l3-network
+org.onap.ccsdk.sli.adaptors.aai.path.l3.network=/aai/v11/network/l3-networks/l3-network/{network-id}
+
+# subnet
+org.onap.ccsdk.sli.adaptors.aai.path.subnet=/aai/v11/network/l3-networks/l3-network/{network-id}/subnets/subnet/{subnet-id}
+
+# multicast-configuration
+org.onap.ccsdk.sli.adaptors.aai.path.multicast.configuration=/aai/v11/network/multicast-configurations/multicast-configuration/{multicast-configuration-id}
+
+# org.onap.ccsdk.sli.adaptors.aai.path.l.interface.ipv4.address.list
+org.onap.ccsdk.sli.adaptors.aai.path.l3-interface.ipv4.address.list=/aai/v11/network/generic-vnfs/generic-vnf/{vnf-id}/l-interfaces/l-interface/{interface-name}/l3-interface-ipv4-address-list/{l3-interface-ipv4-address}
+
+# org.onap.ccsdk.sli.adaptors.aai.path.l.interface.vlan.ipv4.address.list
+org.onap.ccsdk.sli.adaptors.aai.path.l3-interface.vlan.ipv4.address.list=/aai/v11/network/generic-vnfs/generic-vnf/{vnf-id}/l-interfaces/l-interface/{interface-name}/vlans/vlan/{vlan-interface}/l3-interface-ipv4-address-list/{l3-interface-ipv4-address}
+
+# org.onap.ccsdk.sli.adaptors.aai.path.l.interface.ipv6.address.list
+org.onap.ccsdk.sli.adaptors.aai.path.l3-interface.ipv6.address.list=/aai/v11/network/generic-vnfs/generic-vnf/{vnf-id}/l-interfaces/l-interface/{interface-name}/l3-interface-ipv6-address-list/{l3-interface-ipv6-address}
+
+# volume.group
+org.onap.ccsdk.sli.adaptors.aai.path.volume.group=/aai/v11/cloud-infrastructure/cloud-regions/cloud-region/{cloud-owner}/{cloud-region-id}/volume-groups/volume-group/{volume-group-id}
+
+#cloud region
+org.onap.ccsdk.sli.adaptors.aai.path.cloud.region=/aai/v11/cloud-infrastructure/cloud-regions/cloud-region/{cloud-owner}/{cloud-region-id}
+
+# vf-module
+org.onap.ccsdk.sli.adaptors.aai.path.vf.module=/aai/v11/network/generic-vnfs/generic-vnf/{vnf-id}/vf-modules/vf-module/{vf-module-id}
+
+# l-interface through generic-vnf
+org.onap.ccsdk.sli.adaptors.aai.path.generic.vnf.linterface=/aai/v11/network/generic-vnfs/generic-vnf/{vnf-id}/l-interfaces/l-interface/{interface-name}
+
+# network-policy
+org.onap.ccsdk.sli.adaptors.aai.path.network.policy=/aai/v11/network/network-policies/network-policy/{network-policy-id}
+
+# pnf
+org.onap.ccsdk.sli.adaptors.aai.path.pnf=/aai/v11/network/pnfs/pnf/{pnf-name}
+
+#
+# Formatting
+#
+org.onap.ccsdk.sli.adaptors.aai.param.format=filter=%s:%s
+org.onap.ccsdk.sli.adaptors.aai.param.vnf_type=vnf-type
+org.onap.ccsdk.sli.adaptors.aai.param.physical.location.id=physical-location-id
+org.onap.ccsdk.sli.adaptors.aai.param.service.type=service-type
### ###
### Properties for demo ###
### ###
-appc.demo.poolMembers=dmaap.onap-message-router:3904
+appc.demo.poolMembers=dmaap.{{.Values.nsPrefix}}:3904
appc.demo.topic.read=APPC-CL
appc.demo.topic.write=APPC-CL
appc.demo.client.name=appcDemoEventListener
appc.demo.provider.user=admin
appc.demo.provider.pass=Kp8bJ4SXszM0WXlhak3eHlcse2gAw84vaoGGmJvUy2U
appc.demo.provider.url=http://localhost:8181/restconf/operations/appc-provider
-appc.provider.vfodl.url=http://admin:Kp8bJ4SXszM0WXlhak3eHlcse2gAw84vaoGGmJvUy2U@sdnhost.onap-appc:8282/restconf/config/network-topology:network-topology/topology/topology-netconf/node/NODE_NAME/yang-ext:mount/sample-plugin:sample-plugin/pg-streams/
+appc.provider.vfodl.url=http://admin:Kp8bJ4SXszM0WXlhak3eHlcse2gAw84vaoGGmJvUy2U@appc-sdnhost.{{.Values.nsPrefix}}:8282/restconf/config/network-topology:network-topology/topology/topology-netconf/node/NODE_NAME/yang-ext:mount/sample-plugin:sample-plugin/pg-streams/
# The properties right below are needed to properly call the Master DG to serve demo purposes
appc.service.logic.module.name=APPC
appc.topology.dg.version=2.0.0
# TEMP - Properties that might be needed to make the AAI-APPC connection
-org.openecomp.appc.db.url.appcctl=jdbc:mysql://dbhost:3306/appcctl
+org.openecomp.appc.db.url.appcctl=jdbc:mysql://appc-dbhost.{{.Values.nsPrefix}}:3306/appcctl
org.openecomp.appc.db.user.appcctl=appcctl
org.openecomp.appc.db.pass.appcctl=appcctl
-org.openecomp.appc.db.url.sdnctl=jdbc:mysql://dbhost:3306/sdnctl
+org.openecomp.appc.db.url.sdnctl=jdbc:mysql://appc-dbhost.{{.Values.nsPrefix}}:3306/sdnctl
org.openecomp.appc.db.user.sdnctl=sdnctl
org.openecomp.appc.db.pass.sdnctl=gamma
# Property below is valid in appc-command-executor-core, appc-license-manager-core, appc-lifecycle-management-core,
# appc-request-handler-core, appc-workflow-management-core (all from the appc-dispatcher package).
-dmaap.poolMembers=dmaap.onap-message-router:3904
+dmaap.poolMembers=dmaap.{{.Values.nsPrefix}}:3904
# appc-event-listener-bundle properties (only defined in src/test of default.properties)
-appc.LCM.poolMembers=dmaap.onap-message-router:3904
+appc.LCM.poolMembers=dmaap.{{.Values.nsPrefix}}:3904
appc.LCM.topic.read=APPC-LCM-READ
appc.LCM.topic.write=APPC-LCM-WRITE
appc.LCM.client.name=APPC-EVENT-LISTENER-TEST
appc.LCM.provider.url=http://localhost:8181/restconf/operations/appc-provider-lcm
# properties from appc-netconf-adapter-bundle, appc-dg-common, appc-dmaap-adapter-bundle
-poolMembers=dmaap.onap-message-router:3904
-event.pool.members=dmaap.onap-message-router:3904
+poolMembers=dmaap.{{.Values.nsPrefix}}:3904
+event.pool.members=dmaap.{{.Values.nsPrefix}}:3904
restconf.user=admin
restconf.pass=Kp8bJ4SXszM0WXlhak3eHlcse2gAw84vaoGGmJvUy2U
org.openecomp.appc.iaas.skiphypervisorcheck=true
# Properties from default.properties in the src/test and src/main paths of appc-asdc-listener-bundle
-appc.sdc.host=sdc-be.onap-sdc:8443
+appc.sdc.host=sdc-be.{{.Values.nsPrefix}}:8443
appc.sdc.env=APPC-ASDC-ENV
appc.sdc.user=test
appc.sdc.pass=test
DCAE.dmaap.event.topic.write=EventSenderTest
DCAE.dmaap.appc.username=test
DCAE.dmaap.appc.password=test
-DCAE.dmaap.event.pool.members=dmaap.onap-message-router:3904
+DCAE.dmaap.event.pool.members=dmaap.{{.Values.nsPrefix}}:3904
#OAM Listener
appc.OAM.disabled=true
appc.OAM.provider.url=http://localhost:8181/restconf/operations/appc-oam
-appc.OAM.poolMembers=dmaap.onap-message-router:3904
+appc.OAM.poolMembers=dmaap.{{.Values.nsPrefix}}:3904
appc.OAM.service=ueb
appc.OAM.topic.read=testOAM
appc.OAM.topic.write=testOAM
--- /dev/null
+###
+# ============LICENSE_START=======================================================
+# APPC
+# ================================================================================
+# Copyright (C) 2017 AT&T Intellectual Property. All rights reserved.
+# ================================================================================
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+# ============LICENSE_END=========================================================
+###
+
+# dblib.properrties
+org.onap.ccsdk.sli.dbtype=jdbc
+
+org.onap.ccsdk.sli.jdbc.hosts=sdnctldb01
+org.onap.ccsdk.sli.jdbc.url=jdbc:mysql://appc-dbhost.{{.Values.nsPrefix}}:3306/sdnctl
+org.onap.ccsdk.sli.jdbc.driver=org.mariadb.jdbc.Driver
+org.onap.ccsdk.sli.jdbc.database=sdnctl
+org.onap.ccsdk.sli.jdbc.user=sdnctl
+org.onap.ccsdk.sli.jdbc.password=gamma
+org.onap.ccsdk.sli.jdbc.connection.name=sdnctldb01
+org.onap.ccsdk.sli.jdbc.connection.timeout=50
+org.onap.ccsdk.sli.jdbc.request.timeout=100
+org.onap.ccsdk.sli.jdbc.limit.init=10
+org.onap.ccsdk.sli.jdbc.limit.min=10
+org.onap.ccsdk.sli.jdbc.limit.max=20
+org.onap.dblib.connection.recovery=false
+
--- /dev/null
+###
+# ============LICENSE_START=======================================================
+# APPC
+# ================================================================================
+# Copyright (C) 2017 AT&T Intellectual Property. All rights reserved.
+# ================================================================================
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+# ============LICENSE_END=========================================================
+###
+
+org.onap.ccsdk.sli.dbtype = dblib
+#Note : the next 4 fields are only used if org.onap.ccsdk.sli.dbtype = jdbc
+org.onap.ccsdk.sli.jdbc.url=jdbc:mysql://appc-dbhost.{{.Values.nsPrefix}}:3306/sdnctl
+org.onap.ccsdk.sli.jdbc.database=sdnctl
+org.onap.ccsdk.sli.jdbc.user=sdnctl
+org.onap.ccsdk.sli.jdbc.password=gamma
+
+org.xml.sax.driver=org.apache.xerces.parsers.SAXParser
+
--- /dev/null
+###
+# ============LICENSE_START=======================================================
+# APPC
+# ================================================================================
+# Copyright (C) 2017 AT&T Intellectual Property. All rights reserved.
+# ================================================================================
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+# ============LICENSE_END=========================================================
+# ECOMP is a trademark and service mark of AT&T Intellectual Property.
+###
+
+MYSQL_USER=${MYSQL_USER:-sdnctl}
+MYSQL_PWD=${MYSQL_PWD:-gamma}
+MYSQL_DB=${MYSQL_DB:-sdnctl}
+MYSQL_HOST=${MYSQL_HOST:-appc-dbhost.{{.Values.nsPrefix}}}
+
+mysql --user=${MYSQL_USER} --password=${MYSQL_PWD} --host=${MYSQL_HOST} ${MYSQL_DB} <<-END
+SELECT module, rpc, version, mode from SVC_LOGIC where active='Y';
+END
--- /dev/null
+###
+# ============LICENSE_START=======================================================
+# APPC
+# ================================================================================
+# Copyright (C) 2017 AT&T Intellectual Property. All rights reserved.
+# ================================================================================
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+# ============LICENSE_END=========================================================
+# ECOMP is a trademark and service mark of AT&T Intellectual Property.
+###
+
+org.onap.ccsdk.sli.dbtype = jdbc
+org.onap.ccsdk.sli.jdbc.url = jdbc:mysql://appc-sdnctldb01.{{.Values.nsPrefix}}:3306/sdnctl
+org.onap.ccsdk.sli.jdbc.database = sdnctl
+org.onap.ccsdk.sli.jdbc.user = sdnctl
+org.onap.ccsdk.sli.jdbc.password = gamma
+
--- /dev/null
+org.onap.ccsdk.sli.dbtype=jdbc
+org.onap.ccsdk.sli.jdbc.url=jdbc:mysql://appc-dbhost.{{.Values.nsPrefix}}:3306/sdnctl
+org.onap.ccsdk.sli.jdbc.database=sdnctl
+org.onap.ccsdk.sli.jdbc.user=sdnctl
+org.onap.ccsdk.sli.jdbc.password=gamma
#List of logstash server ip addresses with port number.
#But, in our case, this will be the loadbalancer IP address.
#For the below property to work the loadbalancer or logstash should expose 5044 port to listen the filebeat events or port in the property should be changed appropriately.
- hosts: ["logstash.onap-log:5044"]
+ hosts: ["logstash.{{.Values.nsPrefix}}:5044"]
#If enable will do load balancing among availabe Logstash, automatically.
loadbalance: true
--- /dev/null
+#!/bin/bash
+export PATH=$PATH:.
+appDir=$(pwd)
+if [ "$#" != 3 -a "$#" != 4 ]
+then
+ echo "Usage $0 releaseDir loginId emailAddress [gitLocalRepository]"
+ echo "Note: Specify the gitLocalRepository path if you would want to be able to import flows from your local git repository"
+ exit
+fi
+if [ ! -e "releases" ]
+then
+ mkdir releases
+fi
+releaseDir="$1"
+name="Release $releaseDir"
+loginId="$2"
+emailid="$3"
+dbHost="appc-dbhost.{{.Values.nsPrefix}}"
+dbPort="3306"
+dbName="sdnctl"
+dbUser="sdnctl"
+dbPassword="gamma"
+gitLocalRepository="$4"
+
+lastPort=$(find "releases/" -name "customSettings.js" |xargs grep uiPort|cut -d: -f2|sed -e s/,//|sort|tail -1)
+echo $lastPort|grep uiPort >/dev/null 2>&1
+if [ "$?" == "0" ]
+then
+lastPort=$(find "releases/" -name "customSettings.js" |xargs grep uiPort|cut -d: -f3|sed -e s/,//|sort|tail -1)
+fi
+#echo $lastPort
+if [ "${lastPort}" == "" ]
+then
+ lastPort="3099"
+fi
+let nextPort=$(expr $lastPort+1)
+#echo $nextPort
+if [ ! -e "releases/$releaseDir" ]
+then
+mkdir releases/$releaseDir
+cd releases/$releaseDir
+mkdir flows
+mkdir flows/shared
+mkdir flows/shared/backups
+mkdir html
+mkdir xml
+mkdir lib
+mkdir lib/flows
+mkdir logs
+mkdir conf
+mkdir codecloud
+customSettingsFile="customSettings.js"
+if [ ! -e "./$customSettingsFile" ]
+then
+ echo "module.exports = {" >$customSettingsFile
+ echo " 'name' : '$name'," >>$customSettingsFile
+ echo " 'emailAddress' :'$emailid'," >>$customSettingsFile
+ echo " 'uiPort' :$nextPort," >>$customSettingsFile
+ echo " 'mqttReconnectTime': 15000," >>$customSettingsFile
+ echo " 'serialReconnectTime' : 15000," >>$customSettingsFile
+ echo " 'debugMaxLength': 1000," >>$customSettingsFile
+ echo " 'htmlPath': 'releases/$releaseDir/html/'," >>$customSettingsFile
+ echo " 'xmlPath': 'releases/$releaseDir/xml/'," >>$customSettingsFile
+ echo " 'flowFile' : 'releases/$releaseDir/flows/flows.json'," >>$customSettingsFile
+ echo " 'sharedDir': 'releases/$releaseDir/flows/shared'," >>$customSettingsFile
+ echo " 'userDir' : 'releases/$releaseDir'," >>$customSettingsFile
+ echo " 'httpAuth': {user:'$loginId',pass:'cc03e747a6afbbcbf8be7668acfebee5'}," >>$customSettingsFile
+ echo " 'dbHost': '$dbHost'," >>$customSettingsFile
+ echo " 'dbPort': '$dbPort'," >>$customSettingsFile
+ echo " 'dbName': '$dbName'," >>$customSettingsFile
+ echo " 'dbUser': '$dbUser'," >>$customSettingsFile
+ echo " 'dbPassword': '$dbPassword'," >>$customSettingsFile
+ echo " 'gitLocalRepository': '$gitLocalRepository'" >>$customSettingsFile
+ echo " }" >>$customSettingsFile
+fi
+ #echo "Created custom settings file $customSettingsFile"
+ echo "Done ....."
+else
+ echo "ERROR:customSettings file $customSettingsFile already exists for $releaseDir"
+ exit
+fi
+#echo "Content of custom settings file"
+#echo "============================================================================"
+# cat $customSettingsFile
+#echo "============================================================================"
+svclogicPropFile="./conf/svclogic.properties"
+if [ ! -d "${appDir}/yangFiles" ]
+then
+ mkdir -p "${appDir}/yangFiles"
+fi
+if [ ! -d "${appDir}/generatedJS" ]
+then
+ mkdir -p "${appDir}/generatedJS"
+fi
+
+if [ ! -e "./$svclogicPropFile" ]
+then
+ echo "org.onap.ccsdk.sli.dbtype=jdbc" >$svclogicPropFile
+ echo "org.onap.ccsdk.sli.jdbc.url=jdbc:mysql://appc-dbhost.{{.Values.nsPrefix}}:3306/sdnctl" >>$svclogicPropFile
+ echo "org.onap.ccsdk.sli.jdbc.database=sdnctl" >>$svclogicPropFile
+ echo "org.onap.ccsdk.sli.jdbc.user=sdnctl" >>$svclogicPropFile
+ echo "org.onap.ccsdk.sli.jdbc.password=gamma" >>$svclogicPropFile
+fi
+if [ ! -e "${appDir}/flowShareUsers.js" ]
+then
+ echo "module.exports = {\"flowShareUsers\":" >${appDir}/flowShareUsers.js
+ echo " [" >>${appDir}/flowShareUsers.js
+ echo " ]" >>${appDir}/flowShareUsers.js
+ echo "}" >>${appDir}/flowShareUsers.js
+fi
+grep "$releaseDir" ${appDir}/flowShareUsers.js >/dev/null 2>&1
+if [ "$?" != "0" ]
+then
+ num_of_lines=$(cat ${appDir}/flowShareUsers.js|wc -l)
+ if [ $num_of_lines -gt 4 ]
+ then
+ content=$(head -n -2 ${appDir}/flowShareUsers.js)
+ echo "${content}," > ${appDir}/flowShareUsers.js
+ else
+ content=$(head -n -2 ${appDir}/flowShareUsers.js)
+ echo "$content" > ${appDir}/flowShareUsers.js
+ fi
+ echo " {" >> ${appDir}/flowShareUsers.js
+ echo " \"name\" : \"$name\"," >> ${appDir}/flowShareUsers.js
+ echo " \"rootDir\" : \"$releaseDir\"" >> ${appDir}/flowShareUsers.js
+ echo " }" >> ${appDir}/flowShareUsers.js
+ echo " ]" >> ${appDir}/flowShareUsers.js
+ echo "}" >> ${appDir}/flowShareUsers.js
+fi
--- /dev/null
+module.exports={
+ "name": "Release sdnc1.0",
+ "emailAddress": "dguser@onap.org",
+ "uiPort": 3100,
+ "mqttReconnectTime": 15000,
+ "serialReconnectTime": 15000,
+ "debugMaxLength": 1000,
+ "htmlPath": "releases/sdnc1.0/html/",
+ "xmlPath": "releases/sdnc1.0/xml/",
+ "flowFile": "releases/sdnc1.0/flows/flows.json",
+ "sharedDir": "releases/sdnc1.0/flows/shared",
+ "userDir": "releases/sdnc1.0",
+ "httpAuth": {
+ "user": "dguser",
+ "pass": "cc03e747a6afbbcbf8be7668acfebee5"
+ },
+ "dbHost": "appc-dbhost.{{.Values.nsPrefix}}",
+ "dbPort": "3306",
+ "dbName": "sdnctl",
+ "dbUser": "sdnctl",
+ "dbPassword": "gamma",
+ "gitLocalRepository": "",
+ "httpRoot": "/",
+ "disableEditor": false,
+ "httpAdminRoot": "/",
+ "httpAdminAuth": {
+ "user": "dguser",
+ "pass": "cc03e747a6afbbcbf8be7668acfebee5"
+ },
+ "httpNodeRoot": "/",
+ "httpNodeAuth": {
+ "user": "dguser",
+ "pass": "cc03e747a6afbbcbf8be7668acfebee5"
+ },
+ "uiHost": "0.0.0.0",
+ "version": "0.9.1",
+ "performGitPull": "N"
+}
apiVersion: v1
kind: Service
metadata:
- name: dbhost
- namespace: "{{ .Values.nsPrefix }}-appc"
+ name: appc-dbhost
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: appc-dbhost
spec:
apiVersion: v1
kind: Service
metadata:
- name: sdnctldb01
- namespace: "{{ .Values.nsPrefix }}-appc"
+ name: appc-sdnctldb01
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: appc-dbhost
spec:
apiVersion: v1
kind: Service
metadata:
- name: sdnctldb02
- namespace: "{{ .Values.nsPrefix }}-appc"
+ name: appc-sdnctldb02
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: appc-dbhost
spec:
apiVersion: v1
kind: Service
metadata:
- name: sdnhost
- namespace: "{{ .Values.nsPrefix }}-appc"
+ name: appc-sdnhost
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: appc
spec:
apiVersion: v1
kind: Service
metadata:
- name: dgbuilder
- namespace: "{{ .Values.nsPrefix }}-appc"
+ name: appc-dgbuilder
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: appc-dgbuilder
spec:
type: NodePort
selector:
app: appc-dgbuilder
-#{{ end }}
\ No newline at end of file
+#{{ end }}
+++ /dev/null
-#{{ if not .Values.disableAppcAppc }}
-apiVersion: v1
-kind: ConfigMap
-metadata:
- name: appc-conf-configmap
- namespace: {{ .Values.nsPrefix }}-appc
-data:
-{{ tpl (.Files.Glob "resources/config/conf/*").AsConfig . | indent 2 }}
-#{{ end }}
--- /dev/null
+#{{ if not .Values.disableAppcAppc }}
+apiVersion: v1
+kind: ConfigMap
+metadata:
+ name: appc-openecomp-appc-data-properties-configmap
+ namespace: {{ .Values.nsPrefix }}
+data:
+{{ tpl (.Files.Glob "resources/config/appc/opt/openecomp/appc/data/properties/*").AsConfig . | indent 2 }}
+---
+apiVersion: v1
+kind: ConfigMap
+metadata:
+ name: appc-openecomp-appc-svclogic-config-configmap
+ namespace: {{ .Values.nsPrefix }}
+data:
+{{ tpl (.Files.Glob "resources/config/appc/opt/openecomp/appc/svclogic/config/svclogic.properties").AsConfig . | indent 2 }}
+---
+apiVersion: v1
+kind: ConfigMap
+metadata:
+ name: appc-openecomp-appc-svclogic-bin-configmap
+ namespace: {{ .Values.nsPrefix }}
+data:
+{{ tpl (.Files.Glob "resources/config/appc/opt/openecomp/appc/svclogic/bin/showActiveGraphs.sh").AsConfig . | indent 2 }}
+---
+apiVersion: v1
+kind: ConfigMap
+metadata:
+ name: appc-openecomp-appc-bin-configmap
+ namespace: {{ .Values.nsPrefix }}
+data:
+{{ tpl (.Files.Glob "resources/config/appc/opt/openecomp/appc/bin/*").AsConfig . | indent 2 }}
+---
+apiVersion: v1
+kind: ConfigMap
+metadata:
+ name: appc-onap-sdnc-data-properties-configmap
+ namespace: {{ .Values.nsPrefix }}
+data:
+{{ tpl (.Files.Glob "resources/config/appc/opt/onap/sdnc/data/properties/*").AsConfig . | indent 2 }}
+---
+apiVersion: v1
+kind: ConfigMap
+metadata:
+ name: appc-onap-sdnc-svclogic-config-configmap
+ namespace: {{ .Values.nsPrefix }}
+data:
+{{ tpl (.Files.Glob "resources/config/appc/opt/onap/sdnc/svclogic/config/svclogic.properties").AsConfig . | indent 2 }}
+---
+apiVersion: v1
+kind: ConfigMap
+metadata:
+ name: appc-onap-sdnc-svclogic-bin-configmap
+ namespace: {{ .Values.nsPrefix }}
+data:
+{{ tpl (.Files.Glob "resources/config/appc/opt/onap/sdnc/svclogic/bin/showActiveGraphs.sh").AsConfig . | indent 2 }}
+---
+apiVersion: v1
+kind: ConfigMap
+metadata:
+ name: appc-onap-sdnc-bin-configmap
+ namespace: {{ .Values.nsPrefix }}
+data:
+{{ tpl (.Files.Glob "resources/config/appc/opt/onap/sdnc/bin/*").AsConfig . | indent 2 }}
+---
+apiVersion: v1
+kind: ConfigMap
+metadata:
+ name: appc-scripts-configmap
+ namespace: {{ .Values.nsPrefix }}
+data:
+{{ tpl (.Files.Glob "resources/scripts/*").AsConfig . | indent 2 }}
+#{{ end }}
kind: Deployment
metadata:
name: appc
- namespace: "{{ .Values.nsPrefix }}-appc"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.appcReplicas }}
selector:
value: /opt/openecomp/appc/data/properties
- name: DMAAP_TOPIC_ENV
value: SUCCESS
+ - name: MYSQL_HOST
+ value: appc-dbhost
image: {{ .Values.image.appc }}
imagePullPolicy: {{ .Values.pullPolicy }}
name: appc-controller-container
- mountPath: /etc/localtime
name: localtime
readOnly: true
+ - mountPath: /opt/openecomp/appc/data/properties/dblib.properties
+ name: appc-openecomp-appc-data-properties
+ subPath: dblib.properties
+ - mountPath: /opt/openecomp/appc/data/properties/svclogic.properties
+ name: appc-openecomp-appc-data-properties
+ subPath: svclogic.properties
- mountPath: /opt/openecomp/appc/data/properties/appc.properties
- name: appc-conf
+ name: appc-openecomp-appc-data-properties
subPath: appc.properties
- mountPath: /opt/openecomp/appc/data/properties/aaiclient.properties
- name: appc-conf
+ name: appc-openecomp-appc-data-properties
subPath: aaiclient.properties
+ - mountPath: /opt/openecomp/appc/svclogic/config/svclogic.properties
+ name: appc-openecomp-appc-svclogic-config
+ subPath: svclogic.properties
+ - mountPath: /opt/openecomp/appc/svclogic/bin/showActiveGraphs.sh
+ name: appc-openecomp-appc-svclogic-bin
+ subPath: showActiveGraphs.sh
+ - mountPath: /opt/openecomp/appc/bin/startODL.sh
+ name: appc-openecomp-appc-bin
+ subPath: startODL.sh
+ - mountPath: /opt/openecomp/appc/bin/installAppcDb.sh
+ name: appc-openecomp-appc-bin
+ subPath: installAppcDb.sh
+ - mountPath: /opt/onap/sdnc/data/properties/dblib.properties
+ name: appc-onap-sdnc-data-properties
+ subPath: dblib.properties
+ - mountPath: /opt/onap/sdnc/data/properties/svclogic.properties
+ name: appc-onap-sdnc-data-properties
+ subPath: svclogic.properties
- mountPath: /opt/onap/sdnc/data/properties/aaiclient.properties
- name: appc-conf
+ name: appc-onap-sdnc-data-properties
subPath: aaiclient.properties
+ - mountPath: /opt/onap/sdnc/svclogic/config/svclogic.properties
+ name: appc-onap-sdnc-svclogic-config
+ subPath: svclogic.properties
+ - mountPath: /opt/onap/sdnc/svclogic/bin/showActiveGraphs.sh
+ name: appc-onap-sdnc-svclogic-bin
+ subPath: showActiveGraphs.sh
+ - mountPath: /opt/onap/sdnc/bin/startODL.sh
+ name: appc-onap-sdnc-bin
+ subPath: startODL.sh
+ - mountPath: /opt/onap/sdnc/bin/installSdncDb.sh
+ name: appc-onap-sdnc-bin
+ subPath: installSdncDb.sh
- mountPath: /var/log/onap
name: appc-logs
- mountPath: /opt/opendaylight/current/etc/org.ops4j.pax.logging.cfg
emptyDir: {}
- name: appc-data-filebeat
emptyDir: {}
- - name: appc-conf
+ - name: appc-openecomp-appc-data-properties
configMap:
- name: appc-conf-configmap
+ name: appc-openecomp-appc-data-properties-configmap
+ - name: appc-openecomp-appc-svclogic-config
+ configMap:
+ name: appc-openecomp-appc-svclogic-config-configmap
+ - name: appc-openecomp-appc-svclogic-bin
+ configMap:
+ name: appc-openecomp-appc-svclogic-bin-configmap
+ defaultMode: 0755
+ - name: appc-openecomp-appc-bin
+ configMap:
+ name: appc-openecomp-appc-bin-configmap
+ defaultMode: 0755
+ - name: appc-onap-sdnc-data-properties
+ configMap:
+ name: appc-onap-sdnc-data-properties-configmap
+ - name: appc-onap-sdnc-svclogic-config
+ configMap:
+ name: appc-onap-sdnc-svclogic-config-configmap
+ - name: appc-onap-sdnc-svclogic-bin
+ configMap:
+ name: appc-onap-sdnc-svclogic-bin-configmap
+ defaultMode: 0755
+ - name: appc-onap-sdnc-bin
+ configMap:
+ name: appc-onap-sdnc-bin-configmap
+ defaultMode: 0755
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
#{{ end }}
--- /dev/null
+#{{ if not .Values.disableAppcAppc }}
+apiVersion: v1
+kind: ConfigMap
+metadata:
+ name: appc-dgbuilder-conf-configmap
+ namespace: {{ .Values.nsPrefix }}
+data:
+{{ tpl (.Files.Glob "resources/config/dgbuilder/*").AsConfig . | indent 2 }}
+---
+apiVersion: v1
+kind: ConfigMap
+metadata:
+ name: appc-dgbuilder-scripts-configmap
+ namespace: {{ .Values.nsPrefix }}
+data:
+{{ tpl (.Files.Glob "resources/scripts/dgbuilder/*").AsConfig . | indent 2 }}
+#{{ end }}
kind: ConfigMap
metadata:
name: appc-filebeat-configmap
- namespace: {{ .Values.nsPrefix }}-appc
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/log/filebeat/log4j/*").AsConfig . | indent 2 }}
---
kind: ConfigMap
metadata:
name: appc-logging-cfg-configmap
- namespace: {{ .Values.nsPrefix }}-appc
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/log/*").AsConfig . | indent 2 }}
#{{ end }}
apiVersion: v1
kind: PersistentVolume
metadata:
- name: "{{ .Values.nsPrefix }}-appc-db"
- namespace: "{{ .Values.nsPrefix }}-appc"
+ name: "{{ .Values.nsPrefix }}-appc"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
- name: "{{ .Values.nsPrefix }}-appc-db"
+ name: "{{ .Values.nsPrefix }}-appc"
spec:
capacity:
storage: 2Gi
apiVersion: v1
metadata:
name: appc-db
- namespace: "{{ .Values.nsPrefix }}-appc"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
accessModes:
- ReadWriteMany
storage: 2Gi
selector:
matchLabels:
- name: "{{ .Values.nsPrefix }}-appc-db"
+ name: "{{ .Values.nsPrefix }}-appc"
#{{ end }}
kind: Deployment
metadata:
name: appc-dbhost
- namespace: "{{ .Values.nsPrefix }}-appc"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.dbReplicas }}
selector:
kind: Deployment
metadata:
name: appc-dgbuilder
- namespace: "{{ .Values.nsPrefix }}-appc"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.dgbuilderReplicas }}
selector:
- name: localtime
mountPath: /etc/localtime
readOnly: true
+ - name: dgbuilder-conf
+ mountPath: /opt/onap/sdnc/dgbuilder/releases/sdnc1.0/conf/svclogic.properties
+ subPath: svclogic.properties
+ - name: dgbuilder-conf
+ mountPath: /opt/onap/sdnc/dgbuilder/svclogic/svclogic.properties
+ subPath: svclogic.properties
+ - name: dgbuilder-scripts
+ mountPath: /opt/onap/sdnc/dgbuilder/createReleaseDir.sh
+ subPath: createReleaseDir.sh
+ - name: dgbuilder-scripts
+ mountPath: /opt/onap/sdnc/dgbuilder/releases/sdnc1.0/customSettings.js
+ subPath: customSettings.js
volumes:
- name: localtime
hostPath:
path: /etc/localtime
+ - name: dgbuilder-conf
+ configMap:
+ name: appc-dgbuilder-conf-configmap
+ - name: dgbuilder-scripts
+ configMap:
+ name: appc-dgbuilder-scripts-configmap
+ defaultMode: 0755
restartPolicy: Always
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
kind: Service
metadata:
name: clamp-mariadb
- namespace: "{{ .Values.nsPrefix }}-clamp"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: clamp-mariadb
spec:
kind: Service
metadata:
name: clamp
- namespace: "{{ .Values.nsPrefix }}-clamp"
+ namespace: "{{ .Values.nsPrefix }}"
+ annotations:
+ msb.onap.org/service-info: '[
+ {
+ "serviceName": "clamp",
+ "version": "v1",
+ "url": "/restservices/clds/v1",
+ "protocol": "REST",
+ "port": "8080",
+ "visualRange":"1"
+ }
+ ]'
spec:
ports:
- name: clamp
labels:
app: clamp
name: clamp
- namespace: "{{ .Values.nsPrefix }}-clamp"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
+ replicas: {{ .Values.clampReplicas }}
selector:
matchLabels:
app: clamp
- env:
- name: SPRING_APPLICATION_JSON
value: '{
- "spring.datasource.camunda.url": "jdbc:mariadb:sequential://clamp-mariadb.{{ .Values.nsPrefix }}-clamp:3306/camundabpm?verifyServerCertificate=false&useSSL=false&requireSSL=false&autoReconnect=true&retriesAllDown=2147483647&failoverLoopRetries=2147483647",
- "spring.datasource.cldsdb.url": "jdbc:mariadb:sequential://clamp-mariadb.{{ .Values.nsPrefix }}-clamp:3306/cldsdb4?verifyServerCertificate=false&useSSL=false&requireSSL=false&autoReconnect=true&retriesAllDown=2147483647&failoverLoopRetries=2147483647"
+ "spring.datasource.camunda.url": "jdbc:mariadb:sequential://clamp-mariadb.{{ .Values.nsPrefix }}:3306/camundabpm?verifyServerCertificate=false&useSSL=false&requireSSL=false&autoReconnect=true&retriesAllDown=2147483647&failoverLoopRetries=2147483647",
+ "spring.datasource.cldsdb.url": "jdbc:mariadb:sequential://clamp-mariadb.{{ .Values.nsPrefix }}:3306/cldsdb4?verifyServerCertificate=false&useSSL=false&requireSSL=false&autoReconnect=true&retriesAllDown=2147483647&failoverLoopRetries=2147483647"
}'
image: {{ .Values.image.clampImage }}:{{ .Values.image.clampVersion }}
imagePullPolicy: {{ .Values.pullPolicy }}
kind: ConfigMap
metadata:
name: clamp-entrypoint-initdb-configmap
- namespace: {{ .Values.nsPrefix }}-clamp
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/mariadb/docker-entrypoint-initdb.d/*").AsConfig . | indent 2 }}
---
kind: ConfigMap
metadata:
name: clamp-entrypoint-drop-configmap
- namespace: {{ .Values.nsPrefix }}-clamp
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/mariadb/docker-entrypoint-initdb.d/drop/*").AsConfig . | indent 2 }}
---
kind: ConfigMap
metadata:
name: clamp-entrypoint-bulkload-configmap
- namespace: {{ .Values.nsPrefix }}-clamp
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/mariadb/docker-entrypoint-initdb.d/bulkload/*").AsConfig . | indent 2 }}
---
kind: ConfigMap
metadata:
name: clamp-mariadb-conf-configmap
- namespace: {{ .Values.nsPrefix }}-clamp
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/mariadb/conf.d/conf1/*").AsConfig . | indent 2 }}
#{{ end }}
kind: Deployment
metadata:
name: clamp-mariadb
- namespace: "{{ .Values.nsPrefix }}-clamp"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
- replicas: 1
+ replicas: {{ .Values.clampMariaDbReplicas }}
selector:
matchLabels:
app: clamp-mariadb
apiVersion: v1
kind: PersistentVolume
metadata:
- name: "{{ .Values.nsPrefix }}-clamp-db"
- namespace: "{{ .Values.nsPrefix }}-clamp"
+ name: "{{ .Values.nsPrefix }}-clamp"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
- name: "{{ .Values.nsPrefix }}-clamp-db"
+ name: "{{ .Values.nsPrefix }}-clamp"
spec:
capacity:
storage: 2Gi
apiVersion: v1
metadata:
name: clamp-db
- namespace: "{{ .Values.nsPrefix }}-clamp"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
accessModes:
- ReadWriteMany
storage: 2Gi
selector:
matchLabels:
- name: "{{ .Values.nsPrefix }}-clamp-db"
+ name: "{{ .Values.nsPrefix }}-clamp"
#{{ end }}
nodePortPrefix: 302
mysqlPassword: strong_pitchou
dataRootDir: /dockerdata-nfs
+clampReplicas: 1
+clampMariaDbReplicas: 1
image:
readiness: oomk8s/readiness-check:1.1.0
clampImage: nexus3.onap.org:10001/onap/clamp
kind: Service
metadata:
name: cli
- namespace: "{{ .Values.nsPrefix }}-cli"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: cli
spec:
labels:
app: cli
name: cli
- namespace: "{{ .Values.nsPrefix }}-cli"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
selector:
matchLabels:
apiVersion: v1
description: Common templates for inclusion in other charts
-name: common-templates
+name: common
version: 2.0.0
+++ /dev/null
-EXCLUDES := test
-HELM_CHARTS := $(filter-out $(EXCLUDES), $(patsubst %/.,%,$(wildcard */.)))
-
-.PHONY: $(EXCLUDES) $(HELM_CHARTS)
-
-all: $(HELM_CHARTS)
-
-$(HELM_CHARTS):
- @echo "\n[$@]"
- @make lint-$@
-
-make-%:
- @if [ -f $*/Makefile ]; then make -C $*; fi
-
-dep-%: make-%
- @if [ -f $*/requirements.yaml ]; then helm dep up $*; fi
-
-lint-%: dep-%
- @if [ -f $*/Chart.yaml ]; then helm lint $*; fi
-
-clean:
- @rm -f */requirements.lock
- @rm -f *tgz */charts/*tgz
-%:
- @:
\ No newline at end of file
using the following value:
- .Values.nsPrefix : override namespace prefix
- - .Values.nsSuffix : override namespace suffix
*/}}
{{- define "common.namespace" -}}
- {{- $prefix := default .Release.Name .Values.nsPrefix -}}
- {{- $suffix := default .Chart.Name .Values.nsSuffix -}}
- {{- printf "%s-%s" $prefix $suffix -}}
+ {{- default .Release.Namespace .Values.nsPrefix -}}
{{- end -}}
# readiness check
readinessRepository: oomk8s
- readinessImage: readiness-check:1.0.0
+ readinessImage: readiness-check:1.1.0
# logging agent
loggingRepository: docker.elastic.co
mountPath: /dockerdata-nfs
# flag to enable debugging - application support required
- debugEnabled: true
\ No newline at end of file
+ debugEnabled: true
# replace the default 'onap' namespace qualification of K8s hostnames within the config files
SED_NS_PATHS="/config-init/$NAMESPACE/"
SED_NS_STRINGS=(
- "s/\.onap-/\.${NAMESPACE}-/g"
- "s/kubectl -n onap/kubectl -n ${NAMESPACE}/g"
+ "s/\.namespace-placeholder/\.${NAMESPACE}/g"
+ "s/kubectl -n namespace-placeholder/kubectl -n ${NAMESPACE}/g"
)
SED_NS_STRING=$(concat_array "${SED_NS_STRINGS[@]}")
find $SED_NS_PATHS -type f -exec sed -i -e "${SED_NS_STRING}" {} \;
+++ /dev/null
-if curl -s -X PUT http://elasticsearch.onap-aai:9200/searchhealth/stats/testwrite -d @/consul/config/scripts/aai-search-storage-write-doc.txt | grep '\"created\":true'; then
- if curl -s -X DELETE http://elasticsearch.onap-aai:9200/searchhealth/stats/testwrite | grep '\"failed\":0'; then
- if curl -s -X GET http://elasticsearch.onap-aai:9200/searchhealth/stats/testwrite | grep '\"found\":false'; then
- echo Successful PUT, DELETE, GET from Search Document Storage 2>&1
- exit 0
- else
- echo Failed GET from Search Document Storage 2>&1
- exit 1
- fi
- else
- echo Failed DELETE from Search Document Storage 2>&1
- exit 1
- fi
-else
- echo Failed PUT from Search Document Storage 2>&1
- exit 1
-fi
+++ /dev/null
-APPC_DBHOST_POD=$(/consul/config/bin/kubectl -n onap-appc get pod | grep -o "appc-dbhost-[^[:space:]]*")
-if [ -n "$APPC_DBHOST_POD" ]; then
- if /consul/config/bin/kubectl -n onap-appc exec -it $APPC_DBHOST_POD -- ./healthcheck.sh |grep -i "mysqld is alive"; then
- echo Success. APPC DBHost is running. 2>&1
- exit 0
- else
- echo Failed. APPC DBHost is not running. 2>&1
- exit 1
- fi
-else
- echo Failed. APPC DBHost is offline. 2>&1
- exit 1
-fi
+++ /dev/null
-
-NAME=$(/consul/config/bin/kubectl -n onap-aai get pod | grep -o "gremlin[^[:space:]]*")
-
-if [ -n "$NAME" ]; then
- if /consul/config/bin/kubectl -n onap-aai exec -it $NAME -- ps -efww | grep 'java' | grep 'gremlin-server' > /dev/null; then
-
- echo Success. Gremlin Server process is running. 2>&1
- exit 0
- else
- echo Failed. Gremlin Server process is not running. 2>&1
- exit 1
- fi
-else
- echo Failed. Gremlin Server container is offline. 2>&1
- exit 1
-fi
+++ /dev/null
-SDNC_DBHOST_POD=$(/consul/config/bin/kubectl -n onap-sdnc get pod | grep -o "sdnc-dbhost-[^[:space:]]*")
-if [ -n "$SDNC_DBHOST_POD" ]; then
- if /consul/config/bin/kubectl -n onap-sdnc exec -it $SDNC_DBHOST_POD -- ./healthcheck.sh |grep -i "mysqld is alive"; then
- echo Success. SDNC DBHost is running. 2>&1
- exit 0
- else
- echo Failed. SDNC DBHost is not running. 2>&1
- exit 1
- fi
-else
- echo Failed. SDNC DBHost is offline. 2>&1
- exit 1
-fi
##
## Both Cambria and Kafka make use of Zookeeper.
##
-config.zk.servers=zookeeper.onap-dcae:2181
+config.zk.servers=zookeeper.namespace-placeholder:2181
#config.zk.servers=172.17.0.1:2181
-#dmaap.onap-dcae:2181
+#dmaap.namespace-placeholder:2181
#10.208.128.229:2181
#config.zk.root=/fe3c/cambria/config
## configurations (after removing "kafka.")
## if you want to change request.required.acks it can take this one value
#kafka.metadata.broker.list=localhost:9092,localhost:9093
-kafka.metadata.broker.list=kafka.onap-dcae:9092
+kafka.metadata.broker.list=kafka.namespace-placeholder:9092
#kafka.metadata.broker.list=172.17.0.1:9092
-#dmaap.onap-dcae:9092
+#dmaap.namespace-placeholder:9092
#10.208.128.229:9092
##kafka.request.required.acks=-1
#kafka.client.zookeeper=${config.zk.servers}
#List of logstash server ip addresses with port number.
#But, in our case, this will be the loadbalancer IP address.
#For the below property to work the loadbalancer or logstash should expose 5044 port to listen the filebeat events or port in the property should be changed appropriately.
- hosts: ["logstash.onap-log:5044"]
+ hosts: ["logstash.namespace-placeholder:5044"]
#If enable will do load balancing among availabe Logstash, automatically.
loadbalance: true
#List of logstash server ip addresses with port number.
#But, in our case, this will be the loadbalancer IP address.
#For the below property to work the loadbalancer or logstash should expose 5044 port to listen the filebeat events or port in the property should be changed appropriately.
- hosts: ["logstash.onap-log:5044"]
+ hosts: ["logstash.namespace-placeholder:5044"]
#If enable will do load balancing among availabe Logstash, automatically.
loadbalance: true
##
## Both Cambria and Kafka make use of Zookeeper.
##
-config.zk.servers=zookeeper.onap-message-router:2181
+config.zk.servers=zookeeper.namespace-placeholder:2181
#config.zk.servers=172.17.0.1:2181
-#dmaap.onap-message-router:2181
+#dmaap.namespace-placeholder:2181
#10.208.128.229:2181
#config.zk.root=/fe3c/cambria/config
## configurations (after removing "kafka.")
## if you want to change request.required.acks it can take this one value
#kafka.metadata.broker.list=localhost:9092,localhost:9093
-kafka.metadata.broker.list=global-kafka.onap-message-router:9092
+kafka.metadata.broker.list=global-kafka.namespace-placeholder:9092
#kafka.metadata.broker.list=172.17.0.1:9092
-#dmaap.onap-message-router:9092
+#dmaap.namespace-placeholder:9092
#10.208.128.229:9092
##kafka.request.required.acks=-1
#kafka.client.zookeeper=${config.zk.servers}
+++ /dev/null
-# File generated from /opt/config
-#
-GLOBAL_INJECTED_AAI1_IP_ADDR = "aai-service.onap-aai"
-GLOBAL_INJECTED_AAI2_IP_ADDR = "N/A"
-GLOBAL_INJECTED_APPC_IP_ADDR = "sdnhost.onap-appc"
-GLOBAL_INJECTED_ARTIFACTS_VERSION = "1.1.0-SNAPSHOT"
-GLOBAL_INJECTED_CLAMP_IP_ADDR = "clamp.onap-clamp"
-GLOBAL_INJECTED_CLOUD_ENV = "openstack"
-GLOBAL_INJECTED_DCAE_IP_ADDR = "dcae-controller.onap-dcae"
-GLOBAL_INJECTED_DNS_IP_ADDR = "10.0.100.1"
-GLOBAL_INJECTED_DOCKER_VERSION = "1.1-STAGING-latest"
-#GLOBAL_INJECTED_EXTERNAL_DNS = "N/A"
-GLOBAL_INJECTED_GERRIT_BRANCH = "master"
-GLOBAL_INJECTED_KEYSTONE = "OPENSTACK_KEYSTONE_IP_HERE"
-GLOBAL_INJECTED_MR_IP_ADDR = "dmaap.onap-message-router"
-GLOBAL_INJECTED_MSO_IP_ADDR = "mso.onap-mso"
-GLOBAL_INJECTED_NETWORK = "OPENSTACK_NETWORK_ID_WITH_ONAP_ROUTE_HERE"
-GLOBAL_INJECTED_NEXUS_DOCKER_REPO = "nexus3.onap.org:10001"
-GLOBAL_INJECTED_NEXUS_PASSWORD = "docker"
-GLOBAL_INJECTED_NEXUS_REPO = "https://nexus.onap.org/content/sites/raw"
-GLOBAL_INJECTED_NEXUS_USERNAME = "docker"
-GLOBAL_INJECTED_OPENO_IP_ADDR = "msb-iag.onap-msb"
-GLOBAL_INJECTED_OPENSTACK_PASSWORD = "OPENSTACK_PASSWORD_HERE"
-GLOBAL_INJECTED_OPENSTACK_TENANT_ID = "OPENSTACK_TENANT_ID_HERE"
-GLOBAL_INJECTED_OPENSTACK_USERNAME = "OPENSTACK_USERNAME_HERE"
-GLOBAL_INJECTED_POLICY_IP_ADDR = "pypdp.onap-policy"
-GLOBAL_INJECTED_POLICY_HEALTHCHECK_IP_ADDR = "drools.onap-policy"
-GLOBAL_INJECTED_PORTAL_IP_ADDR = "portalapps.onap-portal"
-GLOBAL_INJECTED_REGION = "OPENSTACK_REGION_HERE"
-GLOBAL_INJECTED_REMOTE_REPO = "http://gerrit.onap.org/r/testsuite/properties.git"
-GLOBAL_INJECTED_SDC_IP_ADDR = "sdc-be.onap-sdc"
-GLOBAL_INJECTED_SDC_FE_IP_ADDR = "sdc-fe.onap-sdc"
-GLOBAL_INJECTED_SDC_BE_IP_ADDR = "sdc-be.onap-sdc"
-GLOBAL_INJECTED_SDNC_IP_ADDR = "sdnhost.onap-sdnc"
-GLOBAL_INJECTED_SDNC_PORTAL_IP_ADDR = "sdnc-portal.onap-sdnc"
-GLOBAL_INJECTED_SO_IP_ADDR = "mso.onap-mso"
-GLOBAL_INJECTED_VID_IP_ADDR = "vid-server.onap-vid"
-GLOBAL_INJECTED_VM_FLAVOR = "OPENSTACK_FLAVOUR_MEDIUM_HERE"
-GLOBAL_INJECTED_VM_IMAGE_NAME = "UBUNTU_14_IMAGE_NAME_HERE"
-GLOBAL_INJECTED_PUBLIC_NET_ID = "OPENSTACK_PUBLIC_NET_ID_HERE"
-
-GLOBAL_INJECTED_PROPERTIES = {
- "GLOBAL_INJECTED_AAI1_IP_ADDR" : "aai-service.onap-aai",
- "GLOBAL_INJECTED_APPC_IP_ADDR" : "sdnhost.onap-appc",
- "GLOBAL_INJECTED_ARTIFACTS_VERSION" : "1.1.0-SNAPSHOT",
- "GLOBAL_INJECTED_CLAMP_IP_ADDR" : "clamp.onap-clamp",
- "GLOBAL_INJECTED_CLOUD_ENV" : "openstack",
- "GLOBAL_INJECTED_DCAE_IP_ADDR" : "dcae-controller.onap-dcae",
- "GLOBAL_INJECTED_DNS_IP_ADDR" : "10.0.100.1",
- "GLOBAL_INJECTED_DOCKER_VERSION" : "1.1-STAGING-latest",
- "GLOBAL_INJECTED_GERRIT_BRANCH" : "master",
- "GLOBAL_INJECTED_KEYSTONE" : "OPENSTACK_KEYSTONE_IP_HERE",
- "GLOBAL_INJECTED_MR_IP_ADDR" : "dmaap.onap-message-router",
- "GLOBAL_INJECTED_MSO_IP_ADDR" : "mso.onap-mso",
- "GLOBAL_INJECTED_NETWORK" : "OPENSTACK_NETWORK_ID_WITH_ONAP_ROUTE_HERE",
- "GLOBAL_INJECTED_NEXUS_DOCKER_REPO" : "nexus3.onap.org:10001",
- "GLOBAL_INJECTED_NEXUS_PASSWORD" : "docker",
- "GLOBAL_INJECTED_NEXUS_REPO" : "https://nexus.onap.org/content/sites/raw",
- "GLOBAL_INJECTED_NEXUS_USERNAME" : "docker",
- "GLOBAL_INJECTED_OPENO_IP_ADDR" : "msb-iag.onap-msb",
- "GLOBAL_INJECTED_OPENSTACK_PASSWORD" : "OPENSTACK_PASSWORD_HERE",
- "GLOBAL_INJECTED_OPENSTACK_TENANT_ID" : "OPENSTACK_TENANT_ID_HERE",
- "GLOBAL_INJECTED_OPENSTACK_USERNAME" : "OPENSTACK_USERNAME_HERE",
- "GLOBAL_INJECTED_POLICY_IP_ADDR" : "pypdp.onap-policy",
- "GLOBAL_INJECTED_POLICY_HEALTHCHECK_IP_ADDR" : "drools.onap-policy",
- "GLOBAL_INJECTED_PORTAL_IP_ADDR" : "portalapps.onap-portal",
- "GLOBAL_INJECTED_REGION" : "OPENSTACK_REGION_HERE",
- "GLOBAL_INJECTED_REMOTE_REPO" : "http://gerrit.onap.org/r/testsuite/properties.git",
- "GLOBAL_INJECTED_SDC_FE_IP_ADDR" : "sdc-fe.onap-sdc",
- "GLOBAL_INJECTED_SDC_BE_IP_ADDR" : "sdc-be.onap-sdc",
- "GLOBAL_INJECTED_SDNC_IP_ADDR" : "sdnhost.onap-sdnc",
- "GLOBAL_INJECTED_SDNC_PORTAL_IP_ADDR" : "sdnc-portal.onap-sdnc",
- "GLOBAL_INJECTED_SO_IP_ADDR" : "mso.onap-mso",
- "GLOBAL_INJECTED_VID_IP_ADDR" : "vid-server.onap-vid",
- "GLOBAL_INJECTED_VM_FLAVOR" : "OPENSTACK_FLAVOUR_MEDIUM_HERE",
- "GLOBAL_INJECTED_VM_IMAGE_NAME" : "UBUNTU_14_IMAGE_NAME_HERE",
- "GLOBAL_INJECTED_PUBLIC_NET_ID" : "OPENSTACK_PUBLIC_NET_ID_HERE"
-}
OPENSTACK_FLAVOUR_MEDIUM: ""
OPENSTACK_SERVICE_TENANT_NAME: ""
DMAAP_TOPIC: ""
-DEMO_ARTIFACTS_VERSION: ""
\ No newline at end of file
+DEMO_ARTIFACTS_VERSION: ""
spec:
containers:
- name: {{ .Chart.Name }}
- image: "{{ .Values.image.repository }}:{{ .Values.image.tag }}"
+ image: "{{ .Values.image.repository }}"
imagePullPolicy: {{ .Values.image.pullPolicy }}
envFrom:
- configMapRef:
fieldRef:
fieldPath: metadata.namespace
- name: NAMESPACE_PREFIX
- value: {{ .Values.nsPrefix }}
+ value: {{ .Values.nsPrefix }}
- name: DOCKER_SHARE_PATH
value: {{ .Values.dockerSharePath }}
volumeMounts:
- name: config-init-root
hostPath:
path: {{ .Values.dockerSharePath }}
- restartPolicy: Never
\ No newline at end of file
+ restartPolicy: Never
nsPrefix: onap
dockerSharePath: /dockerdata-nfs
image:
- repository: oomk8s/config-init
- #master => Beijing (major release uprev)
- tag: 2.0.0-SNAPSHOT
+ repository: oomk8s/config-init:2.0.0-SNAPSHOT
pullPolicy: Always
{
"id": "data-router-process",
"name": "Synapse Presence",
- "script": "/consul/config/scripts/data-router-script.sh",
+ "script": "/consul/scripts/data-router-script.sh",
"interval": "15s",
"timeout": "1s"
}
{
"id": "hbase-aai",
"name": "HBase Health Check",
- "http": "http://hbase.onap-aai:8080/status/cluster",
+ "http": "http://hbase.{{ .Values.nsPrefix }}:8080/status/cluster",
"method": "GET",
"header": {
"Cache-Control": ["no-cache"],
{
"id": "model-loader-process",
"name": "Model Loader Presence",
- "script": "/consul/config/scripts/model-loader-script.sh",
+ "script": "/consul/scripts/model-loader-script.sh",
"interval": "15s",
"timeout": "1s"
}
{
"id": "elasticsearch",
"name": "Search Data Service Document Store",
- "http": "http://elasticsearch.onap-aai:9200/_cat/indices?v",
+ "http": "http://aai-elasticsearch.{{ .Values.nsPrefix }}:9200/_cat/indices?v",
"interval": "15s",
"timeout": "1s"
},
{
"id": "elasticsearch-write-health",
"name": "Search Data Service Document Store Write Test",
- "script": "/consul/config/scripts/aai-search-storage-write-script.sh",
+ "script": "/consul/scripts/aai-search-storage-write-script.sh",
"interval": "60s"
},
{
"id": "search-data-service-availability",
"name": "Search Data Service Availability",
- "script": "curl -k --cert /consul/config/certs/client-cert-onap.crt.pem --cert-type PEM --key /consul/config/bin/client-cert-onap.key.pem --key-type PEM https://search-data-service.onap-aai:9509/services/search-data-service/v1/jaxrsExample/jaxrs-services/echo/up 2>&1 | grep 'Up'",
+ "script": "curl -k --cert /consul/certs/client-cert-onap.crt.pem --cert-type PEM --key /consul/certs/client-cert-onap.key.pem --key-type PEM https://search-data-service.{{ .Values.nsPrefix }}:9509/services/search-data-service/v1/jaxrsExample/jaxrs-services/echo/up 2>&1 | grep 'Up'",
"interval": "15s"
},
{
"id": "search-data-service-api",
"name": "Search Data Service Operational Test",
- "script": "/consul/config/scripts/search-data-service-availability.sh",
+ "script": "/consul/scripts/search-data-service-availability.sh",
"interval": "15s",
"timeout": "1s"
}
{
"id": "aai-service",
"name": "Core A&AI",
- "http": "https://aai-service.onap-aai:8443/aai/util/echo",
+ "http": "https://aai-service.{{ .Values.nsPrefix }}:8443/aai/util/echo",
"header": {
"Authorization": ["Basic QUFJOkFBSQ=="],
"X-TransactionId": ["ConsulHealthCheck"],
{
"id": "aai-resources",
"name": "Resources Microservice",
- "http": "https://aai-resources.onap-aai:8447/aai/util/echo",
+ "http": "https://aai-resources.{{ .Values.nsPrefix }}:8447/aai/util/echo",
"header": {
"Authorization": ["Basic QUFJOkFBSQ=="],
"X-TransactionId": ["ConsulHealthCheck"],
{
"id": "aai-traversal",
"name": "Traversal Microservice",
- "http": "https://aai-traversal.onap-aai:8446/aai/util/echo",
+ "http": "https://aai-traversal.{{ .Values.nsPrefix }}:8446/aai/util/echo",
"header": {
"Authorization": ["Basic QUFJOkFBSQ=="],
"X-TransactionId": ["ConsulHealthCheck"],
"tls_skip_verify": true,
"interval": "15s",
"timeout": "1s"
- },
- {
- "id": "gremlin-server",
- "name": "Graph Data Store",
- "script": "/consul/config/scripts/gremlin-script.sh",
- "interval": "15s",
- "timeout": "1s"
}
]
}
{
"id": "sparky-be-process",
"name": "UI Backend Presence",
- "script": "/consul/config/scripts/sparky-be-script.sh",
+ "script": "/consul/scripts/sparky-be-script.sh",
"interval": "15s",
"timeout": "1s"
}
{
"id": "tabular-backend",
"name": "Tabular Data Store Operational Test",
- "script": "/consul/config/scripts/tabular-db-availability.sh",
+ "script": "/consul/scripts/tabular-db-availability.sh",
"interval": "15s",
"timeout": "1s"
}
{
"id": "appc-dgbuilder",
"name": "APPC-Dgbuilder Server Health Check",
- "http": "http://dgbuilder.onap-appc:3000/",
+ "http": "http://appc-dgbuilder.{{ .Values.nsPrefix }}:3000/",
"method": "HEAD",
"header": {
"Authorization": ["Basic ZGd1c2VyOnRlc3QxMjM="],
{
"id": "appc-dbhost-healthcheck",
"name": "APPC DBHost Health Check",
- "script": "/consul/config/scripts/appc-dbhost-script.sh",
+ "script": "/consul/scripts/appc-dbhost-script.sh",
"interval": "10s",
"timeout": "1s"
}
"name": "Health Check: APPC-SDN-CTL-DB-01",
"checks": [
{
- "id": "sdnctldb01.onap-appc",
+ "id": "appc-sdnctldb01.{{ .Values.nsPrefix }}",
"name": "APPC SDNCTLDB01 Health Check",
- "tcp": "sdnctldb01.onap-appc:3306",
+ "tcp": "appc-sdnctldb01.{{ .Values.nsPrefix }}:3306",
"interval": "10s",
"timeout": "1s"
}
"name": "Health Check: APPC-SDN-CTL-DB-02",
"checks": [
{
- "id": "sdnctldb02.onap-appc",
+ "id": "appc-sdnctldb02.{{ .Values.nsPrefix }}",
"name": "APPC SDNCTLDB02 Health Check",
- "tcp": "sdnctldb02.onap-appc:3306",
+ "tcp": "appc-sdnctldb02.{{ .Values.nsPrefix }}:3306",
"interval": "10s",
"timeout": "1s"
}
{
"id": "appc-sdnhost",
"name": "APPC SDN Host Health Check",
- "http": "http://sdnhost.onap-appc:8282/apidoc/explorer/index.html",
+ "http": "http://appc-sdnhost.{{ .Values.nsPrefix }}:8282/apidoc/explorer/index.html",
"method": "HEAD",
"header": {
"Authorization": ["Basic YWRtaW46S3A4Yko0U1hzek0wV1hsaGFrM2VIbGNzZTJnQXc4NHZhb0dHbUp2VXkyVQ=="],
{
"id": "log-elasticsearch-server",
"name": "Log Elastic Search Health Check",
- "http": "http://elasticsearch.onap-log:9200/_cluster/health?pretty",
+ "http": "http://elasticsearch.{{ .Values.nsPrefix }}:9200/_cluster/health?pretty",
"method": "GET",
"tls_skip_verify": true,
"interval": "15s",
{
"id": "log-elasticsearch-tcp",
"name": "Log Elastic Search TCP Health Check",
- "tcp": "elasticsearchtcp.onap-log:9300",
+ "tcp": "elasticsearchtcp.{{ .Values.nsPrefix }}:9300",
"interval": "15s",
"timeout": "1s"
}
{
"id": "log-kibana-server",
"name": "Log kibana Health Check",
- "http": "http://kibana.onap-log:5601/status",
+ "http": "http://kibana.{{ .Values.nsPrefix }}:5601/status",
"method": "HEAD",
"tls_skip_verify": true,
"interval": "15s",
{
"id": "log-logstash-internal-server-gi",
"name": "Log Stash Health Check - General Information",
- "http": "http://logstashinternal.onap-log:9600/?pretty",
+ "http": "http://logstashinternal.{{ .Values.nsPrefix }}:9600/?pretty",
"method": "GET",
"tls_skip_verify": true,
"interval": "15s",
{
"id": "log-logstash-internal-server-node-info",
"name": "Log Stash Health Check - Node Information",
- "http": "http://logstashinternal.onap-log:9600/_node/?pretty",
+ "http": "http://logstashinternal.{{ .Values.nsPrefix }}:9600/_node/?pretty",
"method": "GET",
"tls_skip_verify": true,
"interval": "15s",
{
"id": "log-logstash-internal-server-os-info",
"name": "Log Stash Health Check - OS Information",
- "http": "http://logstashinternal.onap-log:9600/_node/os?pretty",
+ "http": "http://logstashinternal.{{ .Values.nsPrefix }}:9600/_node/os?pretty",
"method": "GET",
"tls_skip_verify": true,
"interval": "15s",
{
"id": "log-logstash-internal-server-jvm-info",
"name": "Log Stash Health Check - JVM Information",
- "http": "http://logstashinternal.onap-log:9600/_node/jvm?pretty",
+ "http": "http://logstashinternal.{{ .Values.nsPrefix }}:9600/_node/jvm?pretty",
"method": "GET",
"tls_skip_verify": true,
"interval": "15s",
{
"id": "log-logstash-internal-server-plugin-info",
"name": "Log Stash Health Check - Plugin Information",
- "http": "http://logstashinternal.onap-log:9600/_node/plugins?pretty",
+ "http": "http://logstashinternal.{{ .Values.nsPrefix }}:9600/_node/plugins?pretty",
"method": "GET",
"tls_skip_verify": true,
"interval": "15s",
{
"id": "log-logstash-internal-server-node-stat",
"name": "Log Stash Health Check - Node Stats",
- "http": "http://logstashinternal.onap-log:9600/_node/stats?pretty",
+ "http": "http://logstashinternal.{{ .Values.nsPrefix }}:9600/_node/stats?pretty",
"method": "GET",
"tls_skip_verify": true,
"interval": "15s",
{
"id": "log-logstash-internal-server-jvm-stat",
"name": "Log Stash Health Check - JVM Stats",
- "http": "http://logstashinternal.onap-log:9600/_node/stats/jvm?pretty",
+ "http": "http://logstashinternal.{{ .Values.nsPrefix }}:9600/_node/stats/jvm?pretty",
"method": "GET",
"tls_skip_verify": true,
"interval": "15s",
{
"id": "log-logstash-internal-server-process-stat",
"name": "Log Stash Health Check - Process Stats",
- "http": "http://logstashinternal.onap-log:9600/_node/stats/process?pretty",
+ "http": "http://logstashinternal.{{ .Values.nsPrefix }}:9600/_node/stats/process?pretty",
"method": "GET",
"tls_skip_verify": true,
"interval": "15s",
{
"id": "log-logstash-internal-server-os-stat",
"name": "Log Stash Health Check - OS Stats",
- "http": "http://logstashinternal.onap-log:9600/_node/stats/os?pretty",
+ "http": "http://logstashinternal.{{ .Values.nsPrefix }}:9600/_node/stats/os?pretty",
"method": "GET",
"tls_skip_verify": true,
"interval": "15s",
{
"id": "log-logstash-tcp",
"name": "Log Stash File Beat TCP Health Check",
- "tcp": "logstash.onap-log:5044",
+ "tcp": "logstash.{{ .Values.nsPrefix }}:5044",
"interval": "15s",
"timeout": "1s"
}
"service": {
"name": "Health Check: Message Router - DMaaP",
"check": {
- "http": "http://dmaap.onap-message-router:3904/topics",
+ "http": "http://dmaap.{{ .Values.nsPrefix }}:3904/topics",
"interval": "30s",
"timeout": "1s"
}
"service": {
"name": "Health Check: Message Router - Kafka",
"check": {
- "script": "/consul/config/scripts/mr-kafka-health.sh",
+ "script": "/consul/scripts/mr-kafka-health.sh",
"interval": "30s",
"timeout": "1s"
}
"service": {
"name": "Health Check: Message Router - ZooKeeper",
"check": {
- "script": "/consul/config/scripts/mr-zookeeper-health.sh",
+ "script": "/consul/scripts/mr-zookeeper-health.sh",
"interval": "30s",
"timeout": "1s"
}
"name": "Health Check: MSB",
"checks": [
{
- "id": "msb-eag.onap-msb",
+ "id": "msb-eag.{{ .Values.nsPrefix }}",
"name": "MSB eag Health Check",
- "http": "http://msb-eag.onap-msb:80/iui/microservices/default.html",
+ "http": "http://msb-eag.{{ .Values.nsPrefix }}:80/iui/microservices/default.html",
"method": "HEAD",
"tls_skip_verify": true,
"interval": "15s",
"timeout": "1s"
},
{
- "id": "msb-iag.onap-msb",
+ "id": "msb-iag.{{ .Values.nsPrefix }}",
"name": "MSB iag Health Check",
- "http": "http://msb-iag.onap-msb:80/iui/microservices/default.html",
+ "http": "http://msb-iag.{{ .Values.nsPrefix }}:80/iui/microservices/default.html",
"method": "HEAD",
"tls_skip_verify": true,
"interval": "15s",
"timeout": "1s"
},
{
- "id": "msb-consul.onap-msb",
+ "id": "msb-consul.{{ .Values.nsPrefix }}",
"name": "MSB consul Health Check",
- "tcp": "msb-consul.onap-msb:8500",
+ "tcp": "msb-consul.{{ .Values.nsPrefix }}:8500",
"interval": "15s",
"timeout": "1s"
},
{
- "id": "msb-discovery.onap-msb",
+ "id": "msb-discovery.{{ .Values.nsPrefix }}",
"name": "MSB discovery Health Check",
- "tcp": "msb-discovery.onap-msb:10081",
+ "tcp": "msb-discovery.{{ .Values.nsPrefix }}:10081",
"interval": "15s",
"timeout": "1s"
}
{
"id": "mso-api-healthcheck",
"name": "MSO API Health Check",
- "script": "/consul/config/scripts/mso-api-script.sh",
+ "script": "/consul/scripts/mso-api-script.sh",
"interval": "10s",
"timeout": "1s"
},
{
"id": "mso-camunda-healthcheck",
"name": "MSO Camunda Health Check",
- "script": "/consul/config/scripts/mso-camunda-script.sh",
+ "script": "/consul/scripts/mso-camunda-script.sh",
"interval": "10s",
"timeout": "1s"
},
{
"id": "mso-jra-healthcheck",
"name": "MSO JRA Health Check",
- "script": "/consul/config/scripts/mso-jra-script.sh",
+ "script": "/consul/scripts/mso-jra-script.sh",
"interval": "10s",
"timeout": "1s"
}
{
"id": "mso-mariadb",
"name": "MSO Mariadb Health Check",
- "script": "/consul/config/scripts/mso-mariadb-script.sh",
+ "script": "/consul/scripts/mso-mariadb-script.sh",
"interval": "10s",
"timeout": "1s"
}
{
"id": "framework",
"name": "Framework Health Check",
- "http": "http://framework.onap-multicloud:9001/api/multicloud/v0/swagger.json",
+ "http": "http://framework.{{ .Values.nsPrefix }}:9001/api/multicloud/v0/swagger.json",
"method": "HEAD",
"header": {
"Cache-Control": ["no-cache"],
{
"id": "multicloud-ocata",
"name": "Multicloud Ocata Health Check",
- "http": "http://multicloud-ocata.onap-multicloud:9006/api/multicloud-ocata/v0/swagger.json",
+ "http": "http://multicloud-ocata.{{ .Values.nsPrefix }}:9006/api/multicloud-ocata/v0/swagger.json",
"method": "HEAD",
"header": {
"Cache-Control": ["no-cache"],
{
"id": "multicloud-vio",
"name": "Multicloud Vio Health Check",
- "http": "http://multicloud-vio.onap-multicloud:9004/api/multicloud-vio/v0/swagger.json",
+ "http": "http://multicloud-vio.{{ .Values.nsPrefix }}:9004/api/multicloud-vio/v0/swagger.json",
"method": "HEAD",
"header": {
"Cache-Control": ["no-cache"],
{
"id": "multicloud-windriver",
"name": "Multicloud Windriver Health Check",
- "http": "http://multicloud-windriver.onap-multicloud:9005/api/multicloud-titanium_cloud/v0/swagger.json",
+ "http": "http://multicloud-windriver.{{ .Values.nsPrefix }}:9005/api/multicloud-titanium_cloud/v0/swagger.json",
"method": "HEAD",
"header": {
"Cache-Control": ["no-cache"],
--- /dev/null
+if curl -s -X PUT http://aai-elasticsearch.{{ .Values.nsPrefix }}:9200/searchhealth/stats/testwrite -d @/consul/scripts/aai-search-storage-write-doc.txt | grep '\"created\":true'; then
+ if curl -s -X DELETE http://aai-elasticsearch.{{ .Values.nsPrefix }}:9200/searchhealth/stats/testwrite | grep '\"failed\":0'; then
+ if curl -s -X GET http://aai-elasticsearch.{{ .Values.nsPrefix }}:9200/searchhealth/stats/testwrite | grep '\"found\":false'; then
+ echo Successful PUT, DELETE, GET from Search Document Storage 2>&1
+ exit 0
+ else
+ echo Failed GET from Search Document Storage 2>&1
+ exit 1
+ fi
+ else
+ echo Failed DELETE from Search Document Storage 2>&1
+ exit 1
+ fi
+else
+ echo Failed PUT from Search Document Storage 2>&1
+ exit 1
+fi
--- /dev/null
+APPC_DBHOST_POD=$(/consul/bin/kubectl -n {{ .Values.nsPrefix }} get pod | grep -o "appc-dbhost-[^[:space:]]*")
+if [ -n "$APPC_DBHOST_POD" ]; then
+ if /consul/bin/kubectl -n {{ .Values.nsPrefix }} exec -it $APPC_DBHOST_POD -- ./healthcheck.sh |grep -i "mysqld is alive"; then
+ echo Success. APPC DBHost is running. 2>&1
+ exit 0
+ else
+ echo Failed. APPC DBHost is not running. 2>&1
+ exit 1
+ fi
+else
+ echo Failed. APPC DBHost is offline. 2>&1
+ exit 1
+fi
-NAME=$(/consul/config/bin/kubectl -n onap-aai get pod | grep -o "data-router[^[:space:]]*")
+NAME=$(/consul/bin/kubectl -n {{ .Values.nsPrefix }} get pod | grep -o "aai-data-router[^[:space:]]*")
if [ -n "$NAME" ]; then
- if /consul/config/bin/kubectl -n onap-aai exec -it $NAME -- ps -efww | grep 'java' | grep 'data-router' > /dev/null; then
+ if /consul/bin/kubectl -n {{ .Values.nsPrefix }} exec -it $NAME -- ps -efww | grep 'java' | grep 'data-router' > /dev/null; then
echo Success. Synapse process is running. 2>&1
exit 0
-NAME=$(/consul/config/bin/kubectl -n onap-aai get pod | grep -o "model-loader[^[:space:]]*")
+NAME=$(/consul/bin/kubectl -n {{ .Values.nsPrefix }} get pod | grep -o "aai-model-loader[^[:space:]]*")
if [ -n "$NAME" ]; then
- if /consul/config/bin/kubectl -n onap-aai exec -it $NAME -- ps -efww | grep 'java' | grep 'model-loader' > /dev/null; then
+ if /consul/bin/kubectl -n {{ .Values.nsPrefix }} exec -it $NAME -- ps -efww | grep 'java' | grep 'model-loader' > /dev/null; then
echo Success. Model Loader process is running. 2>&1
exit 0
-kafkapod=$(/consul/config/bin/kubectl -n onap-message-router get pod | grep -o "global-kafka-[^[:space:]]*")
+kafkapod=$(/consul/bin/kubectl -n {{ .Values.nsPrefix }} get pod | grep -o "message-router-global-kafka-[^[:space:]]*")
if [ -n "$kafkapod" ]; then
- if /consul/config/bin/kubectl -n onap-message-router exec -it $kafkapod -- ps ef | grep -i kafka; then
+ if /consul/bin/kubectl -n {{ .Values.nsPrefix }} exec -it $kafkapod -- ps ef | grep -i kafka; then
echo Success. Kafka process is running. 2>&1
exit 0
else
-zkpod=$(/consul/config/bin/kubectl -n onap-message-router get pod | grep -o "zookeeper-[^[:space:]]*")
+zkpod=$(/consul/bin/kubectl -n {{ .Values.nsPrefix }} get pod | grep -o "message-router-zookeeper-[^[:space:]]*")
if [ -n "$zkpod" ]; then
- if /consul/config/bin/kubectl -n onap-message-router exec -it $zkpod -- ps ef | grep -i zookeeper; then
+ if /consul/bin/kubectl -n {{ .Values.nsPrefix }} exec -it $zkpod -- ps ef | grep -i zookeeper; then
echo Success. Zookeeper process is running. 2>&1
exit 0
else
## Query the health check API.
-HEALTH_CHECK_ENDPOINT="http://mso.onap-mso.svc.cluster.local:8080/ecomp/mso/infra/healthcheck"
+HEALTH_CHECK_ENDPOINT="http://mso.{{ .Values.nsPrefix }}:8080/ecomp/mso/infra/healthcheck"
HEALTH_CHECK_RESPONSE=$(curl -s $HEALTH_CHECK_ENDPOINT)
READY=$(echo $HEALTH_CHECK_RESPONSE | grep "Application ready")
## Query the health check API.
-HEALTH_CHECK_ENDPOINT="http://mso.onap-mso.svc.cluster.local:8080/mso/healthcheck"
+HEALTH_CHECK_ENDPOINT="http://mso.{{ .Values.nsPrefix }}:8080/mso/healthcheck"
HEALTH_CHECK_RESPONSE=$(curl -s $HEALTH_CHECK_ENDPOINT)
READY=$(echo $HEALTH_CHECK_RESPONSE | grep "Application ready")
## Query the health check API.
-HEALTH_CHECK_ENDPOINT="http://mso.onap-mso.svc.cluster.local:8080/networks/rest/healthcheck"
+HEALTH_CHECK_ENDPOINT="http://mso.{{ .Values.nsPrefix }}:8080/networks/rest/healthcheck"
HEALTH_CHECK_RESPONSE=$(curl -s $HEALTH_CHECK_ENDPOINT)
READY=$(echo $HEALTH_CHECK_RESPONSE | grep "Application ready")
-NAME=$(/consul/config/bin/kubectl -n onap-mso get pod | grep -o "mariadb[^[:space:]]*")
+NAME=$(/consul/bin/kubectl -n {{ .Values.nsPrefix }} get pod | grep -o "mso-mariadb[^[:space:]]*")
if [ -n "$NAME" ]; then
- if /consul/config/bin/kubectl -n onap-mso exec -it $NAME -- bash -c 'mysqladmin status -u root -p$MYSQL_ROOT_PASSWORD' > /dev/null; then
+ if /consul/bin/kubectl -n {{ .Values.nsPrefix }} exec -it $NAME -- bash -c 'mysqladmin status -u root -p$MYSQL_ROOT_PASSWORD' > /dev/null; then
echo Success. mariadb process is running. 2>&1
exit 0
else
## Query the health check API.
-HEALTH_CHECK_ENDPOINT="http://sdc-fe.onap-sdc:8181/sdc1/rest/healthCheck"
+HEALTH_CHECK_ENDPOINT="http://sdc-fe.{{ .Values.nsPrefix }}:8181/sdc1/rest/healthCheck"
HEALTH_CHECK_RESPONSE=$(curl -s $HEALTH_CHECK_ENDPOINT)
## Strip out the ON_BOARDING section from the response XML (otherwise we will
## Query the health check API.
-HEALTH_CHECK_ENDPOINT="http://sdc-fe.onap-sdc:8181/sdc1/rest/healthCheck"
+HEALTH_CHECK_ENDPOINT="http://sdc-fe.{{ .Values.nsPrefix }}:8181/sdc1/rest/healthCheck"
HEALTH_CHECK_RESPONSE=$(curl -s $HEALTH_CHECK_ENDPOINT)
## Strip out the ON_BOARDING section from the response XML (otherwise we will
## Query the health check API.
-HEALTH_CHECK_ENDPOINT="http://sdc-fe.onap-sdc:8181/sdc1/rest/healthCheck"
+HEALTH_CHECK_ENDPOINT="http://sdc-fe.{{ .Values.nsPrefix }}:8181/sdc1/rest/healthCheck"
HEALTH_CHECK_RESPONSE=$(curl -s $HEALTH_CHECK_ENDPOINT)
## Strip out the ON_BOARDING section from the response XML (otherwise we will
## Query the health check API.
-HEALTH_CHECK_ENDPOINT="http://sdc-fe.onap-sdc:8181/sdc1/rest/healthCheck"
+HEALTH_CHECK_ENDPOINT="http://sdc-fe.{{ .Values.nsPrefix }}:8181/sdc1/rest/healthCheck"
HEALTH_CHECK_RESPONSE=$(curl -s $HEALTH_CHECK_ENDPOINT)
## Strip out the ON_BOARDING section from the response XML (otherwise we will
--- /dev/null
+SDNC_DBHOST_POD=$(/consul/bin/kubectl -n {{ .Values.nsPrefix }} get pod | grep -o "sdnc-dbhost-[^[:space:]]*")
+if [ -n "$SDNC_DBHOST_POD" ]; then
+ if /consul/bin/kubectl -n {{ .Values.nsPrefix }} exec -it $SDNC_DBHOST_POD -- ./healthcheck.sh |grep -i "mysqld is alive"; then
+ echo Success. SDNC DBHost is running. 2>&1
+ exit 0
+ else
+ echo Failed. SDNC DBHost is not running. 2>&1
+ exit 1
+ fi
+else
+ echo Failed. SDNC DBHost is offline. 2>&1
+ exit 1
+fi
#!/bin/sh
-SEARCH_SERVICE_NAME="search-data-service.onap-aai"
+SEARCH_SERVICE_NAME="search-data-service.{{ .Values.nsPrefix }}"
SEARCH_SERVICE_PORT=9509
HEALTH_CHECK_INDEX="healthcheck"
INDEX_URL="https://$SEARCH_SERVICE_NAME:$SEARCH_SERVICE_PORT/services/search-data-service/v1/search/indexes/$HEALTH_CHECK_INDEX"
INDEX_SCHEMA="{\"fields\":[{\"name\": \"field1\", \"data-type\": \"string\"}]}"
-
-SEARCH_CERT_FILE="/consul/config/certs/client-cert-onap.crt.pem"
-SEARCH_KEY_FILE="/consul/config/certs/client-cert-onap.key.pem"
-
+SEARCH_CERT_FILE="/consul/certs/client-cert-onap.crt.pem"
+SEARCH_KEY_FILE="/consul/certs/client-cert-onap.key.pem"
## Try to create an index via the Search Data Service API.
-CREATE_INDEX_RESP=$(curl -s -o /dev/null -w "%{http_code}" -k --cert $SEARCH_CERT_FILE --cert-type PEM --key $SEARCH_KEY_FILE --key-type PEM -d "$INDEX_SCHEMA" --header "Content-Type: application/json" --header "X-TransactionId: ConsulHealthCheck" -X PUT $INDEX_URL)
+CREATE_INDEX_RESP=$(curl -s -o /dev/null -w "%{http_code}" -k --cert $SEARCH_CERT_FILE --cert-type PEM --key $SEARCH_KEY_FILE --key-type PEM -d "$INDEX_SCHEMA" --header "Content-Type: application/json" --header "X-TransactionId: ConsulHealthCheck" -X PUT $INDEX_URL)
RESULT_STRING=" "
fi
## Now, clean up after ourselves.
-DELETE_INDEX_RESP=$(curl -s -o /dev/null -w "%{http_code}" -k --cert $SEARCH_CERT_FILE --cert-type PEM --key $SEARCH_KEY_FILE --key-type PEM -d "{ }" --header "Content-Type: application/json" --header "X-TransactionId: ConsulHealthCheck" -X DELETE $INDEX_URL)
+DELETE_INDEX_RESP=$(curl -s -o /dev/null -w "%{http_code}" -k --cert $SEARCH_CERT_FILE --cert-type PEM --key $SEARCH_KEY_FILE --key-type PEM -d "{ }" --header "Content-Type: application/json" --header "X-TransactionId: ConsulHealthCheck" -X DELETE $INDEX_URL)
if [ $DELETE_INDEX_RESP -eq 200 ]; then
RESULT_STRING="Service Is Able To Communicate With Back End"
-else
+else
RESULT_STRING="Service API Failure - $DELETE_INDEX_RESP"
- echo $RESULT_STRING
+ echo $RESULT_STRING
exit 1
fi
-NAME=$(/consul/config/bin/kubectl -n onap-aai get pod | grep -o "sparky-be[^[:space:]]*")
+NAME=$(/consul/bin/kubectl -n {{ .Values.nsPrefix }} get pod | grep -o "aai-sparky-be[^[:space:]]*")
if [ -n "$NAME" ]; then
- if /consul/config/bin/kubectl -n onap-aai exec -it $NAME -- ps -efww | grep 'java' | grep 'sparky' > /dev/null; then
+ if /consul/bin/kubectl -n {{ .Values.nsPrefix }} exec -it $NAME -- ps -efww | grep 'java' | grep 'sparky' > /dev/null; then
echo Success. UI Backend Service process is running. 2>&1
exit 0
# Query the Hbase service for the cluster status.
-GET_CLUSTER_STATUS_RESPONSE=$(curl -si -X GET -H "Accept: text/xml" http://hbase.onap-aai:8080/status/cluster)
+GET_CLUSTER_STATUS_RESPONSE=$(curl -si -X GET -H "Accept: text/xml" http://hbase.{{ .Values.nsPrefix }}:8080/status/cluster)
if [ -z "$GET_CLUSTER_STATUS_RESPONSE" ]; then
echo "Tabular store is unreachable."
-NAME=$(/consul/config/bin/kubectl -n onap-vid get pod | grep -o "vid-mariadb[^[:space:]]*")
+NAME=$(/consul/bin/kubectl -n {{ .Values.nsPrefix }} get pod | grep -o "vid-mariadb[^[:space:]]*")
if [ -n "$NAME" ]; then
- if /consul/config/bin/kubectl -n onap-vid exec -it $NAME -- bash -c 'mysqladmin status -u root -p$MYSQL_ROOT_PASSWORD' > /dev/null; then
+ if /consul/bin/kubectl -n {{ .Values.nsPrefix }} exec -it $NAME -- bash -c 'mysqladmin status -u root -p$MYSQL_ROOT_PASSWORD' > /dev/null; then
echo Success. mariadb process is running. 2>&1
exit 0
else
{
"id": "sdc-fe-healthcheck",
"name": "SDC Front End Health Check",
- "script": "/consul/config/scripts/sdc-fe-script.sh",
+ "script": "/consul/scripts/sdc-fe-script.sh",
"interval": "10s",
"timeout": "1s"
},
{
"id": "sdc-be-healthcheck",
"name": "SDC Back End Health Check",
- "script": "/consul/config/scripts/sdc-be-script.sh",
+ "script": "/consul/scripts/sdc-be-script.sh",
"interval": "10s",
"timeout": "1s"
},
{
"id": "sdc-titan-healthcheck",
"name": "SDC Titan Health Check",
- "script": "/consul/config/scripts/sdc-titan-script.sh",
+ "script": "/consul/scripts/sdc-titan-script.sh",
"interval": "10s",
"timeout": "1s"
},
{
"id": "sdc-cs-healthcheck",
"name": "SDC Cassandra Health Check",
- "script": "/consul/config/scripts/sdc-cs-script.sh",
+ "script": "/consul/scripts/sdc-cs-script.sh",
"interval": "10s",
"timeout": "1s"
},
{
"id": "sdc-catalog-healthcheck",
"name": "SDC Catalog Health Check",
- "http": "https://sdc-be.onap-sdc:8443/asdc/v1/catalog/services",
+ "http": "https://sdc-be.{{ .Values.nsPrefix }}:8443/asdc/v1/catalog/services",
"header": {
"Authorization": ["Basic dmlkOktwOGJKNFNYc3pNMFdYbGhhazNlSGxjc2UyZ0F3ODR2YW9HR21KdlV5MlU="],
"X-ECOMP-InstanceID": ["VID"],
{
"id": "sdnc-dbhost-healthcheck",
"name": "SDNC DBHOST Health Check",
- "script": "/consul/config/scripts/sdnc-dbhost-script.sh",
+ "script": "/consul/scripts/sdnc-dbhost-script.sh",
"interval": "10s",
"timeout": "1s"
}
{
"id": "sdnc-dgbuilder",
"name": "SDNC-DGbuilder Health Check",
- "http": "http://sdnc-dgbuilder.onap-sdnc:3000/",
+ "http": "http://sdnc-dgbuilder.{{ .Values.nsPrefix }}:3000/",
"method": "HEAD",
"header": {
"Authorization": ["Basic ZGd1c2VyOnRlc3QxMjM="],
{
"id": "odl-api-healthcheck",
"name": "SDNC API Health Check",
- "http": "http://sdnhost.onap-sdnc:8282/restconf/operations/SLI-API:healthcheck",
+ "http": "http://sdnhost.{{ .Values.nsPrefix }}:8282/restconf/operations/SLI-API:healthcheck",
"method": "POST",
"header": {
"Authorization": ["Basic YWRtaW46S3A4Yko0U1hzek0wV1hsaGFrM2VIbGNzZTJnQXc4NHZhb0dHbUp2VXkyVQ=="],
{
"id": "sdnc-portal",
"name": "SDNC Portal Health Check",
- "http": "http://sdnc-portal.onap-sdnc:8843/login",
+ "http": "http://sdnc-portal.{{ .Values.nsPrefix }}:8843/login",
"method": "HEAD",
"header": {
"Cache-Control": ["no-cache"],
"name": "Health Check: SDNC-SDN-CTL-DB-01",
"checks": [
{
- "id": "sdnctldb01.onap-sdnc",
+ "id": "sdnctldb01.{{ .Values.nsPrefix }}",
"name": "SDNC SDNCTLDB01 Health Check",
- "tcp": "sdnctldb01.onap-sdnc:3306",
+ "tcp": "sdnctldb01.{{ .Values.nsPrefix }}:3306",
"interval": "10s",
"timeout": "1s"
}
"name": "Health Check: SDNC-SDN-CTL-DB-02",
"checks": [
{
- "id": "sdnctldb02.onap-sdnc",
+ "id": "sdnctldb02.{{ .Values.nsPrefix }}",
"name": "SDNC SDNCTLDB02 Health Check",
- "tcp": "sdnctldb02.onap-sdnc:3306",
+ "tcp": "sdnctldb02.{{ .Values.nsPrefix }}:3306",
"interval": "10s",
"timeout": "1s"
}
{
"id": "sdnc-sdnhost",
"name": "SDNC SDN Host Health Check",
- "http": "http://sdnhost.onap-sdnc:8282/apidoc/explorer/index.html",
+ "http": "http://sdnhost.{{ .Values.nsPrefix }}:8282/apidoc/explorer/index.html",
"method": "HEAD",
"header": {
"Authorization": ["Basic YWRtaW46S3A4Yko0U1hzek0wV1hsaGFrM2VIbGNzZTJnQXc4NHZhb0dHbUp2VXkyVQ=="],
"name": "Health Check: VFC",
"checks": [
{
- "id": "vfc-catalog.onap-vfc",
+ "id": "vfc-catalog.{{ .Values.nsPrefix }}",
"name": "VFC catalog Health Check",
- "tcp": "vfc-catalog.onap-vfc:8806",
+ "tcp": "vfc-catalog.{{ .Values.nsPrefix }}:8806",
"interval": "15s",
"timeout": "1s"
},
{
- "id": "vfc-emsdriver.onap-vfc",
+ "id": "vfc-emsdriver.{{ .Values.nsPrefix }}",
"name": "VFC emsdriver Health Check",
- "tcp": "vfc-emsdriver.onap-vfc:8206",
+ "tcp": "vfc-emsdriver.{{ .Values.nsPrefix }}:8206",
"interval": "15s",
"timeout": "1s"
},
{
- "id": "vfc-gvnfmdriver.onap-vfc",
+ "id": "vfc-gvnfmdriver.{{ .Values.nsPrefix }}",
"name": "VFC gvnfmdriver Health Check",
- "tcp": "vfc-gvnfmdriver.onap-vfc:8484",
+ "tcp": "vfc-gvnfmdriver.{{ .Values.nsPrefix }}:8484",
"interval": "15s",
"timeout": "1s"
},
{
- "id": "vfc-hwvnfmdriver.onap-vfc",
+ "id": "vfc-hwvnfmdriver.{{ .Values.nsPrefix }}",
"name": "VFC hwvnfmdriver Health Check",
- "tcp": "vfc-hwvnfmdriver.onap-vfc:8482",
+ "tcp": "vfc-hwvnfmdriver.{{ .Values.nsPrefix }}:8482",
"interval": "15s",
"timeout": "1s"
},
{
- "id": "vfc-jujudriver.onap-vfc",
+ "id": "vfc-jujudriver.{{ .Values.nsPrefix }}",
"name": "VFC jujudriver Health Check",
- "tcp": "vfc-jujudriver.onap-vfc:8483",
+ "tcp": "vfc-jujudriver.{{ .Values.nsPrefix }}:8483",
"interval": "15s",
"timeout": "1s"
},
{
- "id": "vfc-nokiavnfmdriver.onap-vfc",
+ "id": "vfc-nokiavnfmdriver.{{ .Values.nsPrefix }}",
"name": "VFC nokiavnfmdriver Health Check",
- "tcp": "vfc-nokiavnfmdriver.onap-vfc:8486",
+ "tcp": "vfc-nokiavnfmdriver.{{ .Values.nsPrefix }}:8486",
"interval": "15s",
"timeout": "1s"
},
{
- "id": "vfc-nslcm.onap-vfc",
+ "id": "vfc-nslcm.{{ .Values.nsPrefix }}",
"name": "VFC nslcm Health Check",
- "tcp": "vfc-nslcm.onap-vfc:8403",
+ "tcp": "vfc-nslcm.{{ .Values.nsPrefix }}:8403",
"interval": "15s",
"timeout": "1s"
},
{
- "id": "vfc-resmgr.onap-vfc",
+ "id": "vfc-resmgr.{{ .Values.nsPrefix }}",
"name": "VFC resmgr Health Check",
- "tcp": "vfc-resmgr.onap-vfc:8480",
+ "tcp": "vfc-resmgr.{{ .Values.nsPrefix }}:8480",
"interval": "15s",
"timeout": "1s"
},
{
- "id": "vfc-vnflcm.onap-vfc",
+ "id": "vfc-vnflcm.{{ .Values.nsPrefix }}",
"name": "VFC vnflcm Health Check",
- "tcp": "vfc-vnflcm.onap-vfc:8801",
+ "tcp": "vfc-vnflcm.{{ .Values.nsPrefix }}:8801",
"interval": "15s",
"timeout": "1s"
},
{
- "id": "vfc-vnfmgr.onap-vfc",
+ "id": "vfc-vnfmgr.{{ .Values.nsPrefix }}",
"name": "VFC vnfmgr Health Check",
- "tcp": "vfc-vnfmgr.onap-vfc:8803",
+ "tcp": "vfc-vnfmgr.{{ .Values.nsPrefix }}:8803",
"interval": "15s",
"timeout": "1s"
},
{
- "id": "vfc-vnfres.onap-vfc",
+ "id": "vfc-vnfres.{{ .Values.nsPrefix }}",
"name": "VFC vnfres Health Check",
- "tcp": "vfc-vnfres.onap-vfc:8802",
+ "tcp": "vfc-vnfres.{{ .Values.nsPrefix }}:8802",
"interval": "15s",
"timeout": "1s"
},
{
- "id": "vfc-workflow.onap-vfc",
+ "id": "vfc-workflow.{{ .Values.nsPrefix }}",
"name": "VFC workflow Health Check",
- "tcp": "vfc-workflow.onap-vfc:10550",
+ "tcp": "vfc-workflow.{{ .Values.nsPrefix }}:10550",
"interval": "15s",
"timeout": "1s"
},
{
- "id": "vfc-workflowengineactiviti.onap-vfc",
+ "id": "vfc-workflowengineactiviti.{{ .Values.nsPrefix }}",
"name": "VFC workflow-engine Health Check",
- "tcp": "vfc-workflowengineactiviti.onap-vfc:8080",
+ "tcp": "vfc-workflowengineactiviti.{{ .Values.nsPrefix }}:8080",
"interval": "15s",
"timeout": "1s"
},
{
- "id": "vfc-ztesdncdriver.onap-vfc",
+ "id": "vfc-ztesdncdriver.{{ .Values.nsPrefix }}",
"name": "VFC ztesdncdriver Health Check",
- "tcp": "vfc-ztesdncdriver.onap-vfc:8411",
+ "tcp": "vfc-ztesdncdriver.{{ .Values.nsPrefix }}:8411",
"interval": "15s",
"timeout": "1s"
},
{
- "id": "vfc-ztevnfmdriver.onap-vfc",
+ "id": "vfc-ztevnfmdriver.{{ .Values.nsPrefix }}",
"name": "VFC ztevnfmdriver Health Check",
- "tcp": "vfc-ztevnfmdriver.onap-vfc:8410",
+ "tcp": "vfc-ztevnfmdriver.{{ .Values.nsPrefix }}:8410",
"interval": "15s",
"timeout": "1s"
}
{
"id": "vid-server",
"name": "VID Server Health Check",
- "http": "http://vid-server.onap-vid:8080/vid/healthCheck",
+ "http": "http://vid-server.{{ .Values.nsPrefix }}:8080/vid/healthCheck",
"method": "GET",
"header": {
"Authorization": ["Basic YWRtaW46S3A4Yko0U1hzek0wV1hsaGFrM2VIbGNzZTJnQXc4NHZhb0dHbUp2VXkyVQ=="],
{
"id": "vid-mariadb",
"name": "Vid Mariadb Health Check",
- "script": "/consul/config/scripts/vid-mariadb-script.sh",
+ "script": "/consul/scripts/vid-mariadb-script.sh",
"interval": "10s",
"timeout": "1s"
}
--- /dev/null
+#{{ if not .Values.disableConsulConsulAgent }}
+apiVersion: v1
+kind: ConfigMap
+metadata:
+ name: consul-agent-configmap
+ namespace: {{ .Values.nsPrefix }}
+data:
+{{ tpl (.Files.Glob "resources/config/consul-agent-config/*").AsConfig . | indent 2 }}
+---
+apiVersion: v1
+kind: ConfigMap
+metadata:
+ name: consul-agent-scripts-configmap
+ namespace: {{ .Values.nsPrefix }}
+data:
+{{ tpl (.Files.Glob "resources/config/consul-agent-config/scripts/*").AsConfig . | indent 2 }}
+---
+apiVersion: v1
+kind: Secret
+metadata:
+ name: consul-agent-certs-secret
+ namespace: {{ .Values.nsPrefix }}
+data:
+{{ tpl (.Files.Glob "resources/config/consul-agent-config/certs/*").AsSecrets . | indent 2 }}
+#{{ end }}
labels:
app: consul-agent
name: consul-agent
- namespace: "{{ .Values.nsPrefix }}-consul"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
+ replicas: {{ .Values.consulAgentReplicas }}
selector:
matchLabels:
app: consul-agent
name: consul-agent
spec:
containers:
- - image: "{{ .Values.consulimageRegistry }}:{{ .Values.consuldockerTag }}"
+ - image: "{{ .Values.consulagentRegistry }}"
command: ["/usr/local/bin/docker-entrypoint.sh"]
- args: ["agent","-client","0.0.0.0","-enable-script-checks","-join","consul-server.{{ .Values.nsPrefix }}-consul"]
+ args: ["agent","-client","0.0.0.0","-enable-script-checks","-join","consul-server.{{ .Values.nsPrefix }}"]
name: consul-server
volumeMounts:
- mountPath: /consul/config
name: consul-agent-config
+ - mountPath: /consul/scripts
+ name: consul-agent-scripts-config
+ - mountPath: /consul/certs
+ name: consul-agent-certs-config
volumes:
- - hostPath:
- path: {{ .Values.rootHostPath }}/{{ .Values.nsPrefix }}/consul/consul-agent-config
+ - configMap:
+ name: consul-agent-configmap
name: consul-agent-config
-#{{ end }}
\ No newline at end of file
+ - configMap:
+ name: consul-agent-scripts-configmap
+ defaultMode: 0755
+ name: consul-agent-scripts-config
+ - secret:
+ secretName: consul-agent-certs-secret
+ name: consul-agent-certs-config
+#{{ end }}
labels:
app: consul-server
name: consul-server
- namespace: "{{ .Values.nsPrefix }}-consul"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
- replicas: 3
+ replicas: {{ .Values.consulServerReplicas }}
selector:
matchLabels:
app: consul-server
name: consul-server
spec:
containers:
- - image: "{{ .Values.consulimageRegistry }}:{{ .Values.consuldockerTag }}"
+ - image: "{{ .Values.consulserverRegistry }}"
command: ["/usr/local/bin/docker-entrypoint.sh"]
- args: ["agent","-server","-client","0.0.0.0","-enable-script-checks","-bootstrap-expect=3","-ui","-join","consul-server.{{ .Values.nsPrefix }}-consul"]
+ args: ["agent","-server","-client","0.0.0.0","-enable-script-checks","-bootstrap-expect=3","-ui","-join","consul-server.{{ .Values.nsPrefix }}"]
name: consul-server
- volumeMounts:
- - mountPath: /consul/config
- name: consul-server-config
- volumes:
- - hostPath:
- path: {{ .Values.rootHostPath }}/{{ .Values.nsPrefix }}/consul/consul-server-config
- name: consul-server-config
-#{{ end }}
\ No newline at end of file
+#{{ end }}
labels:
app: consul-server
name: consul-server
- namespace: "{{ .Values.nsPrefix }}-consul"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
ports:
- name: consul-ui
nsPrefix: "onap"
nodePortPrefix: 302
-consuldockerTag: "latest"
-rootHostPath: "/dockerdata-nfs"
-consulimageRegistry: "docker.io/consul"
+consulserverRegistry: "docker.io/consul:1.0.6"
+consulagentRegistry: "oomk8s/consul:1.0.0"
+consulAgentReplicas: 1
+consulServerReplicas: 3
service:
type: NodePort
kind: Service
metadata:
name: zookeeper
- namespace: "{{ .Values.nsPrefix }}-dcae"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: zookeeper
spec:
kind: Service
metadata:
name: kafka
- namespace: "{{ .Values.nsPrefix }}-dcae"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: kafka
spec:
kind: Service
metadata:
name: dmaap
- namespace: "{{ .Values.nsPrefix }}-dcae"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: dmaap
version: 1.1.0
kind: Service
metadata:
name: zldciad4vipstg00
- namespace: "{{ .Values.nsPrefix }}-dcae"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: dcae-pgaas
spec:
app: dcae-collector-common-event
version: latest
name: dcae-collector-common-event
- namespace: "{{ .Values.nsPrefix }}-dcae"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
ports:
- name: dcae-ce1
app: dcae-collector-dmaapbc
version: latest
name: dcae-collector-dmaapbc
- namespace: "{{ .Values.nsPrefix }}-dcae"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
ports:
- name: dcae-bc1
labels:
app: cdap0
name: cdap0
- namespace: "{{ .Values.nsPrefix }}-dcae"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
selector:
matchLabels:
labels:
app: cdap1
name: cdap1
- namespace: "{{ .Values.nsPrefix }}-dcae"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
selector:
matchLabels:
labels:
app: cdap2
name: cdap2
- namespace: "{{ .Values.nsPrefix }}-dcae"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
selector:
matchLabels:
kind: Deployment
metadata:
name: dcae-collector-common-event
- namespace: "{{ .Values.nsPrefix }}-dcae"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
selector:
matchLabels:
path: /dockerdata-nfs/{{ .Values.nsPrefix }}/dcae/common-event/config
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
kind: Deployment
metadata:
name: dcae-collector-dmaapbc
- namespace: "{{ .Values.nsPrefix }}-dcae"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
selector:
matchLabels:
path: /dockerdata-nfs/{{ .Values.nsPrefix }}/dcae/dmaapbc/config
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
app: dcae-controller
version: latest
name: dcae-controller
- namespace: "{{ .Values.nsPrefix }}-dcae"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
ports:
- name: dcae-con1
kind: Deployment
metadata:
name: dcae-controller
- namespace: "{{ .Values.nsPrefix }}-dcae"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
selector:
matchLabels:
path: /dockerdata-nfs/{{ .Values.nsPrefix }}/dcae/dcae-controller/config
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
kind: Deployment
metadata:
name: dmaap
- namespace: "{{ .Values.nsPrefix }}-dcae"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
selector:
matchLabels:
kind: Deployment
metadata:
name: kafka
- namespace: "{{ .Values.nsPrefix }}-dcae"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
selector:
matchLabels:
periodSeconds: 10
env:
- name: KAFKA_ZOOKEEPER_CONNECT
- value: "zookeeper.{{ .Values.nsPrefix }}-dcae:2181"
+ value: "zookeeper.{{ .Values.nsPrefix }}"
- name: KAFKA_ADVERTISED_HOST_NAME
value: "kafka"
- name: KAFKA_BROKER_ID
kind: Deployment
metadata:
name: zookeeper
- namespace: "{{ .Values.nsPrefix }}-dcae"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
selector:
matchLabels:
path: /dockerdata-nfs/{{ .Values.nsPrefix }}/dcae/message-router/dcae-startup-vm-message-router/docker_files/data-zookeeper
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
labels:
app: dcae-pgaas
name: dcae-pgaas
- namespace: "{{ .Values.nsPrefix }}-dcae"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: 1
selector:
path: /dockerdata-nfs/{{ .Values.nsPrefix }}/dcae/pgaas/tools
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
kind: Pod
metadata:
name: {{ .Chart.Name }}
- namespace: "{{ .Values.nsPrefix }}-dcaegen2"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
containers:
- name: {{ .Chart.Name }}
- image: "{{ .Values.image.repository }}:{{ .Values.image.tag }}"
+ image: "{{ .Values.image.repository }}"
imagePullPolicy: {{ .Values.image.pullPolicy }}
volumeMounts:
- name: installer-config-mount
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
restartPolicy: Never
-#{{ end }}
\ No newline at end of file
+#{{ end }}
nsPrefix: onap
location: dg2
image:
- repository: nexus3.onap.org:10001/onap/org.onap.dcaegen2.deployments.bootstrap
- tag: v1.1.0
- pullPolicy: Always
\ No newline at end of file
+ repository: nexus3.onap.org:10001/onap/org.onap.dcaegen2.deployments.bootstrap:v1.1.0
+ pullPolicy: Always
labels:
app: esr-esrserver
name: esr-esrserver
- namespace: "{{ .Values.nsPrefix }}-esr"
+ namespace: "{{ .Values.nsPrefix }}"
annotations:
msb.onap.org/service-info: '[
{
"url": "/api/aai-esr-server/v1",
"protocol": "REST",
"port": "{{.Values.esrserver.port}}",
+ "enable_ssl": true,
"visualRange":"1"
}
]'
labels:
app: esr-esrgui
name: esr-esrgui
- namespace: "{{ .Values.nsPrefix }}-esr"
+ namespace: "{{ .Values.nsPrefix }}"
annotations:
msb.onap.org/service-info: '[
{
labels:
app: esr-esrgui
name: esr-esrgui
- namespace: "{{ .Values.nsPrefix }}-esr"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
selector:
matchLabels:
name: esr-esrgui
env:
- name: MSB_ADDR
- value: {{ .Values.msbaddr }}
+ value: {{ tpl .Values.msbaddr . }}
ports:
- containerPort: {{ .Values.esrgui.port }}
readinessProbe:
periodSeconds: 10
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
labels:
app: esr-esrserver
name: esr-esrserver
- namespace: "{{ .Values.nsPrefix }}-esr"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
selector:
matchLabels:
name: esr-esrserver
env:
- name: MSB_ADDR
- value: {{ .Values.msbaddr }}
+ value: {{ tpl .Values.msbaddr . }}
ports:
- containerPort: {{ .Values.esrserver.port }}
readinessProbe:
periodSeconds: 10
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
nsPrefix: onap
pullPolicy: Always
nodePortPrefix: 302
-msbaddr: msb-iag.onap-msb:80
+msbaddr: msb-iag.{{ .Values.nsPrefix }}:80
esrserver:
image: nexus3.onap.org:10001/onap/aai/esr-server:v1.0.0
port: 9518
+++ /dev/null
-apiVersion: v1
-description: A Helm chart for Kubernetes
-name: kube2msb
-version: 0.1.0
+++ /dev/null
-nsPrefix: onap
-pullPolicy: IfNotPresent
-image:
- kube2msb: nexus3.onap.org:10001/onap/oom/kube2msb
-kubeMasterUrl: https://kubernetes.default.svc.cluster.local:443
-discoveryUrl: http://msb-discovery.onap-msb:10081
# By default, Elasticsearch will bind to the available loopback addresses and will scan ports 9300 to 9305 to try
# to connect to other nodes running on the same server.
#
-#discovery.zen.ping.unicast.hosts: ["elasticsearch.onap-log"
+#discovery.zen.ping.unicast.hosts: ["elasticsearch.{{.Values.nsPrefix}}"
#$discovery.zen.ping.unicast.hosts
#
# This setting tells Elasticsearch to not elect a master unless there are enough master-eligible nodes
# The URL of the Elasticsearch instance to use for all your queries.
# OOM-427, OOM-441 hardcoded onap workspace to avoid helm upgrade past 2.3
-elasticsearch.url: "http://elasticsearch.onap-log:9200"
-#elasticsearch-service.onap-log:9200"
+elasticsearch.url: "http://elasticsearch.{{.Values.nsPrefix}}:9200"
+#elasticsearch-service.{{.Values.nsPrefix}}:9200"
#elasticsearch.url: "http://10.247.47.3:9200"
# When this setting's value is true Kibana uses the hostname specified in the server.host
# setting. When the value of this setting is false, Kibana uses the hostname of the host
#can specify one or a list of hosts. If sniffing is set, one is enough and others will be auto-discovered
##Also protocol can be specified like ["http://10.247.186.12:9200"]
## OOM-427, OOM-441 hardcoded onap workspace to avoid helm upgrade past 2.3
- hosts => ["http://elasticsearch.onap-log:9200"]
+ hosts => ["http://elasticsearch.{{.Values.nsPrefix}}:9200"]
## This setting asks Elasticsearch for the list of all cluster nodes and adds them to the hosts list. Default is false.
labels:
app: elasticsearch
name: elasticsearch
- namespace: {{ .Values.nsPrefix }}-log
+ namespace: {{ .Values.nsPrefix }}
spec:
ports:
- name: http
kind: Service
metadata:
name: elasticsearchtcp
- namespace: {{ .Values.nsPrefix }}-log
+ namespace: {{ .Values.nsPrefix }}
labels:
app: elasticsearch
spec:
kind: Service
metadata:
name: logstashinternal
- namespace: {{ .Values.nsPrefix }}-log
+ namespace: {{ .Values.nsPrefix }}
labels:
app: logstash
spec:
kind: Service
metadata:
name: logstash
- namespace: {{ .Values.nsPrefix }}-log
+ namespace: {{ .Values.nsPrefix }}
labels:
app: logstash
spec:
labels:
app: kibana
name: kibana
- namespace: {{ .Values.nsPrefix }}-log
+ namespace: {{ .Values.nsPrefix }}
spec:
ports:
- name: tcp-ks
metadata:
labels:
app: elasticsearch
- name: elasticsearch
- namespace: {{ .Values.nsPrefix }}-log
+ name: log-elasticsearch
+ namespace: {{ .Values.nsPrefix }}
spec:
+ replicas: {{ .Values.elasticsearchReplicas }}
selector:
matchLabels:
app: elasticsearch
metadata:
labels:
app: elasticsearch
- name: elasticsearch
+ name: log-elasticsearch
spec:
initContainers:
- command:
path: {{ .Values.dataRootDir }}/{{ .Values.nsPrefix }}/log/
- name: elasticsearch-conf
configMap:
- name: elasticsearch-configmap
+ name: log-elasticsearch-configmap
items:
- key: elasticsearch.yml
path: elasticsearch.yml
apiVersion: v1
kind: ConfigMap
metadata:
- name: elasticsearch-configmap
- namespace: {{ .Values.nsPrefix }}-log
+ name: log-elasticsearch-configmap
+ namespace: {{ .Values.nsPrefix }}
data:
{{ (.Files.Glob "resources/elasticsearch/conf/elasticsearch.yml").AsConfig | indent 2 }}
#{{ end }}
metadata:
labels:
app: kibana
- name: kibana
- namespace: {{ .Values.nsPrefix }}-log
+ name: log-kibana
+ namespace: {{ .Values.nsPrefix }}
spec:
+ replicas: {{ .Values.kibanaReplicas }}
selector:
matchLabels:
app: kibana
metadata:
labels:
app: kibana
- name: kibana
+ name: log-kibana
spec:
initContainers:
- command:
volumes:
- name: kibana-conf
configMap:
- name: kibana-configmap
+ name: log-kibana-configmap
items:
- key: kibana.yml
path: kibana.yml
apiVersion: v1
kind: ConfigMap
metadata:
- name: kibana-configmap
- namespace: {{ .Values.nsPrefix }}-log
+ name: log-kibana-configmap
+ namespace: {{ .Values.nsPrefix }}
data:
-{{ (.Files.Glob "resources/kibana/conf/kibana.yml").AsConfig | indent 2 }}
+{{ tpl (.Files.Glob "resources/kibana/conf/kibana.yml").AsConfig . | indent 2 }}
#{{ end }}
apiVersion: v1
kind: PersistentVolume
metadata:
- name: elasticsearch-db
- namespace: "{{ .Values.nsPrefix }}-log"
+ name: {{ .Values.nsPrefix }}-elasticsearch-db
+ namespace: "{{ .Values.nsPrefix }}"
labels:
- name: elasticsearch-db
+ name: {{ .Values.nsPrefix }}-elasticsearch-db
spec:
capacity:
storage: 2Gi
apiVersion: v1
metadata:
name: elasticsearch-db
- namespace: "{{ .Values.nsPrefix }}-log"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
accessModes:
- ReadWriteMany
storage: 2Gi
selector:
matchLabels:
- name: elasticsearch-db
+ name: {{ .Values.nsPrefix }}-elasticsearch-db
#{{ end }}
metadata:
labels:
app: logstash
- name: logstash
- namespace: {{ .Values.nsPrefix }}-log
+ name: log-logstash
+ namespace: {{ .Values.nsPrefix }}
spec:
+ replicas: {{ .Values.logstashReplicas }}
selector:
matchLabels:
app: logstash
metadata:
labels:
app: logstash
- name: logstash
+ name: log-logstash
spec:
initContainers:
- command:
volumes:
- name: logstash-conf
configMap:
- name: logstash-configmap
+ name: log-logstash-configmap
items:
- key: logstash.yml
path: logstash.yml
- name: logstash-pipeline
configMap:
- name: logstash-configmap
+ name: log-logstash-configmap
items:
- key: onap-pipeline.conf
path: onap-pipeline.conf
apiVersion: v1
kind: ConfigMap
metadata:
- name: logstash-configmap
- namespace: {{ .Values.nsPrefix }}-log
+ name: log-logstash-configmap
+ namespace: {{ .Values.nsPrefix }}
data:
-{{ (.Files.Glob "resources/logstash/conf/logstash.yml").AsConfig | indent 2 }}
-{{ (.Files.Glob "resources/logstash/pipeline/onap-pipeline.conf").AsConfig | indent 2 }}
+{{ tpl (.Files.Glob "resources/logstash/conf/logstash.yml").AsConfig . | indent 2 }}
+{{ tpl (.Files.Glob "resources/logstash/pipeline/onap-pipeline.conf").AsConfig . | indent 2 }}
#{{ end }}
pullPolicy: Always
nodePortPrefix: 302
dataRootDir: /dockerdata-nfs
+elasticsearchReplicas: 1
+kibanaReplicas: 1
+logstashReplicas: 1
image:
readiness: oomk8s/readiness-check:1.1.0
logstash: docker.elastic.co/logstash/logstash:5.4.3
kind: Service
metadata:
name: zookeeper
- namespace: "{{ .Values.nsPrefix }}-message-router"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: zookeeper
spec:
kind: Service
metadata:
name: global-kafka
- namespace: "{{ .Values.nsPrefix }}-message-router"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: global-kafka
spec:
kind: Service
metadata:
name: dmaap
- namespace: "{{ .Values.nsPrefix }}-message-router"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: dmaap
version: 1.0.0
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
- name: dmaap
- namespace: "{{ .Values.nsPrefix }}-message-router"
+ name: message-router-dmaap
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.dmaapReplicas }}
selector:
metadata:
labels:
app: dmaap
- name: dmaap
+ name: message-router-dmaap
spec:
initContainers:
- command:
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
- name: global-kafka
- namespace: "{{ .Values.nsPrefix }}-message-router"
+ name: message-router-global-kafka
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.kafkaReplicas }}
selector:
metadata:
labels:
app: global-kafka
- name: global-kafka
+ name: message-router-global-kafka
spec:
initContainers:
- command:
periodSeconds: 10
env:
- name: KAFKA_ZOOKEEPER_CONNECT
- value: "zookeeper.{{ .Values.nsPrefix }}-message-router:2181"
+ value: "zookeeper.{{ .Values.nsPrefix }}:2181"
- name: KAFKA_ADVERTISED_HOST_NAME
value: "global-kafka"
- name: KAFKA_BROKER_ID
kind: PersistentVolume
metadata:
name: "{{ .Values.nsPrefix }}-message-router-kafka"
- namespace: "{{ .Values.nsPrefix }}-message-router"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
name: "{{ .Values.nsPrefix }}-message-router-kafka"
spec:
apiVersion: v1
metadata:
name: message-router-kafka
- namespace: "{{ .Values.nsPrefix }}-message-router"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
accessModes:
- ReadWriteMany
kind: PersistentVolume
metadata:
name: "{{ .Values.nsPrefix }}-message-router-zookeeper"
- namespace: "{{ .Values.nsPrefix }}-message-router"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
name: "{{ .Values.nsPrefix }}-message-router-zookeeper"
spec:
apiVersion: v1
metadata:
name: message-router-zookeeper
- namespace: "{{ .Values.nsPrefix }}-message-router"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
accessModes:
- ReadWriteMany
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
- name: zookeeper
- namespace: "{{ .Values.nsPrefix }}-message-router"
+ name: message-router-zookeeper
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.zookeeperReplicas }}
selector:
metadata:
labels:
app: zookeeper
- name: zookeeper
+ name: message-router-zookeeper
spec:
containers:
- image: {{ .Values.image.zookeeper }}
kind: Service
metadata:
name: sniro-emulator
- namespace: "{{ .Values.nsPrefix }}-mock"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: sniro-emulator
spec:
kind: Deployment
metadata:
name: sniro-emulator
- namespace: "{{ .Values.nsPrefix }}-mock"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
selector:
matchLabels:
args: ["--verbose"]
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
kind: Service
metadata:
name: msb-consul
- namespace: "{{ .Values.nsPrefix }}-msb"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: msb-consul
spec:
kind: Service
metadata:
name: msb-discovery
- namespace: "{{ .Values.nsPrefix }}-msb"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: msb-discovery
spec:
kind: Service
metadata:
name: msb-iag
- namespace: "{{ .Values.nsPrefix }}-msb"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: msb-iag
spec:
kind: Service
metadata:
name: msb-eag
- namespace: "{{ .Values.nsPrefix }}-msb"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: msb-eag
spec:
kind: Deployment
metadata:
name: kube2msb-registrator
- namespace: "{{ .Values.nsPrefix }}-kube2msb"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: 1
selector:
name: kube2msb-registrator
spec:
hostname: kube2msb-registrator
+ initContainers:
+ - command:
+ - /root/ready.py
+ args:
+ - --container-name
+ - msb-discovery
+ env:
+ - name: NAMESPACE
+ valueFrom:
+ fieldRef:
+ apiVersion: v1
+ fieldPath: metadata.namespace
+ image: {{ .Values.image.readiness }}
+ imagePullPolicy: {{ .Values.pullPolicy }}
+ name: kube2msb-registrator-readiness
containers:
- args:
image: {{ .Values.image.kube2msb }}
- name: AUTH_TOKEN
value: {{ .Values.kubeMasterAuthToken }}
- name: MSB_URL
- value: {{ .Values.discoveryUrl }}
+ value: {{tpl $.Values.discoveryUrl .}}
imagePullPolicy: {{ .Values.pullPolicy }}
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
kind: Deployment
metadata:
name: msb-consul
- namespace: "{{ .Values.nsPrefix }}-msb"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
- replicas: 1
+ replicas: {{ .Values.msbConsulReplicas }}
selector:
matchLabels:
app: msb-consul
initialDelaySeconds: 5
periodSeconds: 10
imagePullPolicy: {{ .Values.pullPolicy }}
-#{{ end }}
\ No newline at end of file
+#{{ end }}
kind: Deployment
metadata:
name: msb-discovery
- namespace: "{{ .Values.nsPrefix }}-msb"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.discoveryReplicas }}
selector:
name: "msb-discovery"
env:
- name: CONSUL_IP
- value: msb-consul.{{ .Values.nsPrefix }}-msb
+ value: msb-consul.{{ .Values.nsPrefix }}
ports:
- containerPort: {{ .Values.discoveryPort }}
name: msb-discovery
kind: Deployment
metadata:
name: msb-eag
- namespace: "{{ .Values.nsPrefix }}-msb"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.eagReplicas }}
selector:
name: "msb-eag"
env:
- name: CONSUL_IP
- value: msb-consul.{{ .Values.nsPrefix }}-msb
+ value: msb-consul.{{ .Values.nsPrefix }}
- name: SDCLIENT_IP
- value: msb-discovery.{{ .Values.nsPrefix }}-msb
+ value: msb-discovery.{{ .Values.nsPrefix }}
- name: ROUTE_LABELS
value: "visualRange:0"
ports:
kind: Deployment
metadata:
name: msb-iag
- namespace: "{{ .Values.nsPrefix }}-msb"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.iagReplicas }}
selector:
name: "msb-iag"
env:
- name: CONSUL_IP
- value: msb-consul.{{ .Values.nsPrefix }}-msb
+ value: msb-consul.{{ .Values.nsPrefix }}
- name: SDCLIENT_IP
- value: msb-discovery.{{ .Values.nsPrefix }}-msb
+ value: msb-discovery.{{ .Values.nsPrefix }}
- name: ROUTE_LABELS
value: "visualRange:1"
ports:
consul: consul:0.9.3
discovery: nexus3.onap.org:10001/onap/msb/msb_discovery:1.1.0-SNAPSHOT-latest
apigateway: nexus3.onap.org:10001/onap/msb/msb_apigateway:1.1.0-SNAPSHOT-latest
+ kube2msb: nexus3.onap.org:10001/onap/oom/kube2msb
consulPort: 8500
consulNodePort: 30500
+msbConsulReplicas: 1
discoveryPort: 10081
discoveryNodePort: 30081
eagNodePort: 30082
eagNodePort_https: 30446
eagReplicas: 1
+
+kubeMasterUrl: https://kubernetes.default.svc.cluster.local:443
+discoveryUrl: http://msb-discovery.{{ .Values.nsPrefix }}:10081
#List of logstash server ip addresses with port number.
#But, in our case, this will be the loadbalancer IP address.
#For the below property to work the loadbalancer or logstash should expose 5044 port to listen the filebeat events or port in the property should be changed appropriately.
- hosts: ["logstash.{{ .Values.nsPrefix }}-log:5044"]
+ hosts: ["logstash.{{.Values.nsPrefix}}:5044"]
#If enable will do load balancing among availabe Logstash, automatically.
loadbalance: true
--- /dev/null
+<?xml version='1.0' encoding='UTF-8'?>
+
+<server xmlns="urn:jboss:domain:4.2">
+
+ <extensions>
+ <extension module="org.jboss.as.clustering.infinispan"/>
+ <extension module="org.jboss.as.clustering.jgroups"/>
+ <extension module="org.jboss.as.connector"/>
+ <extension module="org.jboss.as.deployment-scanner"/>
+ <extension module="org.jboss.as.ee"/>
+ <extension module="org.jboss.as.ejb3"/>
+ <extension module="org.jboss.as.jaxrs"/>
+ <extension module="org.jboss.as.jdr"/>
+ <extension module="org.jboss.as.jmx"/>
+ <extension module="org.jboss.as.jpa"/>
+ <extension module="org.jboss.as.jsf"/>
+ <extension module="org.jboss.as.jsr77"/>
+ <extension module="org.jboss.as.logging"/>
+ <extension module="org.jboss.as.mail"/>
+ <extension module="org.jboss.as.modcluster"/>
+ <extension module="org.jboss.as.naming"/>
+ <extension module="org.jboss.as.pojo"/>
+ <extension module="org.jboss.as.remoting"/>
+ <extension module="org.jboss.as.sar"/>
+ <extension module="org.jboss.as.security"/>
+ <extension module="org.jboss.as.transactions"/>
+ <extension module="org.jboss.as.webservices"/>
+ <extension module="org.jboss.as.weld"/>
+ <extension module="org.wildfly.extension.batch.jberet"/>
+ <extension module="org.wildfly.extension.bean-validation"/>
+ <extension module="org.wildfly.extension.clustering.singleton"/>
+ <extension module="org.wildfly.extension.io"/>
+ <extension module="org.wildfly.extension.messaging-activemq"/>
+ <extension module="org.wildfly.extension.request-controller"/>
+ <extension module="org.wildfly.extension.security.manager"/>
+ <extension module="org.wildfly.extension.undertow"/>
+ <extension module="org.wildfly.iiop-openjdk"/>
+ </extensions>
+
+
+ <management>
+ <security-realms>
+ <security-realm name="ManagementRealm">
+ <authentication>
+ <local default-user="$local" skip-group-loading="true"/>
+ <properties path="mgmt-users.properties" relative-to="jboss.server.config.dir"/>
+ </authentication>
+ <authorization map-groups-to-roles="false">
+ <properties path="mgmt-groups.properties" relative-to="jboss.server.config.dir"/>
+ </authorization>
+ </security-realm>
+ <security-realm name="ApplicationRealm">
+ <server-identities>
+ <ssl>
+ <keystore path="application.keystore" relative-to="jboss.server.config.dir" keystore-password="password" alias="server" key-password="password" generate-self-signed-certificate-host="localhost"/>
+ </ssl>
+ </server-identities>
+ <authentication>
+ <local default-user="$local" allowed-users="*" skip-group-loading="true"/>
+ <properties path="application-users.properties" relative-to="jboss.server.config.dir"/>
+ </authentication>
+ <authorization>
+ <properties path="application-roles.properties" relative-to="jboss.server.config.dir"/>
+ </authorization>
+ </security-realm>
+ </security-realms>
+ <audit-log>
+ <formatters>
+ <json-formatter name="json-formatter"/>
+ </formatters>
+ <handlers>
+ <file-handler name="file" formatter="json-formatter" path="audit-log.log" relative-to="jboss.server.data.dir"/>
+ </handlers>
+ <logger log-boot="true" log-read-only="false" enabled="false">
+ <handlers>
+ <handler name="file"/>
+ </handlers>
+ </logger>
+ </audit-log>
+ <management-interfaces>
+ <http-interface security-realm="ManagementRealm" http-upgrade-enabled="true">
+ <socket-binding http="management-http"/>
+ </http-interface>
+ </management-interfaces>
+ <access-control provider="simple">
+ <role-mapping>
+ <role name="SuperUser">
+ <include>
+ <user name="$local"/>
+ </include>
+ </role>
+ </role-mapping>
+ </access-control>
+ </management>
+
+ <profile>
+ <subsystem xmlns="urn:jboss:domain:logging:3.0">
+ <console-handler name="CONSOLE">
+ <level name="INFO"/>
+ <formatter>
+ <named-formatter name="COLOR-PATTERN"/>
+ </formatter>
+ </console-handler>
+ <periodic-rotating-file-handler name="FILE" autoflush="true">
+ <formatter>
+ <named-formatter name="PATTERN"/>
+ </formatter>
+ <file relative-to="jboss.server.log.dir" path="server.log"/>
+ <suffix value=".yyyy-MM-dd"/>
+ <append value="true"/>
+ </periodic-rotating-file-handler>
+ <logger category="com.arjuna">
+ <level name="WARN"/>
+ </logger>
+ <logger category="org.jboss.as.config">
+ <level name="DEBUG"/>
+ </logger>
+ <logger category="sun.rmi">
+ <level name="WARN"/>
+ </logger>
+ <root-logger>
+ <level name="INFO"/>
+ <handlers>
+ <handler name="CONSOLE"/>
+ <handler name="FILE"/>
+ </handlers>
+ </root-logger>
+ <formatter name="PATTERN">
+ <pattern-formatter pattern="%d{yyyy-MM-dd HH:mm:ss,SSS} %-5p [%c] (%t) %s%e%n"/>
+ </formatter>
+ <formatter name="COLOR-PATTERN">
+ <pattern-formatter pattern="%K{level}%d{HH:mm:ss,SSS} %-5p [%c] (%t) %s%e%n"/>
+ </formatter>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:batch-jberet:1.0">
+ <default-job-repository name="in-memory"/>
+ <default-thread-pool name="batch"/>
+ <job-repository name="in-memory">
+ <in-memory/>
+ </job-repository>
+ <thread-pool name="batch">
+ <max-threads count="10"/>
+ <keepalive-time time="30" unit="seconds"/>
+ </thread-pool>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:bean-validation:1.0"/>
+ <subsystem xmlns="urn:jboss:domain:datasources:4.0">
+ <datasources>
+ <datasource jndi-name="java:jboss/datasources/mso-requests" pool-name="mso-requests" enabled="true" use-ccm="true">
+ <connection-url>jdbc:mariadb://mso-mariadb.{{.Values.nsPrefix}}:3306/mso_requests?autoReconnect=true&connectTimeout=60000&socketTimeout=60000</connection-url>
+ <driver>mariadb</driver>
+ <transaction-isolation>TRANSACTION_READ_COMMITTED</transaction-isolation>
+ <pool>
+ <min-pool-size>10</min-pool-size>
+ <max-pool-size>100</max-pool-size>
+ <prefill>true</prefill>
+ <use-strict-min>false</use-strict-min>
+ <flush-strategy>FailingConnectionOnly</flush-strategy>
+ </pool>
+ <security>
+ <user-name>mso</user-name>
+ <password>mso123</password>
+ </security>
+ <validation>
+ <valid-connection-checker class-name="org.jboss.jca.adapters.jdbc.extensions.mysql.MySQLValidConnectionChecker"/>
+ <validate-on-match>false</validate-on-match>
+ <background-validation>true</background-validation>
+ <background-validation-millis>20000</background-validation-millis>
+ <use-fast-fail>true</use-fast-fail>
+ <exception-sorter class-name="org.jboss.jca.adapters.jdbc.extensions.mysql.MySQLExceptionSorter"/>
+ </validation>
+ <timeout>
+ <idle-timeout-minutes>15</idle-timeout-minutes>
+ <query-timeout>30</query-timeout>
+ <allocation-retry>1</allocation-retry>
+ <allocation-retry-wait-millis>3000</allocation-retry-wait-millis>
+ </timeout>
+ </datasource>
+ <datasource jndi-name="java:jboss/datasources/mso-catalog" pool-name="mso-catalog" enabled="true" use-ccm="true">
+ <connection-url>jdbc:mariadb://mso-mariadb.{{.Values.nsPrefix}}:3306/mso_catalog?autoReconnect=true&connectTimeout=60000&socketTimeout=60000</connection-url>
+ <driver>mariadb</driver>
+ <transaction-isolation>TRANSACTION_READ_COMMITTED</transaction-isolation>
+ <pool>
+ <min-pool-size>10</min-pool-size>
+ <max-pool-size>100</max-pool-size>
+ <prefill>true</prefill>
+ <use-strict-min>false</use-strict-min>
+ <flush-strategy>FailingConnectionOnly</flush-strategy>
+ </pool>
+ <security>
+ <user-name>catalog</user-name>
+ <password>catalog123</password>
+ </security>
+ <validation>
+ <valid-connection-checker class-name="org.jboss.jca.adapters.jdbc.extensions.mysql.MySQLValidConnectionChecker"/>
+ <validate-on-match>false</validate-on-match>
+ <background-validation>true</background-validation>
+ <background-validation-millis>20000</background-validation-millis>
+ <use-fast-fail>true</use-fast-fail>
+ <exception-sorter class-name="org.jboss.jca.adapters.jdbc.extensions.mysql.MySQLExceptionSorter"/>
+ </validation>
+ <timeout>
+ <idle-timeout-minutes>15</idle-timeout-minutes>
+ <query-timeout>30</query-timeout>
+ <allocation-retry>1</allocation-retry>
+ <allocation-retry-wait-millis>3000</allocation-retry-wait-millis>
+ </timeout>
+ </datasource>
+ <datasource jta="true" jndi-name="java:jboss/datasources/ProcessEngine" pool-name="ProcessEngine" enabled="true" use-java-context="true" use-ccm="true">
+ <connection-url>jdbc:mariadb://mso-mariadb.{{.Values.nsPrefix}}:3306/camundabpmn?autoReconnect=true&connectTimeout=60000&socketTimeout=60000</connection-url>
+ <driver>mariadb</driver>
+ <new-connection-sql>set autocommit=1</new-connection-sql>
+ <transaction-isolation>TRANSACTION_READ_COMMITTED</transaction-isolation>
+ <pool>
+ <min-pool-size>10</min-pool-size>
+ <max-pool-size>100</max-pool-size>
+ <prefill>true</prefill>
+ <use-strict-min>false</use-strict-min>
+ <flush-strategy>FailingConnectionOnly</flush-strategy>
+ </pool>
+ <security>
+ <user-name>camunda</user-name>
+ <password>camunda123</password>
+ </security>
+ <validation>
+ <valid-connection-checker class-name="org.jboss.jca.adapters.jdbc.extensions.mysql.MySQLValidConnectionChecker"/>
+ <exception-sorter class-name="org.jboss.jca.adapters.jdbc.extensions.mysql.MySQLExceptionSorter"/>
+ </validation>
+ <statement>
+ <prepared-statement-cache-size>32</prepared-statement-cache-size>
+ <share-prepared-statements>true</share-prepared-statements>
+ </statement>
+ </datasource>
+ <datasource jndi-name="java:jboss/datasources/ExampleDS" pool-name="ExampleDS" enabled="true" use-java-context="true">
+ <connection-url>jdbc:h2:mem:test;DB_CLOSE_DELAY=-1;DB_CLOSE_ON_EXIT=FALSE</connection-url>
+ <driver>h2</driver>
+ <security>
+ <user-name>sa</user-name>
+ <password>sa</password>
+ </security>
+ </datasource>
+ <drivers>
+ <driver name="h2" module="com.h2database.h2">
+ <xa-datasource-class>org.h2.jdbcx.JdbcDataSource</xa-datasource-class>
+ </driver>
+ <driver name="mariadb" module="mariadb">
+ <xa-datasource-class>org.mariadb.jdbc.MySQLDataSource</xa-datasource-class>
+ </driver>
+ </drivers>
+ </datasources>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:deployment-scanner:2.0">
+ <deployment-scanner path="deployments" relative-to="jboss.server.base.dir" scan-interval="5000" runtime-failure-causes-rollback="${jboss.deployment.scanner.rollback.on.failure:false}"/>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:ee:4.0">
+ <spec-descriptor-property-replacement>false</spec-descriptor-property-replacement>
+ <concurrent>
+ <context-services>
+ <context-service name="default" jndi-name="java:jboss/ee/concurrency/context/default" use-transaction-setup-provider="true"/>
+ </context-services>
+ <managed-thread-factories>
+ <managed-thread-factory name="default" jndi-name="java:jboss/ee/concurrency/factory/default" context-service="default"/>
+ </managed-thread-factories>
+ <managed-executor-services>
+ <managed-executor-service name="default" jndi-name="java:jboss/ee/concurrency/executor/default" context-service="default" hung-task-threshold="60000" keepalive-time="5000"/>
+ </managed-executor-services>
+ <managed-scheduled-executor-services>
+ <managed-scheduled-executor-service name="default" jndi-name="java:jboss/ee/concurrency/scheduler/default" context-service="default" hung-task-threshold="60000" keepalive-time="3000"/>
+ </managed-scheduled-executor-services>
+ </concurrent>
+ <default-bindings context-service="java:jboss/ee/concurrency/context/default" datasource="java:jboss/datasources/ExampleDS" jms-connection-factory="java:jboss/DefaultJMSConnectionFactory" managed-executor-service="java:jboss/ee/concurrency/executor/default" managed-scheduled-executor-service="java:jboss/ee/concurrency/scheduler/default" managed-thread-factory="java:jboss/ee/concurrency/factory/default"/>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:ejb3:4.0">
+ <session-bean>
+ <stateless>
+ <bean-instance-pool-ref pool-name="slsb-strict-max-pool"/>
+ </stateless>
+ <stateful default-access-timeout="5000" cache-ref="distributable" passivation-disabled-cache-ref="simple"/>
+ <singleton default-access-timeout="5000"/>
+ </session-bean>
+ <mdb>
+ <resource-adapter-ref resource-adapter-name="${ejb.resource-adapter-name:activemq-ra.rar}"/>
+ <bean-instance-pool-ref pool-name="mdb-strict-max-pool"/>
+ </mdb>
+ <pools>
+ <bean-instance-pools>
+ <strict-max-pool name="slsb-strict-max-pool" derive-size="from-worker-pools" instance-acquisition-timeout="5" instance-acquisition-timeout-unit="MINUTES"/>
+ <strict-max-pool name="mdb-strict-max-pool" derive-size="from-cpu-count" instance-acquisition-timeout="5" instance-acquisition-timeout-unit="MINUTES"/>
+ </bean-instance-pools>
+ </pools>
+ <caches>
+ <cache name="simple"/>
+ <cache name="distributable" passivation-store-ref="infinispan" aliases="passivating clustered"/>
+ </caches>
+ <passivation-stores>
+ <passivation-store name="infinispan" cache-container="ejb" max-size="10000"/>
+ </passivation-stores>
+ <async thread-pool-name="default"/>
+ <timer-service thread-pool-name="default" default-data-store="default-file-store">
+ <data-stores>
+ <file-data-store name="default-file-store" path="timer-service-data" relative-to="jboss.server.data.dir"/>
+ </data-stores>
+ </timer-service>
+ <remote connector-ref="http-remoting-connector" thread-pool-name="default"/>
+ <thread-pools>
+ <thread-pool name="default">
+ <max-threads count="10"/>
+ <keepalive-time time="100" unit="milliseconds"/>
+ </thread-pool>
+ </thread-pools>
+ <iiop enable-by-default="false" use-qualified-name="false"/>
+ <default-security-domain value="other"/>
+ <default-missing-method-permissions-deny-access value="true"/>
+ <log-system-exceptions value="true"/>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:io:1.1">
+ <worker name="default"/>
+ <buffer-pool name="default"/>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:infinispan:4.0">
+ <cache-container name="server" aliases="singleton cluster" default-cache="default" module="org.wildfly.clustering.server">
+ <transport lock-timeout="60000"/>
+ <replicated-cache name="default" mode="SYNC">
+ <transaction mode="BATCH"/>
+ </replicated-cache>
+ </cache-container>
+ <cache-container name="web" default-cache="dist" module="org.wildfly.clustering.web.infinispan">
+ <transport lock-timeout="60000"/>
+ <distributed-cache name="dist" mode="ASYNC" l1-lifespan="0" owners="2">
+ <locking isolation="REPEATABLE_READ"/>
+ <transaction mode="BATCH"/>
+ <file-store/>
+ </distributed-cache>
+ <distributed-cache name="concurrent" mode="SYNC" l1-lifespan="0" owners="2">
+ <file-store/>
+ </distributed-cache>
+ </cache-container>
+ <cache-container name="ejb" aliases="sfsb" default-cache="dist" module="org.wildfly.clustering.ejb.infinispan">
+ <transport lock-timeout="60000"/>
+ <distributed-cache name="dist" mode="ASYNC" l1-lifespan="0" owners="2">
+ <locking isolation="REPEATABLE_READ"/>
+ <transaction mode="BATCH"/>
+ <file-store/>
+ </distributed-cache>
+ </cache-container>
+ <cache-container name="hibernate" default-cache="local-query" module="org.hibernate.infinispan">
+ <transport lock-timeout="60000"/>
+ <local-cache name="local-query">
+ <eviction strategy="LRU" max-entries="10000"/>
+ <expiration max-idle="100000"/>
+ </local-cache>
+ <invalidation-cache name="entity" mode="SYNC">
+ <transaction mode="NON_XA"/>
+ <eviction strategy="LRU" max-entries="10000"/>
+ <expiration max-idle="100000"/>
+ </invalidation-cache>
+ <replicated-cache name="timestamps" mode="ASYNC"/>
+ </cache-container>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:iiop-openjdk:1.0">
+ <orb socket-binding="iiop" ssl-socket-binding="iiop-ssl"/>
+ <initializers security="identity" transactions="spec"/>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:jaxrs:1.0"/>
+ <subsystem xmlns="urn:jboss:domain:jca:4.0">
+ <archive-validation enabled="true" fail-on-error="true" fail-on-warn="false"/>
+ <bean-validation enabled="true"/>
+ <default-workmanager>
+ <short-running-threads>
+ <core-threads count="50"/>
+ <queue-length count="50"/>
+ <max-threads count="50"/>
+ <keepalive-time time="10" unit="seconds"/>
+ </short-running-threads>
+ <long-running-threads>
+ <core-threads count="50"/>
+ <queue-length count="50"/>
+ <max-threads count="50"/>
+ <keepalive-time time="10" unit="seconds"/>
+ </long-running-threads>
+ </default-workmanager>
+ <cached-connection-manager/>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:jdr:1.0"/>
+ <subsystem xmlns="urn:jboss:domain:jgroups:4.0">
+ <channels default="ee">
+ <channel name="ee" stack="udp"/>
+ </channels>
+ <stacks>
+ <stack name="udp">
+ <transport type="UDP" socket-binding="jgroups-udp"/>
+ <protocol type="PING"/>
+ <protocol type="MERGE3"/>
+ <protocol type="FD_SOCK" socket-binding="jgroups-udp-fd"/>
+ <protocol type="FD_ALL"/>
+ <protocol type="VERIFY_SUSPECT"/>
+ <protocol type="pbcast.NAKACK2"/>
+ <protocol type="UNICAST3"/>
+ <protocol type="pbcast.STABLE"/>
+ <protocol type="pbcast.GMS"/>
+ <protocol type="UFC"/>
+ <protocol type="MFC"/>
+ <protocol type="FRAG2"/>
+ </stack>
+ <stack name="tcp">
+ <transport type="TCP" socket-binding="jgroups-tcp"/>
+ <protocol type="MPING" socket-binding="jgroups-mping"/>
+ <protocol type="MERGE3"/>
+ <protocol type="FD_SOCK" socket-binding="jgroups-tcp-fd"/>
+ <protocol type="FD"/>
+ <protocol type="VERIFY_SUSPECT"/>
+ <protocol type="pbcast.NAKACK2"/>
+ <protocol type="UNICAST3"/>
+ <protocol type="pbcast.STABLE"/>
+ <protocol type="pbcast.GMS"/>
+ <protocol type="MFC"/>
+ <protocol type="FRAG2"/>
+ </stack>
+ </stacks>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:jmx:1.3">
+ <expose-resolved-model/>
+ <expose-expression-model/>
+ <remoting-connector/>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:jpa:1.1">
+ <jpa default-datasource="" default-extended-persistence-inheritance="DEEP"/>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:jsf:1.0"/>
+ <subsystem xmlns="urn:jboss:domain:jsr77:1.0"/>
+ <subsystem xmlns="urn:jboss:domain:mail:2.0">
+ <mail-session name="default" jndi-name="java:jboss/mail/Default">
+ <smtp-server outbound-socket-binding-ref="mail-smtp"/>
+ </mail-session>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:messaging-activemq:1.0">
+ <server name="default">
+ <cluster password="${jboss.messaging.cluster.password:CHANGE ME!!}"/>
+ <security-setting name="#">
+ <role name="guest" send="true" consume="true" create-non-durable-queue="true" delete-non-durable-queue="true"/>
+ </security-setting>
+ <address-setting name="#" dead-letter-address="jms.queue.DLQ" expiry-address="jms.queue.ExpiryQueue" max-size-bytes="10485760" page-size-bytes="2097152" message-counter-history-day-limit="10" redistribution-delay="1000"/>
+ <http-connector name="http-connector" socket-binding="http" endpoint="http-acceptor"/>
+ <http-connector name="http-connector-throughput" socket-binding="http" endpoint="http-acceptor-throughput">
+ <param name="batch-delay" value="50"/>
+ </http-connector>
+ <in-vm-connector name="in-vm" server-id="0"/>
+ <http-acceptor name="http-acceptor" http-listener="default"/>
+ <http-acceptor name="http-acceptor-throughput" http-listener="default">
+ <param name="batch-delay" value="50"/>
+ <param name="direct-deliver" value="false"/>
+ </http-acceptor>
+ <in-vm-acceptor name="in-vm" server-id="0"/>
+ <broadcast-group name="bg-group1" jgroups-channel="activemq-cluster" connectors="http-connector"/>
+ <discovery-group name="dg-group1" jgroups-channel="activemq-cluster"/>
+ <cluster-connection name="my-cluster" address="jms" connector-name="http-connector" discovery-group="dg-group1"/>
+ <jms-queue name="ExpiryQueue" entries="java:/jms/queue/ExpiryQueue"/>
+ <jms-queue name="DLQ" entries="java:/jms/queue/DLQ"/>
+ <connection-factory name="InVmConnectionFactory" entries="java:/ConnectionFactory" connectors="in-vm"/>
+ <connection-factory name="RemoteConnectionFactory" entries="java:jboss/exported/jms/RemoteConnectionFactory" connectors="http-connector" ha="true" block-on-acknowledge="true" reconnect-attempts="-1"/>
+ <pooled-connection-factory name="activemq-ra" entries="java:/JmsXA java:jboss/DefaultJMSConnectionFactory" connectors="in-vm" transaction="xa"/>
+ </server>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:modcluster:2.0">
+ <mod-cluster-config advertise-socket="modcluster" connector="ajp">
+ <dynamic-load-provider>
+ <load-metric type="cpu"/>
+ </dynamic-load-provider>
+ </mod-cluster-config>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:naming:2.0">
+ <remote-naming/>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:pojo:1.0"/>
+ <subsystem xmlns="urn:jboss:domain:remoting:3.0">
+ <endpoint/>
+ <http-connector name="http-remoting-connector" connector-ref="default" security-realm="ApplicationRealm"/>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:resource-adapters:4.0"/>
+ <subsystem xmlns="urn:jboss:domain:request-controller:1.0"/>
+ <subsystem xmlns="urn:jboss:domain:sar:1.0"/>
+ <subsystem xmlns="urn:jboss:domain:security-manager:1.0">
+ <deployment-permissions>
+ <maximum-set>
+ <permission class="java.security.AllPermission"/>
+ </maximum-set>
+ </deployment-permissions>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:security:1.2">
+ <security-domains>
+ <security-domain name="other" cache-type="default">
+ <authentication>
+ <login-module code="Remoting" flag="optional">
+ <module-option name="password-stacking" value="useFirstPass"/>
+ </login-module>
+ <login-module code="RealmDirect" flag="required">
+ <module-option name="password-stacking" value="useFirstPass"/>
+ </login-module>
+ </authentication>
+ </security-domain>
+ <security-domain name="jboss-web-policy" cache-type="default">
+ <authorization>
+ <policy-module code="Delegating" flag="required"/>
+ </authorization>
+ </security-domain>
+ <security-domain name="jboss-ejb-policy" cache-type="default">
+ <authorization>
+ <policy-module code="Delegating" flag="required"/>
+ </authorization>
+ </security-domain>
+ <security-domain name="jaspitest" cache-type="default">
+ <authentication-jaspi>
+ <login-module-stack name="dummy">
+ <login-module code="Dummy" flag="optional"/>
+ </login-module-stack>
+ <auth-module code="Dummy"/>
+ </authentication-jaspi>
+ </security-domain>
+ </security-domains>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:singleton:1.0">
+ <singleton-policies default="default">
+ <singleton-policy name="default" cache-container="server">
+ <simple-election-policy/>
+ </singleton-policy>
+ </singleton-policies>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:transactions:3.0">
+ <core-environment>
+ <process-id>
+ <uuid/>
+ </process-id>
+ </core-environment>
+ <recovery-environment socket-binding="txn-recovery-environment" status-socket-binding="txn-status-manager"/>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:undertow:3.1">
+ <buffer-cache name="default"/>
+ <server name="default-server">
+ <ajp-listener name="ajp" socket-binding="ajp"/>
+ <http-listener name="default" socket-binding="http" enable-http2="true"/>
+ <https-listener name="https" socket-binding="https" security-realm="ApplicationRealm" enable-http2="true"/>
+ <host name="default-host" alias="localhost">
+ <location name="/" handler="welcome-content"/>
+ <filter-ref name="server-header"/>
+ <filter-ref name="x-powered-by-header"/>
+ </host>
+ </server>
+ <servlet-container name="default">
+ <jsp-config/>
+ <websockets/>
+ </servlet-container>
+ <handlers>
+ <file name="welcome-content" path="${jboss.home.dir}/welcome-content"/>
+ </handlers>
+ <filters>
+ <response-header name="server-header" header-name="Server" header-value="WildFly/10"/>
+ <response-header name="x-powered-by-header" header-name="X-Powered-By" header-value="Undertow/1"/>
+ </filters>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:webservices:2.0">
+ <wsdl-host>${jboss.bind.address:127.0.0.1}</wsdl-host>
+ <endpoint-config name="Standard-Endpoint-Config"/>
+ <endpoint-config name="Recording-Endpoint-Config">
+ <pre-handler-chain name="recording-handlers" protocol-bindings="##SOAP11_HTTP ##SOAP11_HTTP_MTOM ##SOAP12_HTTP ##SOAP12_HTTP_MTOM">
+ <handler name="RecordingHandler" class="org.jboss.ws.common.invocation.RecordingServerHandler"/>
+ </pre-handler-chain>
+ </endpoint-config>
+ <client-config name="Standard-Client-Config"/>
+ </subsystem>
+ <subsystem xmlns="urn:jboss:domain:weld:3.0"/>
+ </profile>
+
+ <interfaces>
+ <interface name="management">
+ <inet-address value="${jboss.bind.address.management:127.0.0.1}"/>
+ </interface>
+ <interface name="public">
+ <inet-address value="${jboss.bind.address:127.0.0.1}"/>
+ </interface>
+ <interface name="private">
+ <inet-address value="${jboss.bind.address.private:127.0.0.1}"/>
+ </interface>
+ <interface name="unsecure">
+ <inet-address value="${jboss.bind.address.unsecure:127.0.0.1}"/>
+ </interface>
+ </interfaces>
+
+ <socket-binding-group name="standard-sockets" default-interface="public" port-offset="${jboss.socket.binding.port-offset:0}">
+ <socket-binding name="management-http" interface="management" port="${jboss.management.http.port:9990}"/>
+ <socket-binding name="management-https" interface="management" port="${jboss.management.https.port:9993}"/>
+ <socket-binding name="ajp" port="${jboss.ajp.port:8009}"/>
+ <socket-binding name="http" port="${jboss.http.port:8080}"/>
+ <socket-binding name="https" port="${jboss.https.port:8443}"/>
+ <socket-binding name="iiop" interface="unsecure" port="3528"/>
+ <socket-binding name="iiop-ssl" interface="unsecure" port="3529"/>
+ <socket-binding name="jgroups-mping" interface="private" port="0" multicast-address="${jboss.default.multicast.address:230.0.0.4}" multicast-port="45700"/>
+ <socket-binding name="jgroups-tcp" interface="private" port="7600"/>
+ <socket-binding name="jgroups-tcp-fd" interface="private" port="57600"/>
+ <socket-binding name="jgroups-udp" interface="private" port="55200" multicast-address="${jboss.default.multicast.address:230.0.0.4}" multicast-port="45688"/>
+ <socket-binding name="jgroups-udp-fd" interface="private" port="54200"/>
+ <socket-binding name="modcluster" port="0" multicast-address="224.0.1.105" multicast-port="23364"/>
+ <socket-binding name="txn-recovery-environment" port="4712"/>
+ <socket-binding name="txn-status-manager" port="4713"/>
+ <outbound-socket-binding name="mail-smtp">
+ <remote-destination host="localhost" port="25"/>
+ </outbound-socket-binding>
+ </socket-binding-group>
+
+</server>
"mso-api-handler-infra-config":
{
- "bpelURL": "http://mso.{{ .Values.nsPrefix }}-mso.svc.cluster.local:8080",
+ "bpelURL": "http://localhost:8080",
"bpelAuth": "786864AA53D0DCD881AED1154230C0C3058D58B9339D2EFB6193A0F0D82530E1",
- "camundaURL": "http://mso.{{ .Values.nsPrefix }}-mso.svc.cluster.local:8080",
+ "camundaURL": "http://localhost:8080",
"camundaAuth": "5119D1AF37F671FC01FFAD2151D93EFB2BBB503E879FD07104D024EDDF118FD1"
},
"consumerGroup": "sdc-OpenSource-Env1",
"consumerId": "sdc-COpenSource-Env11",
"environmentName": "{{ .Values.dmaapTopic }}",
- "asdcAddress": "sdc-be.onap-sdc.svc.cluster.local:8443",
+ "asdcAddress": "sdc-be.{{.Values.nsPrefix}}:8443",
"password": "613AF3483E695524F9857643B697FA51C7A9A0951094F53791485BF3458F9EADA37DBACCCEBD0CB242B85B4062745247",
"pollingInterval": 60,
"pollingTimeout": 60,
{
"sdncurls":
[
- "http://sdnhost.{{ .Values.nsPrefix }}-sdnc.svc.cluster.local:8282/restconf/operations/L3SDN-API:",
- "http://sdnhost.{{ .Values.nsPrefix }}-sdnc.svc.cluster.local:8282/restconf/config/L3SDN-API:",
- "http://sdnhost.{{ .Values.nsPrefix }}-sdnc.svc.cluster.local:8282/restconf/operations/Firewall-API:",
- "http://sdnhost.{{ .Values.nsPrefix }}-sdnc.svc.cluster.local:8282/restconf/config",
- "http://sdnhost.{{ .Values.nsPrefix }}-sdnc.svc.cluster.local:8282/restconf/operations/VNF-API:",
- "http://sdnhost.{{ .Values.nsPrefix }}-sdnc.svc.cluster.local:8282/restconf/operations/NBNC-API:",
- "http://sdnhost.{{ .Values.nsPrefix }}-sdnc.svc.cluster.local:8282/restconf/operations/NORTHBOUND-API:service-topology-operation",
- "http://sdnhost.{{ .Values.nsPrefix }}-sdnc.svc.cluster.local:8282/restconf/operations/GENERIC-RESOURCE-API:",
- "http://sdnhost.{{ .Values.nsPrefix }}-sdnc.svc.cluster.local:8282/restconf/operations/VNFTOPOLOGYAIC-API:"
+ "http://sdnhost.{{.Values.nsPrefix}}:8282/restconf/operations/L3SDN-API:",
+ "http://sdnhost.{{.Values.nsPrefix}}:8282/restconf/config/L3SDN-API:",
+ "http://sdnhost.{{.Values.nsPrefix}}:8282/restconf/operations/Firewall-API:",
+ "http://sdnhost.{{.Values.nsPrefix}}:8282/restconf/config",
+ "http://sdnhost.{{.Values.nsPrefix}}:8282/restconf/operations/VNF-API:",
+ "http://sdnhost.{{.Values.nsPrefix}}:8282/restconf/operations/NBNC-API:",
+ "http://sdnhost.{{.Values.nsPrefix}}:8282/restconf/operations/NORTHBOUND-API:service-topology-operation",
+ "http://sdnhost.{{.Values.nsPrefix}}:8282/restconf/operations/GENERIC-RESOURCE-API:",
+ "http://sdnhost.{{.Values.nsPrefix}}:8282/restconf/operations/VNFTOPOLOGYAIC-API:"
],
- "bpelurl": "http://mso.{{ .Values.nsPrefix }}-mso.svc.cluster.local:8080/mso/SDNCAdapterCallbackService",
- "restbpelurl": "http://mso.{{ .Values.nsPrefix }}-mso.svc.cluster.local:8080/mso/WorkflowMessage",
- "myurl": "http://mso.{{ .Values.nsPrefix }}-mso.svc.cluster.local:8080/adapters/rest/SDNCNotify",
+ "bpelurl": "http://localhost:8080/mso/SDNCAdapterCallbackService",
+ "restbpelurl": "http://localhost:8080/mso/WorkflowMessage",
+ "myurl": "http://localhost:8080/adapters/rest/SDNCNotify",
"sdncauth": "263f7d5f944d4d0c76db74b4148bec67d0bc796a874bc0d2a2a12aae89a866aa69133f700f391f784719a37f6a68d29bf5a2fbae1dab0402db7788c800c5ba73",
"bpelauth": "5119D1AF37F671FC01FFAD2151D93EFB2BBB503E879FD07104D024EDDF118FD1",
"sdncconnecttime": "5000"
"mso-workflow-message-adapter-config":
{
- "wmbpelurl": "http://mso.{{ .Values.nsPrefix }}-mso.svc.cluster.local:8080/mso/WorkflowMessage",
+ "wmbpelurl": "http://localhost:8080/mso/WorkflowMessage",
"wmbpelauth": "5119D1AF37F671FC01FFAD2151D93EFB2BBB503E879FD07104D024EDDF118FD1"
},
"versionIdL3ToHigherLayerDeleteBonding": "52dbec20-47aa-42e4-936c-331d8e350d44",
"infraCustomerId": "21014aa2-526b-11e6-beb8-9e71128cae77",
"sniroAuth": "test:testpwd",
- "sniroEndpoint": "http://sniro-emulator.{{ .Values.nsPrefix }}-mock.svc.cluster.local:8080/sniro/api/v2/placement",
+ "sniroEndpoint": "http://sniro-emulator.{{.Values.nsPrefix}}:8080/sniro/api/v2/placement",
"sniroTimeout": "PT30M",
- "serviceAgnosticSniroHost": "http://sniro-emulator.{{ .Values.nsPrefix }}-mock.svc.cluster.local:8080",
+ "serviceAgnosticSniroHost": "http://sniro-emulator.{{.Values.nsPrefix}}:8080",
"serviceAgnosticSniroEndpoint": "/sniro/api/v2/placement",
- "aaiEndpoint": "https://aai-service.{{ .Values.nsPrefix }}-aai.svc.cluster.local:8443",
+ "aaiEndpoint": "https://aai-service.{{.Values.nsPrefix}}:8443",
"aaiAuth": "2630606608347B7124C244AB0FE34F6F",
"adaptersNamespace": "http://org.openecomp.mso",
- "adaptersCompletemsoprocessEndpoint": "http://mso.{{ .Values.nsPrefix }}-mso.svc.cluster.local:8080/CompleteMsoProcess",
- "adaptersDbEndpoint": "http://mso.{{ .Values.nsPrefix }}-mso.svc.cluster.local:8080/dbadapters/MsoRequestsDbAdapter",
- "adaptersOpenecompDbEndpoint": "http://mso.{{ .Values.nsPrefix }}-mso.svc.cluster.local:8080/dbadapters/RequestsDbAdapter",
- "catalogDbEndpoint": "http://mso.{{ .Values.nsPrefix }}-mso.svc.cluster.local:8080/ecomp/mso/catalog",
- "adaptersSdncEndpoint": "http://mso.{{ .Values.nsPrefix }}-mso.svc.cluster.local:8080/adapters/SDNCAdapter",
- "adaptersSdncRestEndpoint": "http://mso.{{ .Values.nsPrefix }}-mso.svc.cluster.local:8080/adapters/rest/v1/sdnc",
- "adaptersTenantEndpoint": "http://mso.{{ .Values.nsPrefix }}-mso.svc.cluster.local:8080/tenants/TenantAdapter",
+ "adaptersCompletemsoprocessEndpoint": "http://localhost:8080/CompleteMsoProcess",
+ "adaptersDbEndpoint": "http://localhost:8080/dbadapters/MsoRequestsDbAdapter",
+ "adaptersOpenecompDbEndpoint": "http://localhost:8080/dbadapters/RequestsDbAdapter",
+ "catalogDbEndpoint": "http://localhost:8080/ecomp/mso/catalog",
+ "adaptersSdncEndpoint": "http://localhost:8080/adapters/SDNCAdapter",
+ "adaptersSdncRestEndpoint": "http://localhost:8080/adapters/rest/v1/sdnc",
+ "adaptersTenantEndpoint": "http://localhost:8080/tenants/TenantAdapter",
"adaptersDbAuth": "6B0E6863FB8EE010AB6F191B3C0489437601E81DC7C86305CB92DB98AFC53D74",
- "adaptersWorkflowMessageEndpoint": "http://mso.{{ .Values.nsPrefix }}-mso.svc.cluster.local:8080/workflows/messages/message",
- "workflowMessageEndpoint": "http://mso.{{ .Values.nsPrefix }}-mso.svc.cluster.local:8080/mso/WorkflowMessage",
- "workflowSdncAdapterCallback": "http://mso.{{ .Values.nsPrefix }}-mso.svc.cluster.local:8080/mso/SDNCAdapterCallbackService",
+ "adaptersWorkflowMessageEndpoint": "http://localhost:8080/workflows/messages/message",
+ "workflowMessageEndpoint": "http://localhost:8080/mso/WorkflowMessage",
+ "workflowSdncAdapterCallback": "http://localhost:8080/mso/SDNCAdapterCallbackService",
"workflowSdncReplicationDelay": "PT5S",
"workflowAaiDistributionDelay": "PT30S",
"msoKey": "07a7159d3bf51a0e53be7a8f89699be7",
"adaptersPoAuth": "6B0E6863FB8EE010AB6F191B3C0489437601E81DC7C86305CB92DB98AFC53D74",
"sdncTimeout": "PT5M",
"rollback": "true",
- "adaptersNetworkEndpoint": "http://mso.{{ .Values.nsPrefix }}-mso.svc.cluster.local:8080/networks/NetworkAdapter",
- "adaptersNetworkRestEndpoint": "http://mso.{{ .Values.nsPrefix }}-mso.svc.cluster.local:8080/networks/rest/v1/networks",
- "adaptersVnfAsyncEndpoint": "http://mso.{{ .Values.nsPrefix }}-mso.svc.cluster.local:8080/vnfs/VnfAdapterAsync",
- "workflowVnfAdapterDeleteCallback": "http://mso.{{ .Values.nsPrefix }}-mso.svc.cluster.local:8080/mso/vnfAdapterNotify",
- "workflowVnfAdapterCreateCallback": "http://mso.{{ .Values.nsPrefix }}-mso.svc.cluster.local:8080/mso/vnfAdapterNotify",
- "adaptersVnfRestEndpoint": "http://mso.{{ .Values.nsPrefix }}-mso.svc.cluster.local:8080/vnfs/rest/v1/vnfs",
- "workflowVnfAdapterRestCallback": "http://mso.{{ .Values.nsPrefix }}-mso.svc.cluster.local:8080/mso/vnfAdapterRestNotify",
+ "adaptersNetworkEndpoint": "http://localhost:8080/networks/NetworkAdapter",
+ "adaptersNetworkRestEndpoint": "http://localhost:8080/networks/rest/v1/networks",
+ "adaptersVnfAsyncEndpoint": "http://localhost:8080/vnfs/VnfAdapterAsync",
+ "workflowVnfAdapterDeleteCallback": "http://localhost:8080/mso/vnfAdapterNotify",
+ "workflowVnfAdapterCreateCallback": "http://localhost:8080/mso/vnfAdapterNotify",
+ "adaptersVnfRestEndpoint": "http://localhost:8080/vnfs/rest/v1/vnfs",
+ "workflowVnfAdapterRestCallback": "http://localhost:8080/mso/vnfAdapterRestNotify",
"poTimeout": "PT5M",
"sdncFirewallYangModel": "http://com/att/svc/mis/firewall-lite-gui",
"sdncFirewallYangModelVersion": "2015-05-15",
apiVersion: v1
kind: Service
metadata:
- name: mariadb
- namespace: "{{ .Values.nsPrefix }}-mso"
+ name: mso-mariadb
+ namespace: "{{ .Values.nsPrefix }}"
labels:
- app: mariadb
+ app: mso-mariadb
spec:
ports:
- - port: 3306
- nodePort: {{ .Values.nodePortPrefix }}52
+ - port: 3306
+ name: mso-mariadb-3306
selector:
- app: mariadb
- type: NodePort
+ app: mso-mariadb
+ clusterIP: None
#{{ end }}
#{{ if not .Values.disableMsoMso }}
---
kind: Service
metadata:
name: mso
- namespace: "{{ .Values.nsPrefix }}-mso"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: mso
annotations:
port: 8787
nodePort: {{ .Values.nodePortPrefix }}50
type: NodePort
-#{{ end }}
\ No newline at end of file
+#{{ end }}
kind: ConfigMap
metadata:
name: mso-confd-configmap
- namespace: {{ .Values.nsPrefix }}-mso
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/mariadb/conf.d/*").AsConfig . | indent 2 }}
---
kind: ConfigMap
metadata:
name: mso-docker-entry-initd-configmap
- namespace: {{ .Values.nsPrefix }}-mso
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/mariadb/docker-entrypoint-initdb.d/*").AsConfig . | indent 2 }}
---
kind: ConfigMap
metadata:
name: mso-automated-tests-configmap
- namespace: {{ .Values.nsPrefix }}-mso
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/mariadb/docker-entrypoint-initdb.d/db-sql-scripts/bulkload-files/automated-tests/*").AsConfig . | indent 2 }}
---
kind: ConfigMap
metadata:
name: mso-bulkload-default-configmap
- namespace: {{ .Values.nsPrefix }}-mso
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/mariadb/docker-entrypoint-initdb.d/db-sql-scripts/bulkload-files/default/*").AsConfig . | indent 2 }}
---
kind: ConfigMap
metadata:
name: mso-demo-dns-configmap
- namespace: {{ .Values.nsPrefix }}-mso
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/mariadb/docker-entrypoint-initdb.d/db-sql-scripts/bulkload-files/demo-dns/*").AsConfig . | indent 2 }}
---
kind: ConfigMap
metadata:
name: mso-demo-vfw-configmap
- namespace: {{ .Values.nsPrefix }}-mso
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/mariadb/docker-entrypoint-initdb.d/db-sql-scripts/bulkload-files/demo-vfw/*").AsConfig . | indent 2 }}
---
kind: ConfigMap
metadata:
name: mso-scripts-camunda-configmap
- namespace: {{ .Values.nsPrefix }}-mso
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/mariadb/docker-entrypoint-initdb.d/db-sql-scripts/camunda/*").AsConfig . | indent 2 }}
---
kind: ConfigMap
metadata:
name: mso-main-schemas-configmap
- namespace: {{ .Values.nsPrefix }}-mso
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/mariadb/docker-entrypoint-initdb.d/db-sql-scripts/main-schemas/*").AsConfig . | indent 2 }}
#{{ end }}
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
- name: mariadb
- namespace: "{{ .Values.nsPrefix }}-mso"
+ name: mso-mariadb
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.dbReplicas }}
selector:
matchLabels:
- app: mariadb
+ app: mso-mariadb
template:
metadata:
labels:
- app: mariadb
- name: mariadb
+ app: mso-mariadb
+ name: mso-mariadb
spec:
- hostname: mariadb
+ hostname: mso-mariadb
containers:
- args:
image: {{ .Values.image.mariadb }}
imagePullPolicy: {{ .Values.pullPolicy }}
- name: "mariadb"
+ name: "mso-mariadb"
env:
- name: MYSQL_ROOT_PASSWORD
value: password
name: mso-mariadb-data
ports:
- containerPort: 3306
- name: mariadb
+ name: mso-mariadb
readinessProbe:
tcpSocket:
port: 3306
kind: ConfigMap
metadata:
name: mso-config-mso-configmap
- namespace: {{ .Values.nsPrefix }}-mso
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/mso/*").AsConfig . | indent 2 }}
---
kind: ConfigMap
metadata:
name: mso-docker-file-configmap
- namespace: {{ .Values.nsPrefix }}-mso
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/docker-files/scripts/start-jboss-server.sh").AsConfig . | indent 2 }}
+---
+apiVersion: v1
+kind: ConfigMap
+metadata:
+ name: mso-jboss-configmap
+ namespace: {{ .Values.nsPrefix }}
+data:
+{{ tpl (.Files.Glob "resources/config/mso/jboss/standalone-full-ha-mso.xml").AsConfig . | indent 2 }}
#{{ end }}
kind: Deployment
metadata:
name: mso
- namespace: "{{ .Values.nsPrefix }}-mso"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.msoReplicas }}
selector:
- /root/ready.py
args:
- --container-name
- - mariadb
+ - mso-mariadb
env:
- name: NAMESPACE
valueFrom:
- mountPath: /tmp/start-jboss-server.sh
name: mso-docker-files
subPath: start-jboss-server.sh
+ - mountPath: /opt/jboss/standalone/configuration/standalone-full-ha-mso.xml
+ name: mso-jboss-config
+ subPath: standalone-full-ha-mso.xml
- mountPath: /var/log/onap
name: mso-logs
- mountPath: /var/berks-cookbooks/mso-config/files/default/mso-po-adapter-config/logback.network.xml
- key: start-jboss-server.sh
path: start-jboss-server.sh
mode: 0755
+ - name: mso-jboss-config
+ configMap:
+ name: mso-jboss-configmap
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
#{{ end }}
kind: ConfigMap
metadata:
name: mso-log-configmap
- namespace: {{ .Values.nsPrefix }}-mso
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/log/*").AsConfig . | indent 2 }}
---
kind: ConfigMap
metadata:
name: mso-filebeat-configmap
- namespace: {{ .Values.nsPrefix }}-mso
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/log/filebeat/*").AsConfig . | indent 2 }}
#{{ end }}
apiVersion: v1
kind: PersistentVolume
metadata:
- name: "{{ .Values.nsPrefix }}-mso-db"
- namespace: "{{ .Values.nsPrefix }}-mso"
+ name: "{{ .Values.nsPrefix }}-mso"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
- name: "{{ .Values.nsPrefix }}-mso-db"
+ name: "{{ .Values.nsPrefix }}-mso"
spec:
capacity:
storage: 2Gi
apiVersion: v1
metadata:
name: mso-db
- namespace: "{{ .Values.nsPrefix }}-mso"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
accessModes:
- ReadWriteMany
storage: 2Gi
selector:
matchLabels:
- name: "{{ .Values.nsPrefix }}-mso-db"
+ name: "{{ .Values.nsPrefix }}-mso"
#{{ end }}
dataRootDir: /dockerdata-nfs
image:
readiness: oomk8s/readiness-check:1.1.0
- mso: nexus3.onap.org:10001/openecomp/mso:v1.1.1
+ mso: nexus3.onap.org:10001/openecomp/mso:v1.1.2
mariadb: nexus3.onap.org:10001/mariadb:10.1.11
filebeat: docker.elastic.co/beats/filebeat:5.5.0
--- /dev/null
+filebeat.prospectors:
+#it is mandatory, in our case it's log
+- input_type: log
+ #This is the canolical path as mentioned in logback.xml, *.* means it will monitor all files in the directory.
+ paths:
+ - /var/log/onap/*/*/*/*.log
+ - /var/log/onap/*/*/*.log
+ - /var/log/onap/*/*.log
+ #Files older than this should be ignored.In our case it will be 48 hours i.e. 2 days. It is a helping flag for clean_inactive
+ ignore_older: 24h
+ # Remove the registry entry for a file that is more than the specified time. In our case it will be 96 hours, i.e. 4 days. It will help to keep registry records with in limit
+ clean_inactive: 48h
+
+
+# Name of the registry file. If a relative path is used, it is considered relative to the
+# data path. Else full qualified file name.
+#filebeat.registry_file: ${path.data}/registry
+
+
+output.logstash:
+ #List of logstash server ip addresses with port number.
+ #But, in our case, this will be the loadbalancer IP address.
+ #For the below property to work the loadbalancer or logstash should expose 5044 port to listen the filebeat events or port in the property should be changed appropriately.
+ hosts: ["logstash.{{.Values.nsPrefix}}:5044"]
+ #If enable will do load balancing among availabe Logstash, automatically.
+ loadbalance: true
+
+ #The list of root certificates for server verifications.
+ #If certificate_authorities is empty or not set, the trusted
+ #certificate authorities of the host system are used.
+ #ssl.certificate_authorities: $ssl.certificate_authorities
+
+ #The path to the certificate for SSL client authentication. If the certificate is not specified,
+ #client authentication is not available.
+ #ssl.certificate: $ssl.certificate
+
+ #The client certificate key used for client authentication.
+ #ssl.key: $ssl.key
+
+ #The passphrase used to decrypt an encrypted key stored in the configured key file
+ #ssl.key_passphrase: $ssl.key_passphrase
--- /dev/null
+version: 1
+disable_existing_loggers: False
+
+loggers:
+ multivimbroker:
+ handlers: [multivimbroker_handler]
+ level: "DEBUG"
+ propagate: False
+
+handlers:
+ multivimbroker_handler:
+ level: "DEBUG"
+ class: "logging.handlers.RotatingFileHandler"
+ filename: "/var/log/onap/multicloud/multivimbroker/multivimbroker.log"
+ formatter: "mdcFormat"
+ maxBytes: 1024*1024*50
+ backupCount: 10
+
+formatters:
+ standard:
+ format: "%(asctime)s:[%(name)s]:[%(filename)s]-[%(lineno)d] [%(levelname)s]:%(message)s"
+ mdcFormat:
+ format: "%(asctime)s:[%(name)s]:[%(filename)s]-[%(lineno)d] [%(levelname)s]:[%(mdc)s]: %(message)s"
+ mdcfmt: "{requestID}"
+ datefmt: "%Y-%m-%d %H:%M:%S"
+ (): onaplogging.mdcformatter.MDCFormatter
\ No newline at end of file
--- /dev/null
+version: 1
+disable_existing_loggers: False
+
+loggers:
+ vio:
+ handlers: [vio_handler]
+ level: "DEBUG"
+ propagate: False
+
+handlers:
+ vio_handler:
+ level: "DEBUG"
+ class: "logging.handlers.RotatingFileHandler"
+ filename: "/var/log/onap/multicloud/vio/vio.log"
+ formatter: "mdcFormat"
+ maxBytes: 1024*1024*50
+ backupCount: 10
+
+formatters:
+ standard:
+ format: "%(asctime)s:[%(name)s]:[%(filename)s]-[%(lineno)d] [%(levelname)s]:%(message)s"
+ mdcFormat:
+ format: "%(asctime)s:[%(name)s]:[%(filename)s]-[%(lineno)d] [%(levelname)s]:[%(mdc)s]: %(message)s"
+ mdcfmt: "{requestID}"
+ datefmt: "%Y-%m-%d %H:%M:%S"
+ (): onaplogging.mdcformatter.MDCFormatter
\ No newline at end of file
kind: Service
metadata:
name: framework
- namespace: "{{ .Values.nsPrefix }}-multicloud"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
name: framework
annotations:
kind: Service
metadata:
name: multicloud-vio
- namespace: "{{ .Values.nsPrefix }}-multicloud"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
name: multicloud-vio
annotations:
kind: Service
metadata:
name: multicloud-ocata
- namespace: "{{ .Values.nsPrefix }}-multicloud"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
name: multicloud-ocata
annotations:
kind: Service
metadata:
name: multicloud-windriver
- namespace: "{{ .Values.nsPrefix }}-multicloud"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
name: multicloud-windriver
annotations:
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
- name: framework
- namespace: "{{ .Values.nsPrefix }}-multicloud"
+ name: multicloud-framework
+ namespace: "{{ .Values.nsPrefix }}"
spec:
+ replicas: {{ .Values.frameworkReplicas }}
selector:
matchLabels:
app: framework
metadata:
labels:
app: framework
- name: framework
+ name: multicloud-framework
spec:
- hostname: framework
containers:
- env:
- name: MSB_ADDR
- value: {{ .Values.msbgateway }}
+ value: {{ tpl .Values.msbgateway . }}
- name: MSB_PORT
value: "{{ .Values.msbPort }}"
- name: AAI_ADDR
- value: aai-service.{{ .Values.nsPrefix }}-aai
+ value: aai-service.{{ .Values.nsPrefix }}
- name: AAI_PORT
value: "8443"
- name: AAI_SCHEMA_VERSION
value: "AAI"
- name: AAI_PASSWORD
value: "AAI"
+ image: {{ .Values.image.framework }}
+ imagePullPolicy: {{ .Values.pullPolicy }}
name: framework
volumeMounts:
- - mountPath: /service/multivimbroker/logs
+ - mountPath: /var/log/onap
name: framework-log
- image: {{ .Values.image.framework }}
- imagePullPolicy: {{ .Values.pullPolicy }}
+ - mountPath: /opt/multivimbroker/multivimbroker/pub/config/log.yml
+ name: framework-logconfig
+ subPath: log.yml
ports:
- containerPort: 9001
livenessProbe:
timeoutSeconds: 10
successThreshold: 1
failureThreshold: 5
- restartPolicy: Always
+ - image: {{ .Values.image.filebeat }}
+ imagePullPolicy: {{ .Values.pullPolicy }}
+ name: filebeat-onap
+ volumeMounts:
+ - mountPath: /usr/share/filebeat/filebeat.yml
+ name: filebeat-conf
+ subPath: filebeat.yml
+ - mountPath: /var/log/onap
+ name: framework-log
+ - mountPath: /usr/share/filebeat/data
+ name: framework-data-filebeat
volumes:
- name: framework-log
- hostPath:
- path: {{ .Values.dataRootDir }}/{{ .Values.nsPrefix }}/multicloud/framework/logs
+ emptyDir: {}
+ - name: framework-data-filebeat
+ emptyDir: {}
+ - name: filebeat-conf
+ configMap:
+ name: multicloud-filebeat-configmap
+ - name: framework-logconfig
+ configMap:
+ name: multicloud-framework-log-configmap
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
+ restartPolicy: Always
#{{ end }}
--- /dev/null
+#{{ if not .Values.disableVidVidServer }}
+apiVersion: v1
+kind: ConfigMap
+metadata:
+ name: multicloud-filebeat-configmap
+ namespace: {{ .Values.nsPrefix }}
+data:
+{{ tpl (.Files.Glob "resources/config/log/filebeat/*").AsConfig . | indent 2 }}
+---
+apiVersion: v1
+kind: ConfigMap
+metadata:
+ name: multicloud-framework-log-configmap
+ namespace: {{ .Values.nsPrefix }}
+data:
+{{ tpl (.Files.Glob "resources/config/log/framework/*").AsConfig . | indent 2 }}
+---
+apiVersion: v1
+kind: ConfigMap
+metadata:
+ name: multicloud-vio-log-configmap
+ namespace: {{ .Values.nsPrefix }}
+data:
+{{ tpl (.Files.Glob "resources/config/log/vio/*").AsConfig . | indent 2 }}
+#{{ end }}
kind: Deployment
metadata:
name: multicloud-ocata
- namespace: "{{ .Values.nsPrefix }}-multicloud"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
+ replicas: {{ .Values.ocataReplicas }}
selector:
matchLabels:
app: multicloud-ocata
containers:
- env:
- name: MSB_ADDR
- value: "{{ .Values.msbgateway }}"
+ value: "{{ tpl .Values.msbgateway . }}"
- name: MSB_PORT
value: "{{ .Values.msbPort }}"
- name: AAI_ADDR
- value: aai-service.{{ .Values.nsPrefix }}-aai
+ value: aai-service.{{ .Values.nsPrefix }}
- name: AAI_PORT
value: "8443"
- name: AAI_SCHEMA_VERSION
kind: Deployment
metadata:
name: multicloud-vio
- namespace: "{{ .Values.nsPrefix }}-multicloud"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
+ replicas: {{ .Values.vioReplicas }}
selector:
matchLabels:
app: multicloud-vio
app: multicloud-vio
name: multicloud-vio
spec:
- hostname: multicloud-vio
containers:
- env:
- name: MSB_ADDR
- value: "{{ .Values.msbgateway }}"
+ value: "{{ tpl .Values.msbgateway . }}"
- name: MSB_PORT
value: "{{ .Values.msbPort }}"
- name: AAI_ADDR
- value: aai-service.{{ .Values.nsPrefix }}-aai
+ value: aai-service.{{ .Values.nsPrefix }}
- name: AAI_PORT
value: "8443"
- name: AAI_SCHEMA_VERSION
value: "AAI"
name: multicloud-vio
volumeMounts:
- - mountPath: /service/vio/logs
+ - mountPath: /var/log/onap
name: vio-log
+ - mountPath: /opt/vio/vio/pub/config/log.yml
+ name: vio-logconfig
+ subPath: log.yml
image: {{ .Values.image.vio }}
imagePullPolicy: {{ .Values.pullPolicy }}
ports:
timeoutSeconds: 10
successThreshold: 1
failureThreshold: 5
- restartPolicy: Always
+ - image: {{ .Values.image.filebeat }}
+ imagePullPolicy: {{ .Values.pullPolicy }}
+ name: filebeat-onap
+ volumeMounts:
+ - mountPath: /usr/share/filebeat/filebeat.yml
+ name: filebeat-conf
+ subPath: filebeat.yml
+ - mountPath: /var/log/onap
+ name: vio-log
+ - mountPath: /usr/share/filebeat/data
+ name: vio-data-filebeat
volumes:
- name: vio-log
- hostPath:
- path: {{ .Values.dataRootDir }}/{{ .Values.nsPrefix }}/multicloud/vio/logs
+ emptyDir: {}
+ - name: vio-data-filebeat
+ emptyDir: {}
+ - name: filebeat-conf
+ configMap:
+ name: multicloud-filebeat-configmap
+ - name: vio-logconfig
+ configMap:
+ name: multicloud-vio-log-configmap
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
+ restartPolicy: Always
#{{ end }}
kind: Deployment
metadata:
name: multicloud-windriver
- namespace: "{{ .Values.nsPrefix }}-multicloud"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
+ replicas: {{ .Values.windRiverReplicas }}
selector:
matchLabels:
app: multicloud-windriver
containers:
- env:
- name: MSB_ADDR
- value: "{{ .Values.msbgateway }}"
+ value: "{{ tpl .Values.msbgateway . }}"
- name: MSB_PORT
value: "{{ .Values.msbPort }}"
- name: AAI_ADDR
- value: aai-service.{{ .Values.nsPrefix }}-aai
+ value: aai-service.{{ .Values.nsPrefix }}
- name: AAI_PORT
value: "8443"
- name: AAI_SCHEMA_VERSION
pullPolicy: Always
nodePortPrefix: 302
dataRootDir: /dockerdata-nfs
+frameworkReplicas: 1
+ocataReplicas: 1
+vioReplicas: 1
+windRiverReplicas: 1
image:
readiness: oomk8s/readiness-check:1.1.0
- framework: nexus3.onap.org:10001/onap/multicloud/framework:v1.0.0
- vio: nexus3.onap.org:10001/onap/multicloud/vio:v1.0.0
+ framework: nexus3.onap.org:10001/onap/multicloud/framework:latest
+ vio: nexus3.onap.org:10001/onap/multicloud/vio:latest
ocata: nexus3.onap.org:10001/onap/multicloud/openstack-ocata:v1.0.0
windriver: nexus3.onap.org:10001/onap/multicloud/openstack-windriver:v1.0.0
-
+ filebeat: docker.elastic.co/beats/filebeat:5.5.0
# domain name of msb gateway
-msbgateway: msb-iag.onap-msb
+msbgateway: msb-iag.{{ .Values.nsPrefix }}
msbPort: 80
version: ~0.1.0
repository: '@local'
condition: cli.enabled
+ - name: common
+ version: ~2.0.0
+ repository: '@local'
- name: consul
version: ~1.1.0
repository: '@local'
version: ~1.1.0
repository: '@local'
condition: esr.enabled
- - name: kube2msb
- version: ~0.1.0
- repository: '@local'
- condition: kube2msb.enabled
- name: log
version: ~0.1.0
repository: '@local'
repository: '@local'
condition: portal.enabled
- name: robot
- version: ~0.1.0
+ version: ~2.0.0
repository: '@local'
condition: robot.enabled
- name: sdc
version: ~0.1.0
repository: '@local'
condition: sdnc.enabled
+ - name: setup
+ version: ~2.0.0
+ repository: '@local'
- name: so
version: ~2.0.0
repository: '@local'
- name: vnfsdk
version: ~1.1.0
repository: '@local'
- condition: vnfsdk.enabled
\ No newline at end of file
+ condition: vnfsdk.enabled
# flag to enable debugging - application support required
debugEnabled: false
+
#################################################################
# Enable/disable and configure helm charts (ie. applications)
# to customize the ONAP deployment.
enabled: false
esr:
enabled: false
-kube2msb:
- enabled: false
log:
enabled: false
message-router:
portal:
enabled: false
robot: # Robot Health Check
- enabled: false
+ enabled: true
sdc:
enabled: false
sdnc:
vid:
enabled: false
vnfsdk:
- enabled: false
\ No newline at end of file
+ enabled: false
}
create_service_account() {
- cmd=`echo kubectl create clusterrolebinding $1-$2-admin-binding --clusterrole=cluster-admin --serviceaccount=$1-$2:default`
+ cmd=`echo kubectl create clusterrolebinding $1-admin-binding --clusterrole=cluster-admin --serviceaccount=$1:default`
eval ${cmd}
check_return_code $cmd
}
create_namespace() {
- cmd=`echo kubectl create namespace $1-$2`
+ cmd=`echo kubectl create namespace $1`
eval ${cmd}
- check_return_code $cmd
}
create_registry_key() {
- cmd=`echo kubectl --namespace $1-$2 create secret docker-registry $3 --docker-server=$4 --docker-username=$5 --docker-password=$6 --docker-email=$7`
+cmd=`echo kubectl --namespace $1 create secret docker-registry $2 --docker-server=$3 --docker-username=$4 --docker-password=$5 --docker-email=$6`
eval ${cmd}
check_return_code $cmd
}
# assign default auth token
if [[ -z $ONAP_DEFAULT_AUTH_TOKEN ]]; then
- DEFAULT_SECRET=`kubectl get secrets -n $1-$2 | grep default-token | awk '{ print $1}'`
- ONAP_DEFAULT_AUTH_TOKEN=`kubectl get secrets $DEFAULT_SECRET -n $1-$2 -o yaml | grep 'token:' | awk '{ print $2}' | base64 --decode`
+ DEFAULT_SECRET=`kubectl get secrets -n $1 | grep default-token | awk '{ print $1}'`
+ ONAP_DEFAULT_AUTH_TOKEN=`kubectl get secrets $DEFAULT_SECRET -n $1 -o yaml | grep 'token:' | awk '{ print $2}' | base64 --decode`
fi
cmd=`echo helm install $LOCATION/$2/ --name $1-$2 --namespace $1 --set nsPrefix=$1,nodePortPrefix=$3,kubeMasterAuthToken=$ONAP_DEFAULT_AUTH_TOKEN ${HELM_VALUES_ADDITION}`
DU=$ONAP_DOCKER_USER
DP=$ONAP_DOCKER_PASS
+SINGLE_COMPONENT=false
+
while getopts ":n:u:s:i:a:du:dp:l:v:" PARAM; do
case $PARAM in
u)
LOCATION=${OPTARG}
;;
a)
+ SINGLE_COMPONENT=true
APP=${OPTARG}
if [[ -z $APP ]]; then
usage
printf "\n********** Creating ONAP: ${ONAP_APPS[*]}\n"
+if [ "$SINGLE_COMPONENT" == "false" ]
+then
+ printf "\nCreating namespace **********\n"
+ create_namespace $NS
-printf "\n\n********** Creating deployments for ${HELM_APPS[*]} ********** \n"
+ printf "\nCreating registry secret **********\n"
+ create_registry_key $NS ${NS}-docker-registry-key $ONAP_DOCKER_REGISTRY $DU $DP $ONAP_DOCKER_MAIL
-for i in ${HELM_APPS[@]}; do
- printf "\nCreating namespace **********\n"
- create_namespace $NS $i
+ printf "\nCreating service account **********\n"
+ create_service_account $NS
+fi
- printf "\nCreating service account **********\n"
- create_service_account $NS $i
+printf "\n\n********** Creating deployments for ${HELM_APPS[*]} ********** \n"
- printf "\nCreating registry secret **********\n"
- create_registry_key $NS $i ${NS}-docker-registry-key $ONAP_DOCKER_REGISTRY $DU $DP $ONAP_DOCKER_MAIL
+for i in ${HELM_APPS[@]}; do
printf "\nCreating deployments and services **********\n"
create_onap_helm $NS $i $start
. $(dirname "$0")/setenv.bash
delete_namespace() {
- _NS=$1-$2
- kubectl delete namespace $_NS
+ kubectl delete namespace $1
}
delete_service_account() {
- kubectl delete clusterrolebinding $1-$2-admin-binding
- printf "Service account $1-$2-admin-binding deleted.\n\n"
+ kubectl delete clusterrolebinding $1-admin-binding
}
delete_registry_key() {
- kubectl --namespace $1-$2 delete secret ${1}-docker-registry-key
+ kubectl --namespace $1 delete secret ${1}-docker-registry-key
}
delete_app_helm() {
while true; do
declare -i _STATUS=0
for i in ${HELM_APPS[@]}; do
- kubectl get namespaces $1-$i > /dev/null 2>&1
- if [ "$?" -eq "0" ]; then
+ kubectl get pods --namespace $1 | grep -w " $i" > /dev/null 2>&1
+ if [ "$?" -ne "0" ]; then
_STATUS=1
break
fi
done
- if [ "$_STATUS" -eq "0" ]; then
+
+ if [ "$SINGLE_COMPONENT" == "false" ]; then
+ kubectl get namespaces $1 > /dev/null 2>&1
+ _STATUS=$?
+ fi
+ if [ "$_STATUS" -ne "0" ]; then
break
fi
sleep 2
WAIT_TERMINATE=true
SKIP_INTERACTIVE_CONFIRMATION=no
KUBECTL_CONTEXT=
-
+SINGLE_COMPONENT=false
while getopts ":c:n:u:s:a:yN" PARAM; do
case $PARAM in
u)
usage
exit 1
fi
+ SINGLE_COMPONENT=true
;;
N)
WAIT_TERMINATE=false
printf "\n********** Cleaning up ONAP: ${ONAP_APPS[*]}\n"
for i in ${HELM_APPS[@]}; do
-
delete_app_helm $NS $i
- delete_namespace $NS $i
- delete_service_account $NS $i
-
done
-delete_app_helm $NS "config"
-kubectl delete namespace $NS
+if [ "$SINGLE_COMPONENT" == "false" ]
+then
+ delete_app_helm $NS "config"
+ delete_namespace $NS
+ delete_registry_key $NS
+ delete_service_account $NS
+fi
if $WAIT_TERMINATE; then
wait_terminate $NS
#!/bin/bash
-# Deploying MSB first and kube2msb last will ensure all the ONAP services can be registered to MSB
-HELM_APPS=('consul' 'msb' 'mso' 'message-router' 'sdnc' 'vid' 'robot' 'portal' 'policy' 'appc' 'aai' 'sdc' 'dcaegen2' 'log' 'cli' 'multicloud' 'clamp' 'vnfsdk' 'uui' 'aaf' 'vfc' 'kube2msb' 'esr')
+# Deploying MSB first so the started ONAP services can be registered to MSB
+HELM_APPS=('consul' 'msb' 'mso' 'message-router' 'sdnc' 'vid' 'robot' 'portal' 'policy' 'appc' 'aai' 'sdc' 'dcaegen2' 'log' 'cli' 'multicloud' 'clamp' 'vnfsdk' 'uui' 'aaf' 'vfc' 'esr')
ONAP_DOCKER_REGISTRY=${ONAP_DOCKER_REGISTRY:-nexus3.onap.org:10001}
ONAP_DOCKER_USER=${ONAP_DOCKER_USER:-docker}
ONAP_DOCKER_PASS=${ONAP_DOCKER_PASS:-docker}
#List of logstash server ip addresses with port number.
#But, in our case, this will be the loadbalancer IP address.
#For the below property to work the loadbalancer or logstash should expose 5044 port to listen the filebeat events or port in the property should be changed appropriately.
- hosts: ["logstash.onap-log:5044"]
+ hosts: ["logstash.{{.Values.nsPrefix}}:5044"]
#If enable will do load balancing among availabe Logstash, automatically.
loadbalance: true
# Relational (SQL) DB access
-SQL_HOST=mariadb.onap-policy
+SQL_HOST=mariadb.{{.Values.nsPrefix}}
SQL_USER=policy_user
SQL_PASSWORD=policy_user
# PDP-D DMaaP configuration channel
PDPD_CONFIGURATION_TOPIC=PDPD-CONFIGURATION
-PDPD_CONFIGURATION_SERVERS=dmaap.onap-message-router
+PDPD_CONFIGURATION_SERVERS=dmaap.{{.Values.nsPrefix}}
PDPD_CONFIGURATION_API_KEY=
PDPD_CONFIGURATION_API_SECRET=
PDPD_CONFIGURATION_CONSUMER_GROUP=
# PAP
-PAP_HOST=pap.onap-policy
+PAP_HOST=pap.{{.Values.nsPrefix}}
PAP_USERNAME=testpap
PAP_PASSWORD=alpha123
# PDP-X
-PDP_HOST=pdp.onap-policy
+PDP_HOST=pdp.{{.Values.nsPrefix}}
PDP_USERNAME=testpdp
PDP_PASSWORD=alpha123
PDP_CLIENT_USERNAME=python
# DCAE DMaaP
DCAE_TOPIC=unauthenticated.DCAE_CL_OUTPUT
-DCAE_SERVERS=dmaap.onap-message-router
+DCAE_SERVERS=dmaap.{{.Values.nsPrefix}}
# Open DMaaP
-DMAAP_SERVERS=dmaap.onap-message-router
+DMAAP_SERVERS=dmaap.{{.Values.nsPrefix}}
# AAI
# MSO
-SO_URL=http://mso.onap-mso:8080/ecomp/mso/infra
+SO_URL=http://mso.{{.Values.nsPrefix}}:8080/ecomp/mso/infra
SO_USERNAME=InfraPortalClient
SO_PASSWORD=password1$
VFC_URL=
VFC_USERNAME=
-VFC_PASSWORD=
\ No newline at end of file
+VFC_PASSWORD=
KEYSTORE_PASSWD=PolicyR0ck$
JDBC_DRIVER=org.mariadb.jdbc.Driver
-JDBC_URL=jdbc:mariadb://mariadb.onap-policy:3306/onap_sdk?failOverReadOnly=false&autoReconnect=true
-JDBC_LOG_URL=jdbc:mariadb://mariadb.onap-policy:3306/log?failOverReadOnly=false&autoReconnect=true
+JDBC_URL=jdbc:mariadb://mariadb.{{.Values.nsPrefix}}:3306/onap_sdk?failOverReadOnly=false&autoReconnect=true
+JDBC_LOG_URL=jdbc:mariadb://mariadb.{{.Values.nsPrefix}}:3306/log?failOverReadOnly=false&autoReconnect=true
JDBC_USER=policy_user
JDBC_PASSWORD=policy_user
#Micro Service Model Properties
policy_msOnapName=
-policy_msPolicyName=
\ No newline at end of file
+policy_msPolicyName=
COMPONENT_X_MX_MB=1024
COMPONENT_X_MS_MB=1024
-REST_PAP_URL=http://pap.onap-policy:9091/pap/
-REST_PDP_ID=http://pdp.onap-policy:8081/pdp/
+REST_PAP_URL=http://pap.{{.Values.nsPrefix}}:9091/pap/
+REST_PDP_ID=http://pdp.{{.Values.nsPrefix}}:8081/pdp/
PDP_HTTP_USER_ID=testpdp
PDP_HTTP_PASSWORD=alpha123
M2_HOME=/usr/share/maven
snapshotRepositoryID=policy-nexus-snapshots
snapshotRepositoryName=Snapshots
-snapshotRepositoryURL=http://nexus.onap-policy:8081/nexus/content/repositories/snapshots
+snapshotRepositoryURL=http://nexus.{{.Values.nsPrefix}}:8081/nexus/content/repositories/snapshots
releaseRepositoryID=policy-nexus-releases
releaseRepositoryName=Releases
-releaseRepositoryURL=http://nexus.onap-policy:8081/nexus/content/repositories/releases
+releaseRepositoryURL=http://nexus.{{.Values.nsPrefix}}:8081/nexus/content/repositories/releases
repositoryUsername=admin
repositoryPassword=admin123
-UEB_URL=dmaap.onap-message-router
+UEB_URL=dmaap.{{.Values.nsPrefix}}
UEB_TOPIC=PDPD-CONFIGURATION
UEB_API_KEY=
UEB_API_SECRET=
#Notification Properties... type can be either websocket, ueb, or dmaap
BRMS_NOTIFICATION_TYPE=websocket
-BRMS_UEB_URL=dmaap.onap-message-router
+BRMS_UEB_URL=dmaap.{{.Values.nsPrefix}}
BRMS_UEB_TOPIC=PDPD-CONFIGURATION
BRMS_UEB_DELAY=
BRMS_CLIENT_ID=python
#
# Location where all the user workspaces are located.
#
-REST_ADMIN_WORKSPACE=${{POLICY_HOME}}/servers/console/bin/workspace
+REST_ADMIN_WORKSPACE=/opt/app/policy/servers/console/bin/workspace
#
# These can be set so the Admin Console knows who is logged on. Ideally, you can run the console in a J2EE
#
-REST_PAP_URL=http://pap.onap-policy:9091/pap/
+REST_PAP_URL=http://pap.{{.Values.nsPrefix}}:9091/pap/
#
# Config/Action Properties location.
#
-REST_CONFIG_HOME=${{POLICY_HOME}}/servers/pap/webapps/Config/
-REST_ACTION_HOME=${{POLICY_HOME}}/servers/pap/webapps/Action/
-REST_CONFIG_URL=http://pap.onap-policy:9091/
-REST_CONFIG_WEBAPPS=${{POLICY_HOME}}/servers/pap/webapps/
+REST_CONFIG_HOME=/opt/app/policy/servers/pap/webapps/Config/
+REST_ACTION_HOME=/opt/app/policy/servers/pap/webapps/Action/
+REST_CONFIG_URL=http://pap.{{.Values.nsPrefix}}:9091/
+REST_CONFIG_WEBAPPS=/opt/app/policy/servers/pap/webapps/
# PAP account information
CONSOLE_PAP_HTTP_USER_ID=testpap
#-----------------------ONAP-PORTAL-Properties----------------------
-ONAP_REDIRECT_URL=http://portalapps.onap-portal:8989/ONAPPORTAL/login.htm
-ONAP_REST_URL=http://portalapps.onap-portal:8989/ONAPPORTAL/auxapi
+ONAP_REDIRECT_URL=http://portalapps.{{.Values.nsPrefix}}:8989/ONAPPORTAL/login.htm
+ONAP_REST_URL=http://portalapps.{{.Values.nsPrefix}}:8989/ONAPPORTAL/auxapi
ONAP_UEB_URL_LIST=
ONAP_PORTAL_INBOX_NAME=
ONAP_UEB_APP_KEY=
ONAP_UEB_APP_SECRET=
ONAP_UEB_APP_MAILBOX_NAME=
APP_DISPLAY_NAME=ONAP Policy
-ONAP_SHARED_CONTEXT_REST_URL=http://portalapps.onap-portal:8989/ONAPPORTAL/context
+ONAP_SHARED_CONTEXT_REST_URL=http://portalapps.{{.Values.nsPrefix}}:8989/ONAPPORTAL/context
# pap properties
-PAP_PDPS=${{POLICY_HOME}}/servers/pap/bin/pdps
-PAP_URL=http://pap.onap-policy:9091/pap/
+PAP_PDPS=/opt/app/policy/servers/pap/bin/pdps
+PAP_URL=http://pap.{{.Values.nsPrefix}}:9091/pap/
PAP_INITIATE_PDP=true
PAP_HEARTBEAT_INTERVAL=10000
# PDP related properties
-PAP_PDP_URL=http://pdp.onap-policy:8081/pdp/
+PAP_PDP_URL=http://pdp.{{.Values.nsPrefix}}:8081/pdp/
PAP_PDP_HTTP_USER_ID=testpdp
PAP_PDP_HTTP_PASSWORD=alpha123
LOGPARSER_X_MS_MB=1024
LOGPARSER_X_MX_MB=1024
-SERVER=http://pap.onap-policy:9091/pap/
-LOGPATH=${{POLICY_HOME}}/servers/pap/logs/pap-rest.log
+SERVER=http://pap.{{.Values.nsPrefix}}:9091/pap/
+LOGPATH=/opt/app/policy/servers/pap/logs/pap-rest.log
PARSERLOGPATH=IntegrityMonitor.log
node_type=logparser
# pdp properties
-UEB_CLUSTER=dmaap.onap-message-router
+UEB_CLUSTER=dmaap.{{.Values.nsPrefix}}
-REST_PAP_URL=http://pap.onap-policy:9091/pap/
-REST_PDP_ID=http://pdp.onap-policy:8081/pdp/
-REST_PDP_CONFIG=${{POLICY_HOME}}/servers/pdp/bin/config
-REST_PDP_WEBAPPS=${{POLICY_HOME}}/servers/pdp/webapps
+REST_PAP_URL=http://pap.{{.Values.nsPrefix}}:9091/pap/
+REST_PDP_ID=http://pdp.{{.Values.nsPrefix}}:8081/pdp/
+REST_PDP_CONFIG=/opt/app/policy/servers/pdp/bin/config
+REST_PDP_WEBAPPS=/opt/app/policy/servers/pdp/webapps
REST_PDP_REGISTER=true
REST_PDP_REGISTER_SLEEP=15
REST_PDP_REGISTER_RETRIES=-1
POLICY_AAF_RESOURCE=
# Indeterminate resolution
-DECISION_INDETERMINATE_RESPONSE=PERMIT
\ No newline at end of file
+DECISION_INDETERMINATE_RESPONSE=PERMIT
LOGPARSER_X_MS_MB=1024
LOGPARSER_X_MX_MB=1024
-SERVER=http://pdp.onap-policy:8081/pdp/
-LOGPATH=${{POLICY_HOME}}/servers/pdp/logs/pdp-rest.log
+SERVER=http://pdp.{{.Values.nsPrefix}}:8081/pdp/
+LOGPATH=/opt/app/policy/servers/pdp/logs/pdp-rest.log
PARSERLOGPATH=IntegrityMonitor.log
node_type=logparser
sleep 2
-curl -v --silent -X POST --header 'Content-Type: multipart/form-data' --header 'Accept: text/plain' --header 'ClientAuth: cHl0aG9uOnRlc3Q=' --header 'Authorization: Basic dGVzdHBkcDphbHBoYTEyMw==' --header 'Environment: TEST' -F "file=@cl-amsterdam-template.drl" -F "importParametersJson={\"serviceName\":\"ClosedLoopControlName\",\"serviceType\":\"BRMSPARAM\"}" 'http://pdp.onap-policy:8081/pdp/api/policyEngineImport'
+curl -v --silent -X POST --header 'Content-Type: multipart/form-data' --header 'Accept: text/plain' --header 'ClientAuth: cHl0aG9uOnRlc3Q=' --header 'Authorization: Basic dGVzdHBkcDphbHBoYTEyMw==' --header 'Environment: TEST' -F "file=@cl-amsterdam-template.drl" -F "importParametersJson={\"serviceName\":\"ClosedLoopControlName\",\"serviceType\":\"BRMSPARAM\"}" 'http://pdp.{{.Values.nsPrefix}}:8081/pdp/api/policyEngineImport'
echo "PRELOAD_POLICIES is $PRELOAD_POLICIES"
"controlLoopYaml": "controlLoop%3A%0D%0A++version%3A+2.0.0%0D%0A++controlLoopName%3A+ControlLoop-vFirewall-d0a1dfc6-94f5-4fd4-a5b5-4630b438850a%0D%0A++trigger_policy%3A+unique-policy-id-1-modifyConfig%0D%0A++timeout%3A+1200%0D%0A++abatement%3A+false%0D%0A+%0D%0Apolicies%3A%0D%0A++-+id%3A+unique-policy-id-1-modifyConfig%0D%0A++++name%3A+modify+packet+gen+config%0D%0A++++description%3A%0D%0A++++actor%3A+APPC%0D%0A++++recipe%3A+ModifyConfig%0D%0A++++target%3A%0D%0A++++++%23+TBD+-+Cannot+be+known+until+instantiation+is+done%0D%0A++++++resourceID%3A+Eace933104d443b496b8.nodes.heat.vpg%0D%0A++++++type%3A+VNF%0D%0A++++retry%3A+0%0D%0A++++timeout%3A+300%0D%0A++++success%3A+final_success%0D%0A++++failure%3A+final_failure%0D%0A++++failure_timeout%3A+final_failure_timeout%0D%0A++++failure_retries%3A+final_failure_retries%0D%0A++++failure_exception%3A+final_failure_exception%0D%0A++++failure_guard%3A+final_failure_guard"
}
}
-}' 'http://pdp.onap-policy:8081/pdp/api/createPolicy'
+}' 'http://pdp.{{.Values.nsPrefix}}:8081/pdp/api/createPolicy'
sleep 2
"controlLoopYaml": "controlLoop%3A%0D%0A++version%3A+2.0.0%0D%0A++controlLoopName%3A+ControlLoop-vDNS-6f37f56d-a87d-4b85-b6a9-cc953cf779b3%0D%0A++trigger_policy%3A+unique-policy-id-1-scale-up%0D%0A++timeout%3A+1200%0D%0A++abatement%3A+false%0D%0Apolicies%3A%0D%0A++-+id%3A+unique-policy-id-1-scale-up%0D%0A++++name%3A+Create+a+new+VF+Module%0D%0A++++description%3A%0D%0A++++actor%3A+SO%0D%0A++++recipe%3A+VF+Module+Create%0D%0A++++target%3A%0D%0A++++++type%3A+VNF%0D%0A++++retry%3A+0%0D%0A++++timeout%3A+1200%0D%0A++++success%3A+final_success%0D%0A++++failure%3A+final_failure%0D%0A++++failure_timeout%3A+final_failure_timeout%0D%0A++++failure_retries%3A+final_failure_retries%0D%0A++++failure_exception%3A+final_failure_exception%0D%0A++++failure_guard%3A+final_failure_guard"
}
}
-}' 'http://pdp.onap-policy:8081/pdp/api/createPolicy'
+}' 'http://pdp.{{.Values.nsPrefix}}:8081/pdp/api/createPolicy'
sleep 2
"controlLoopYaml": "controlLoop%3A%0D%0A++version%3A+2.0.0%0D%0A++controlLoopName%3A+ControlLoop-VOLTE-2179b738-fd36-4843-a71a-a8c24c70c55b%0D%0A++trigger_policy%3A+unique-policy-id-1-restart%0D%0A++timeout%3A+3600%0D%0A++abatement%3A+false%0D%0A+%0D%0Apolicies%3A%0D%0A++-+id%3A+unique-policy-id-1-restart%0D%0A++++name%3A+Restart+the+VM%0D%0A++++description%3A%0D%0A++++actor%3A+VFC%0D%0A++++recipe%3A+Restart%0D%0A++++target%3A%0D%0A++++++type%3A+VM%0D%0A++++retry%3A+3%0D%0A++++timeout%3A+1200%0D%0A++++success%3A+final_success%0D%0A++++failure%3A+final_failure%0D%0A++++failure_timeout%3A+final_failure_timeout%0D%0A++++failure_retries%3A+final_failure_retries%0D%0A++++failure_exception%3A+final_failure_exception%0D%0A++++failure_guard%3A+final_failure_guard"
}
}
-}' 'http://pdp.onap-policy:8081/pdp/api/createPolicy'
+}' 'http://pdp.{{.Values.nsPrefix}}:8081/pdp/api/createPolicy'
sleep 2
"controlLoopYaml": "controlLoop%3A%0D%0A++version%3A+2.0.0%0D%0A++controlLoopName%3A+ControlLoop-vCPE-48f0c2c3-a172-4192-9ae3-052274181b6e%0D%0A++trigger_policy%3A+unique-policy-id-1-restart%0D%0A++timeout%3A+3600%0D%0A++abatement%3A+true%0D%0A+%0D%0Apolicies%3A%0D%0A++-+id%3A+unique-policy-id-1-restart%0D%0A++++name%3A+Restart+the+VM%0D%0A++++description%3A%0D%0A++++actor%3A+APPC%0D%0A++++recipe%3A+Restart%0D%0A++++target%3A%0D%0A++++++type%3A+VM%0D%0A++++retry%3A+3%0D%0A++++timeout%3A+1200%0D%0A++++success%3A+final_success%0D%0A++++failure%3A+final_failure%0D%0A++++failure_timeout%3A+final_failure_timeout%0D%0A++++failure_retries%3A+final_failure_retries%0D%0A++++failure_exception%3A+final_failure_exception%0D%0A++++failure_guard%3A+final_failure_guard"
}
}
-}' 'http://pdp.onap-policy:8081/pdp/api/createPolicy'
+}' 'http://pdp.{{.Values.nsPrefix}}:8081/pdp/api/createPolicy'
#########################################Create Micro Service Config policies##########################################
"policyConfigType": "MicroService",
"policyName": "com.MicroServicevFirewall",
"onapName": "DCAE"
-}' 'http://pdp.onap-policy:8081/pdp/api/createPolicy'
+}' 'http://pdp.{{.Values.nsPrefix}}:8081/pdp/api/createPolicy'
sleep 2
"policyConfigType": "MicroService",
"policyName": "com.MicroServicevDNS",
"onapName": "DCAE"
-}' 'http://pdp.onap-policy:8081/pdp/api/createPolicy'
+}' 'http://pdp.{{.Values.nsPrefix}}:8081/pdp/api/createPolicy'
sleep 2
"policyConfigType": "MicroService",
"policyName": "com.MicroServicevCPE",
"onapName": "DCAE"
-}' 'http://pdp.onap-policy:8081/pdp/api/createPolicy'
+}' 'http://pdp.{{.Values.nsPrefix}}:8081/pdp/api/createPolicy'
#########################################Creating Decision Guard policy#########################################
"guardActiveEnd": "00:00:00-05:00"
}
}
-}' 'http://pdp.onap-policy:8081/pdp/api/createPolicy'
+}' 'http://pdp.{{.Values.nsPrefix}}:8081/pdp/api/createPolicy'
#########################################Push Decision policy#########################################
"pdpGroup": "default",
"policyName": "com.AllPermitGuard",
"policyType": "DECISION"
-}' 'http://pdp.onap-policy:8081/pdp/api/pushPolicy'
+}' 'http://pdp.{{.Values.nsPrefix}}:8081/pdp/api/pushPolicy'
#########################################Pushing BRMS Param policies##########################################
"pdpGroup": "default",
"policyName": "com.BRMSParamvFirewall",
"policyType": "BRMS_Param"
-}' 'http://pdp.onap-policy:8081/pdp/api/pushPolicy'
+}' 'http://pdp.{{.Values.nsPrefix}}:8081/pdp/api/pushPolicy'
sleep 2
"pdpGroup": "default",
"policyName": "com.BRMSParamvDNS",
"policyType": "BRMS_Param"
-}' 'http://pdp.onap-policy:8081/pdp/api/pushPolicy'
+}' 'http://pdp.{{.Values.nsPrefix}}:8081/pdp/api/pushPolicy'
sleep 2
"pdpGroup": "default",
"policyName": "com.BRMSParamVOLTE",
"policyType": "BRMS_Param"
-}' 'http://pdp.onap-policy:8081/pdp/api/pushPolicy'
+}' 'http://pdp.{{.Values.nsPrefix}}:8081/pdp/api/pushPolicy'
sleep 2
"pdpGroup": "default",
"policyName": "com.BRMSParamvCPE",
"policyType": "BRMS_Param"
-}' 'http://pdp.onap-policy:8081/pdp/api/pushPolicy'
+}' 'http://pdp.{{.Values.nsPrefix}}:8081/pdp/api/pushPolicy'
#########################################Pushing MicroService Config policies##########################################
"pdpGroup": "default",
"policyName": "com.MicroServicevFirewall",
"policyType": "MicroService"
-}' 'http://pdp.onap-policy:8081/pdp/api/pushPolicy'
+}' 'http://pdp.{{.Values.nsPrefix}}:8081/pdp/api/pushPolicy'
sleep 10
"pdpGroup": "default",
"policyName": "com.MicroServicevDNS",
"policyType": "MicroService"
-}' 'http://pdp.onap-policy:8081/pdp/api/pushPolicy'
+}' 'http://pdp.{{.Values.nsPrefix}}:8081/pdp/api/pushPolicy'
sleep 10
"pdpGroup": "default",
"policyName": "com.MicroServicevCPE",
"policyType": "MicroService"
-}' 'http://pdp.onap-policy:8081/pdp/api/pushPolicy'
\ No newline at end of file
+}' 'http://pdp.{{.Values.nsPrefix}}:8081/pdp/api/pushPolicy'
kind: Service
metadata:
name: mariadb
- namespace: "{{ .Values.nsPrefix }}-policy"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: mariadb
spec:
labels:
app: nexus
name: nexus
- namespace: "{{ .Values.nsPrefix }}-policy"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
ports:
- name: nexus-port
labels:
app: drools
name: drools
- namespace: "{{ .Values.nsPrefix }}-policy"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
ports:
- name: "drools-port"
kind: Service
metadata:
name: pap
- namespace: "{{ .Values.nsPrefix }}-policy"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: pap
spec:
kind: Service
metadata:
name: pdp
- namespace: "{{ .Values.nsPrefix }}-policy"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: pdp
annotations:
kind: Service
metadata:
name: brmsgw
- namespace: "{{ .Values.nsPrefix }}-policy"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: brmsgw
spec:
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
- name: brmsgw
- namespace: "{{ .Values.nsPrefix }}-policy"
+ name: policy-brmsgw
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.brmsgwReplicas }}
selector:
metadata:
labels:
app: brmsgw
- name: brmsgw
+ name: policy-brmsgw
spec:
initContainers:
- command:
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
- name: drools
- namespace: "{{ .Values.nsPrefix }}-policy"
+ name: policy-drools
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.droolsReplicas }}
selector:
metadata:
labels:
app: drools
- name: drools
+ name: policy-drools
spec:
initContainers:
- command:
- mountPath: /etc/localtime
name: localtime
readOnly: true
- - mountPath: /tmp/policy-install/config
- name: drools
+ - mountPath: /tmp/policy-install/config/policy-keystore
+ name: drools-secret
+ subPath: policy-keystore
+ - mountPath: /tmp/policy-install/config/feature-healthcheck.conf
+ name: drools-secret
+ subPath: feature-healthcheck.conf
+ - mountPath: /tmp/policy-install/config/base.conf
+ name: drools-config
+ subPath: base.conf
+ - mountPath: /tmp/policy-install/config/policy-management.conf
+ name: drools-config
+ subPath: policy-management.conf
+ - mountPath: /tmp/policy-install/config/drools-tweaks.sh
+ name: drools-config
+ subPath: drools-tweaks.sh
- mountPath: /usr/share/maven/conf/settings.xml
name: drools-settingsxml
subPath: settings.xml
- name: drools-settingsxml
configMap:
name: policy-dep-drools-settings-configmap
- - name: drools
- secret:
- secretName: policy-dep-drools-secret
+ - name: drools-config
+ configMap:
+ name: policy-dep-drools-configmap
items:
- - key: policy-keystore
- path: policy-keystore
- mode: 0644
- - key: feature-healthcheck.conf
- path: feature-healthcheck.conf
- mode: 0644
- key: base.conf
path: base.conf
mode: 0755
- key: drools-tweaks.sh
path: drools-tweaks.sh
mode: 0755
+ - name: drools-secret
+ secret:
+ secretName: policy-dep-drools-secret
+ items:
+ - key: policy-keystore
+ path: policy-keystore
+ mode: 0644
+ - key: feature-healthcheck.conf
+ path: feature-healthcheck.conf
+ mode: 0644
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
#{{ end }}
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
- name: mariadb
- namespace: "{{ .Values.nsPrefix }}-policy"
+ name: policy-mariadb
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.mariadbReplicas }}
selector:
metadata:
labels:
app: mariadb
+ name: policy-mariadb
spec:
hostname: mariadb
containers:
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
- name: nexus
- namespace: "{{ .Values.nsPrefix }}-policy"
+ name: policy-nexus
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.nexusReplicas }}
selector:
metadata:
labels:
app: nexus
- name: nexus
+ name: policy-nexus
spec:
initContainers:
- command:
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
- name: pap
- namespace: "{{ .Values.nsPrefix }}-policy"
+ name: policy-pap
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.papReplicas }}
selector:
metadata:
labels:
app: pap
- name: pap
+ name: policy-pap
spec:
initContainers:
- command:
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
- name: pdp
- namespace: "{{ .Values.nsPrefix }}-policy"
+ name: policy-pdp
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.pdpReplicas }}
selector:
metadata:
labels:
app: pdp
- name: pdp
+ name: policy-pdp
spec:
initContainers:
- command:
kind: ConfigMap
metadata:
name: policy-dep-drools-settings-configmap
- namespace: {{ .Values.nsPrefix }}-policy
+ namespace: {{ .Values.nsPrefix }}
data:
-{{ (.Files.Glob "resources/config/drools/settings.xml").AsConfig | indent 2 }}
+{{ tpl (.Files.Glob "resources/config/drools/settings.xml").AsConfig . | indent 2 }}
#{{ end }}
---
#{{ if not .Values.disablePolicyPdp }}
kind: ConfigMap
metadata:
name: policy-dep-pe-configmap
- namespace: {{ .Values.nsPrefix }}-policy
+ namespace: {{ .Values.nsPrefix }}
data:
-{{ (.Files.Glob "resources/config/opt/policy/config/pe/*").AsConfig | indent 2 }}
+{{ tpl (.Files.Glob "resources/config/opt/policy/config/pe/*").AsConfig . | indent 2 }}
+---
+apiVersion: v1
+kind: ConfigMap
+metadata:
+ name: policy-dep-drools-configmap
+ namespace: {{ .Values.nsPrefix }}
+data:
+{{ tpl (.Files.Glob "resources/config/opt/policy/config/drools/*").AsConfig . | indent 2 }}
#{{ end }}
kind: Secret
metadata:
name: policy-dep-drools-secret
- namespace: {{ .Values.nsPrefix }}-policy
+ namespace: {{ .Values.nsPrefix }}
type: Opaque
data:
-{{ tpl (.Files.Glob "resources/config/opt/policy/config/drools/*").AsSecrets . | indent 2 }}
+{{ (.Files.Glob "resources/config/opt/policy/config/drools/keys/*").AsSecrets | indent 2 }}
#{{ end }}
kind: ConfigMap
metadata:
name: policy-pap-log-configmap
- namespace: {{ .Values.nsPrefix }}-policy
+ namespace: {{ .Values.nsPrefix }}
data:
-{{ (.Files.Glob "resources/config/log/xacml-pap-rest/*").AsConfig | indent 2 }}
+{{ tpl (.Files.Glob "resources/config/log/xacml-pap-rest/*").AsConfig . | indent 2 }}
---
apiVersion: v1
kind: ConfigMap
metadata:
name: policy-sdk-log-configmap
- namespace: {{ .Values.nsPrefix }}-policy
+ namespace: {{ .Values.nsPrefix }}
data:
-{{ (.Files.Glob "resources/config/log/ep_sdk_app/*").AsConfig | indent 2 }}
+{{ tpl (.Files.Glob "resources/config/log/ep_sdk_app/*").AsConfig . | indent 2 }}
---
apiVersion: v1
kind: ConfigMap
metadata:
name: policy-filebeat-configmap
- namespace: {{ .Values.nsPrefix }}-policy
+ namespace: {{ .Values.nsPrefix }}
data:
-{{ (.Files.Glob "resources/config/log/filebeat/*").AsConfig | indent 2 }}
+{{ tpl (.Files.Glob "resources/config/log/filebeat/*").AsConfig . | indent 2 }}
#{{ end }}
#{{ if not .Values.disablePolicyPdp }}
---
kind: ConfigMap
metadata:
name: policy-pdp-log-configmap
- namespace: {{ .Values.nsPrefix }}-policy
+ namespace: {{ .Values.nsPrefix }}
data:
-{{ (.Files.Glob "resources/config/log/xacml-pdp-rest/*").AsConfig | indent 2 }}
+{{ tpl (.Files.Glob "resources/config/log/xacml-pdp-rest/*").AsConfig . | indent 2 }}
#{{ end }}
#{{ if not .Values.disablePolicyDrools }}
---
kind: ConfigMap
metadata:
name: policy-drools-log-configmap
- namespace: {{ .Values.nsPrefix }}-policy
+ namespace: {{ .Values.nsPrefix }}
data:
-{{ (.Files.Glob "resources/config/log/drools/*").AsConfig | indent 2 }}
+{{ tpl (.Files.Glob "resources/config/log/drools/*").AsConfig . | indent 2 }}
#{{ end }}
#List of logstash server ip addresses with port number.
#But, in our case, this will be the loadbalancer IP address.
#For the below property to work the loadbalancer or logstash should expose 5044 port to listen the filebeat events or port in the property should be changed appropriately.
- hosts: ["logstash.onap-log:5044"]
+ hosts: ["logstash.{{.Values.nsPrefix}}:5044"]
#If enable will do load balancing among availabe Logstash, automatically.
loadbalance: true
ueb_app_secret = uCYgKjWKK5IxPGNNZzYSSWo9
ueb_app_mailbox_name = ECOMP-PORTAL-INBOX
-ueb_url_list = dmaap.onap-message-router
+ueb_url_list = dmaap.{{.Values.nsPrefix}}
ecomp_portal_inbox_name = ECOMP-PORTAL-INBOX
# Consumer group name for UEB topic.
#mysql
db.driver = org.mariadb.jdbc.Driver
-db.connectionURL = jdbc:mariadb:failover://portaldb.onap-portal:3306/portal
+db.connectionURL = jdbc:mariadb:failover://portaldb.{{.Values.nsPrefix}}:3306/portal
db.userName =root
db.password =Aa123456
db.hib.dialect = org.hibernate.dialect.MySQLDialect
#Microservices Related Properties for Portal
microservices.widget.protocol = http
-microservices.widget.hostname = portalwidgets.onap-portal
+microservices.widget.hostname = portalwidgets.{{.Values.nsPrefix}}
microservices.widget.username = widget_user
microservices.widget.password = KpuqIB08YHg+btG+pjX+sA==
#This property won't be needed after consul is functional on VMs -
ueb_listeners_enable = false
# UEB Configuration
-ueb_url_list = dmaap.onap-message-router
+ueb_url_list = dmaap.{{.Values.nsPrefix}}
# ECOMP Portal listens on this UEB topic
ecomp_portal_inbox_name = ECOMP-PORTAL-INBOX
# Replace these 3 default values with the ones for your specific App,
##########################################################################
db.driver = org.mariadb.jdbc.Driver
-db.connectionURL = jdbc:mariadb://portaldb.onap-portal:3306/ecomp_sdk
+db.connectionURL = jdbc:mariadb://portaldb.{{.Values.nsPrefix}}:3306/ecomp_sdk
db.userName = root
db.password = Aa123456
db.min_pool_size = 5
guard_notebook_url=
#authenticate user server
-authenticate_user_server=http://portalapps.onap-portal:8383/openid-connect-server-webapp/allUsers
+authenticate_user_server=http://portalapps.{{.Values.nsPrefix}}:8383/openid-connect-server-webapp/allUsers
microservice.widget.location=/tmp
## App DB Properties
-spring.datasource.url=jdbc:mysql://portaldb.onap-portal:3306/portal
+spring.datasource.url=jdbc:mysql://portaldb.{{.Values.nsPrefix}}:3306/portal
spring.datasource.username=root
spring.datasource.password=Aa123456
spring.jpa.properties.hibernate.dialect=org.hibernate.dialect.MySQLDialect
kind: Service
metadata:
name: portaldb
- namespace: "{{ .Values.nsPrefix }}-portal"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: portaldb
spec:
labels:
app: portalapps
name: portalapps
- namespace: "{{ .Values.nsPrefix }}-portal"
+ namespace: "{{ .Values.nsPrefix }}"
annotations:
msb.onap.org/service-info: '[
{
kind: Service
metadata:
labels:
- app: vnc-portal
- name: vnc-portal
- namespace: "{{ .Values.nsPrefix }}-portal"
+ app: portal-vnc
+ name: portal-vnc
+ namespace: "{{ .Values.nsPrefix }}"
spec:
ports:
- name: tcp-1
targetPort: 5900
nodePort: {{ .Values.nodePortPrefix }}12
selector:
- app: vnc-portal
+ app: portal-vnc
type: NodePort
#{{ end }}
#{{ if not .Values.disablePortalPortalwidgets }}
kind: Service
metadata:
name: portalwidgets
- namespace: "{{ .Values.nsPrefix }}-portal"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: portalwidgets
spec:
selector:
app: portalwidgets
clusterIP: None
-#{{ end }}
\ No newline at end of file
+#{{ end }}
kind: ConfigMap
metadata:
name: portal-onap-portal-sdk-configmap
- namespace: {{ .Values.nsPrefix }}-portal
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/portal-fe/webapps/etc/ONAPPORTALSDK/*").AsConfig . | indent 2 }}
---
kind: ConfigMap
metadata:
name: portal-onap-portal-configmap
- namespace: {{ .Values.nsPrefix }}-portal
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/portal-fe/webapps/etc/ONAPPORTAL/*").AsConfig . | indent 2 }}
---
kind: ConfigMap
metadata:
name: portal-mariadb-configmap
- namespace: {{ .Values.nsPrefix }}-portal
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/mariadb/oom_updates.sql").AsConfig . | indent 2 }}
#{{ end }}
kind: ConfigMap
metadata:
name: portal-onapwidgetms-configmap
- namespace: {{ .Values.nsPrefix }}-portal
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/portal-fe/webapps/etc/ONAPWIDGETMS/application.properties").AsConfig . | indent 2 }}
#{{ end }}
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
- name: portalapps
- namespace: "{{ .Values.nsPrefix }}-portal"
+ name: portal-apps
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.portalAppsReplicas }}
selector:
metadata:
labels:
app: portalapps
- name: portalapps
+ name: portal-apps
spec:
initContainers:
- command:
kind: ConfigMap
metadata:
name: portal-filebeat-configmap
- namespace: {{ .Values.nsPrefix }}-portal
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/log/filebeat/filebeat.yml").AsConfig . | indent 2 }}
---
kind: ConfigMap
metadata:
name: portal-onapportal-log-configmap
- namespace: {{ .Values.nsPrefix }}-portal
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/log/portal/onapportal/logback.xml").AsConfig . | indent 2 }}
---
kind: ConfigMap
metadata:
name: portal-onapportalsdk-log-configmap
- namespace: {{ .Values.nsPrefix }}-portal
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/log/portal/onapportalsdk/logback.xml").AsConfig . | indent 2 }}
#{{ end }}
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
- name: portaldb
- namespace: "{{ .Values.nsPrefix }}-portal"
+ name: portal-db
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.portalDbReplicas }}
selector:
metadata:
labels:
app: portaldb
- name: portaldb
+ name: portal-db
spec:
containers:
- image: {{ .Values.image.portaldb}}
name: "portaldb"
env:
- name: MYSQL_HOST
- value: "portaldb.{{ .Values.nsPrefix }}-portal"
+ value: "portaldb.{{ .Values.nsPrefix }}"
- name: MYSQL_ROOT_PASSWORD
value: Aa123456
volumeMounts:
apiVersion: v1
kind: PersistentVolume
metadata:
- name: "{{ .Values.nsPrefix }}-portal-db"
- namespace: "{{ .Values.nsPrefix }}-portal"
+ name: "{{ .Values.nsPrefix }}-portal"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
- name: "{{ .Values.nsPrefix }}-portal-db"
+ name: "{{ .Values.nsPrefix }}-portal"
spec:
capacity:
storage: 2Gi
apiVersion: v1
metadata:
name: portal-db
- namespace: "{{ .Values.nsPrefix }}-portal"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
accessModes:
- ReadWriteMany
storage: 2Gi
selector:
matchLabels:
- name: "{{ .Values.nsPrefix }}-portal-db"
+ name: "{{ .Values.nsPrefix }}-portal"
#{{ end }}
#{{ if not .Values.disablePortalVncPortal }}
kind: ConfigMap
metadata:
- name: vnc-profiles-ini
- namespace: "{{ .Values.nsPrefix }}-portal"
+ name: portal-vnc-profiles-ini
+ namespace: "{{ .Values.nsPrefix }}"
data:
{{ (.Files.Glob "resources/vnc/init_profile/profiles.ini").AsConfig | indent 2 }}
---
apiVersion: v1
kind: ConfigMap
metadata:
- name: vnc-update-hosts
- namespace: "{{ .Values.nsPrefix }}-portal"
+ name: portal-vnc-update-hosts
+ namespace: "{{ .Values.nsPrefix }}"
data:
{{ (.Files.Glob "resources/scripts/update_hosts.sh").AsConfig | indent 2 }}
---
kind: Deployment
metadata:
labels:
- app: vnc-portal
- name: vnc-portal
- namespace: "{{ .Values.nsPrefix }}-portal"
+ app: portal-vnc
+ name: portal-vnc
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.vncPortalReplicas }}
selector:
matchLabels:
- app: vnc-portal
+ app: portal-vnc
template:
metadata:
labels:
- app: vnc-portal
- name: vnc-portal
+ app: portal-vnc
+ name: portal-vnc
spec:
initContainers:
- command:
fieldPath: metadata.namespace
image: {{ .Values.image.readiness }}
imagePullPolicy: {{ .Values.pullPolicy }}
- name: vnc-portal-readiness
+ name: portal-vnc-readiness
- command:
- /root/ready.py
args:
- pap
env:
- name: NAMESPACE
- value: {{ .Values.nsPrefix }}-policy
+ value: {{ .Values.nsPrefix }}
image: {{ .Values.image.readiness }}
imagePullPolicy: {{ .Values.pullPolicy }}
name: vnc-pap-readiness
- sdc-fe
env:
- name: NAMESPACE
- value: {{ .Values.nsPrefix }}-sdc
+ value: {{ .Values.nsPrefix }}
image: {{ .Values.image.readiness }}
imagePullPolicy: {{ .Values.pullPolicy }}
name: vnc-sdc-readiness
- vid-server
env:
- name: NAMESPACE
- value: {{ .Values.nsPrefix }}-vid
+ value: {{ .Values.nsPrefix }}
image: {{ .Values.image.readiness }}
imagePullPolicy: {{ .Values.pullPolicy }}
name: vnc-vid-readiness
lifecycle:
postStart:
exec:
- command: ["/bin/sh", "-c", "mkdir -p /root/.mozilla/firefox/onap.default; cp /root/.init_profile/profiles.ini /root/.mozilla/firefox/; echo 'user_pref(\"browser.tabs.remote.autostart.2\", false);' > /root/.mozilla/firefox/onap.default/prefs.js; (while true; do /tmp/update_hosts.sh sdc-be.{{ .Values.nsPrefix }}-sdc sdc.api.be.simpledemo.onap.org; /tmp/update_hosts.sh portalapps.{{ .Values.nsPrefix }}-portal portal.api.simpledemo.onap.org; /tmp/update_hosts.sh pap.{{ .Values.nsPrefix }}-policy policy.api.simpledemo.onap.org; /tmp/update_hosts.sh sdc-fe.{{ .Values.nsPrefix }}-sdc sdc.api.simpledemo.onap.org; /tmp/update_hosts.sh vid-server.{{ .Values.nsPrefix }}-vid vid.api.simpledemo.onap.org; /tmp/update_hosts.sh sparky-be.{{ .Values.nsPrefix }}-aai aai.api.simpledemo.onap.org; /tmp/update_hosts.sh cli.{{ .Values.nsPrefix }}-cli cli.api.simpledemo.onap.org; sleep 10; done) > update_hosts.log 2>&1 &"]
+ command: ["/bin/sh", "-c", "mkdir -p /root/.mozilla/firefox/onap.default; cp /root/.init_profile/profiles.ini /root/.mozilla/firefox/; echo 'user_pref(\"browser.tabs.remote.autostart.2\", false);' > /root/.mozilla/firefox/onap.default/prefs.js; (while true; do /tmp/update_hosts.sh sdc-be.{{ .Values.nsPrefix }} sdc.api.be.simpledemo.onap.org; /tmp/update_hosts.sh portalapps.{{ .Values.nsPrefix }} portal.api.simpledemo.onap.org; /tmp/update_hosts.sh pap.{{ .Values.nsPrefix }} policy.api.simpledemo.onap.org; /tmp/update_hosts.sh sdc-fe.{{ .Values.nsPrefix }} sdc.api.simpledemo.onap.org; /tmp/update_hosts.sh vid-server.{{ .Values.nsPrefix }} vid.api.simpledemo.onap.org; /tmp/update_hosts.sh sparky-be.{{ .Values.nsPrefix }} aai.api.simpledemo.onap.org; /tmp/update_hosts.sh cli.{{ .Values.nsPrefix }} cli.api.simpledemo.onap.org; sleep 10; done) > update_hosts.log 2>&1 &"]
env:
- name: VNC_PASSWORD
value: password
- name: vnc-portal
+ name: portal-vnc
volumeMounts:
- mountPath: /etc/localtime
name: localtime
path: /etc/localtime
- name: vnc-profiles-ini
configMap:
- name: vnc-profiles-ini
+ name: portal-vnc-profiles-ini
- name: vnc-update-hosts
configMap:
- name: vnc-update-hosts
+ name: portal-vnc-update-hosts
defaultMode: 0755
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
- name: portalwidgets
- namespace: "{{ .Values.nsPrefix }}-portal"
+ name: portal-widgets
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.portalWidgetsReplicas }}
selector:
metadata:
labels:
app: portalwidgets
- name: portalwidgets
+ name: portal-widgets
spec:
initContainers:
- command:
- hbase
command:
- /root/ready.py
- image: oomk8s/readiness-check:1.0.0
+ image: oomk8s/readiness-check:1.1.0
imagePullPolicy: Always
name: healthy
env:
# Patterns to ignore when building packages.
# This supports shell glob matching, relative path matching, and
# negation (prefixed with !). Only one pattern per line.
-
-# k8s scripts
-ete-k8s.sh
-demo-k8s.sh
+.DS_Store
+# Common VCS dirs
+.git/
+.gitignore
+.bzr/
+.bzrignore
+.hg/
+.hgignore
+.svn/
+# Common backup files
+*.swp
+*.bak
+*.tmp
+*~
+# Various IDEs
+.project
+.idea/
+*.tmproj
apiVersion: v1
-description: A Helm chart for Kubernetes
+description: A helm Chart for kubernetes-ONAP Robot
name: robot
-version: 0.1.0
+version: 2.0.0
+++ /dev/null
-apiVersion: v1
-kind: Service
-metadata:
- labels:
- app: robot
- name: robot
-spec:
- ports:
- - port: 88
- nodePort: {{ .Values.nodePortPrefix }}09
- selector:
- app: robot
- type: NodePort
-#!/bin/bash
+#!/bin/bash -x
#
# Execute tags built to support the hands on demo,
#
function usage
{
- echo "Usage: demo.sh <command> [<parameters>]"
+ echo "Usage: demo.sh namespace <command> [<parameters>]"
echo " "
- echo " demo.sh init"
+ echo " demo.sh <namespace> init"
echo " - Execute both init_customer + distribute"
echo " "
- echo " demo.sh init_customer"
+ echo " demo.sh <namespace> init_customer"
echo " - Create demo customer (Demonstration) and services, etc."
echo " "
- echo " demo.sh distribute [<prefix>]"
+ echo " demo.sh <namespace> distribute [<prefix>]"
echo " - Distribute demo models (demoVFW and demoVLB)"
echo " "
- echo " demo.sh preload <vnf_name> <module_name>"
+ echo " demo.sh <namespace> preload <vnf_name> <module_name>"
echo " - Preload data for VNF for the <module_name>"
echo " "
- echo " demo.sh appc <module_name>"
+ echo " demo.sh <namespace> appc <module_name>"
echo " - provide APPC with vFW module mount point for closed loop"
echo " "
- echo " demo.sh init_robot"
+ echo " demo.sh <namespace> init_robot"
echo " - Initialize robot after all ONAP VMs have started"
echo " "
- echo " demo.sh instantiateVFW"
+ echo " demo.sh <namespace> instantiateVFW"
echo " - Instantiate vFW module for the a demo customer (DemoCust<uuid>)"
echo " "
- echo " demo.sh deleteVNF <module_name from instantiateVFW>"
+ echo " demo.sh <namespace> deleteVNF <module_name from instantiateVFW>"
echo " - Delete the module created by instantiateVFW"
echo " "
- echo " demo.sh heatbridge <stack_name> <service_instance_id> <service>"
+ echo " demo.sh <namespace> heatbridge <stack_name> <service_instance_id> <service>"
echo " - Run heatbridge against the stack for the given service instance and service"
}
# Set the defaults
-if [ $# -eq 0 ];then
+if [ $# -le 1 ];then
usage
exit
fi
+
+NAMESPACE=$1
+shift
+
##
## if more than 1 tag is supplied, the must be provided with -i or -e
##
ETEHOME=/var/opt/OpenECOMP_ETE
VARIABLEFILES="-V /share/config/vm_properties.py -V /share/config/integration_robot_properties.py -V /share/config/integration_preload_parameters.py"
-POD=$(kubectl --namespace onap-robot get pods | sed 's/ .*//'| grep robot)
-kubectl --namespace onap-robot exec ${POD} -- ${ETEHOME}/runTags.sh ${VARIABLEFILES} ${VARIABLES} -d /share/logs/demo/${TAG} -i ${TAG} --display 89 2> ${TAG}.out
+POD=$(kubectl --namespace $NAMESPACE get pods | sed 's/ .*//'| grep robot)
+kubectl --namespace $NAMESPACE exec ${POD} -- ${ETEHOME}/runTags.sh ${VARIABLEFILES} ${VARIABLES} -d /share/logs/demo/${TAG} -i ${TAG} --display 89 2> ${TAG}.out
# Please clean up logs when you are done...
# Note: Do not run multiple concurrent ete.sh as the --display is not parameterized and tests will collide
#
-if [ "$1" == "" ];then
- echo "Usage: ete.sh [ health | ete | closedloop | instantiate | distribute ]"
+if [ "$1" == "" ] || [ "$2" == "" ]; then
+ echo "Usage: ete.sh [namespace] [ health | ete | closedloop | instantiate | distribute ]"
exit
fi
-export TAGS="-i $1"
+export NAMESPACE="$1"
+export TAGS="-i $2"
export ETEHOME=/var/opt/OpenECOMP_ETE
export OUTPUT_FOLDER=ETE_$$
VARIABLEFILES="-V /share/config/vm_properties.py -V /share/config/integration_robot_properties.py -V /share/config/integration_preload_parameters.py"
VARIABLES="-v GLOBAL_BUILD_NUMBER:$$"
-#docker exec openecompete_container ${ETEHOME}/runTags.sh ${VARIABLEFILES} ${VARIABLES} -d /share/logs/${OUTPUT_FOLDER} ${TAGS} --display 88
-
-POD=$(kubectl --namespace onap-robot get pods | sed 's/ .*//'| grep robot)
-kubectl --namespace onap-robot exec ${POD} -- ${ETEHOME}/runTags.sh ${VARIABLEFILES} ${VARIABLES} -d /share/logs/${OUTPUT_FOLDER} ${TAGS} --display 88
+POD=$(kubectl --namespace $NAMESPACE get pods | sed 's/ .*//'| grep robot)
+kubectl --namespace $NAMESPACE exec ${POD} -- ${ETEHOME}/runTags.sh ${VARIABLEFILES} ${VARIABLES} -d /share/logs/${OUTPUT_FOLDER} ${TAGS} --display 88
--- /dev/null
+dependencies:
+ - name: common
+ version: ~2.0.0
+ # local reference to common chart, as it is
+ # a part of this chart's package and will not
+ # be published independently to a repo (at this point)
+ repository: '@local'
\ No newline at end of file
"pub_key" : "ssh-rsa AAAAB3NzaC1yc2EAAAABJQAAAQEAqqnA9BAiMLtjOPSYBfhzLu4CiBolWoskDg4KVwhTJVTTeB6CqrQNcadlGXxOHhCYuNCKkUmIVF4WTOisVOJ75Z1c4OMoZLL85xVPKSIeH63kgVugwgPYQu55NbbWX+rsbUha3LnElDhNviMM3iTPbD5nnhKixNERAJMTLKXvZZZGqxW94bREknYPQTT2qrk3YRqwldncopp6Nkgv3AnSJz2gc9tjxnWF0poTQnQm/3D6hiJICrzKfAV0EaPN0KdtYjPhKrYoy6Qb/tKOVaaqsvwfKBJGrT9LfcA7D7M/yj292RT1XN63hI84WC383LsaPJ6eWdDTE6zUP1eGTWCoOw== rsa-key-20161026",
"repo_url_blob" : "https://nexus.onap.org/content/repositories/raw",
"repo_url_artifacts" : "https://nexus.onap.org/content/groups/staging",
- "demo_artifacts_version" : "DEMO_ARTIFACTS_VERSION_HERE",
- "onap_private_net_id" : "OPENSTACK_NETWORK_ID_WITH_ONAP_ROUTE_HERE",
- "onap_private_subnet_id" : "OPENSTACK_SUBNET_ID_WITH_ONAP_ROUTE_HERE",
- "onap_private_net_cidr" : "NETWORK_CIDR_WITH_ONAP_ROUTE_HERE",
+ "demo_artifacts_version" : "{{ .Values.demoArtifactsVersion }}",
+ "onap_private_net_id" : "{{ .Values.openStackPrivateNetId }}",
+ "onap_private_subnet_id" : "{{ .Values.openStackPrivateSubnetId }}",
+ "onap_private_net_cidr" : "{{ .Values.openStackPrivateNetCidr }}",
"dcae_collector_ip" : "10.0.4.102",
"dcae_collector_port" : "8080",
- "public_net_id" : "OPENSTACK_PUBLIC_NET_ID_HERE",
+ "public_net_id" : "{{ .Values.openStackPublicNetId }}",
"cloud_env" : "${cloud_env}",
- "install_script_version" : "${install_script_version}",
+ "install_script_version" : "${install_script_version}",
###
# vims_preload same for every instantiation
###
"ralf_image_name" : "${vm_image_name}",
"ellis_image_name" : "${vm_image_name}",
"dns_image_name" : "${vm_image_name}",
- "bono_flavor_name" : "${vm_flavor_name}",
- "sprout_flavor_name" : "${vm_flavor_name}",
- "homer_flavor_name" : "${vm_flavor_name}",
- "homestead_flavor_name" : "${vm_flavor_name}",
- "ralf_flavor_name" : "${vm_flavor_name}",
- "ellis_flavor_name" : "${vm_flavor_name}",
- "dns_flavor_name" : "${vm_flavor_name}",
- "repo_url" : "http://repo.cw-ngv.com/stable",
- "zone" : "me.cw-ngv.com",
- "dn_range_start" : "2425550000",
- "dn_range_length" : "10000",
- "dnssec_key" : "9FPdYTWhk5+LbhrqtTPQKw==",
+ "bono_flavor_name" : "${vm_flavor_name}",
+ "sprout_flavor_name" : "${vm_flavor_name}",
+ "homer_flavor_name" : "${vm_flavor_name}",
+ "homestead_flavor_name" : "${vm_flavor_name}",
+ "ralf_flavor_name" : "${vm_flavor_name}",
+ "ellis_flavor_name" : "${vm_flavor_name}",
+ "dns_flavor_name" : "${vm_flavor_name}",
+ "repo_url" : "http://repo.cw-ngv.com/stable",
+ "zone" : "me.cw-ngv.com",
+ "dn_range_start" : "2425550000",
+ "dn_range_length" : "10000",
+ "dnssec_key" : "9FPdYTWhk5+LbhrqtTPQKw==",
###
# vlb_preload same for every instantiation
###
- "vlb_image_name" : "${vm_image_name}",
- "vlb_flavor_name" : "${vm_flavor_name}",
+ "vlb_image_name" : "${vm_image_name}",
+ "vlb_flavor_name" : "${vm_flavor_name}",
###
# vlb_preload same for every instantiation
###
- "vfw_image_name" : "${vm_image_name}",
- "vfw_flavor_name" : "${vm_flavor_name}",
+ "vfw_image_name" : "${vm_image_name}",
+ "vfw_flavor_name" : "${vm_flavor_name}",
###
},
"protected_private_net_cidr" : "192.168.20.0/24",
"vfw_private_ip_0" : "192.168.10.100",
"vfw_private_ip_1" : "192.168.20.100",
- "vfw_private_ip_2" : "OPENSTACK_OAM_NETWORK_CIDR_PREFIX_HERE.1",
+ "vfw_private_ip_2" : "{{ .Values.openStackOamNetworkCidrPrefix }}.1",
"vpg_private_ip_0" : "192.168.10.200",
- "vpg_private_ip_1" : "OPENSTACK_OAM_NETWORK_CIDR_PREFIX_HERE.2",
+ "vpg_private_ip_1" : "{{ .Values.openStackOamNetworkCidrPrefix }}.2",
"vsn_private_ip_0" : "192.168.20.250",
- "vsn_private_ip_1" : "OPENSTACK_OAM_NETWORK_CIDR_PREFIX_HERE.3",
+ "vsn_private_ip_1" : "{{ .Values.openStackOamNetworkCidrPrefix }}.3",
'vfw_name_0':'vofwl01fwl${hostid}',
'vpg_name_0':'vofwl01pgn${hostid}',
'vsn_name_0':'vofwl01snk${hostid}'
"vlb_private_net_id" : "volb01_private${hostid}",
"vlb_private_net_cidr" : "192.168.30.0/24",
"vlb_private_ip_0" : "192.168.30.100",
- "vlb_private_ip_1" : "OPENSTACK_OAM_NETWORK_CIDR_PREFIX_HERE.4",
+ "vlb_private_ip_1" : "{{ .Values.openStackOamNetworkCidrPrefix }}.4",
"vdns_private_ip_0" : "192.168.30.110",
- "vdns_private_ip_1" : "OPENSTACK_OAM_NETWORK_CIDR_PREFIX_HERE.5",
+ "vdns_private_ip_1" : "{{ .Values.openStackOamNetworkCidrPrefix }}.5",
'vlb_name_0':'vovlblb${hostid}',
'vdns_name_0':'vovlbdns${hostid}',
- "vlb_private_net_cidr" : "192.168.10.0/24",
- "pktgen_private_net_cidr" : "192.168.9.0/24"
-
+ "vlb_private_net_cidr" : "192.168.10.0/24",
+ "pktgen_private_net_cidr" : "192.168.9.0/24"
+
},
"dnsscaling_preload.template" : {
"vlb_private_net_id" : "volb01_private${hostid}",
"vlb_private_ip_0" : "192.168.30.100",
- "vlb_private_ip_1" : "OPENSTACK_OAM_NETWORK_CIDR_PREFIX_HERE.4",
+ "vlb_private_ip_1" : "{{ .Values.openStackOamNetworkCidrPrefix }}.4",
"vdns_private_ip_0" : "192.168.30.222",
- "vdns_private_ip_1" : "OPENSTACK_OAM_NETWORK_CIDR_PREFIX_HERE.6",
+ "vdns_private_ip_1" : "{{ .Values.openStackOamNetworkCidrPrefix }}.6",
'scaling_vdns_name_0':'vovlbscaling${hostid}',
- "vlb_private_net_cidr" : "192.168.10.0/24"
+ "vlb_private_net_cidr" : "192.168.10.0/24"
},
"vvg_preload.template" : {
}
},
# heat template parameter values for heat template instances created during Closed-Loop test cases
"Closed-Loop" : {
- "vfw_preload.template": {
+ "vfw_preload.template": {
"unprotected_private_net_id" : "clfwl01_unprotected${hostid}",
"unprotected_private_net_cidr" : "192.168.110.0/24",
"protected_private_net_id" : "clfwl01_protected${hostid}",
"protected_private_net_cidr" : "192.168.120.0/24",
"vfw_private_ip_0" : "192.168.110.100",
"vfw_private_ip_1" : "192.168.120.100",
- "vfw_private_ip_2" : "OPENSTACK_OAM_NETWORK_CIDR_PREFIX_HERE.11",
+ "vfw_private_ip_2" : "{{ .Values.openStackOamNetworkCidrPrefix }}.11",
"vpg_private_ip_0" : "192.168.110.200",
- "vpg_private_ip_1" : "OPENSTACK_OAM_NETWORK_CIDR_PREFIX_HERE.12",
+ "vpg_private_ip_1" : "{{ .Values.openStackOamNetworkCidrPrefix }}.12",
"vsn_private_ip_0" : "192.168.120.250",
- "vsn_private_ip_1" : "OPENSTACK_OAM_NETWORK_CIDR_PREFIX_HERE.13",
+ "vsn_private_ip_1" : "{{ .Values.openStackOamNetworkCidrPrefix }}.13",
'vfw_name_0':'clfwl01fwl${hostid}',
'vpg_name_0':'clfwl01pgn${hostid}',
'vsn_name_0':'clfwl01snk${hostid}'
"vlb_private_net_id" : "cllb01_private${hostid}",
"vlb_private_net_cidr" : "192.168.130.0/24",
"vlb_private_ip_0" : "192.168.130.100",
- "vlb_private_ip_1" : "OPENSTACK_OAM_NETWORK_CIDR_PREFIX_HERE.14",
+ "vlb_private_ip_1" : "{{ .Values.openStackOamNetworkCidrPrefix }}.14",
"vdns_private_ip_0" : "192.168.130.110",
- "vdns_private_ip_1" : "OPENSTACK_OAM_NETWORK_CIDR_PREFIX_HERE.15",
+ "vdns_private_ip_1" : "{{ .Values.openStackOamNetworkCidrPrefix }}.15",
'vlb_name_0':'clvlblb${hostid}',
'vdns_name_0':'clvlbdns${hostid}',
- "vlb_private_net_cidr" : "192.168.10.0/24",
- "pktgen_private_net_cidr" : "192.168.9.0/24"
+ "vlb_private_net_cidr" : "192.168.10.0/24",
+ "pktgen_private_net_cidr" : "192.168.9.0/24"
},
"dnsscaling_preload.template" : {
"vlb_private_net_id" : "cllb01_private${hostid}",
"vlb_private_ip_0" : "192.168.130.100",
- "vlb_private_ip_1" : "OPENSTACK_OAM_NETWORK_CIDR_PREFIX_HERE.14",
+ "vlb_private_ip_1" : "{{ .Values.openStackOamNetworkCidrPrefix }}.14",
"vdns_private_ip_0" : "192.168.130.222",
- "vdns_private_ip_1" : "OPENSTACK_OAM_NETWORK_CIDR_PREFIX_HERE.16",
+ "vdns_private_ip_1" : "{{ .Values.openStackOamNetworkCidrPrefix }}.16",
'scaling_vdns_name_0':'clvlbscaling${hostid}',
- "vlb_private_net_cidr" : "192.168.10.0/24"
+ "vlb_private_net_cidr" : "192.168.10.0/24"
},
"vvg_preload.template" : {
}
"protected_private_net_cidr" : "192.168.120.0/24",
"vfw_private_ip_0" : "192.168.110.100",
"vfw_private_ip_1" : "192.168.120.100",
- "vfw_private_ip_2" : "OPENSTACK_OAM_NETWORK_CIDR_PREFIX_HERE.11",
+ "vfw_private_ip_2" : "{{ .Values.openStackOamNetworkCidrPrefix }}.11",
"vpg_private_ip_0" : "192.168.110.200",
- "vpg_private_ip_1" : "OPENSTACK_OAM_NETWORK_CIDR_PREFIX_HERE.12",
+ "vpg_private_ip_1" : "{{ .Values.openStackOamNetworkCidrPrefix }}.12",
"vsn_private_ip_0" : "192.168.120.250",
- "vsn_private_ip_1" : "OPENSTACK_OAM_NETWORK_CIDR_PREFIX_HERE.13",
+ "vsn_private_ip_1" : "{{ .Values.openStackOamNetworkCidrPrefix }}.13",
'vfw_name_0':'demofwl01fwl',
'vpg_name_0':'demofwl01pgn',
'vsn_name_0':'demofwl01snk'
"vlb_private_net_id" : "demolb_private",
"vlb_private_net_cidr" : "192.168.130.0/24",
"vlb_private_ip_0" : "192.168.130.100",
- "vlb_private_ip_1" : "OPENSTACK_OAM_NETWORK_CIDR_PREFIX_HERE.14",
+ "vlb_private_ip_1" : "{{ .Values.openStackOamNetworkCidrPrefix }}.14",
"vdns_private_ip_0" : "192.168.130.110",
- "vdns_private_ip_1" : "OPENSTACK_OAM_NETWORK_CIDR_PREFIX_HERE.15",
+ "vdns_private_ip_1" : "{{ .Values.openStackOamNetworkCidrPrefix }}.15",
'vlb_name_0':'demovlblb',
'vdns_name_0':'demovlbdns',
- "vlb_private_net_cidr" : "192.168.10.0/24",
- "pktgen_private_net_cidr" : "192.168.9.0/24"
+ "vlb_private_net_cidr" : "192.168.10.0/24",
+ "pktgen_private_net_cidr" : "192.168.9.0/24"
},
"dnsscaling_preload.template" : {
"vlb_private_net_id" : "demolb_private",
"vlb_private_ip_0" : "192.168.130.100",
- "vlb_private_ip_1" : "OPENSTACK_OAM_NETWORK_CIDR_PREFIX_HERE.14",
+ "vlb_private_ip_1" : "{{ .Values.openStackOamNetworkCidrPrefix }}.14",
"vdns_private_ip_0" : "192.168.130.222",
- "vdns_private_ip_1" : "OPENSTACK_OAM_NETWORK_CIDR_PREFIX_HERE.16",
+ "vdns_private_ip_1" : "{{ .Values.openStackOamNetworkCidrPrefix }}.16",
'scaling_vdns_name_0':'demovlbscaling',
- "vlb_private_net_cidr" : "192.168.10.0/24"
+ "vlb_private_net_cidr" : "192.168.10.0/24"
},
"vvg_preload.template" : {
}
}
}
-
GLOBAL_PROXY_WARNING_TITLE = ""
GLOBAL_PROXY_WARNING_CONTINUE_XPATH = ""
# settings for vm to attach vvg too
-GLOBAL_VVGSERVER_IMAGE = "UBUNTU_14_IMAGE_NAME_HERE"
-GLOBAL_VVGSERVER_FLAVOR = "OPENSTACK_FLAVOUR_MEDIUM_HERE"
+GLOBAL_VVGSERVER_IMAGE = "{{ .Values.ubuntuImage }}"
+GLOBAL_VVGSERVER_FLAVOR = "{{ .Values.openStackFlavourMedium }}"
# dns info
GLOBAL_DNS_TRAFFIC_DURATION = "600"
GLOBAL_HEAT_TEMPLATES_FOLDER = "/share/heat"
--- /dev/null
+# File generated from /opt/config
+#
+GLOBAL_INJECTED_AAI1_IP_ADDR = "aai-service.{{ .Values.nsPrefix }}"
+GLOBAL_INJECTED_AAI2_IP_ADDR = "N/A"
+GLOBAL_INJECTED_APPC_IP_ADDR = "appc-sdnhost.{{ .Values.nsPrefix }}"
+GLOBAL_INJECTED_ARTIFACTS_VERSION = "1.1.0-SNAPSHOT"
+GLOBAL_INJECTED_CLAMP_IP_ADDR = "clamp.{{ .Values.nsPrefix }}"
+GLOBAL_INJECTED_CLOUD_ENV = "openstack"
+GLOBAL_INJECTED_DCAE_IP_ADDR = "dcae-controller.{{ .Values.nsPrefix }}"
+GLOBAL_INJECTED_DNS_IP_ADDR = "10.0.100.1"
+GLOBAL_INJECTED_DOCKER_VERSION = "1.1-STAGING-latest"
+#GLOBAL_INJECTED_EXTERNAL_DNS = "N/A"
+GLOBAL_INJECTED_GERRIT_BRANCH = "master"
+GLOBAL_INJECTED_KEYSTONE = "{{ .Values.openStackKeyStoneUrl }}"
+GLOBAL_INJECTED_MR_IP_ADDR = "dmaap.{{ .Values.nsPrefix }}"
+GLOBAL_INJECTED_MSO_IP_ADDR = "mso.{{ .Values.nsPrefix }}"
+GLOBAL_INJECTED_NETWORK = "{{ .Values.openStackPrivateNetId }}"
+GLOBAL_INJECTED_NEXUS_DOCKER_REPO = "nexus3.onap.org:10001"
+GLOBAL_INJECTED_NEXUS_PASSWORD = "docker"
+GLOBAL_INJECTED_NEXUS_REPO = "https://nexus.onap.org/content/sites/raw"
+GLOBAL_INJECTED_NEXUS_USERNAME = "docker"
+GLOBAL_INJECTED_OPENO_IP_ADDR = "msb-iag.{{ .Values.nsPrefix }}"
+GLOBAL_INJECTED_OPENSTACK_PASSWORD = "{{ .Values.openStackEncryptedPassword }}"
+GLOBAL_INJECTED_OPENSTACK_TENANT_ID = "{{ .Values.openStackTenantId }}"
+GLOBAL_INJECTED_OPENSTACK_USERNAME = "{{ .Values.openStackUserName }}"
+GLOBAL_INJECTED_POLICY_IP_ADDR = "pypdp.{{ .Values.nsPrefix }}"
+GLOBAL_INJECTED_POLICY_HEALTHCHECK_IP_ADDR = "drools.{{ .Values.nsPrefix }}"
+GLOBAL_INJECTED_PORTAL_IP_ADDR = "portalapps.{{ .Values.nsPrefix }}"
+GLOBAL_INJECTED_REGION = "{{ .Values.openStackRegion }}"
+GLOBAL_INJECTED_REMOTE_REPO = "http://gerrit.onap.org/r/testsuite/properties.git"
+GLOBAL_INJECTED_SDC_IP_ADDR = "sdc-be.{{ .Values.nsPrefix }}"
+GLOBAL_INJECTED_SDC_FE_IP_ADDR = "sdc-fe.{{ .Values.nsPrefix }}"
+GLOBAL_INJECTED_SDC_BE_IP_ADDR = "sdc-be.{{ .Values.nsPrefix }}"
+GLOBAL_INJECTED_SDNC_IP_ADDR = "sdnhost.{{ .Values.nsPrefix }}"
+GLOBAL_INJECTED_SDNC_PORTAL_IP_ADDR = "sdnc-portal.{{ .Values.nsPrefix }}"
+GLOBAL_INJECTED_SO_IP_ADDR = "mso.{{ .Values.nsPrefix }}"
+GLOBAL_INJECTED_VID_IP_ADDR = "vid-server.{{ .Values.nsPrefix }}"
+GLOBAL_INJECTED_VM_FLAVOR = "{{ .Values.openStackFlavourMedium }}"
+GLOBAL_INJECTED_VM_IMAGE_NAME = "{{ .Values.ubuntuImage }}"
+GLOBAL_INJECTED_PUBLIC_NET_ID = "{{ .Values.openStackPublicNetId }}"
+GLOBAL_INJECTED_PROPERTIES = {
+ "GLOBAL_INJECTED_AAI1_IP_ADDR" : "aai-service.{{ .Values.nsPrefix }}",
+ "GLOBAL_INJECTED_APPC_IP_ADDR" : "appc-sdnhost.{{ .Values.nsPrefix }}",
+ "GLOBAL_INJECTED_ARTIFACTS_VERSION" : "1.1.0-SNAPSHOT",
+ "GLOBAL_INJECTED_CLAMP_IP_ADDR" : "clamp.{{ .Values.nsPrefix }}",
+ "GLOBAL_INJECTED_CLOUD_ENV" : "openstack",
+ "GLOBAL_INJECTED_DCAE_IP_ADDR" : "dcae-controller.{{ .Values.nsPrefix }}",
+ "GLOBAL_INJECTED_DNS_IP_ADDR" : "10.0.100.1",
+ "GLOBAL_INJECTED_DOCKER_VERSION" : "1.1-STAGING-latest",
+ "GLOBAL_INJECTED_GERRIT_BRANCH" : "master",
+ "GLOBAL_INJECTED_KEYSTONE" : "{{ .Values.openStackKeyStoneUrl }}",
+ "GLOBAL_INJECTED_MR_IP_ADDR" : "dmaap.{{ .Values.nsPrefix }}",
+ "GLOBAL_INJECTED_MSO_IP_ADDR" : "mso.{{ .Values.nsPrefix }}",
+ "GLOBAL_INJECTED_NETWORK" : "{{ .Values.openStackPrivateNetId }}",
+ "GLOBAL_INJECTED_NEXUS_DOCKER_REPO" : "nexus3.onap.org:10001",
+ "GLOBAL_INJECTED_NEXUS_PASSWORD" : "docker",
+ "GLOBAL_INJECTED_NEXUS_REPO" : "https://nexus.onap.org/content/sites/raw",
+ "GLOBAL_INJECTED_NEXUS_USERNAME" : "docker",
+ "GLOBAL_INJECTED_OPENO_IP_ADDR" : "msb-iag.{{ .Values.nsPrefix }}",
+ "GLOBAL_INJECTED_OPENSTACK_PASSWORD" : "{{ .Values.openStackEncryptedPassword }}",
+ "GLOBAL_INJECTED_OPENSTACK_TENANT_ID" : "{{ .Values.openStackTenantId }}",
+ "GLOBAL_INJECTED_OPENSTACK_USERNAME" : "{{ .Values.openStackUserName }}",
+ "GLOBAL_INJECTED_POLICY_IP_ADDR" : "pypdp.{{ .Values.nsPrefix }}",
+ "GLOBAL_INJECTED_POLICY_HEALTHCHECK_IP_ADDR" : "drools.{{ .Values.nsPrefix }}",
+ "GLOBAL_INJECTED_PORTAL_IP_ADDR" : "portalapps.{{ .Values.nsPrefix }}",
+ "GLOBAL_INJECTED_REGION" : "{{ .Values.openStackRegion }}",
+ "GLOBAL_INJECTED_REMOTE_REPO" : "http://gerrit.onap.org/r/testsuite/properties.git",
+ "GLOBAL_INJECTED_SDC_FE_IP_ADDR" : "sdc-fe.{{ .Values.nsPrefix }}",
+ "GLOBAL_INJECTED_SDC_BE_IP_ADDR" : "sdc-be.{{ .Values.nsPrefix }}",
+ "GLOBAL_INJECTED_SDNC_IP_ADDR" : "sdnhost.{{ .Values.nsPrefix }}",
+ "GLOBAL_INJECTED_SDNC_PORTAL_IP_ADDR" : "sdnc-portal.{{ .Values.nsPrefix }}",
+ "GLOBAL_INJECTED_SO_IP_ADDR" : "mso.{{ .Values.nsPrefix }}",
+ "GLOBAL_INJECTED_VID_IP_ADDR" : "vid-server.{{ .Values.nsPrefix }}",
+ "GLOBAL_INJECTED_VM_FLAVOR" : "{{ .Values.openStackFlavourMedium }}",
+ "GLOBAL_INJECTED_VM_IMAGE_NAME" : "{{ .Values.ubuntuImage }}",
+ "GLOBAL_INJECTED_PUBLIC_NET_ID" : "{{ .Values.openStackPublicNetId }}"
+}
[Arguments] ${model_zip_path} ${catalog_service_name}=
${catalog_service_id}= Add ASDC Catalog Service ${catalog_service_name}
${catalog_resource_ids}= Create List
+ ${catalog_resources}= Create Dictionary
: FOR ${zip} IN @{model_zip_path}
\ ${loop_catalog_resource_id}= Setup ASDC Catalog Resource ${zip}
\ Append To List ${catalog_resource_ids} ${loop_catalog_resource_id}
\ ${loop_catalog_resource_resp}= Get ASDC Catalog Resource ${loop_catalog_resource_id}
\ Add ASDC Resource Instance ${catalog_service_id} ${loop_catalog_resource_id} ${loop_catalog_resource_resp['name']}
+ \ Set To Dictionary ${catalog_resources} ${loop_catalog_resource_id}=${loop_catalog_resource_resp}
${catalog_service_resp}= Get ASDC Catalog Service ${catalog_service_id}
Checkin ASDC Catalog Service ${catalog_service_id}
Request Certify ASDC Catalog Service ${catalog_service_id}
${catalog_service_resp}= Get ASDC Catalog Service ${catalog_service_id}
${vf_module}= Find Element In Array ${loop_catalog_resource_resp['groups']} type org.openecomp.groups.VfModule
Check Catalog Service Distributed ${catalog_service_resp['uuid']}
- [Return] ${catalog_service_resp['name']} ${loop_catalog_resource_resp['name']} ${vf_module} ${catalog_resource_ids} ${catalog_service_id}
+ [Return] ${catalog_service_resp['name']} ${loop_catalog_resource_resp['name']} ${vf_module} ${catalog_resource_ids} ${catalog_service_id} ${catalog_resources}
Setup ASDC Catalog Resource
[Documentation] Creates all the steps a vf needs for an asdc catalog resource and returns the id
[Arguments] ${catalog_service_uuid}
${dist_resp}= Get Catalog Service Distribution ${catalog_service_uuid}
Should Be Equal As Strings ${dist_resp['distributionStatusOfServiceList'][0]['deployementStatus']} Distributed
+ Sleep 3 minutes
${det_resp}= Get Catalog Service Distribution Details ${dist_resp['distributionStatusOfServiceList'][0]['distributionID']}
@{ITEMS}= Copy List ${det_resp['distributionStatusList']}
:FOR ${ELEMENT} IN @{ITEMS}
Library RequestsLibrary
Library String
Library JSONUtils
-Library Collections
+Library Collections
Resource global_properties.robot
*** Variables ***
Run Policy Health Check
[Documentation] Runs Policy Health check
- ${auth}= Create List ${GLOBAL_POLICY_USERNAME} ${GLOBAL_POLICY_PASSWORD}
+ ${auth}= Create List ${GLOBAL_POLICY_USERNAME} ${GLOBAL_POLICY_PASSWORD}
Log Creating session ${POLICY_ENDPOINT}
${session}= Create Session policy ${POLICY_HEALTHCHECK_ENDPOINT} auth=${auth}
${headers}= Create Dictionary Accept=application/json Content-Type=application/json
:FOR ${ELEMENT} IN @{ITEMS}
\ Should Be Equal As Strings ${ELEMENT['code']} 200
\ Should Be True ${ELEMENT['healthy']}
-
Run Policy Put Request
[Documentation] Runs Policy Put request
[Arguments] ${data_path} ${data}
${resp}= Put Request policy ${data_path} data=${data} headers=${headers}
Log Received response from policy ${resp.text}
[Return] ${resp}
-
Run Policy Delete Request
[Documentation] Runs Policy Delete request
[Arguments] ${data_path} ${data}
${resp}= Delete Request policy ${data_path} data=${data} headers=${headers}
Log Received response from policy ${resp.text}
[Return] ${resp}
-
Run Policy Get Configs Request
[Documentation] Runs Policy Get Configs request
[Arguments] ${data_path} ${data}
${headers}= Create Dictionary Accept=application/json Content-Type=application/json Authorization=Basic ${GLOBAL_POLICY_AUTH} ClientAuth=${GLOBAL_POLICY_CLIENTAUTH}
${resp}= Post Request policy ${data_path} data=${data} headers=${headers}
Log Received response from policy ${resp.text}
- [Return] ${resp}
\ No newline at end of file
+ [Return] ${resp}
*** Settings ***
Documentation The main interface for interacting with SDN-GC. It handles low level stuff like managing the http request library and SDN-GC required fields
Library RequestsLibrary
-Library UUID
+Library UUID
Library OperatingSystem
Library ExtendedSelenium2Library
Library Collections
Set To Dictionary ${valuemap} network=${GLOBAL_INJECTED_NETWORK}
Set To Dictionary ${valuemap} public_net_id=${GLOBAL_INJECTED_PUBLIC_NET_ID}
Set To Dictionary ${valuemap} cloud_env=${GLOBAL_INJECTED_CLOUD_ENV}
- Set To Dictionary ${valuemap} install_script_version=${GLOBAL_INJECTED_INSTALL_SCRIPT_VERSION}
+ Set To Dictionary ${valuemap} install_script_version=${GLOBAL_INJECTED_SCRIPT_VERSION}
Set To Dictionary ${valuemap} vm_image_name=${GLOBAL_INJECTED_VM_IMAGE_NAME}
- Set To Dictionary ${valuemap} vm_flavor_name=${GLOBAL_INJECTED_VM_FLAVOR_NAME}
+ Set To Dictionary ${valuemap} vm_flavor_name=${GLOBAL_INJECTED_VM_FLAVOR}
# update the value map with unique values.
Set To Dictionary ${valuemap} uuid=${uuid} hostid=${hostid} ecompnet=${ecompnet}
${parameters}= Create Dictionary
## Setup Browser is now being managed by the test case
## Setup Browser
Go To ${SDNGC_ADMIN_SIGNUP_URL}
- Maximize Browser Window
+ ##Maximize Browser Window
Set Selenium Speed ${GLOBAL_SELENIUM_DELAY}
Set Browser Implicit Wait ${GLOBAL_SELENIUM_BROWSER_IMPLICIT_WAIT}
Log Logging in to ${SDNGC_ADMIN_LOGIN_URL}
Input Password xpath=//input[@id='password'] ${shortened_uuid}
Click Button xpath=//button[@type='submit']
Title Should Be SDN-C AdminPortal
- Log Logged in to ${SDNGC_ADMIN_LOGIN_URL}
\ No newline at end of file
+ Log Logged in to ${SDNGC_ADMIN_LOGIN_URL}
--- /dev/null
+1. Get the application URL by running these commands:
+{{- if .Values.ingress.enabled }}
+{{- range .Values.ingress.hosts }}
+ http://{{ . }}
+{{- end }}
+{{- else if contains "NodePort" .Values.service.type }}
+ export NODE_PORT=$(kubectl get --namespace {{ include "common.namespace" . }} -o jsonpath="{.spec.ports[0].nodePort}" services {{ include "common.name" . }})
+ export NODE_IP=$(kubectl get nodes --namespace {{ include "common.namespace" . }} -o jsonpath="{.items[0].status.addresses[0].address}")
+ echo http://$NODE_IP:$NODE_PORT
+{{- else if contains "LoadBalancer" .Values.service.type }}
+ NOTE: It may take a few minutes for the LoadBalancer IP to be available.
+ You can watch the status of by running 'kubectl get svc -w {{ include "common.name" . }}'
+ export SERVICE_IP=$(kubectl get svc --namespace {{ include "common.namespace" . }} {{ include "common.name" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')
+ echo http://$SERVICE_IP:{{ .Values.service.externalPort }}
+{{- else if contains "ClusterIP" .Values.service.type }}
+ export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "so.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
+ echo "Visit http://127.0.0.1:8080 to use your application"
+ kubectl port-forward $POD_NAME 8080:{{ .Values.service.internalPort }}
+{{- end }}
+++ /dev/null
-#{{ if not .Values.disableRobotRobot }}
-apiVersion: v1
-kind: Service
-metadata:
- labels:
- app: robot
- name: robot
- namespace: "{{ .Values.nsPrefix }}-robot"
-spec:
- ports:
- - port: 88
- nodePort: {{ .Values.nodePortPrefix }}09
- selector:
- app: robot
- type: NodePort
-#{{ end }}
\ No newline at end of file
--- /dev/null
+apiVersion: v1
+kind: ConfigMap
+metadata:
+ name: {{ include "common.name" . }}-eteshare-configmap
+ namespace: {{ include "common.namespace" . }}
+data:
+{{ tpl (.Files.Glob "resources/config/eteshare/config/*").AsConfig . | indent 2 }}
+---
+apiVersion: v1
+kind: ConfigMap
+metadata:
+ name: {{ include "common.name" . }}-resources-configmap
+ namespace: {{ include "common.namespace" . }}
+data:
+{{ tpl (.Files.Glob "resources/config/robot/resources/*").AsConfig . | indent 2 }}
+---
+apiVersion: v1
+kind: ConfigMap
+metadata:
+ name: {{ include "common.name" . }}-lighttpd-authorization-configmap
+ namespace: {{ include "common.namespace" . }}
+data:
+{{ tpl (.Files.Glob "resources/config/authorization").AsConfig . | indent 2 }}
--- /dev/null
+apiVersion: extensions/v1beta1
+kind: Deployment
+metadata:
+ name: {{ include "common.name" . }}
+ namespace: {{ include "common.namespace" . }}
+ labels:
+ app: {{ include "common.name" . }}
+ chart: {{ .Chart.Name }}-{{ .Chart.Version | replace "+" "_" }}
+ release: {{ .Release.Name }}
+ heritage: {{ .Release.Service }}
+spec:
+ replicas: {{ .Values.robotReplicas }}
+ template:
+ metadata:
+ labels:
+ app: {{ include "common.name" . }}
+ release: {{ .Release.Name }}
+ spec:
+ containers:
+ - name: {{ .Chart.Name }}
+ image: "{{ .Values.global.repository | default .Values.repository }}/{{ .Values.image }}"
+ imagePullPolicy: {{ .Values.global.pullPolicy | default .Values.pullPolicy }}
+ ports:
+ - containerPort: {{ .Values.service.internalPort }}
+ # disable liveness probe when breakpoints set in debugger
+ # so K8s doesn't restart unresponsive container
+ {{- if eq .Values.liveness.enabled true }}
+ livenessProbe:
+ tcpSocket:
+ port: {{ .Values.service.internalPort }}
+ initialDelaySeconds: {{ .Values.liveness.initialDelaySeconds }}
+ periodSeconds: {{ .Values.liveness.periodSeconds }}
+ {{ end -}}
+ readinessProbe:
+ tcpSocket:
+ port: {{ .Values.service.internalPort }}
+ initialDelaySeconds: {{ .Values.readiness.initialDelaySeconds }}
+ periodSeconds: {{ .Values.readiness.periodSeconds }}
+ volumeMounts:
+ - name: localtime
+ mountPath: /etc/localtime
+ readOnly: true
+ - name: robot-eteshare
+ mountPath: /share/config
+ - name: robot-resources
+ mountPath: /var/opt/OpenECOMP_ETE/robot/resources/asdc_interface.robot
+ subPath: asdc_interface.robot
+ - name: robot-resources
+ mountPath: /var/opt/OpenECOMP_ETE/robot/resources/policy_interface.robot
+ subPath: policy_interface.robot
+ - name: robot-resources
+ mountPath: /var/opt/OpenECOMP_ETE/robot/resources/sdngc_interface.robot
+ subPath: sdngc_interface.robot
+ - name: robot-lighttpd-authorization
+ mountPath: /etc/lighttpd/authorization
+ subPath: authorization
+ resources:
+{{ toYaml .Values.resources | indent 12 }}
+ {{- if .Values.nodeSelector }}
+ nodeSelector:
+{{ toYaml .Values.nodeSelector | indent 10 }}
+ {{- end -}}
+ {{- if .Values.affinity }}
+ affinity:
+{{ toYaml .Values.affinity | indent 10 }}
+ {{- end }}
+ volumes:
+ - name: localtime
+ hostPath:
+ path: /etc/localtime
+ - name: robot-eteshare
+ configMap:
+ name: {{ include "common.name" . }}-eteshare-configmap
+ defaultMode: 0755
+ - name: robot-resources
+ configMap:
+ name: {{ include "common.name" . }}-resources-configmap
+ - name: robot-lighttpd-authorization
+ configMap:
+ name: {{ include "common.name" . }}-lighttpd-authorization-configmap
+ imagePullSecrets:
+ - name: "{{ include "common.namespace" . }}-docker-registry-key"
+++ /dev/null
-#{{ if not .Values.disableRobotRobot }}
-apiVersion: extensions/v1beta1
-kind: Deployment
-metadata:
- name: robot
- namespace: "{{ .Values.nsPrefix }}-robot"
-spec:
- selector:
- matchLabels:
- app: robot
- template:
- metadata:
- labels:
- app: robot
- name: robot
- spec:
- containers:
- - image: {{ .Values.image.testsuite }}
- imagePullPolicy: {{ .Values.pullPolicy }}
- name: robot
- volumeMounts:
- - name: localtime
- mountPath: /etc/localtime
- readOnly: true
- - name: robot-eteshare
- mountPath: /share
- - name: robot-resources-asdc-interface
- mountPath: /var/opt/OpenECOMP_ETE/robot/resources/asdc_interface.robot
- - name: robot-resources-policy-interface
- mountPath: /var/opt/OpenECOMP_ETE/robot/resources/policy_interface.robot
- - name: robot-resources-sdngc-interface
- mountPath: /var/opt/OpenECOMP_ETE/robot/resources/sdngc_interface.robot
- - name: lighttpd-authorization
- mountPath: /etc/lighttpd/authorization
- ports:
- - containerPort: 88
- readinessProbe:
- tcpSocket:
- port: 88
- initialDelaySeconds: 5
- periodSeconds: 10
- volumes:
- - name: localtime
- hostPath:
- path: /etc/localtime
- - name: robot-eteshare
- hostPath:
- path: /dockerdata-nfs/{{ .Values.nsPrefix }}/robot/eteshare
- - name: robot-resources-asdc-interface
- hostPath:
- path: /dockerdata-nfs/{{ .Values.nsPrefix }}/robot/robot/resources/asdc_interface.robot
- - name: robot-resources-policy-interface
- hostPath:
- path: /dockerdata-nfs/{{ .Values.nsPrefix }}/robot/robot/resources/policy_interface.robot
- - name: robot-resources-sdngc-interface
- hostPath:
- path: /dockerdata-nfs/{{ .Values.nsPrefix }}/robot/robot/resources/sdngc_interface.robot
- - name: lighttpd-authorization
- hostPath:
- path: /dockerdata-nfs/{{ .Values.nsPrefix }}/robot/authorization
- imagePullSecrets:
- - name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
--- /dev/null
+apiVersion: v1
+kind: Service
+metadata:
+ name: {{ include "common.name" . }}
+ namespace: {{ include "common.namespace" . }}
+ labels:
+ app: {{ include "common.name" . }}
+ chart: {{ .Chart.Name }}-{{ .Chart.Version | replace "+" "_" }}
+ release: {{ .Release.Name }}
+ heritage: {{ .Release.Service }}
+spec:
+ type: {{ .Values.service.type }}
+ ports:
+ {{if eq .Values.service.type "NodePort" -}}
+ - port: {{ .Values.service.internalPort }}
+ nodePort: {{ .Values.global.nodePortPrefix | default .Values.nodePortPrefix }}{{ .Values.service.externalPort }}
+ name: {{ .Values.service.name }}
+ {{- else -}}
+ - port: {{ .Values.service.externalPort }}
+ targetPort: {{ .Values.service.internalPort }}
+ name: {{ .Values.service.name }}
+ {{- end}}
+ selector:
+ app: {{ include "common.name" . }}
+ release: {{ .Release.Name }}
-nsPrefix: onap
+#################################################################
+# Global configuration defaults.
+#################################################################
+global: # global defaults
+ nodePortPrefix: 302
+ repositorySecret: eyJuZXh1czMub25hcC5vcmc6MTAwMDEiOnsidXNlcm5hbWUiOiJkb2NrZXIiLCJwYXNzd29yZCI6ImRvY2tlciIsImVtYWlsIjoiQCIsImF1dGgiOiJaRzlqYTJWeU9tUnZZMnRsY2c9PSJ9fQ==
+ readinessRepository: oomk8s
+ readinessImage: readiness-check:1.0.0
+ loggingRepository: docker.elastic.co
+ loggingImage: beats/filebeat:5.5.0
+
+subChartsOnly:
+ enabled: true
+
+# application image
+repository: nexus3.onap.org:10001
+image: openecomp/testsuite:1.1-STAGING-latest
pullPolicy: Always
-nodePortPrefix: 302
-image:
- testsuite: nexus3.onap.org:10001/openecomp/testsuite:1.1-STAGING-latest
+# flag to enable debugging - application support required
+debugEnabled: false
+
+#################################################################
+# Application configuration defaults.
+#################################################################
+
+# openstack configuration
+
+demoArtifactsVersion: "1.1.0-SNAPSHOT"
+openStackPrivateNetCidr: "192.168.30.0"
+openStackFlavourMedium: "m1.medium"
+openStackKeyStoneUrl: "http://1.2.3.4:5000"
+openStackPublicNetId: "e8f51958045716781ffc"
+openStackPrivateNetId: "e8f51956-00dd-4425-af36-045716781ffc"
+openStackOamNetworkCidrPrefix: "192.168.120"
+openStackEncryptedPassword: "c124921a3a0efbe579782cde8227681e"
+openStackRegion: "RegionOne"
+openStackPrivateSubnetId: "e8f51956-00dd-4425-af36-045716781ffc"
+openStackTenantId: "47899782ed714295b1151681fdfd51f5"
+openStackUserName: "vnf_user"
+ubuntuImage: "Ubuntu_14.04.5_LTS"
+
+# default number of instances
+robotReplicas: 1
+
+nodeSelector: {}
+
+affinity: {}
+
+# probe configuration parameters
+liveness:
+ initialDelaySeconds: 10
+ periodSeconds: 10
+ # necessary to disable liveness probe when setting breakpoints
+ # in debugger so K8s doesn't restart unresponsive container
+ enabled: true
+
+readiness:
+ initialDelaySeconds: 10
+ periodSeconds: 10
+
+
+service:
+ type: NodePort
+ name: robot
+ externalPort: "09"
+ internalPort: 88
+
+
+ingress:
+ enabled: false
+
+
+resources: {}
+ # We usually recommend not to specify default resources and to leave this as a conscious
+ # choice for the user. This also increases chances charts run on environments with little
+ # resources, such as Minikube. If you do want to specify resources, uncomment the following
+ # lines, adjust them as necessary, and remove the curly braces after 'resources:'.
+ #
+ # Example:
+ # Configure resource requests and limits
+ # ref: http://kubernetes.io/docs/user-guide/compute-resources/
+ # Minimum memory for development is 2 CPU cores and 4GB memory
+ # Minimum memory for production is 4 CPU cores and 8GB memory
+#resources:
+# limits:
+# cpu: 2
+# memory: 4Gi
+# requests:
+# cpu: 2
+# memory: 4Gi
"chef_type": "environment",
"default_attributes": {
- "CS_VIP": "sdc-cs.onap-sdc",
- "BE_VIP": "sdc-be.onap-sdc",
- "FE_VIP": "sdc-fe.onap-sdc",
- "ES_VIP": "sdc-es.onap-sdc",
+ "CS_VIP": "sdc-cs.{{.Values.nsPrefix}}",
+ "BE_VIP": "sdc-be.{{.Values.nsPrefix}}",
+ "FE_VIP": "sdc-fe.{{.Values.nsPrefix}}",
+ "ES_VIP": "sdc-es.{{.Values.nsPrefix}}",
"interfaces": {
"application": "eth0",
"private": "eth0"
},
"ECompP": {
- "ecomp_rest_url": "http://portalapps.onap-portal:8989/ONAPPORTAL/auxapi",
- "ueb_url_list": "dmaap.onap-message-router, dmaap.onap-message-router",
+ "ecomp_rest_url": "http://portalapps.{{.Values.nsPrefix}}:8989/ONAPPORTAL/auxapi",
+ "ueb_url_list": "dmaap.{{.Values.nsPrefix}}, dmaap.{{.Values.nsPrefix}}",
"app_secret": "XftIATw9Jr3VzAcPqt3NnJOu",
"app_key": "x9UfO7JsDn8BESVX",
"inbox_name": "ECOMP-PORTAL-INBOX",
- "ecomp_redirect_url": "http://portalapps.onap-portal:8989/ONAPPORTAL/login.htm",
+ "ecomp_redirect_url": "http://portalapps.{{.Values.nsPrefix}}:8989/ONAPPORTAL/login.htm",
"app_topic_name": "ECOMP-PORTAL-OUTBOX-SDC1",
"decryption_key": "AGLDdG4D04BKm2IxIWEr8o=="
},
"UEB": {
"PublicKey": "iPIxkpAMI8qTcQj8",
"SecretKey": "Ehq3WyT4bkif4zwgEbvshGal",
- "fqdn": ["dmaap.onap-message-router", "dmaap.onap-message-router"]
+ "fqdn": ["dmaap.{{.Values.nsPrefix}}", "dmaap.{{.Values.nsPrefix}}"]
},
"Nodes": {
- "CS": "sdc-cs.onap-sdc",
- "BE": "sdc-be.onap-sdc",
- "FE": "sdc-fe.onap-sdc",
- "ES": "sdc-es.onap-sdc"
+ "CS": "sdc-cs.{{.Values.nsPrefix}}",
+ "BE": "sdc-be.{{.Values.nsPrefix}}",
+ "FE": "sdc-fe.{{.Values.nsPrefix}}",
+ "ES": "sdc-es.{{.Values.nsPrefix}}"
}
},
"override_attributes": {
"private": "eth1"
},
"ECompP": {
- "ecomp_rest_url": "http://portalapps.onap-portal:8989/ONAPPORTAL/auxapi",
- "ueb_url_list": "dmaap.onap-message-router,dmaap.onap-message-router",
+ "ecomp_rest_url": "http://portalapps.{{.Values.nsPrefix}}:8989/ONAPPORTAL/auxapi",
+ "ueb_url_list": "dmaap.{{.Values.nsPrefix}},dmaap.{{.Values.nsPrefix}}",
"app_secret": "XftIATw9Jr3VzAcPqt3NnJOu",
"app_key": "x9UfO7JsDn8BESVX",
"inbox_name": "ECOMP-PORTAL-INBOX",
- "ecomp_redirect_url": "http://portalapps.onap-portal:8989/ONAPPORTAL/login.htm",
+ "ecomp_redirect_url": "http://portalapps.{{.Values.nsPrefix}}:8989/ONAPPORTAL/login.htm",
"app_topic_name": "ECOMP-PORTAL-OUTBOX-SDC1",
"decryption_key": "AGLDdG4D04BKm2IxIWEr8o=="
},
"UEB": {
"PublicKey": "iPIxkpAMI8qTcQj8",
"SecretKey": "Ehq3WyT4bkif4zwgEbvshGal",
- "fqdn": ["dmaap.onap-message-router", "dmaap.onap-message-router"]
+ "fqdn": ["dmaap.{{.Values.nsPrefix}}", "dmaap.{{.Values.nsPrefix}}"]
},
"Nodes": {
"CS": "yyy",
#List of logstash server ip addresses with port number.
#But, in our case, this will be the loadbalancer IP address.
#For the below property to work the loadbalancer or logstash should expose 5044 port to listen the filebeat events or port in the property should be changed appropriately.
- hosts: ["logstash.onap-log:5044"]
+ hosts: ["logstash.{{.Values.nsPrefix}}:5044"]
#If enable will do load balancing among availabe Logstash, automatically.
loadbalance: true
mode "0755"
variables({
:fe_host_ip => node['HOST_IP'],
- :be_host_ip => "sdc-be.onap-sdc",
+ :be_host_ip => "sdc-be.{{.Values.nsPrefix}}",
:catalog_port => node['BE'][:http_port],
:ssl_port => node['BE'][:https_port]
})
labels:
app: sdc-es
name: sdc-es
- namespace: "{{ .Values.nsPrefix }}-sdc"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
ports:
- name: sdc-es-port-9200
labels:
app: sdc-cs
name: sdc-cs
- namespace: "{{ .Values.nsPrefix }}-sdc"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
ports:
- name: sdc-cs-port-9042
labels:
app: sdc-kb
name: sdc-kb
- namespace: "{{ .Values.nsPrefix }}-sdc"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
ports:
- name: sdc-kb-port-5601
labels:
app: sdc-be
name: sdc-be
- namespace: "{{ .Values.nsPrefix }}-sdc"
+ namespace: "{{ .Values.nsPrefix }}"
annotations:
msb.onap.org/service-info: '[
{
labels:
app: sdc-fe
name: sdc-fe
- namespace: "{{ .Values.nsPrefix }}-sdc"
+ namespace: "{{ .Values.nsPrefix }}"
annotations:
msb.onap.org/service-info: '[
{
labels:
app: sdc-be
name: sdc-be
- namespace: "{{ .Values.nsPrefix }}-sdc"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
selector:
matchLabels:
- dmaap
env:
- name: NAMESPACE
- value: {{ .Values.nsPrefix }}-message-router
+ value: {{ .Values.nsPrefix }}
image: {{ .Values.image.readiness }}
imagePullPolicy: {{ .Values.pullPolicy }}
name: sdc-dmaap-readiness
labels:
app: sdc-cs
name: sdc-cs
- namespace: "{{ .Values.nsPrefix }}-sdc"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
selector:
matchLabels:
kind: ConfigMap
metadata:
name: sdc-environments-configmap
- namespace: {{ .Values.nsPrefix }}-sdc
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/environments/*").AsConfig . | indent 2 }}
#{{ end }}
labels:
app: sdc-es
name: sdc-es
- namespace: "{{ .Values.nsPrefix }}-sdc"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
selector:
matchLabels:
kind: ConfigMap
metadata:
name: sdc-fe-configmap
- namespace: {{ .Values.nsPrefix }}-sdc
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/sdc-fe/*").AsConfig . | indent 2 }}
#{{ end }}
labels:
app: sdc-fe
name: sdc-fe
- namespace: "{{ .Values.nsPrefix }}-sdc"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
selector:
matchLabels:
labels:
app: sdc-kb
name: sdc-kb
- namespace: "{{ .Values.nsPrefix }}-sdc"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
selector:
matchLabels:
kind: ConfigMap
metadata:
name: sdc-log-be-configmap
- namespace: {{ .Values.nsPrefix }}-sdc
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/log/be/*").AsConfig . | indent 2 }}
---
kind: ConfigMap
metadata:
name: sdc-filebeat-configmap
- namespace: {{ .Values.nsPrefix }}-sdc
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/log/filebeat/*").AsConfig . | indent 2 }}
#{{ end }}
kind: ConfigMap
metadata:
name: sdc-log-fe-configmap
- namespace: {{ .Values.nsPrefix }}-sdc
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/log/fe/*").AsConfig . | indent 2 }}
#{{ end }}
apiVersion: v1
kind: PersistentVolume
metadata:
- name: "{{ .Values.nsPrefix }}-sdc-cs-db"
- namespace: "{{ .Values.nsPrefix }}-sdc"
+ name: "{{ .Values.nsPrefix }}-sdc"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
- name: "{{ .Values.nsPrefix }}-sdc-cs-db"
+ name: "{{ .Values.nsPrefix }}-sdc"
spec:
capacity:
storage: 2Gi
apiVersion: v1
metadata:
name: sdc-cs-db
- namespace: "{{ .Values.nsPrefix }}-sdc"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
accessModes:
- ReadWriteMany
storage: 2Gi
selector:
matchLabels:
- name: "{{ .Values.nsPrefix }}-sdc-cs-db"
+ name: "{{ .Values.nsPrefix }}-sdc"
#{{ end }}
#
# Configuration file for A&AI Client
#
-org.onap.ccsdk.sli.adaptors.aai.uri=https://aai-service.onap-aai:8443
+org.onap.ccsdk.sli.adaptors.aai.uri=https://aai-service.{{.Values.nsPrefix}}:8443
# query
# UBB Notify
org.onap.ccsdk.sli.adaptors.aai.path.notify=/aai/v11/actions/notify
-org.onap.ccsdk.sli.adaptors.aai.notify.selflink.fqdn=https://aai-service.onap-aai:8443/restconf/config/L3SDN-API:services/layer3-service-list/{service-instance-id}
-org.onap.ccsdk.sli.adaptors.aai.notify.selflink.avpn=https://aai-service.onap-aai:8543/restconf/config/L3AVPN-EVC-API:services/service-list/{service-instance-id}/service-data/avpn-logicalchannel-information
+org.onap.ccsdk.sli.adaptors.aai.notify.selflink.fqdn=https://aai-service.{{.Values.nsPrefix}}:8443/restconf/config/L3SDN-API:services/layer3-service-list/{service-instance-id}
+org.onap.ccsdk.sli.adaptors.aai.notify.selflink.avpn=https://aai-service.{{.Values.nsPrefix}}:8543/restconf/config/L3AVPN-EVC-API:services/service-list/{service-instance-id}/service-data/avpn-logicalchannel-information
# Service
org.onap.ccsdk.sli.adaptors.aai.path.service=/aai/v11/service-design-and-creation/services/service/{service-id}
"dbPassword": "gamma",
"dbName": "sdnctl",
"odlProtocol": "http",
- "odlHost": "sdnhost.onap-sdnc",
- "odlConexusHost": "sdnhost.onap-sdnc",
+ "odlHost": "sdnhost.{{.Values.nsPrefix}}",
+ "odlConexusHost": "sdnhost.{{.Values.nsPrefix}}",
"odlPort": "8181",
"odlConexusPort": "8181",
"odlUser": "admin",
contenttype =application/json
authKey=fxoW4jZrO7mdLWWa:f4KxkoBtToyoEG7suMoV8KhnkwM=
authDate=2016-02-18T13:57:37-0800
-host=dmaap.onap-message-router:3904
+host=dmaap.{{.Values.nsPrefix}}:3904
topic=VCPE-DHCP-EVENT
group=jmsgrp
id=sdnc1
DME2preferredRouterFilePath=/opt/onap/sdnc/data/properties/dmaap-listener.preferredRoute.txt
sdnc.odl.user=admin
sdnc.odl.password=admin
-sdnc.odl.url-base=https://sdnhost.onap-sdnc:8443/restconf/operations
\ No newline at end of file
+sdnc.odl.url-base=https://sdnhost.{{.Values.nsPrefix}}:8443/restconf/operations
#List of logstash server ip addresses with port number.
#But, in our case, this will be the loadbalancer IP address.
#For the below property to work the loadbalancer or logstash should expose 5044 port to listen the filebeat events or port in the property should be changed appropriately.
- hosts: ["logstash.onap-log:5044"]
+ hosts: ["logstash.{{.Values.nsPrefix}}:5044"]
#If enable will do load balancing among availabe Logstash, automatically.
loadbalance: true
-org.onap.ccsdk.sli.northbound.uebclient.asdc-address=sdc-be.onap-sdc:8443
+org.onap.ccsdk.sli.northbound.uebclient.asdc-address=sdc-be.{{.Values.nsPrefix}}:8443
org.onap.ccsdk.sli.northbound.uebclient.consumer-group=sdc-OpenSource-Env1-sdnc-dockero
org.onap.ccsdk.sli.northbound.uebclient.consumer-id=sdc-COpenSource-Env11-sdnc-dockero
org.onap.ccsdk.sli.northbound.uebclient.environment-name=AUTO
org.onap.ccsdk.sli.northbound.uebclient.user=sdnc
org.onap.ccsdk.sli.northbound.uebclient.sdnc-user=admin
org.onap.ccsdk.sli.northbound.uebclient.sdnc-passwd=Kp8bJ4SXszM0WXlhak3eHlcse2gAw84vaoGGmJvUy2U
-org.onap.ccsdk.sli.northbound.uebclient.asdc-api-base-url=http://sdnhost.onap-sdnc:8282/restconf/operations/
+org.onap.ccsdk.sli.northbound.uebclient.asdc-api-base-url=http://sdnhost.{{.Values.nsPrefix}}:8282/restconf/operations/
org.onap.ccsdk.sli.northbound.uebclient.asdc-api-namespace=org:onap:ccsdk
org.onap.ccsdk.sli.northbound.uebclient.spool.incoming=/opt/onap/sdnc/ueb-listener/spool/incoming
org.onap.ccsdk.sli.northbound.uebclient.spool.archive=/opt/onap/sdnc/ueb-listener/spool/archive
org.onap.ccsdk.sli.northbound.uebclient.keystore-path=
org.onap.ccsdk.sli.northbound.uebclient.keystore-password=
org.onap.ccsdk.sli.northbound.uebclient.xslt-path-list=
-org.onap.ccsdk.sli.northbound.uebclient.artifact-map=/opt/onap/sdnc/data/properties/artifact.map
\ No newline at end of file
+org.onap.ccsdk.sli.northbound.uebclient.artifact-map=/opt/onap/sdnc/data/properties/artifact.map
kind: Service
metadata:
name: dbhost
- namespace: "{{ .Values.nsPrefix }}-sdnc"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: sdnc-dbhost
spec:
kind: Service
metadata:
name: dbhost-read
- namespace: "{{ .Values.nsPrefix }}-sdnc"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: sdnc-dbhost
spec:
kind: Service
metadata:
name: sdnctldb01
- namespace: "{{ .Values.nsPrefix }}-sdnc"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: sdnc-dbhost
spec:
kind: Service
metadata:
name: sdnctldb02
- namespace: "{{ .Values.nsPrefix }}-sdnc"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: sdnc-dbhost
spec:
kind: Service
metadata:
name: sdnc-dgbuilder
- namespace: "{{ .Values.nsPrefix }}-sdnc"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: sdnc-dgbuilder
spec:
kind: Service
metadata:
name: sdnhost
- namespace: "{{ .Values.nsPrefix }}-sdnc"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: sdnc
annotations:
apiVersion: v1
metadata:
name: nfs-provisioner
- namespace: "{{ .Values.nsPrefix }}-sdnc"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: nfs-provisioner
spec:
kind: Service
metadata:
name: sdnc-portal
- namespace: "{{ .Values.nsPrefix }}-sdnc"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: sdnc-portal
annotations:
kind: Service
metadata:
name: sdnhost-cluster
- namespace: "{{ .Values.nsPrefix }}-sdnc"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: sdnc
annotations:
kind: StatefulSet
metadata:
name: sdnc-dbhost
- namespace: "{{ .Values.nsPrefix }}-sdnc"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
serviceName: "dbhost"
replicas: {{ .Values.numberOfDbReplicas }}
echo ${BASH_REMATCH}
[[ $ordinal -eq 0 ]] && exit 0
# Clone data from previous peer.
- ncat --recv-only sdnc-dbhost-$(($ordinal-1)).dbhost.{{ .Values.nsPrefix }}-sdnc 3307 | xbstream -x -C /var/lib/mysql
+ ncat --recv-only sdnc-dbhost-$(($ordinal-1)).dbhost.{{ .Values.nsPrefix }} 3307 | xbstream -x -C /var/lib/mysql
# Prepare the backup.
xtrabackup --user=root --password=$MYSQL_ROOT_PASSWORD --prepare --target-dir=/var/lib/mysql
ls -l /var/lib/mysql
mv change_master_to.sql.in change_master_to.sql.orig
mysql --user=root --password=$MYSQL_ROOT_PASSWORD -h 127.0.0.1 <<EOF
$(<change_master_to.sql.orig),
- MASTER_HOST="sdnc-dbhost-0.dbhost.{{ .Values.nsPrefix }}-sdnc",
+ MASTER_HOST="sdnc-dbhost-0.dbhost.{{ .Values.nsPrefix }}",
MASTER_USER="root",
MASTER_PASSWORD="$MYSQL_ROOT_PASSWORD",
MASTER_CONNECT_RETRY=10;
emptyDir: {}
- name: config-map
configMap:
- name: mysql
+ name: sdnc-mysql
- name: localtime
hostPath:
path: /etc/localtime
+#{{ if .Values.disableNfsProvisioner }}
+ - name: sdnc-data
+ hostPath:
+ path: /dockerdata-nfs/{{ .Values.nsPrefix }}/sdnc/data
+#{{ else }}
volumeClaimTemplates:
- metadata:
name: sdnc-data
requests:
storage: 1Gi
#{{ end }}
+#{{ end }}
kind: Deployment
metadata:
name: sdnc-dgbuilder
- namespace: "{{ .Values.nsPrefix }}-sdnc"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.dgbuilderReplicas }}
selector:
kind: ConfigMap
metadata:
name: sdnc-dmaap-configmap
- namespace: {{ .Values.nsPrefix }}-sdnc
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/dmaap/*").AsConfig . | indent 2 }}
#{{ end }}
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
- name: dmaap-listener
- namespace: "{{ .Values.nsPrefix }}-sdnc"
+ name: sdnc-dmaap-listener
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.dmaapReplicas }}
selector:
metadata:
labels:
app: dmaap-listener
- name: dmaap-listener
+ name: sdnc-dmaap-listener
spec:
initContainers:
- command:
apiVersion: v1
kind: ConfigMap
metadata:
- name: mysql
- namespace: "{{ .Values.nsPrefix }}-sdnc"
+ name: sdnc-mysql
+ namespace: "{{ .Values.nsPrefix }}"
labels:
app: mysql
data:
#{{ if not .Values.disableSdncSdncDbhost }}
+#{{ if not .Values.disableNfsProvisioner }}
kind: Deployment
apiVersion: extensions/v1beta1
metadata:
- name: nfs-provisioner
- namespace: "{{ .Values.nsPrefix }}-sdnc"
+ name: sdnc-nfs-provisioner
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.nfsReplicas }}
strategy:
metadata:
labels:
app: nfs-provisioner
+ name: sdnc-nfs-provisioner
spec:
containers:
- name: nfs-provisioner
hostPath:
path: {{ .Values.dataRootDir }}/{{ .Values.nsPrefix }}/sdnc/data
#{{ end }}
+#{{ end }}
kind: ConfigMap
metadata:
name: sdnc-conf-configmap
- namespace: {{ .Values.nsPrefix }}-sdnc
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/conf/*").AsConfig . | indent 2 }}
#{{ end }}
#{{ if not .Values.disableSdncSdncDbhost }}
+#{{ if not .Values.disableNfsProvisioner }}
kind: StorageClass
apiVersion: storage.k8s.io/v1
metadata:
name: "{{ .Values.nsPrefix }}-sdnc-data"
- namespace: "{{ .Values.nsPrefix }}-sdnc"
+ namespace: "{{ .Values.nsPrefix }}"
provisioner: sdnc/nfs
#{{ end }}
+#{{ end }}
kind: ConfigMap
metadata:
name: sdnc-log-configmap
- namespace: {{ .Values.nsPrefix }}-sdnc
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/log/filebeat/log4j/*").AsConfig . | indent 2 }}
---
kind: ConfigMap
metadata:
name: sdnc-logging-cfg-configmap
- namespace: {{ .Values.nsPrefix }}-sdnc
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/log/*").AsConfig . | indent 2 }}
#{{ end }}
kind: StatefulSet
metadata:
name: sdnc
- namespace: "{{ .Values.nsPrefix }}-sdnc"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
serviceName: "sdnhost-cluster"
replicas: {{ .Values.numberOfODLReplicas }}
- bash
- "-c"
- |
- sed -i 's/dbhost/sdnc-dbhost-0.dbhost.onap-sdnc.svc.cluster.local/g' /opt/onap/sdnc/data/properties/svclogic.properties
- sed -i 's/^\(org.onap.ccsdk.sli.jdbc.hosts=\).*/\1sdnc-dbhost-0.dbhost.onap-sdnc.svc.cluster.local/' /opt/onap/sdnc/data/properties/dblib.properties
+ sed -i 's/dbhost/sdnc-dbhost-0.dbhost.{{ .Values.nsPrefix }}.svc.cluster.local/g' /opt/onap/sdnc/data/properties/svclogic.properties
+ sed -i 's/^\(org.onap.ccsdk.sli.jdbc.hosts=\).*/\1sdnc-dbhost-0.dbhost.{{ .Values.nsPrefix }}.svc.cluster.local/' /opt/onap/sdnc/data/properties/dblib.properties
/opt/onap/sdnc/bin/startODL.sh
env:
- name: MYSQL_ROOT_PASSWORD
kind: ConfigMap
metadata:
name: sdnc-ueb-configmap
- namespace: {{ .Values.nsPrefix }}-sdnc
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/ueb/*").AsConfig . | indent 2 }}
#{{ end }}
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
- name: ueb-listener
- namespace: "{{ .Values.nsPrefix }}-sdnc"
+ name: sdnc-ueb-listener
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.uebReplicas }}
selector:
metadata:
labels:
app: ueb-listener
- name: ueb-listener
+ name: sdnc-ueb-listener
spec:
initContainers:
- command:
kind: Deployment
metadata:
name: sdnc-portal
- namespace: "{{ .Values.nsPrefix }}-sdnc"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.portalReplicas }}
selector:
portalReplicas: 1
disableSdncSdncDgbuilder: false
disableSdncSdncPortal: false
+disableNfsProvisioner: false
+
--- /dev/null
+# Patterns to ignore when building packages.
+# This supports shell glob matching, relative path matching, and
+# negation (prefixed with !). Only one pattern per line.
+.DS_Store
+# Common VCS dirs
+.git/
+.gitignore
+.bzr/
+.bzrignore
+.hg/
+.hgignore
+.svn/
+# Common backup files
+*.swp
+*.bak
+*.tmp
+*~
+# Various IDEs
+.project
+.idea/
+*.tmproj
--- /dev/null
+apiVersion: v1
+description: An ONAP environment setup chart
+name: setup
+version: 2.0.0
--- /dev/null
+dependencies:
+ - name: common
+ version: ~2.0.0
+ # local reference to common chart, as it is
+ # a part of this chart's package and will not
+ # be published independently to a repo (at this point)
+ repository: '@local'
\ No newline at end of file
-
apiVersion: rbac.authorization.k8s.io/v1beta1
kind: ClusterRoleBinding
metadata:
- name: {{ include "common.fullname" . }}
+ name: {{ include "common.namespace" . }}-binding
namespace: {{ include "common.namespace" . }}
labels:
app: {{ include "common.name" . }}
apiVersion: v1
kind: Secret
metadata:
- name: {{ include "common.name" . }}-docker-registry-key
+ name: {{ include "common.namespace" . }}-docker-registry-key
namespace: {{ include "common.namespace" . }}
labels:
- app: {{ include "common.fullname" . }}
+ app: {{ include "common.name" . }}
chart: {{ .Chart.Name }}-{{ .Chart.Version | replace "+" "_" }}
release: {{ .Release.Name }}
heritage: {{ .Release.Service }}
--- /dev/null
+global:
+ # image repositories
+ repository: nexus3.onap.org:10001
+ repositorySecret: eyJuZXh1czMub25hcC5vcmc6MTAwMDEiOnsidXNlcm5hbWUiOiJkb2NrZXIiLCJwYXNzd29yZCI6ImRvY2tlciIsImVtYWlsIjoiQCIsImF1dGgiOiJaRzlqYTJWeU9tUnZZMnRsY2c9PSJ9fQ==
http://{{ . }}
{{- end }}
{{- else if contains "NodePort" .Values.service.type }}
- export NODE_PORT=$(kubectl get --namespace {{ .Release.Namespace }} -o jsonpath="{.spec.ports[0].nodePort}" services {{ include "common.fullname" . }})
- export NODE_IP=$(kubectl get nodes --namespace {{ .Release.Namespace }} -o jsonpath="{.items[0].status.addresses[0].address}")
+ export NODE_PORT=$(kubectl get --namespace {{ include "common.namespace" . }} -o jsonpath="{.spec.ports[0].nodePort}" services {{ include "common.name" . }})
+ export NODE_IP=$(kubectl get nodes --namespace {{ include "common.namespace" . }} -o jsonpath="{.items[0].status.addresses[0].address}")
echo http://$NODE_IP:$NODE_PORT
{{- else if contains "LoadBalancer" .Values.service.type }}
NOTE: It may take a few minutes for the LoadBalancer IP to be available.
- You can watch the status of by running 'kubectl get svc -w {{ include "common.fullname" . }}'
- export SERVICE_IP=$(kubectl get svc --namespace {{ .Release.Namespace }} {{ include "common.fullname" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')
+ You can watch the status of by running 'kubectl get svc -w {{ include "common.name" . }}'
+ export SERVICE_IP=$(kubectl get svc --namespace {{ include "common.namespace" . }} {{ include "common.name" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')
echo http://$SERVICE_IP:{{ .Values.service.externalPort }}
{{- else if contains "ClusterIP" .Values.service.type }}
- export POD_NAME=$(kubectl get pods --namespace {{ .Release.Namespace }} -l "app={{ include "common.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
+ export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ include "common.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
echo "Visit http://127.0.0.1:8080 to use your application"
kubectl port-forward $POD_NAME 8080:{{ .Values.service.internalPort }}
{{- end }}
hostPath:
path: /etc/localtime
imagePullSecrets:
- - name: "{{ include "common.name" . }}-docker-registry-key"
+ - name: "{{ include "common.namespace" . }}-docker-registry-key"
name: {{ include "common.fullname" . }}
namespace: {{ include "common.namespace" . }}
labels:
- app: {{ include "common.fullname" . }}
+ app: {{ include "common.name" . }}
chart: "{{ .Chart.Name }}-{{ .Chart.Version | replace "+" "_" }}"
release: "{{ .Release.Name }}"
heritage: "{{ .Release.Service }}"
+ name: {{ include "common.fullname" . }}
spec:
capacity:
storage: {{ .Values.persistence.size}}
persistentVolumeReclaimPolicy: {{ .Values.persistence.volumeReclaimPolicy }}
hostPath:
path: {{ .Values.global.persistence.mountPath | default .Values.persistence.mountPath }}/{{ .Release.Name }}/{{ .Values.persistence.mountSubPath }}
-{{- end -}}
\ No newline at end of file
+{{- end -}}
name: {{ include "common.fullname" . }}
namespace: {{ include "common.namespace" . }}
labels:
- app: {{ include "common.fullname" . }}
+ app: {{ include "common.name" . }}
chart: "{{ .Chart.Name }}-{{ .Chart.Version }}"
release: "{{ .Release.Name }}"
heritage: "{{ .Release.Service }}"
{{ toYaml .Values.persistence.annotations | indent 4 }}
{{- end }}
spec:
+ selector:
+ matchLabels:
+ name: {{ include "common.fullname" . }}
accessModes:
- {{ .Values.persistence.accessMode }}
resources:
name: {{ include "common.fullname" . }}
namespace: {{ include "common.namespace" . }}
labels:
- app: {{ include "common.fullname" . }}
+ app: {{ include "common.name" . }}
chart: {{ .Chart.Name }}-{{ .Chart.Version | replace "+" "_" }}
release: {{ .Release.Name }}
heritage: {{ .Release.Service }}
type: Opaque
data:
db-root-password: {{ .Values.config.mariadbRootPassword | b64enc | quote }}
----
-apiVersion: v1
-kind: Secret
-metadata:
- name: {{ include "common.name" . }}-docker-registry-key
- namespace: {{ include "common.namespace" . }}
- labels:
- app: {{ include "common.fullname" . }}
- chart: {{ .Chart.Name }}-{{ .Chart.Version | replace "+" "_" }}
- release: {{ .Release.Name }}
- heritage: {{ .Release.Service }}
-data:
- .dockercfg: {{ .Values.global.repositorySecret | default .Values.repositorySecret }}
-type: kubernetes.io/dockercfg
\ No newline at end of file
dependencies:
- - name: common-templates
+ - name: common
version: ~2.0.0
# local reference to common chart, as it is
# a part of this chart's package and will not
# be published independently to a repo (at this point)
- repository: file://../common/common-templates
\ No newline at end of file
+ repository: '@local'
\ No newline at end of file
http://{{ . }}
{{- end }}
{{- else if contains "NodePort" .Values.service.type }}
- export NODE_PORT=$(kubectl get --namespace {{ .Release.Namespace }} -o jsonpath="{.spec.ports[0].nodePort}" services {{ include "common.fullname" . }})
- export NODE_IP=$(kubectl get nodes --namespace {{ .Release.Namespace }} -o jsonpath="{.items[0].status.addresses[0].address}")
+ export NODE_PORT=$(kubectl get --namespace {{ include "common.namespace" . }} -o jsonpath="{.spec.ports[0].nodePort}" services {{ include "common.name" . }})
+ export NODE_IP=$(kubectl get nodes --namespace {{ include "common.namespace" . }} -o jsonpath="{.items[0].status.addresses[0].address}")
echo http://$NODE_IP:$NODE_PORT
{{- else if contains "LoadBalancer" .Values.service.type }}
NOTE: It may take a few minutes for the LoadBalancer IP to be available.
- You can watch the status of by running 'kubectl get svc -w {{ include "common.fullname" . }}'
- export SERVICE_IP=$(kubectl get svc --namespace {{ .Release.Namespace }} {{ include "common.fullname" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')
+ You can watch the status of by running 'kubectl get svc -w {{ include "common.name" . }}'
+ export SERVICE_IP=$(kubectl get svc --namespace {{ include "common.namespace" . }} {{ include "common.name" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')
echo http://$SERVICE_IP:{{ .Values.service.externalPort }}
{{- else if contains "ClusterIP" .Values.service.type }}
- export POD_NAME=$(kubectl get pods --namespace {{ .Release.Namespace }} -l "app={{ template "so.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
+ export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "so.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
echo "Visit http://127.0.0.1:8080 to use your application"
kubectl port-forward $POD_NAME 8080:{{ .Values.service.internalPort }}
{{- end }}
+++ /dev/null
-
-apiVersion: rbac.authorization.k8s.io/v1beta1
-kind: ClusterRoleBinding
-metadata:
- name: {{ include "common.fullname" . }}
- namespace: {{ include "common.namespace" . }}
- labels:
- app: {{ include "common.name" . }}
- chart: {{ .Chart.Name }}-{{ .Chart.Version | replace "+" "_" }}
- release: {{ .Release.Name }}
- heritage: {{ .Release.Service }}
-roleRef:
- apiGroup: rbac.authorization.k8s.io
- kind: ClusterRole
- name: cluster-admin
-subjects:
- - kind: ServiceAccount
- name: default
- namespace: {{ include "common.namespace" . }}
\ No newline at end of file
path: start-jboss-server.sh
mode: 0755
imagePullSecrets:
- - name: "{{ include "common.name" . }}-docker-registry-key"
\ No newline at end of file
+ - name: "{{ include "common.namespace" . }}-docker-registry-key"
+++ /dev/null
-apiVersion: v1
-kind: Namespace
-metadata:
- name: {{ include "common.namespace" . }}
\ No newline at end of file
nodePortPrefix: 302
repositorySecret: eyJuZXh1czMub25hcC5vcmc6MTAwMDEiOnsidXNlcm5hbWUiOiJkb2NrZXIiLCJwYXNzd29yZCI6ImRvY2tlciIsImVtYWlsIjoiQCIsImF1dGgiOiJaRzlqYTJWeU9tUnZZMnRsY2c9PSJ9fQ==
readinessRepository: oomk8s
- readinessImage: readiness-check:1.0.0
+ readinessImage: readiness-check:1.1.0
loggingRepository: docker.elastic.co
loggingImage: beats/filebeat:5.5.0
+subChartsOnly:
+ enabled: true
# application image
repository: nexus3.onap.org:10001
labels:
app: uui
name: uui
- namespace: "{{ .Values.nsPrefix }}-uui"
+ namespace: "{{ .Values.nsPrefix }}"
annotations:
msb.onap.org/service-info: '[
{
"serviceName": "usecaseui-gui",
"version": "v1",
- "url": "/iui/usecaseui",
+ "url": "/usecase-ui",
+ "path":"/iui/usecaseui",
"protocol": "UI"
"port": "8080",
"visualRange":"1|0"
labels:
app: uui-server
name: uui-server
- namespace: "{{ .Values.nsPrefix }}-uui"
+ namespace: "{{ .Values.nsPrefix }}"
annotations:
msb.onap.org/service-info: '[
{
labels:
app: uui
name: uui
- namespace: "{{ .Values.nsPrefix }}-uui"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
+ replicas: {{ .Values.uuiReplicas }}
selector:
matchLabels:
app: uui
name: uui
env:
- name: MSB_ADDR
- value: {{ .Values.msbaddr }}
+ value: {{ tpl .Values.msbaddr . }}
ports:
- containerPort: 8080
readinessProbe:
periodSeconds: 10
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
labels:
app: uui-server
name: uui-server
- namespace: "{{ .Values.nsPrefix }}-uui"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
+ replicas: {{ .Values.uuiServerReplicas }}
selector:
matchLabels:
app: uui-server
name: uui-server
env:
- name: MSB_ADDR
- value: {{ .Values.msbaddr }}
+ value: {{tpl .Values.msbaddr .}}
- name: MR_ADDR
- value: {{ .Values.mraddr }}
+ value: {{tpl .Values.mraddr .}}
ports:
- containerPort: 8082
readinessProbe:
periodSeconds: 10
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
nsPrefix: onap
pullPolicy: Always
uuiPortPrefix: 303
-msbaddr: msb-iag.onap-msb:80
-mraddr: dmaap.onap-message-router:3904
+msbaddr: msb-iag.{{ .Values.nsPrefix }}:80
+mraddr: dmaap.{{ .Values.nsPrefix }}:3904
+uuiReplicas: 1
+uuiServerReplicas: 1
image:
uuiImage: nexus3.onap.org:10001/onap/usecase-ui
uuiVersion: v1.0.1
labels:
app: vfc-nslcm
name: vfc-nslcm
- namespace: "{{ .Values.nsPrefix }}-vfc"
+ namespace: "{{ .Values.nsPrefix }}"
annotations:
msb.onap.org/service-info: '[
{
labels:
app: vfc-resmgr
name: vfc-resmgr
- namespace: "{{ .Values.nsPrefix }}-vfc"
+ namespace: "{{ .Values.nsPrefix }}"
annotations:
msb.onap.org/service-info: '[
{
labels:
app: vfc-gvnfmdriver
name: vfc-gvnfmdriver
- namespace: "{{ .Values.nsPrefix }}-vfc"
+ namespace: "{{ .Values.nsPrefix }}"
annotations:
msb.onap.org/service-info: '[
{
labels:
app: vfc-ztevnfmdriver
name: vfc-ztevnfmdriver
- namespace: "{{ .Values.nsPrefix }}-vfc"
+ namespace: "{{ .Values.nsPrefix }}"
annotations:
msb.onap.org/service-info: '[
{
labels:
app: vfc-nokiavnfmdriver
name: vfc-nokiavnfmdriver
- namespace: "{{ .Values.nsPrefix }}-vfc"
+ namespace: "{{ .Values.nsPrefix }}"
annotations:
msb.onap.org/service-info: '[
{
labels:
app: vfc-hwvnfmdriver
name: vfc-hwvnfmdriver
- namespace: "{{ .Values.nsPrefix }}-vfc"
+ namespace: "{{ .Values.nsPrefix }}"
annotations:
msb.onap.org/service-info: '[
{
labels:
app: vfc-ztesdncdriver
name: vfc-ztesdncdriver
- namespace: "{{ .Values.nsPrefix }}-vfc"
+ namespace: "{{ .Values.nsPrefix }}"
annotations:
msb.onap.org/service-info: '[
{
labels:
app: vfc-jujudriver
name: vfc-jujudriver
- namespace: "{{ .Values.nsPrefix }}-vfc"
+ namespace: "{{ .Values.nsPrefix }}"
annotations:
msb.onap.org/service-info: '[
{
labels:
app: vfc-vnflcm
name: vfc-vnflcm
- namespace: "{{ .Values.nsPrefix }}-vfc"
+ namespace: "{{ .Values.nsPrefix }}"
annotations:
msb.onap.org/service-info: '[
{
labels:
app: vfc-vnfres
name: vfc-vnfres
- namespace: "{{ .Values.nsPrefix }}-vfc"
+ namespace: "{{ .Values.nsPrefix }}"
annotations:
msb.onap.org/service-info: '[
{
labels:
app: vfc-vnfmgr
name: vfc-vnfmgr
- namespace: "{{ .Values.nsPrefix }}-vfc"
+ namespace: "{{ .Values.nsPrefix }}"
annotations:
msb.onap.org/service-info: '[
{
labels:
app: vfc-emsdriver
name: vfc-emsdriver
- namespace: "{{ .Values.nsPrefix }}-vfc"
+ namespace: "{{ .Values.nsPrefix }}"
annotations:
msb.onap.org/service-info: '[
{
labels:
app: vfc-workflow
name: vfc-workflow
- namespace: "{{ .Values.nsPrefix }}-vfc"
+ namespace: "{{ .Values.nsPrefix }}"
annotations:
msb.onap.org/service-info: '[
{
labels:
app: vfc-workflowengineactiviti
name: vfc-workflowengineactiviti
- namespace: "{{ .Values.nsPrefix }}-vfc"
+ namespace: "{{ .Values.nsPrefix }}"
annotations:
msb.onap.org/service-info: '[
{
labels:
app: vfc-catalog
name: vfc-catalog
- namespace: "{{ .Values.nsPrefix }}-vfc"
+ namespace: "{{ .Values.nsPrefix }}"
annotations:
msb.onap.org/service-info: '[
{
kind: Deployment
metadata:
name: vfc-catalog
- namespace: "{{ .Values.nsPrefix }}-{{ .Chart.Name }}"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.catalog.replicas }}
selector:
name: "vfc-catalog"
env:
- name: MSB_ADDR
- value: {{ .Values.msbaddr }}
+ value: {{ tpl .Values.msbaddr . }}
ports:
- containerPort: {{ .Values.catalog.port }}
readinessProbe:
imagePullPolicy: "{{ .Values.pullPolicy }}"
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
kind: Deployment
metadata:
name: vfc-emsdriver
- namespace: "{{ .Values.nsPrefix }}-{{ .Chart.Name }}"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.emsdriver.replicas }}
selector:
name: "vfc-emsdriver"
env:
- name: MSB_ADDR
- value: {{ .Values.msbaddr }}
+ value: {{ tpl .Values.msbaddr . }}
ports:
- containerPort: {{ .Values.emsdriver.port }}
readinessProbe:
imagePullPolicy: "{{ .Values.pullPolicy }}"
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
kind: Deployment
metadata:
name: vfc-gvnfmdriver
- namespace: "{{ .Values.nsPrefix }}-{{ .Chart.Name }}"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.gvnfmdriver.replicas }}
selector:
name: "vfc-gvnfmdriver"
env:
- name: MSB_ADDR
- value: {{ .Values.msbaddr }}
+ value: {{ tpl .Values.msbaddr . }}
ports:
- containerPort: {{ .Values.gvnfmdriver.port }}
readinessProbe:
imagePullPolicy: "{{ .Values.pullPolicy }}"
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
kind: Deployment
metadata:
name: vfc-hwvnfmdriver
- namespace: "{{ .Values.nsPrefix }}-{{ .Chart.Name }}"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.hwvnfmdriver.replicas }}
selector:
name: "vfc-hwvnfmdriver"
env:
- name: MSB_ADDR
- value: {{ .Values.msbaddr }}
+ value: {{ tpl .Values.msbaddr . }}
ports:
- containerPort: {{ .Values.hwvnfmdriver.port }}
readinessProbe:
imagePullPolicy: "{{ .Values.pullPolicy }}"
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
kind: Deployment
metadata:
name: vfc-jujudriver
- namespace: "{{ .Values.nsPrefix }}-{{ .Chart.Name }}"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.jujudriver.replicas }}
selector:
name: "vfc-jujudriver"
env:
- name: MSB_ADDR
- value: {{ .Values.msbaddr }}
+ value: {{ tpl .Values.msbaddr . }}
ports:
- containerPort: {{ .Values.jujudriver.port }}
readinessProbe:
imagePullPolicy: "{{ .Values.pullPolicy }}"
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
kind: Deployment
metadata:
name: vfc-nokiavnfmdriver
- namespace: "{{ .Values.nsPrefix }}-{{ .Chart.Name }}"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.nokiavnfmdriver.replicas }}
selector:
name: "vfc-nokiavnfmdriver"
env:
- name: MSB_ADDR
- value: {{ .Values.msbaddr }}
+ value: {{ tpl .Values.msbaddr . }}
ports:
- containerPort: {{ .Values.nokiavnfmdriver.port }}
readinessProbe:
imagePullPolicy: "{{ .Values.pullPolicy }}"
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
kind: Deployment
metadata:
name: vfc-nslcm
- namespace: "{{ .Values.nsPrefix }}-{{ .Chart.Name }}"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.nslcm.replicas }}
selector:
name: "vfc-nslcm"
env:
- name: MSB_ADDR
- value: {{ .Values.msbaddr }}
+ value: {{ tpl .Values.msbaddr . }}
ports:
- containerPort: {{ .Values.nslcm.port }}
readinessProbe:
imagePullPolicy: "{{ .Values.pullPolicy }}"
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
kind: Deployment
metadata:
name: vfc-resmgr
- namespace: "{{ .Values.nsPrefix }}-{{ .Chart.Name }}"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.resmgr.replicas }}
selector:
name: "vfc-resmgr"
env:
- name: MSB_ADDR
- value: {{ .Values.msbaddr }}
+ value: {{ tpl .Values.msbaddr . }}
ports:
- containerPort: {{ .Values.resmgr.port }}
readinessProbe:
imagePullPolicy: "{{ .Values.pullPolicy }}"
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
kind: Deployment
metadata:
name: vfc-vnflcm
- namespace: "{{ .Values.nsPrefix }}-{{ .Chart.Name }}"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.vnflcm.replicas }}
selector:
name: "vfc-vnflcm"
env:
- name: MSB_ADDR
- value: {{ .Values.msbaddr }}
+ value: {{ tpl .Values.msbaddr . }}
ports:
- containerPort: {{ .Values.vnflcm.port }}
readinessProbe:
imagePullPolicy: "{{ .Values.pullPolicy }}"
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
kind: Deployment
metadata:
name: vfc-vnfmgr
- namespace: "{{ .Values.nsPrefix }}-{{ .Chart.Name }}"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.vnfmgr.replicas }}
selector:
name: "vfc-vnfmgr"
env:
- name: MSB_ADDR
- value: {{ .Values.msbaddr }}
+ value: {{ tpl .Values.msbaddr . }}
ports:
- containerPort: {{ .Values.vnfmgr.port }}
readinessProbe:
imagePullPolicy: "{{ .Values.pullPolicy }}"
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
kind: Deployment
metadata:
name: vfc-vnfres
- namespace: "{{ .Values.nsPrefix }}-{{ .Chart.Name }}"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.vnfres.replicas }}
selector:
name: "vfc-vnfres"
env:
- name: MSB_ADDR
- value: {{ .Values.msbaddr }}
+ value: {{ tpl .Values.msbaddr . }}
ports:
- containerPort: {{ .Values.vnfres.port }}
readinessProbe:
imagePullPolicy: "{{ .Values.pullPolicy }}"
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
kind: Deployment
metadata:
name: vfc-workflow
- namespace: "{{ .Values.nsPrefix }}-{{ .Chart.Name }}"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.workflow.replicas }}
selector:
- name: SERVICE_PORT
value: "{{ .Values.workflow.port }}"
- name: OPENPALETTE_MSB_IP
- value: msb-iag.{{ .Values.nsPrefix }}-msb
+ value: msb-iag.{{ .Values.nsPrefix }}
- name: OPENPALETTE_MSB_PORT
value: "80"
ports:
imagePullPolicy: "{{ .Values.pullPolicy }}"
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
kind: Deployment
metadata:
name: vfc-workflowengineactiviti
- namespace: "{{ .Values.nsPrefix }}-{{ .Chart.Name }}"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.workflowengineactiviti.replicas }}
selector:
- name: SERVICE_PORT
value: "{{ .Values.workflow.port }}"
- name: OPENPALETTE_MSB_IP
- value: msb-iag.{{ .Values.nsPrefix }}-msb
+ value: msb-iag.{{ .Values.nsPrefix }}
- name: OPENPALETTE_MSB_PORT
value: "80"
ports:
imagePullPolicy: "{{ .Values.pullPolicy }}"
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
kind: Deployment
metadata:
name: vfc-ztesdncdriver
- namespace: "{{ .Values.nsPrefix }}-{{ .Chart.Name }}"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.ztesdncdriver.replicas }}
selector:
name: "vfc-ztesdncdriver"
env:
- name: MSB_ADDR
- value: {{ .Values.msbaddr }}
+ value: {{ tpl .Values.msbaddr . }}
ports:
- containerPort: {{ .Values.ztesdncdriver.port }}
readinessProbe:
imagePullPolicy: "{{ .Values.pullPolicy }}"
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
kind: Deployment
metadata:
name: vfc-ztevnfmdriver
- namespace: "{{ .Values.nsPrefix }}-{{ .Chart.Name }}"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.ztevnfmdriver.replicas }}
selector:
name: "vfc-ztevnfmdriver"
env:
- name: MSB_ADDR
- value: {{ .Values.msbaddr }}
+ value: {{ tpl .Values.msbaddr . }}
ports:
- containerPort: {{ .Values.ztevnfmdriver.port }}
readinessProbe:
imagePullPolicy: "{{ .Values.pullPolicy }}"
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
-#{{ end }}
\ No newline at end of file
+#{{ end }}
nsPrefix: onap
pullPolicy: IfNotPresent
-msbaddr: msb-iag.onap-msb:80
+msbaddr: msb-iag.{{ .Values.nsPrefix }}:80
nslcm:
image: nexus3.onap.org:10001/onap/vfc/nslcm:v1.0.2
port: 8403
image: nexus3.onap.org:10001/onap/vfc/catalog:v1.0.2
port: 8806
nodePort: 30806
- replicas: 1
\ No newline at end of file
+ replicas: 1
#List of logstash server ip addresses with port number.
#But, in our case, this will be the loadbalancer IP address.
#For the below property to work the loadbalancer or logstash should expose 5044 port to listen the filebeat events or port in the property should be changed appropriately.
- hosts: ["logstash.onap-log:5044"]
+ hosts: ["logstash.{{.Values.nsPrefix}}:5044"]
#If enable will do load balancing among availabe Logstash, automatically.
loadbalance: true
kind: Service
metadata:
name: vid-mariadb
- namespace: "{{ .Values.nsPrefix }}-vid"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
ports:
- name: vid-mariadb
labels:
app: vid-server
name: vid-server
- namespace: "{{ .Values.nsPrefix }}-vid"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
ports:
- name: vid-server
kind: ConfigMap
metadata:
name: vid-lfconfig-configmap
- namespace: {{ .Values.nsPrefix }}-vid
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/lf_config/*").AsConfig . | indent 2 }}
#{{ end }}
kind: ConfigMap
metadata:
name: vid-log-configmap
- namespace: {{ .Values.nsPrefix }}-vid
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/log/vid/*").AsConfig . | indent 2 }}
---
kind: ConfigMap
metadata:
name: vid-filebeat-configmap
- namespace: {{ .Values.nsPrefix }}-vid
+ namespace: {{ .Values.nsPrefix }}
data:
{{ tpl (.Files.Glob "resources/config/log/filebeat/*").AsConfig . | indent 2 }}
#{{ end }}
labels:
app: vid-mariadb
name: vid-mariadb
- namespace: "{{ .Values.nsPrefix }}-vid"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.vidMariaDbReplicas }}
selector:
apiVersion: v1
kind: PersistentVolume
metadata:
- name: "{{ .Values.nsPrefix }}-vid-db"
- namespace: "{{ .Values.nsPrefix }}-vid"
+ name: "{{ .Values.nsPrefix }}-vid"
+ namespace: "{{ .Values.nsPrefix }}"
labels:
- name: "{{ .Values.nsPrefix }}-vid-db"
+ name: "{{ .Values.nsPrefix }}-vid"
spec:
capacity:
storage: 2Gi
apiVersion: v1
metadata:
name: vid-db
- namespace: "{{ .Values.nsPrefix }}-vid"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
accessModes:
- ReadWriteMany
storage: 2Gi
selector:
matchLabels:
- name: "{{ .Values.nsPrefix }}-vid-db"
+ name: "{{ .Values.nsPrefix }}-vid"
#{{ end }}
labels:
app: vid-server
name: vid-server
- namespace: "{{ .Values.nsPrefix }}-vid"
+ namespace: "{{ .Values.nsPrefix }}"
spec:
replicas: {{ .Values.vidServerReplicas }}
selector:
containers:
- env:
- name: ASDC_CLIENT_REST_HOST
- value: sdc-be.{{ .Values.nsPrefix }}-sdc
+ value: sdc-be.{{ .Values.nsPrefix }}
- name: ASDC_CLIENT_REST_AUTH
value: Basic dmlkOktwOGJKNFNYc3pNMFdYbGhhazNlSGxjc2UyZ0F3ODR2YW9HR21KdlV5MlU=
- name: ASDC_CLIENT_REST_PORT
value: "8080"
- name: VID_AAI_HOST
- value: aai-service.{{ .Values.nsPrefix }}-aai
+ value: aai-service.{{ .Values.nsPrefix }}
- name: VID_AAI_PORT
value: "8443"
- name: VID_ECOMP_SHARED_CONTEXT_REST_URL
- value: http://portalapps.{{ .Values.nsPrefix }}-portal:8989/ONAPPORTAL/context
+ value: http://portalapps.{{ .Values.nsPrefix }}:8989/ONAPPORTAL/context
- name: VID_MSO_SERVER_URL
- value: http://mso.{{ .Values.nsPrefix }}-mso:8080
+ value: http://mso.{{ .Values.nsPrefix }}:8080
- name: VID_MSO_PASS
value: OBF:1ih71i271vny1yf41ymf1ylz1yf21vn41hzj1icz
- name: MSO_DME2_SERVER_URL
- name: MSO_DME2_ENABLED
value: "false"
- name: VID_ECOMP_REDIRECT_URL
- value: http://portalapps.{{ .Values.nsPrefix }}-portal:8989/ONAPPORTAL/login.htm
+ value: http://portalapps.{{ .Values.nsPrefix }}:8989/ONAPPORTAL/login.htm
- name: VID_ECOMP_REST_URL
- value: http://portalapps.{{ .Values.nsPrefix }}-portal:8989/ONAPPORTAL/auxapi
+ value: http://portalapps.{{ .Values.nsPrefix }}:8989/ONAPPORTAL/auxapi
- name: VID_CONTACT_US_LINK
value: https://todo_contact_us_link.com
- name: VID_UEB_URL_LIST
- value: dmaap.{{ .Values.nsPrefix }}-message-router
+ value: dmaap.{{ .Values.nsPrefix }}
- name: VID_MYSQL_HOST
value: vid-mariadb
- name: VID_MYSQL_PORT
kind: Service
metadata:
name: postgres
- namespace: {{ .Values.nsPrefix }}-vnfsdk
+ namespace: {{ .Values.nsPrefix }}
labels:
app: postgres
spec:
labels:
app: refrepo
name: refrepo
- namespace: {{ .Values.nsPrefix }}-vnfsdk
+ namespace: {{ .Values.nsPrefix }}
spec:
ports:
- name: "refrepo-port"
apiVersion: extensions/v1beta1
kind: Deployment
metadata:
- name: postgres
- namespace: {{ .Values.nsPrefix }}-vnfsdk
+ name: vnfsdk-postgres
+ namespace: {{ .Values.nsPrefix }}
spec:
replicas: 1
selector:
metadata:
labels:
app: postgres
- name: postgres
+ name: vnfsdk-postgres
spec:
hostname: postgres
containers:
name: "postgres"
imagePullSecrets:
- name: {{ .Values.nsPrefix }}-docker-registry-key
-#{{ end }}
\ No newline at end of file
+#{{ end }}
metadata:
labels:
app: refrepo
- name: refrepo
- namespace: "{{ .Values.nsPrefix }}-vnfsdk"
+ name: vnfsdk-refrepo
+ namespace: "{{ .Values.nsPrefix }}"
spec:
selector:
matchLabels:
metadata:
labels:
app: refrepo
- name: refrepo
+ name: vnfsdk-refrepo
spec:
initContainers:
- command: