The main goal of the Amsterdam release was to:
- - Support Flexible Platform Deployment via Kubernetes of fully containerized OMAP components - on any type of environment.
+ - Support Flexible Platform Deployment via Kubernetes of fully containerized ONAP components - on any type of environment.
- Support State Management of ONAP platform components.
- Support full production ONAP deployment and any variation of component level deployment for development.
- Platform Operations Orchestration / Control Loop Actions.
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/aaf/aaf_cm:2.1.5-SNAPSHOT
+image: onap/aaf/aaf_cm:2.1.5
aaf_register_as: "aaf-cm.onap"
pullPolicy: Always
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/aaf/aaf_fs:2.1.5-SNAPSHOT
+image: onap/aaf/aaf_fs:2.1.5
aaf_register_as: "aaf-fs.onap"
pullPolicy: Always
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/aaf/aaf_gui:2.1.5-SNAPSHOT
+image: onap/aaf/aaf_gui:2.1.5
aaf_register_as: "aaf-gui.onap"
pullPolicy: Always
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/aaf/aaf_hello:2.1.5-SNAPSHOT
+image: onap/aaf/aaf_hello:2.1.5
aaf_register_as: "aaf-hello.onap"
pullPolicy: Always
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/aaf/aaf_locate:2.1.5-SNAPSHOT
+image: onap/aaf/aaf_locate:2.1.5
aaf_register_as: "aaf-locate.onap"
pullPolicy: Always
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/aaf/aaf_oauth:2.1.5-SNAPSHOT
+image: onap/aaf/aaf_oauth:2.1.5
aaf_register_as: "aaf-oauth.onap"
pullPolicy: Always
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/aaf/aaf_service:2.1.5-SNAPSHOT
+image: onap/aaf/aaf_service:2.1.5
aaf_register_as: "aaf-service.onap"
pullPolicy: Always
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/aaf/smsquorumclient:latest
+image: onap/aaf/smsquorumclient:3.0.0
pullPolicy: Always
# flag to enable debugging - application support required
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/aaf/sms:latest
+image: onap/aaf/sms:3.0.0
pullPolicy: Always
# flag to enable debugging - application support required
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/aaf/abrmd:latest
+image: onap/aaf/abrmd:3.0.0
pullPolicy: Always
# flag to enable debugging - application support required
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/aaf/distcenter:latest
+image: onap/aaf/distcenter:3.0.0
pullPolicy: Always
# flag to enable debugging - application support required
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/aaf/testcaservice:latest
+image: onap/aaf/testcaservice:3.0.0
pullPolicy: Always
# flag to enable debugging - application support required
#################################################################
# application image
-image: onap/babel:1.3-STAGING-latest
+image: onap/babel:1.3.0
flavor: small
#################################################################
# application image
-image: onap/champ:1.3-STAGING-latest
+image: onap/champ:1.3.0
flavor: small
type: NodePort
portName: aai-champ
internalPort: 9522
+ externalPort: 9522
nodePort: 78
ingress:
# application image
repository: nexus3.onap.org:10001
-image: onap/data-router:1.3-STAGING-latest
+image: onap/data-router:1.3.0
pullPolicy: Always
restartPolicy: Always
flavor: small
- /bin/sh
- -c
- |
+ sysctl -w vm.max_map_count=262144
mkdir -p /logroot/elasticsearch/logs
mkdir -p /logroot/elasticsearch/data
chmod -R 777 /logroot/elasticsearch
export SERVICE_IP=$(kubectl get svc --namespace {{ .Release.Namespace }} {{ include "common.fullname" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')
echo http://$SERVICE_IP:{{ .Values.service.externalPort }}
{{- else if contains "ClusterIP" .Values.service.type }}
- export POD_NAME=$(kubectl get pods --namespace {{ .Release.Namespace }} -l "app={{ template "so.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
+ export POD_NAME=$(kubectl get pods --namespace {{ .Release.Namespace }} -l "app={{ template "common.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
echo "Visit http://127.0.0.1:8080 to use your application"
kubectl port-forward $POD_NAME 8080:{{ .Values.service.internalPort }}
{{- end }}
#################################################################
# application image
-image: onap/gizmo:1.3-STAGING-latest
+image: onap/gizmo:1.3.0
flavor: small
# application configuration
config:
name: aai-crud-service
portName: aai-crud-service
internalPort: 9520
+ externalPort: 9520
nodePort: 68
ingress:
# JMS bind address host port
jms.bind.address=tcp://localhost:61649
dmaap.ribbon.listOfServers=message-router.{{.Release.Namespace}}:3905
+dmaap.ribbon.transportType=https
# Schema related attributes for the oxm and edges
# Any additional schema related attributes should start with prefix schema
imagePullPolicy: {{ .Values.global.pullPolicy | default .Values.pullPolicy }}
name: {{ include "common.name" . }}-readiness
containers:
- - name: {{ include "common.name" . }}
+ - name: {{ include "common.name" . }}-job
image: "{{ include "common.repository" . }}/{{ .Values.image }}"
imagePullPolicy: {{ .Values.global.pullPolicy | default .Values.pullPolicy }}
command:
# application image
repository: nexus3.onap.org:10001
-image: onap/aai-graphadmin:1.0-STAGING-latest
+image: onap/aai-graphadmin:1.0.0
pullPolicy: Always
restartPolicy: Always
flavor: small
# application image
repository: nexus3.onap.org:10001
-image: onap/model-loader:1.3-STAGING-latest
+image: onap/model-loader:1.3.0
pullPolicy: Always
restartPolicy: Always
flavor: small
# JMS bind address host port
jms.bind.address=tcp://localhost:61647
dmaap.ribbon.listOfServers=message-router.{{.Release.Namespace}}:3905
+dmaap.ribbon.transportType=https
# Schema related attributes for the oxm and edges
# Any additional schema related attributes should start with prefix schema
# application image
repository: nexus3.onap.org:10001
-image: onap/aai-resources:1.3-STAGING-latest
+image: onap/aai-resources:1.3.0
pullPolicy: Always
restartPolicy: Always
flavor: small
# application image
repository: nexus3.onap.org:10001
-image: onap/search-data-service:1.3-STAGING-latest
+image: onap/search-data-service:1.3.1
pullPolicy: Always
restartPolicy: Always
flavor: small
# application image
repository: nexus3.onap.org:10001
-image: onap/sparky-be:1.3-STAGING-latest
+image: onap/sparky-be:1.3.0
pullPolicy: Always
restartPolicy: Always
flavor: small
# JMS bind address host port
jms.bind.address=tcp://localhost:61647
dmaap.ribbon.listOfServers=message-router.{{ include "common.namespace" . }}:3905
+dmaap.ribbon.transportType=https
# Schema related attributes for the oxm and edges
# Any additional schema related attributes should start with prefix schema
- mountPath: /opt/app/aai-traversal/resources/etc/appprops/aaiconfig.properties
name: {{ include "common.fullname" . }}-aaiconfig-conf
subPath: aaiconfig.properties
- - mountPath: /var/log/onap
+ - mountPath: /opt/aai/logroot/AAI-GQ/
name: {{ include "common.fullname" . }}-logs
- mountPath: /opt/app/aai-traversal/resources/logback.xml
name: {{ include "common.fullname" . }}-log-conf
# application image
repository: nexus3.onap.org:10001
-image: onap/aai-traversal:1.3-STAGING-latest
+image: onap/aai-traversal:1.3.0
pullPolicy: Always
restartPolicy: Always
flavor: small
- /root/ready.py
args:
- --container-name
+ - aai-resources
+ - --container-name
- aai-traversal
+ - --container-name
+ - aai-graphadmin
env:
- name: NAMESPACE
valueFrom:
periodSeconds: {{ .Values.liveness.periodSeconds }}
{{ end -}}
readinessProbe:
- tcpSocket:
+ httpGet:
+ path: /aai/util/echo
port: {{ .Values.service.internalPort2 }}
+ scheme: HTTPS
+ httpHeaders:
+ - name: X-FromAppId
+ value: OOM_ReadinessCheck
+ - name: X-TransactionId
+ value: {{ uuidv4 }}
+ - name: Accept
+ value: application/json
initialDelaySeconds: {{ .Values.readiness.initialDelaySeconds }}
periodSeconds: {{ .Values.readiness.periodSeconds }}
resources:
name: {{ .Values.service.portName }}
- port: {{ .Values.service.externalPort2 }}
targetPort: {{ .Values.service.internalPort2 }}
- name: {{ .Values.service.portName }}
+ name: {{ .Values.service.portName2 }}
{{- end}}
type: {{ .Values.service.type }}
selector:
# application image
repository: nexus3.onap.org:10001
-image: onap/sdnc-ansible-server-image:1.4-STAGING-latest
+image: onap/sdnc-ansible-server-image:1.4.1
pullPolicy: Always
# flag to enable debugging - application support required
# application image
repository: nexus3.onap.org:10001
-image: onap/appc-cdt-image:1.4.0-SNAPSHOT-latest
+image: onap/appc-cdt-image:1.4.0
pullPolicy: Always
# application configuration
export SERVICE_IP=$(kubectl get svc --namespace {{ .Release.Namespace }} {{ include "common.fullname" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')
echo http://$SERVICE_IP:{{ .Values.service.externalPort }}
{{- else if contains "ClusterIP" .Values.service.type }}
- export POD_NAME=$(kubectl get pods --namespace {{ .Release.Namespace }} -l "app={{ template "so.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
+ export POD_NAME=$(kubectl get pods --namespace {{ .Release.Namespace }} -l "app={{ template "common.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
echo "Visit http://127.0.0.1:8080 to use your application"
kubectl port-forward $POD_NAME 8080:{{ .Values.service.internalPort }}
{{- end }}
flavor: small
# application image
repository: nexus3.onap.org:10001
-image: onap/appc-image:1.4.0-SNAPSHOT-latest
+image: onap/appc-image:1.4.0
pullPolicy: Always
# flag to enable debugging - application support required
name: {{ include "common.name" . }}-readiness
containers:
- name: {{ include "common.name" . }}
- image: "{{ .Values.global.imageRepository | default .Values.imageRepository }}/{{ .Values.image }}"
+ image: "{{ include "common.repository" . }}/{{ .Values.image }}"
imagePullPolicy: {{ .Values.global.pullPolicy | default .Values.pullPolicy }}
ports:
- containerPort: {{ .Values.service.internalPort }}
busyboxImage: library/busybox:latest
# application image
-imageRepository: nexus3.onap.org:10001
-image: onap/clamp-dashboard-kibana:3.0-STAGING-latest
+repository: nexus3.onap.org:10001
+image: onap/clamp-dashboard-kibana:3.0.0
pullPolicy: Always
# flag to enable debugging - application support required
name: {{ include "common.name" . }}-readiness
containers:
- name: {{ include "common.name" . }}
- image: "{{ .Values.global.imageRepository| default .Values.imageRepository }}/{{ .Values.image }}"
+ image: "{{ include "common.repository" . }}/{{ .Values.image }}"
imagePullPolicy: {{ .Values.global.pullPolicy | default .Values.pullPolicy }}
env:
- name: dmaap_consumer_group
#################################################################
# application image
-imageRepository: nexus3.onap.org:10001
-image: onap/clamp-dashboard-logstash:3.0-STAGING-latest
+repository: nexus3.onap.org:10001
+image: onap/clamp-dashboard-logstash:3.0.0
pullPolicy: Always
# flag to enable debugging - application support required
export SERVICE_IP=$(kubectl get svc --namespace {{ include "common.namespace" . }} {{ include "common.name" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')
echo http://$SERVICE_IP:{{ .Values.service.externalPort }}
{{- else if contains "ClusterIP" .Values.service.type }}
- export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "so.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
+ export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "common.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
echo "Visit http://127.0.0.1:8080 to use your application"
kubectl port-forward $POD_NAME 8080:{{ .Values.service.internalPort }}
{{- end }}
# application image
repository: nexus3.onap.org:10001
-image: onap/clamp:3.0-STAGING-latest
+image: onap/clamp:3.0.0
pullPolicy: Always
# flag to enable debugging - application support required
"clamp.config.sdc.servicePassword": "b7acccda32b98c5bb7acccda32b98c5b05D511BD6D93626E90D18E9D24D9B78CD34C7EE8012F0A189A28763E82271E50A5D4EC10C7D93E06E0A2D27CAE66B981",
"clamp.config.files.sdcController": "file:/opt/clamp/sdc-controllers-config.json",
"clamp.config.dcae.inventory.url": "http://inventory.{{ include "common.namespace" . }}:8080",
- "clamp.config.dcae.dispatcher.url": "http://deployment-handler.{{ include "common.namespace" . }}:8443",
- "clamp.config.policy.pdpUrl1": "http://pdp.{{ include "common.namespace" . }}:8081/pdp/ , testpdp, alpha123",
- "clamp.config.policy.pdpUrl2": "http://pdp.{{ include "common.namespace" . }}:8081/pdp/ , testpdp, alpha123",
- "clamp.config.policy.papUrl": "http://pap.{{ include "common.namespace" . }}:9091/pap/ , testpap, alpha123",
+ "clamp.config.dcae.dispatcher.url": "https://deployment-handler.{{ include "common.namespace" . }}:8443",
+ "clamp.config.policy.pdpUrl1": "https://pdp.{{ include "common.namespace" . }}:8081/pdp/ , testpdp, alpha123",
+ "clamp.config.policy.pdpUrl2": "https://pdp.{{ include "common.namespace" . }}:8081/pdp/ , testpdp, alpha123",
+ "clamp.config.policy.papUrl": "https://pap.{{ include "common.namespace" . }}:9091/pap/ , testpap, alpha123",
"clamp.config.policy.clientKey": "dGVzdA==",
- "clamp.config.cadi.aafLocateUrl": "https://aaf-onap-test.osaaf.org:8095",
+ "clamp.config.cadi.aafLocateUrl": "https://aaf-locate.{{ include "common.namespace" . }}:8095",
"com.att.eelf.logging.path": "/opt/clamp",
"com.att.eelf.logging.file": "logback.xml"
}
name: clamp
portName: clamp
internalPort: 8080
+ externalPort: 8080
nodePort: 95
portName2: clamp-ssl
internalPort2: 8443
+ externalPort2: 8443
# as of 20180904 port 58 is reserved for clamp from log/logdemonode
# see https://wiki.onap.org/display/DW/OOM+NodePort+List
nodePort2: 58
export SERVICE_IP=$(kubectl get svc --namespace {{ include "common.namespace" . }} {{ include "common.name" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')
echo http://$SERVICE_IP:{{ .Values.service.externalPort }}
{{- else if contains "ClusterIP" .Values.service.type }}
- export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "so.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
+ export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "common.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
echo "Visit http://127.0.0.1:8080 to use your application"
kubectl port-forward $POD_NAME 8080:{{ .Values.service.internalPort }}
{{- end }}
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/cli:2.0.2
+image: onap/cli:2.0.4
pullPolicy: Always
flavor: small
*/}}
{{- define "common.repository.secret" -}}
{{- $repo := include "common.repository" . }}
+ {{- $repo := default "nexus3.onap.org:10001" $repo }}
{{- $cred := .Values.global.repositoryCred }}
{{- $mail := default "@" $cred.mail }}
{{- $auth := printf "%s:%s" $cred.user $cred.password | b64enc }}
export SERVICE_IP=$(kubectl get svc --namespace {{ include "common.namespace" . }} {{ include "common.fullname" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')
echo http://$SERVICE_IP:{{ .Values.service.externalPort }}
{{- else if contains "ClusterIP" .Values.service.type }}
- export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "so.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
+ export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "common.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
echo "Visit http://127.0.0.1:8080 to use your application"
kubectl port-forward $POD_NAME 8080:{{ .Values.service.internalPort }}
{{- end }}
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/ccsdk-dgbuilder-image:0.3-STAGING-latest
+image: onap/ccsdk-dgbuilder-image:0.3.1
pullPolicy: Always
# flag to enable debugging - application support required
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/ccsdk-apps-ms-neng:latest
+image: onap/ccsdk-apps-ms-neng:0.3.1
pullPolicy: IfNotPresent
# application configuration
tag_version: {{ include "common.repository" . }}/{{ .Values.componentImages.datafile_collector }}
{{ end }}
host_port: 30223
-host_port_secure: 30224
+host_port_secure: 30262
dmaap_mr_host: "{{ .Values.config.address.message_router }}"
dmaap_mr_port: 3904
dmaap_mr_user: "admin"
config_binding_service: onap/org.onap.dcaegen2.platform.configbinding.app-app:2.2.3
datafile_collector: onap/org.onap.dcaegen2.collectors.datafile.datafile-app-server:1.0.3
deployment_handler: onap/org.onap.dcaegen2.platform.deployment-handler:3.0.3
- holmes_rules: onap/holmes/rule-management:1.2.0-STAGING-latest
- holmes_engine: onap/holmes/engine-management:1.2.0-STAGING-latest
+ holmes_rules: onap/holmes/rule-management:1.2.0
+ holmes_engine: onap/holmes/engine-management:1.2.0
inventory: onap/org.onap.dcaegen2.platform.inventory-api:3.0.4
policy_handler: onap/org.onap.dcaegen2.platform.policy-handler:4.4.0
service_change_handler: onap/org.onap.dcaegen2.platform.servicechange-handler:1.1.5
tca: onap/org.onap.dcaegen2.deployments.tca-cdap-container:1.1.0
ves: onap/org.onap.dcaegen2.collectors.ves.vescollector:1.3.1
snmptrap: onap/org.onap.dcaegen2.collectors.snmptrap:1.4.0
- prh: onap/org.onap.dcaegen2.services.prh.prh-app-server:1.0.0
- hv_ves: onap/org.onap.dcaegen2.collectors.hv-ves.hv-collector-main:1.0.0-SNAPSHOT
+ prh: onap/org.onap.dcaegen2.services.prh.prh-app-server:1.1.1
+ hv_ves: onap/org.onap.dcaegen2.collectors.hv-ves.hv-collector-main:1.0.0
# Resource Limit flavor -By Default using small
flavor: small
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/org.onap.dcaegen2.deployments.cm-container:1.4.1
+image: onap/org.onap.dcaegen2.deployments.cm-container:1.4.2
pullPolicy: Always
# probe configuration parameters
export SERVICE_IP=$(kubectl get svc --namespace {{ include "common.namespace" . }} {{ include "common.fullname" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')
echo http://$SERVICE_IP:{{ .Values.service.externalPort }}
{{- else if contains "ClusterIP" .Values.service.type }}
- export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "so.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
+ export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "common.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
echo "Visit http://127.0.0.1:8080 to use your application"
kubectl port-forward $POD_NAME 8080:{{ .Values.service.internalPort }}
{{- end }}
--- /dev/null
+# Copyright © 2018 AT&T, Amdocs, Bell Canada Intellectual Property. All rights reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+#####################################################
+#
+# ONAP Casablanca oom configurable deployment params:
+#
+#####################################################
+
+
+#####################################################
+#
+# Hooks for specific environment configurations
+#
+#####################################################
+# Indicator for whether to use AAF
+UseAAF: false
+
+# csit: stubs out some southbound APIs for csit
+csit: No
+
+# name of this DMaaP instance (deprecated)
+DmaapName: demo
+
+#####################################################
+#
+# Settings for Southbound API: Datarouter
+#
+#####################################################
+# FQDN of DR Prov Server (deprecated)
+DR.provhost: dcae-drps.domain.not.set
+
+# URI to retrieve dynamic DR configuration
+ProvisioningURI: /internal/prov
+
+# indicator for handling feed delete:
+# DeleteOnDR - means use the DR API to DELETE a feed. (default for backwards compatibility)
+# SimulateDelete - means preserve the feed on DR (after cleaning it up), and mark as DELETED in DBCL. Better for cfy environments.
+Feed.deleteHandling: SimulateDelete
+
+#####################################################
+#
+# Settings for Soutbound API: Postgresql
+#
+#####################################################
+# flag indieonapdemodbates if we are using postgresql
+UsePGSQL: true
+
+# postgres host name
+# Need to connect to PG primary service, designated by service.name2
+DB.host: {{ .Values.postgres.service.name2 }}
+
+# postgres schema name
+#DB.schema: {{ .Values.postgres.config.pgDatabase }}
+
+# postgres user name
+#DB.user: {{ .Values.postgres.config.pgUserName }}
+
+# postgres user password
+DB.cred: {{ .Values.postgres.config.pgUserPassword }}
+
+
+#####################################################
+#
+# Settings for Soutbound API: Message Router
+#
+#####################################################
+# indicator for multi-site (locations) deployment. Give clue to buscontroller whether
+# there is a need for message replication between edge and central.
+# ONAP Casablanca is a single site deployment
+MR.multisite: false
+
+# FQDN of primary message router.
+# In ONAP Casablanca, there is only 1 message router service, so use that.
+# In a multi-site, MR cluster deployment, use the CNAME DNS entry which resolves to the primary central MR
+MR.CentralCname: {{ .Values.dmaapMessageRouterService }}
+
+# MR Client Delete Level thoroughness:
+# 0 = don't delete
+# 1 = delete from persistent store
+# 2 = delete from persistent store (DB) and authorization store (AAF)
+MR.ClientDeleteLevel: 1
+
+# namespace of MR Topic Factory
+MR.TopicFactoryNS: org.onap.dmaap.mr.topicFactory
+
+# AAF Role assigned to Topic Manager Identity
+MR.TopicMgrRole: org.onap.dmaap-bc-topic-mgr.client
+
+# MR topic ProjectID (used in certain topic name generation formats)
+MR.projectID: ONAP
+
+
+#####################################################
+#
+# Settings for Southbound API: CADI
+#
+#####################################################
+# path to cadi.properties
+cadi.properties: /opt/app/osaaf/lcoal/org.onap.dmaap-bc.props
+
+#####################################################
+#
+# Settings for Southbound API: AAF proxy
+#
+#####################################################
+# URL of the AAF server
+aaf.URL: {{ .Values.aafURL }}
+
+# TopicMgr Identity
+aaf.TopicMgrUser: {{ .Values.topicMgrUser }}
+
+# Password for TopicMgr identity
+aaf.TopicMgrPassword: {{ .Values.topicMgrPwd }}
+
+# Buscontroller Admin Identity
+aaf.AdminUser: {{ .Values.adminUser }}
+
+# Admin Password
+aaf.AdminPassword: {{ .Values.adminPwd }}
+
+# location of the codec keyfile used to decrypt passwords in this properties file before they are passed to AAF
+CredentialCodeKeyfile: etc/LocalKey
+
+# this overrides the Class used for Decryption.
+# This allows for a plugin encryption/decryption method if needed.
+# Call this Class for decryption at runtime.
+#AafDecryption.Class: com.company.proprietaryDecryptor
+
+
+#####################################################
+#
+# Settings for authorization of DBCAPI
+#
+#####################################################
+# Namespace for URI values for the API used to create AAF permissions
+# e.g. if ApiNamespace is X.Y.dmaapbc.api then for URI /mr_clients we create AAF perm X.Y.dmaapbc.api.mr_clients
+ApiNamespace: org.onap.dmaap-bc.api
+
+# If API authorization is required, then implement a class to enforce it.
+# This overrides the Class used for API permission check.
+#ApiPermission.Class: com.company.policy.DecisionPolicy
+
+#####################################################
+#
+# Settings for Southbound API: MirrorMaker provisioning
+#
+#####################################################
+# AAF Role of client publishing MM prov cmds
+MM.ProvRole: org.onap.dmaap-bc-mm-prov.prov
+
+# AAF identity when publishing MM prov cmds
+MM.ProvUserMechId: dmaap-bc-mm-prov@dmaap-bc-mm-prov.onap.org
+
+# pwd for Identity used to publish MM prov cmds
+MM.ProvUserPwd: demo123456!
+
+# AAF Role of MirrorMaker agent subscribed to prov cmds
+MM.AgentRole: org.onal.dmaap-bc-mm-prov.agent
+
+#####################################################
+#
+# HTTP Server Configuration
+#
+#####################################################
+
+# Allow http access to dbcapi
+HttpAllowed: true
+
+# listen to http port within this container (server)
+IntHttpPort: 8080
+
+# listen to https port within this container (server)
+# set to 0 if no certificates are available.
+IntHttpsPort: 8443
+
+# external port number for https taking port mapping into account
+ExtHttpsPort: 443
+
+# the type of keystore for https
+KeyStoreType: jks
+
+# path to the keystore file
+KeyStoreFile: etc/keystore
+
+# password for the https keystore
+KeyStorePassword: *j&Z*Ma;.4My4M]W0eB*fal$
+
+# password for the private key in the https keystore
+KeyPassword: *j&Z*Ma;.4My4M]W0eB*fal$
+
+# type of truststore for https
+TrustStoreType: jks
+
+# path to the truststore for https
+TrustStoreFile: etc/org.onap.dmaap-bc.trust.jks
+
+# password for the https truststore
+TrustStorePassword: pi8HuuSbN03MtQQ7(5TcyQ6;
+
+# path to the file used to trigger an orderly shutdown
+QuiesceFile: etc/SHUTDOWN
+
"topicName": "PNF_READY",
"topicDescription": "This topic will be used to publish the PNF_READY events generated by the PNF REgistration Handler service in the DCAE platform.",
"owner": "PNFRegistrationHandler",
- "txenabled": false
+ "txenabled": false,
+ "clients": [
+ {
+ "dcaeLocationName": "san-francisco",
+ "clientRole": "org.onap.dmaap.mr.PNF_READY.pub",
+ "action": [
+ "pub",
+ "view"
+ ]
+
+ }
+ ]
}
"topicName": "PNF_REGISTRATION",
"topicDescription": "the VES collector will be publishing pnfRegistration events in this topic",
"owner": "VEScollector",
- "txenabled": false
+ "txenabled": false,
+ "clients": [
+ {
+ "dcaeLocationName": "san-francisco",
+ "clientRole": "org.onap.dmaap.mr.PNF_REGISTRATION.sub",
+ "action": [
+ "sub",
+ "view"
+ ]
+
+ }
+ ]
}
periodSeconds: {{ .Values.liveness.periodSeconds }}
readinessProbe:
httpGet:
- port: {{ .Values.service.internalPort }}
- path: /webapi/info
+ host: {{ .Values.dmaapMessageRouterService }}
+ port: 3904
+ path: /topics
initialDelaySeconds: {{ .Values.readiness.initialDelaySeconds }}
periodSeconds: {{ .Values.readiness.periodSeconds }}
{{ end -}}
# application images
repository: nexus3.onap.org:10001
-image: onap/dmaap/buscontroller:1.0.22
+image: onap/dmaap/buscontroller:1.0.23
# application configuration
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/dmaap/datarouter-node:1.0.2
+image: onap/dmaap/datarouter-node:1.0.3
pullPolicy: Always
# flag to enable debugging - application support required
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/dmaap/datarouter-prov:1.0.2
+image: onap/dmaap/datarouter-prov:1.0.3
pullPolicy: Always
# flag to enable debugging - application support required
-# Copyright © 2017 Amdocs, Bell Canada, AT&T
-#
-# Licensed under the Apache License, Version 2.0 (the "License");
-# you may not use this file except in compliance with the License.
-# You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
+###############################################################################
+# ============LICENSE_START=======================================================
+# org.onap.dmaap
+# ================================================================================
+# Copyright © 2017 AT&T Intellectual Property. All rights reserved.
+# ================================================================================
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+# ============LICENSE_END=========================================================
#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
+# ECOMP is a trademark and service mark of AT&T Intellectual Property.
+#
+###############################################################################
###############################################################################
##
## Cambria API Server config
## - Default values are shown as commented settings.
##
-##############################################################################
+###############################################################################
##
## HTTP service
##
##
## Both Cambria and Kafka make use of Zookeeper.
##
+#config.zk.servers=172.18.1.1
config.zk.servers={{.Values.zookeeper.name}}:{{.Values.zookeeper.port}}
+#config.zk.root=/fe3c/cambria/config
###############################################################################
## Items below are passed through to Kafka's producer and consumer
## configurations (after removing "kafka.")
## if you want to change request.required.acks it can take this one value
+#kafka.metadata.broker.list=localhost:9092,localhost:9093
kafka.metadata.broker.list={{.Values.kafka.name}}:{{.Values.kafka.port}}
##kafka.request.required.acks=-1
#kafka.client.zookeeper=${config.zk.servers}
consumer.timeout.ms=100
zookeeper.connection.timeout.ms=6000
-zookeeper.session.timeout.ms=6000
+zookeeper.session.timeout.ms=20000
zookeeper.sync.time.ms=2000
auto.commit.interval.ms=1000
fetch.message.max.bytes =1000000
auto.commit.enable=false
+#(backoff*retries > zksessiontimeout)
+kafka.rebalance.backoff.ms=10000
+kafka.rebalance.max.retries=6
+
###############################################################################
##
## which would substantially impact a high volume consumer's performance.
##
## This complicates Cambria server failover, because we often need server
-## A to close its connection before server B brings up the replacement.
+## A to close its connection before server B brings up the replacement.
##
## The consumer cache is normally enabled.
## consumers every sweepFreqSeconds and will clean up any connections that are
## dormant for touchFreqMs.
#cambria.consumer.cache.sweepFreqSeconds=15
-#cambria.consumer.cache.touchFreqMs=120000
-
+cambria.consumer.cache.touchFreqMs=120000
+##stickforallconsumerrequests=false
## The cache is managed through ZK. The default value for the ZK connection
## string is the same as config.zk.servers.
#cambria.consumer.cache.zkConnect=${config.zk.servers}
##
#cambria.api.node.identifier=<use-something-unique-to-this-instance>
+#cambria.rateLimit.maxEmptyPollsPerMinute=30
+#cambria.rateLimitActual.delay.ms=10
+
###############################################################################
##
## Metrics Reporting
## This server can report its metrics periodically on a topic.
##
#metrics.send.cambria.enabled=true
-#metrics.send.cambria.topic=cambria.apinode.metrics #msgrtr.apinode.metrics.dmaap
+#metrics.send.cambria.topic=cambria.apinode.metrics #msgrtr.apinode.metrics.dmaap
#metrics.send.cambria.sendEverySeconds=60
cambria.consumer.cache.zkBasePath=/fe3c/cambria/consumerCache
##############################################################################
#AAF Properties
-msgRtr.namespace.aaf=org.openecomp.dcae.dmaap.mtnje2.mr.topic
-msgRtr.topicfactory.aaf=org.openecomp.dcae.dmaap.topicFactory|:org.openecomp.dcae.dmaap.mtnje2.mr.topic:
-enforced.topic.name.AAF=org.openecomp
+msgRtr.namespace.aaf=org.onap.dmaap.mr.topic
+msgRtr.topicfactory.aaf=org.onap.dmaap.mr.topicFactory|:org.onap.dmaap.mr.topic:
+enforced.topic.name.AAF=org.onap
forceAAF=false
transidUEBtopicreqd=false
-defaultNSforUEB=org.openecomp.dmaap.mr.ueb
+defaultNSforUEB=org.onap.dmaap.mr
##############################################################################
#Mirror Maker Agent
-msgRtr.mirrormakeradmin.aaf=org.openecomp.dmaap.mr.dev.mirrormaker|*|admin
-msgRtr.mirrormakeruser.aaf=org.openecomp.dmaap.mr.dev.mirrormaker|*|user
-msgRtr.mirrormakeruser.aaf.create=org.openecomp.dmaap.mr.dev.topicFactory|:org.openecomp.dmaap.mr.dev.topic:
+msgRtr.mirrormakeradmin.aaf=com.onap.dmaap.mr.dev.mirrormaker|*|admin
+msgRtr.mirrormakeruser.aaf=com.onap.dmaap.mr.dev.mirrormaker|*|user
+msgRtr.mirrormakeruser.aaf.create=com.onap.dmaap.mr.dev.topicFactory|:com.onap.dmaap.mr.dev.topic:
msgRtr.mirrormaker.timeout=15000
-msgRtr.mirrormaker.topic=org.openecomp.dmaap.mr.prod.mm.agent
+msgRtr.mirrormaker.topic=com.onap.dmaap.mr.prod.mm.agent
msgRtr.mirrormaker.consumergroup=mmagentserver
msgRtr.mirrormaker.consumerid=1
+
+kafka.max.poll.interval.ms=300000
+kafka.heartbeat.interval.ms=60000
+kafka.session.timeout.ms=240000
+kafka.max.poll.records=1000
\ No newline at end of file
-# Copyright © 2017 Amdocs, Bell Canada, AT&T
-#
-# Licensed under the Apache License, Version 2.0 (the "License");
-# you may not use this file except in compliance with the License.
-# You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
+aaf_locate_url=https://aaf-onap-test.osaaf.org:8095
+aaf_url=https://AAF_LOCATE_URL/AAF_NS.service:2.1
+aaf_env=DEV
+aaf_lur=org.onap.aaf.cadi.aaf.v2_0.AAFLurPerm
-basic_realm=openecomp.org
-basic_warn=TRUE
+cadi_truststore=/appl/dmaapMR1/etc/truststoreONAPall.jks
+cadi_truststore_password=changeit
-cadi_loglevel=DEBUG
-#cadi_keyfile=target/swm/package/nix/dist_files/appl/${artifactId}/etc/keyfile2
-cadi_keyfile=/appl/dmaapMR1/etc/keyfile
-# Configure AAF
-aaf_url=https://DME2RESOLVE/service=org.openecomp.authz.AuthorizationService/version=2.0/envContext=DEV/routeOffer=BAU_SE
+cadi_keyfile=/appl/dmaapMR1/etc/keyfilenew
-aaf_id=dgl@openecomp.org
-aaf_password=enc:f2u5br1mh29M02-
-aaf_timeout=5000
-aaf_clean_interval=1200000
-aaf_user_expires=60000
-aaf_high_count=1000000
+cadi_alias=dmaapmr@mr.dmaap.onap.org
+cadi_keystore=/appl/dmaapMR1/etc/org.onap.dmaap.mr.p12
+cadi_keystore_password=Messaging for All
+cadi_x509_issuers=CN=intermediateCA_1, OU=OSAAF, O=ONAP, C=US
-# The following properties are being set by the AJSC Container and should NOT need to be set here.
-AFT_LATITUDE=33.823589
-AFT_LONGITUDE=-84.366982
-AFT_ENVIRONMENT=AFTUAT
+cadi_loglevel=INFO
+cadi_protocols=TLSv1.1,TLSv1.2
+cadi_latitude=37.78187
+cadi_longitude=-122.26147
\ No newline at end of file
port: {{ .Values.service.externalPort }}
initialDelaySeconds: {{ .Values.readiness.initialDelaySeconds }}
periodSeconds: {{ .Values.readiness.periodSeconds }}
+ env:
+ - name: enableCadi
+ value: "false"
volumeMounts:
- mountPath: /etc/localtime
name: localtime
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/dmaap/dmaap-mr:1.1.6
+image: onap/dmaap/dmaap-mr:1.1.7
pullPolicy: Always
kafka:
# probe configuration parameters
liveness:
- initialDelaySeconds: 10
+ initialDelaySeconds: 120
periodSeconds: 10
# necessary to disable liveness probe when setting breakpoints
# in debugger so K8s doesn't restart unresponsive container
enabled: true
readiness:
- initialDelaySeconds: 10
+ initialDelaySeconds: 120
periodSeconds: 10
service:
--set stringArray set values on the command line (can specify multiple or separate values with commas: key1=val1,key2=val2)
--set-string stringArray set STRING values on the command line (can specify multiple or separate values with commas: key1=val1,key2=val2)
-f, --values valueFiles specify values in a YAML file or a URL(can specify multiple) (default [])
+ --verbose enables full helm install/upgrade output during deploy
EOF
}
-parse_yaml() {
- local prefix=$2
- local s='[[:space:]]*' w='[a-zA-Z0-9_]*' fs=$(echo @|tr @ '\034')
- sed -ne "s|^\($s\)\($w\)$s:$s\"\(.*\)\"$s\$|\1$fs\2$fs\3|p" \
- -e "s|^\($s\)\($w\)$s:$s\(.*\)$s\$|\1$fs\2$fs\3|p" $1 |
- awk -F$fs '{
- indent = length($1)/2;
- vname[indent] = $2;
- for (i in vname) {if (i > indent) {delete vname[i]}}
- if (length($3) > 0) {
- vn=""; for (i=0; i<indent; i++) {vn=(vn)(vname[i])("_")}
- printf("%s%s%s=\"%s\"\n", "'$prefix'",vn, $2, $3);
- }
- }'
-}
-
generate_overrides() {
SUBCHART_NAMES=($(cat $COMPUTED_OVERRIDES | grep -v '^\s\s'))
if [[ $START == "global:" ]]; then
echo "global:" > $GLOBAL_OVERRIDES
cat $COMPUTED_OVERRIDES | sed '/common:/,/consul:/d' \
- | sed -n '/'"$START"'/,/'log:'/p' | sed '1d;$d' >> $GLOBAL_OVERRIDES
+ | sed -n '/'"$START"'/,/'log:'/p' | sed '1d;$d' >> $GLOBAL_OVERRIDES
else
SUBCHART_DIR="$CACHE_SUBCHART_DIR/$(cut -d':' -f1 <<<"$START")"
if [[ -d "$SUBCHART_DIR" ]]; then
- cat $COMPUTED_OVERRIDES | sed -n '/'"$START"'/,/'"$END"'/p' \
+ cat $COMPUTED_OVERRIDES | sed -n '/^'"$START"'/,/^'"$END"'/p' \
| sed '1d;$d' | cut -c3- > $SUBCHART_DIR/subchart-overrides.yaml
fi
fi
done
}
+resolve_deploy_flags() {
+ flags=($1)
+ n=${#flags[*]}
+ for (( i = 0; i < n; i++ )); do
+ PARAM=${flags[i]}
+ if [[ $PARAM == "-f" || \
+ $PARAM == "--values" || \
+ $PARAM == "--set" || \
+ $PARAM == "--set-string" ]]; then
+ # skip param and its value
+ i=$((i + 1))
+ else
+ DEPLOY_FLAGS="$DEPLOY_FLAGS $PARAM"
+ fi
+ done
+ echo "$DEPLOY_FLAGS"
+}
+
deploy() {
+ # validate params
+ if [[ -z "$1" || -z "$2" ]]; then
+ usage
+ exit 0
+ fi
+
RELEASE=$1
CHART_URL=$2
FLAGS=${@:3}
CACHE_DIR=~/.helm/plugins/deploy/cache
CHART_DIR=$CACHE_DIR/$CHART_NAME
CACHE_SUBCHART_DIR=$CHART_DIR-subcharts
+ LOG_DIR=$CHART_DIR/logs
+
+ # determine if verbose output is enabled
+ VERBOSE="false"
+ if [[ $FLAGS = *"--verbose"* ]]; then
+ FLAGS="$(echo $FLAGS| sed -n 's/--verbose//p')"
+ VERBOSE="true"
+ fi
+ if [[ $FLAGS = *"--dry-run"* ]]; then
+ VERBOSE="true"
+ FLAGS="$FLAGS --debug"
+ fi
+
# should pass all flags instead
- NAMESPACE="$(echo $FLAGS | sed -n 's/.*\(namespace\).\s*/\1/p' | cut -c10-)"
-
+ NAMESPACE="$(echo $FLAGS | sed -n 's/.*\(namespace\).\s*/\1/p' | cut -c10- | cut -d' ' -f1)"
+
+ # Remove all override values passed in as arguments. These will be used during dry run
+ # to resolve computed override values. Remaining flags will be passed on during
+ # actual upgrade/install of parent and subcharts.
+ DEPLOY_FLAGS=$(resolve_deploy_flags "$FLAGS")
+
+ # determine if upgrading individual subchart or entire parent + subcharts
+ SUBCHART_RELEASE="$(cut -d'-' -f2 <<<"$RELEASE")"
+ if [[ ! -d "$CACHE_SUBCHART_DIR/$SUBCHART_RELEASE" ]]; then
+ SUBCHART_RELEASE=
+ else
+ # update specified subchart without parent
+ RELEASE="$(cut -d'-' -f1 <<<"$RELEASE")"
+ fi
+
# clear previously cached charts
rm -rf $CACHE_DIR
+ # create log driectory
+ mkdir -p $LOG_DIR
+
# fetch umbrella chart (parent chart containing subcharts)
if [[ -d "$CHART_URL" ]]; then
mkdir -p $CHART_DIR
cp -R $CHART_URL/* $CHART_DIR/
- cd $CHART_DIR/charts/
- for subchart in * ; do
- tar xzf ${subchart}
+ charts=$CHART_DIR/charts/*
+ for subchart in $charts ; do
+ tar xzf ${subchart} -C $CHART_DIR/charts/
done
- rm -rf *.tgz
+ rm -rf $CHART_DIR/charts/*.tgz
else
+ echo "fetching $CHART_URL"
helm fetch $CHART_URL --untar --untardir $CACHE_DIR
fi
generate_overrides $COMPUTED_OVERRIDES $GLOBAL_OVERRIDES
# upgrade/install parent chart first
- helm upgrade -i $RELEASE $CHART_DIR --namespace $NAMESPACE -f $COMPUTED_OVERRIDES
+ if [[ -z "$SUBCHART_RELEASE" ]]; then
+ LOG_FILE=$LOG_DIR/${RELEASE}.log
+ :> $LOG_FILE
+
+ helm upgrade -i $RELEASE $CHART_DIR $DEPLOY_FLAGS -f $COMPUTED_OVERRIDES \
+ > $LOG_FILE.log 2>&1
- # parse computed overrides - will use to determine if a subchart is "enabled"
- eval $(parse_yaml $COMPUTED_OVERRIDES "computed_")
+ if [[ $VERBOSE == "true" ]]; then
+ cat $LOG_FILE
+ else
+ echo "release \"$RELEASE\" deployed"
+ fi
+ fi
# upgrade/install each "enabled" subchart
- cd $CACHE_SUBCHART_DIR
+ cd $CACHE_SUBCHART_DIR/
for subchart in * ; do
- VAR="computed_${subchart}_enabled"
- COMMAND="$"$VAR
- eval "SUBCHART_ENABLED=$COMMAND"
- if [[ $SUBCHART_ENABLED == "true" ]]; then
- SUBCHART_OVERRIDES=$CACHE_SUBCHART_DIR/$subchart/subchart-overrides.yaml
- helm upgrade -i "${RELEASE}-${subchart}" $CACHE_SUBCHART_DIR/$subchart \
- --namespace $NAMESPACE -f $GLOBAL_OVERRIDES -f $SUBCHART_OVERRIDES
+ SUBCHART_OVERRIDES=$CACHE_SUBCHART_DIR/$subchart/subchart-overrides.yaml
+
+ SUBCHART_ENABLED=0
+ if [[ -f $SUBCHART_OVERRIDES ]]; then
+ SUBCHART_ENABLED=$(cat $SUBCHART_OVERRIDES | grep -c "^enabled: true")
+ fi
+
+ if [[ $SUBCHART_ENABLED -eq 1 ]]; then
+ if [[ -z "$SUBCHART_RELEASE" || $SUBCHART_RELEASE == "$subchart" ]]; then
+ LOG_FILE=$LOG_DIR/"${RELEASE}-${subchart}".log
+ :> $LOG_FILE
+
+ helm upgrade -i "${RELEASE}-${subchart}" $CACHE_SUBCHART_DIR/$subchart \
+ $DEPLOY_FLAGS -f $GLOBAL_OVERRIDES -f $SUBCHART_OVERRIDES \
+ > $LOG_FILE 2>&1
+
+ if [[ $VERBOSE == "true" ]]; then
+ cat $LOG_FILE
+ else
+ echo "release \"${RELEASE}-${subchart}\" deployed"
+ fi
+ fi
+ else
+ array=($(helm ls -q | grep "${RELEASE}-${subchart}"))
+ n=${#array[*]}
+ for (( i = n-1; i >= 0; i-- )); do
+ helm del "${array[i]}" --purge
+ done
fi
done
-}
-if [[ $# < 2 ]]; then
- usage
- exit 0
-fi
+ # report on success/failures of installs/upgrades
+ helm ls | grep FAILED | grep $RELEASE
+}
case "${1:-"help"}" in
"help")
RELEASE=$1
FLAGS=$2
- array=($(helm ls -q | grep $RELEASE))
+ array=($(helm ls -q --all | grep $RELEASE))
n=${#array[*]}
for (( i = n-1; i >= 0; i-- ))
do
export SERVICE_IP=$(kubectl get svc --namespace {{ include "common.namespace" . }} {{ include "common.fullname" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')
echo http://$SERVICE_IP:{{ .Values.service.externalPort }}
{{- else if contains "ClusterIP" .Values.service.type }}
- export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "so.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
+ export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "common.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
echo "Visit http://127.0.0.1:8080 to use your application"
kubectl port-forward $POD_NAME 8080:{{ .Values.service.internalPort }}
{{- end }}
export SERVICE_IP=$(kubectl get svc --namespace {{ include "common.namespace" . }} {{ include "common.fullname" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')
echo http://$SERVICE_IP:{{ .Values.service.externalPort }}
{{- else if contains "ClusterIP" .Values.service.type }}
- export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "so.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
+ export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "common.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
echo "Visit http://127.0.0.1:8080 to use your application"
kubectl port-forward $POD_NAME 8080:{{ .Values.service.internalPort }}
{{- end }}
export SERVICE_IP=$(kubectl get svc --namespace {{ include "common.namespace" . }} {{ include "common.fullname" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')
echo http://$SERVICE_IP:{{ .Values.service.externalPort }}
{{- else if contains "ClusterIP" .Values.service.type }}
- export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "so.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
+ export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "common.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
echo "Visit http://127.0.0.1:8080 to use your application"
kubectl port-forward $POD_NAME 8080:{{ .Values.service.internalPort }}
{{- end }}
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/msb/msb_discovery:1.2.0-SNAPSHOT-latest
+image: onap/msb/msb_discovery:1.2.1
pullPolicy: Always
istioSidecar: true
export SERVICE_IP=$(kubectl get svc --namespace {{ include "common.namespace" . }} {{ include "common.fullname" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')
echo http://$SERVICE_IP:{{ .Values.service.externalPort }}
{{- else if contains "ClusterIP" .Values.service.type }}
- export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "so.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
+ export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "common.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
echo "Visit http://127.0.0.1:8080 to use your application"
kubectl port-forward $POD_NAME 8080:{{ .Values.service.internalPort }}
{{- end }}
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/msb/msb_apigateway:1.2.0-SNAPSHOT-latest
+image: onap/msb/msb_apigateway:1.2.1
pullPolicy: Always
istioSidecar: true
export SERVICE_IP=$(kubectl get svc --namespace {{ include "common.namespace" . }} {{ include "common.fullname" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')
echo http://$SERVICE_IP:{{ .Values.service.externalPort }}
{{- else if contains "ClusterIP" .Values.service.type }}
- export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "so.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
+ export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "common.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
echo "Visit http://127.0.0.1:8080 to use your application"
kubectl port-forward $POD_NAME 8080:{{ .Values.service.internalPort }}
{{- end }}
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/msb/msb_apigateway:1.2.0-SNAPSHOT-latest
+image: onap/msb/msb_apigateway:1.2.1
pullPolicy: Always
istioSidecar: true
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/multicloud/openstack-ocata:1.2.0-STAGING
+image: onap/multicloud/openstack-ocata:1.2.0
pullPolicy: Always
#Istio sidecar injection policy
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/multicloud/openstack-pike:1.2.0-STAGING
+image: onap/multicloud/openstack-pike:1.2.0
pullPolicy: Always
#Istio sidecar injection policy
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/multicloud/vio:1.1.2
+image: onap/multicloud/vio:1.2.0
pullPolicy: Always
#Istio sidecar injection policy
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/multicloud/openstack-windriver:1.2.0-STAGING
+image: onap/multicloud/openstack-windriver:1.2.0
pullPolicy: Always
#Istio sidecar injection policy
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/multicloud/framework:1.2.0-STAGING
+image: onap/multicloud/framework:1.2.0
pullPolicy: Always
#Istio sidecar injection policy
initialDelaySeconds: 120
readiness:
initialDelaySeconds: 120
+ dmaap-dr-prov:
+ liveness:
+ initialDelaySeconds: 120
+ readiness:
+ initialDelaySeconds: 120
+ dmaap-dr-node:
+ liveness:
+ initialDelaySeconds: 120
+ readiness:
+ initialDelaySeconds: 120
+
portal:
portal-app:
liveness:
persistence:
mountPath: /dockerdata-nfs
- # flag to enable debugging - application support required
- debugEnabled: false
-
# override default resource limit flavor for all charts
flavor: unlimited
-# Repository for creation of nexus3.onap.org secret
-repository: nexus3.onap.org:10001
-
+ # flag to enable debugging - application support required
+ debugEnabled: false
#################################################################
# Enable/disable and configure helm charts (ie. applications)
export SERVICE_IP=$(kubectl get svc --namespace {{ include "common.namespace" . }} {{ include "common.fullname" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')
echo http://$SERVICE_IP:{{ .Values.service.externalPort }}
{{- else if contains "ClusterIP" .Values.service.type }}
- export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "so.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
+ export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "common.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
echo "Visit http://127.0.0.1:8080 to use your application"
kubectl port-forward $POD_NAME 8080:{{ .Values.service.internalPort }}
{{- end }}
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/policy-pe:1.3-SNAPSHOT-latest
+image: onap/policy-pe:1.3.0
pullPolicy: Always
# flag to enable debugging - application support required
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/policy-drools:1.3-SNAPSHOT-latest
+image: onap/policy-drools:1.3.0
pullPolicy: Always
# flag to enable debugging - application support required
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/policy-pe:1.3-SNAPSHOT-latest
+image: onap/policy-pe:1.3.0
pullPolicy: Always
# flag to enable debugging - application support required
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/policy-apex-pdp:2.0-SNAPSHOT-latest
+image: onap/policy-apex-pdp:2.0.0
pullPolicy: Always
# flag to enable debugging - application support required
spec:
containers:
- name: {{ include "common.name" . }}
- image: "{{ .Values.repository }}/{{ .Values.image }}"
+ image: "{{ include "common.repository" . }}/{{ .Values.image }}"
imagePullPolicy: {{ .Values.global.pullPolicy | default .Values.pullPolicy }}
command: ["/opt/app/policy/distribution/bin/policy-dist.sh"]
args: ["/opt/app/policy/distribution/etc/mounted/config.json"]
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/policy-distribution:2.0.0-SNAPSHOT-latest
+image: onap/policy-distribution:2.0.0
pullPolicy: Always
# flag to enable debugging - application support required
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/policy-pe:1.3-SNAPSHOT-latest
+image: onap/policy-pe:1.3.0
pullPolicy: Always
subChartsOnly:
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/pomba-aai-context-builder:latest
+image: onap/pomba-aai-context-builder:1.3.1
pullPolicy: Always
# flag to enable debugging - application support required
# Application configuration defaults.
#################################################################
# application image
-image: onap/pomba-context-aggregator:latest
+image: onap/pomba-context-aggregator:1.3.2
pullPolicy: Always
# flag to enable debugging - application support required
# application image
repository: nexus3.onap.org:10001
-image: onap/data-router:1.3-STAGING-latest
+image: onap/data-router:1.3.0
pullPolicy: Always
restartPolicy: Always
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/pomba-network-discovery-context-builder:latest
+image: onap/pomba-network-discovery-context-builder:1.3.0
pullPolicy: Always
# flag to enable debugging - application support required
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/pomba-sdc-context-builder:latest
+image: onap/pomba-sdc-context-builder:1.3.1
pullPolicy: Always
# flag to enable debugging - application support required
# application image
repository: nexus3.onap.org:10001
-image: onap/search-data-service:1.3-STAGING-latest
+image: onap/search-data-service:1.3.1
pullPolicy: Always
restartPolicy: Always
# application image
#repository: nexus3.onap.org:10001
repository: nexus3.onap.org:10001
-image: onap/validation:1.3-STAGING-latest
+image: onap/validation:1.3.0
#pullPolicy: Always
pullPolicy: IfNotPresent
sessiontimeout_feed_cron = 0 0/5 * * * ? *
#Front end URL
-frontend_url = http://{{.Values.global.portalHostName}}:{{.Values.global.portalFEPort}}/ONAPPORTAL/applicationsHome
+frontend_url = https://{{.Values.global.portalHostName}}:{{.Values.global.portalFEPort}}/ONAPPORTAL/applicationsHome
# An Unqiue 128-bit value defined to indentify a specific version of
export SERVICE_IP=$(kubectl get svc --namespace {{ include "common.namespace" . }} {{ include "common.fullname" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')\r
echo http://$SERVICE_IP:{{ .Values.service.externalPort }}\r
{{- else if contains "ClusterIP" .Values.service.type }}\r
- export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "so.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")\r
+ export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "common.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")\r
echo "Visit http://127.0.0.1:8080 to use your application"\r
kubectl port-forward $POD_NAME 8080:{{ .Values.service.internalPort }}\r
{{- end }}\r
spec:
type: {{ .Values.service.type }}
ports:
- {{- if or (eq .Values.service.type "NodePort") (eq .Values.service.type "LoadBalancer") }}
+ {{ if or (eq .Values.service.type "NodePort") (eq .Values.service.type "LoadBalancer") -}}
- port: {{ .Values.service.externalPort }}
targetPort: {{ .Values.service.internalPort }}
nodePort: {{ .Values.global.nodePortPrefix | default .Values.nodePortPrefix }}{{ .Values.service.nodePort }}
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/portal-app:2.3.0-STAGING-latest
+image: onap/portal-app:2.3.0
pullPolicy: Always
# default number of instances
*/\r
-- app_url is the FE, app_rest_endpoint is the BE\r
--portal-sdk => TODO: doesn't open a node port yet\r
-update fn_app set app_url = 'http://{{.Values.config.portalSdkHostName}}:{{.Values.config.portalSdkPort}}/ONAPPORTALSDK/welcome.htm', app_rest_endpoint = 'http://portal-sdk:8990/ONAPPORTALSDK/api/v2' where app_name = 'xDemo App';\r
+update fn_app set app_url = 'http://{{.Values.config.portalSdkHostName}}:{{.Values.config.portalSdkPort}}/ONAPPORTALSDK/welcome.htm', app_rest_endpoint = 'http://portal-sdk:8080/ONAPPORTALSDK/api/v3' where app_name = 'xDemo App';\r
--dmaap-bc => the dmaap-bc doesn't open a node port..\r
update fn_app set app_url = 'http://{{.Values.config.dmaapBcHostName}}:{{.Values.config.dmaapBcPort}}/ECOMPDBCAPP/dbc#/dmaap', app_rest_endpoint = 'http://dmaap-bc:8989/ECOMPDBCAPP/api/v2' where app_name = 'DMaaP Bus Ctrl';\r
--sdc-be => 8443:30204, 8080:30205\r
--sdc-fe => 8181:30206, 9443:30207\r
-update fn_app set app_url = 'http://{{.Values.config.sdcFeHostName}}:{{.Values.config.sdcFePort}}/sdc1/portal', app_rest_endpoint = 'http://sdc-be:8080/api/v2' where app_name = 'SDC';\r
+update fn_app set app_url = 'http://{{.Values.config.sdcFeHostName}}:{{.Values.config.sdcFePort}}/sdc1/portal', app_rest_endpoint = 'http://sdc-be:8080/api/v3' where app_name = 'SDC';\r
--pap => 8443:30219\r
-update fn_app set app_url = 'https://{{.Values.config.papHostName}}:{{.Values.config.papPort}}/onap/policy', app_rest_endpoint = 'https://pap:8443/onap/api/v2' where app_name = 'Policy';\r
+update fn_app set app_url = 'https://{{.Values.config.papHostName}}:{{.Values.config.papPort}}/onap/policy', app_rest_endpoint = 'https://pap:8443/onap/api/v3' where app_name = 'Policy';\r
--vid => 8080:30200\r
-update fn_app set app_url = 'https://{{.Values.config.vidHostName}}:{{.Values.config.vidPort}}/vid/welcome.htm', app_rest_endpoint = 'https://vid:8443/vid/api/v2' where app_name = 'Virtual Infrastructure Deployment';\r
+update fn_app set app_url = 'https://{{.Values.config.vidHostName}}:{{.Values.config.vidPort}}/vid/welcome.htm', app_rest_endpoint = 'https://vid:8443/vid/api/v3' where app_name = 'Virtual Infrastructure Deployment';\r
--sparky => TODO: sparky doesn't open a node port yet\r
update fn_app set app_url = 'http://{{.Values.config.aaiSparkyHostName}}:{{.Values.config.aaiSparkyPort}}/services/aai/webapp/index.html#/viewInspect', app_rest_endpoint = 'http://aai-sparky-be.{{.Release.Namespace}}:9517/api/v2' where app_name = 'A&AI UI';\r
--cli => 8080:30260\r
update fn_app set app_url = 'http://{{.Values.config.cliHostName}}:{{.Values.config.cliPort}}/', app_type = 1 where app_name = 'CLI';\r
---msb-discovery => 10081:30281 this is clearly incorrect\r
-update fn_app set app_url = 'http://{{.Values.config.msbDiscoveryHostName}}:{{.Values.config.msbDiscoveryPort}}/iui/microservices/default.html' where app_name = 'MSB';\r
+--msb-iag => 80:30280\r
+update fn_app set app_url = 'http://{{.Values.config.msbHostName}}:{{.Values.config.msbPort}}/iui/microservices/default.html' where app_name = 'MSB';\r
+\r
+\r
+/*\r
+Additionally, some more update statments; these should be refactored to another SQL file in future releases \r
+*/\r
+\r
+-- portal\r
+update fn_app set auth_central = 'Y' , auth_namespace = 'org.onap.portal' where app_id = 1;\r
+-- portal-sdk\r
+update fn_app set app_username='Default', app_password='2VxipM8Z3SETg32m3Gp0FvKS6zZ2uCbCw46WDyK6T5E=', ueb_key='ueb_key' where app_id = 2;\r
+-- SDC\r
+update fn_app set app_username='sdc', app_password='j85yNhyIs7zKYbR1VlwEfNhS6b7Om4l0Gx5O8931sCI=', ueb_key='ueb_key' where app_id = 4;\r
+-- policy\r
+update fn_app set app_username='Default', app_password='2VxipM8Z3SETg32m3Gp0FvKS6zZ2uCbCw46WDyK6T5E=', ueb_key='ueb_key_5', auth_central = 'Y' , auth_namespace = 'org.onap.policy' where app_id = 5;\r
+-- vid\r
+update fn_app set app_username='Default', app_password='2VxipM8Z3SETg32m3Gp0FvKS6zZ2uCbCw46WDyK6T5E=', ueb_key='2Re7Pvdkgw5aeAUD', auth_central = 'Y' , auth_namespace = 'org.onap.vid' where app_id = 6;\r
+-- aai sparky\r
+update fn_app set app_username='aaiui', app_password='4LK69amiIFtuzcl6Gsv97Tt7MLhzo03aoOx7dTvdjKQ=', ueb_key='ueb_key' where app_id = 7;\r
+\r
+\r
+\r
# application image
repository: nexus3.onap.org:10001
-image: onap/portal-db:2.3.0-STAGING-latest
+image: onap/portal-db:2.3.0
pullPolicy: Always
readinessImage: readiness-check:2.0.0
dmaapBcPort: "" # TODO: populate with
# application's front end hostname. Must be resolvable on the client side environment
dmaapBcHostName: "dmaap-bc.simpledemo.onap.org"
- # msb discovery ui assignment for port ?
- msbDiscoveryPort: "30281"
+ # msb IAG ui assignment for port 80
+ msbPort: "30280"
# application's front end hostname. Must be resolvable on the client side environment
- msbDiscoveryHostName: "msb.api.discovery.simpledemo.onap.org"
+ msbHostName: "msb.api.simpledemo.onap.org"
# default number of instances
replicaCount: 1
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/portal-sdk:2.3.0-STAGING-latest
+image: onap/portal-sdk:2.3.0
pullPolicy: Always
# flag to enable debugging - application support required
name: portal-sdk
portName: portal-sdk
internalPort: 8080
+ externalPort: 8080
nodePort: 12
mariadb:
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/portal-wms:2.3.0-STAGING-latest
+image: onap/portal-wms:2.3.0
pullPolicy: Always
# flag to enable debugging - application support required
FROM boxfuse/flyway:5.0.7-alpine
-ARG branch=2.0.0-ONAP
+ARG branch=release-2.3.0
ENV no_proxy "localhost,127.0.0.1,.cluster.local,$KUBERNETES_SERVICE_HOST"
# Setup Corporate proxy
ENV https_proxy ${HTTP_PROXY}
"onap_private_net_id" : "${GLOBAL_INJECTED_NETWORK}",
"onap_private_subnet_id" : "{{ .Values.openStackPrivateSubnetId }}",
"onap_private_net_cidr" : "{{ .Values.openStackPrivateNetCidr }}",
+ "sec_group" : "{{ .Values.openStackSecurityGroup }}",
"dcae_collector_ip" : "{{ .Values.dcaeCollectorIp }}",
"dcae_collector_port" : "30235",
"public_net_id" : "${GLOBAL_INJECTED_PUBLIC_NET_ID}",
"vpg_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.2",
"vsn_private_ip_0" : "192.168.20.250",
"vsn_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.3",
+ "sec_group" : "{{ .Values.openStackSecurityGroup }}",
'vfw_name_0':'vofwl01fwl${hostid}',
'vpg_name_0':'vofwl01pgn${hostid}',
"vfw_image_name" : "${GLOBAL_INJECTED_UBUNTU_1404_IMAGE}",
"vpg_private_ip_0" : "192.168.10.200",
"vsn_private_ip_0" : "192.168.20.250",
"vsn_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.102",
+ "sec_group" : "{{ .Values.openStackSecurityGroup }}",
'vfw_name_0':'vofwl01fwl${hostid}',
'vsn_name_0':'vofwl01snk${hostid}',
"image_name" : "${GLOBAL_INJECTED_UBUNTU_1404_IMAGE}",
"vpg_private_ip_0" : "192.168.10.200",
"vpg_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.103",
"vsn_private_ip_0" : "192.168.20.250",
+ "sec_group" : "{{ .Values.openStackSecurityGroup }}",
'vpg_name_0':'vofwl01pgn${hostid}',
"image_name" : "${GLOBAL_INJECTED_UBUNTU_1404_IMAGE}",
"flavor_name" : "${GLOBAL_INJECTED_VM_FLAVOR}",
"vlb_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.4",
"vdns_private_ip_0" : "192.168.30.110",
"vdns_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.5",
+ "sec_group" : "{{ .Values.openStackSecurityGroup }}",
'vlb_name_0':'vovlblb${hostid}',
'vdns_name_0':'vovlbdns${hostid}',
"pktgen_private_net_cidr" : "192.168.9.0/24",
"vlb_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.4",
"vdns_private_ip_0" : "192.168.30.222",
"vdns_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.6",
+ "sec_group" : "{{ .Values.openStackSecurityGroup }}",
'scaling_vdns_name_0':'vovlbscaling${hostid}',
"vlb_private_net_cidr" : "192.168.10.0/24"
},
"vpg_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.12",
"vsn_private_ip_0" : "192.168.120.250",
"vsn_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.13",
+ "sec_group" : "{{ .Values.openStackSecurityGroup }}",
'vfw_name_0':'clfwl01fwl${hostid}',
'vpg_name_0':'clfwl01pgn${hostid}',
"vfw_image_name" : "${GLOBAL_INJECTED_UBUNTU_1404_IMAGE}",
"vpg_private_ip_0" : "192.168.10.200",
"vsn_private_ip_0" : "192.168.20.250",
"vsn_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.112",
+ "sec_group" : "{{ .Values.openStackSecurityGroup }}",
'vfw_name_0':'clfwl01fwl${hostid}',
'vsn_name_0':'clfwl01snk${hostid}',
"image_name" : "${GLOBAL_INJECTED_UBUNTU_1404_IMAGE}",
"vpg_private_ip_0" : "192.168.10.200",
"vpg_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.113",
"vsn_private_ip_0" : "192.168.20.250",
+ "sec_group" : "{{ .Values.openStackSecurityGroup }}",
'vpg_name_0':'clfwl01pgn${hostid}',
"image_name" : "${GLOBAL_INJECTED_UBUNTU_1404_IMAGE}",
"flavor_name" : "${GLOBAL_INJECTED_VM_FLAVOR}",
"vlb_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.14",
"vdns_private_ip_0" : "192.168.30.110",
"vdns_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.15",
+ "sec_group" : "{{ .Values.openStackSecurityGroup }}",
'vlb_name_0':'clvlblb${hostid}',
'vdns_name_0':'clvlbdns${hostid}',
"pktgen_private_net_cidr" : "192.168.9.0/24",
"vlb_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.14",
"vdns_private_ip_0" : "192.168.130.222",
"vdns_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.16",
+ "sec_group" : "{{ .Values.openStackSecurityGroup }}",
'scaling_vdns_name_0':'clvlbscaling${hostid}',
"vlb_private_net_cidr" : "192.168.10.0/24"
},
"vpg_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.12",
"vsn_private_ip_0" : "192.168.120.250",
"vsn_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.13",
+ "sec_group" : "{{ .Values.openStackSecurityGroup }}",
'vfw_name_0':'demofwl01fwl',
'vpg_name_0':'demofwl01pgn',
"vfw_image_name" : "${GLOBAL_INJECTED_UBUNTU_1404_IMAGE}",
"vpg_private_ip_0" : "192.168.10.200",
"vsn_private_ip_0" : "192.168.20.250",
"vsn_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.122",
+ "sec_group" : "{{ .Values.openStackSecurityGroup }}",
'vfw_name_0':'${generic_vnf_name}',
'vsn_name_0':'demofwl01snk${hostid}',
"image_name" : "${GLOBAL_INJECTED_UBUNTU_1404_IMAGE}",
"vpg_private_ip_0" : "192.168.10.200",
"vpg_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.123",
"vsn_private_ip_0" : "192.168.20.250",
+ "sec_group" : "{{ .Values.openStackSecurityGroup }}",
'vpg_name_0':'demofwl01pgn${hostid}',
"image_name" : "${GLOBAL_INJECTED_UBUNTU_1404_IMAGE}",
"flavor_name" : "${GLOBAL_INJECTED_VM_FLAVOR}"
"vlb_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.24",
"vdns_private_ip_0" : "192.168.30.110",
"vdns_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.25",
+ "sec_group" : "{{ .Values.openStackSecurityGroup }}",
'vlb_name_0':'demovlblb${hostid}',
'vdns_name_0':'demovlbdns${hostid}',
"pktgen_private_net_cidr" : "192.168.9.0/24",
"vlb_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.24",
"vdns_private_ip_0" : "192.168.130.222",
"vdns_private_ip_1" : "{{.Values.openStackOamNetworkCidrPrefix}}.${ecompnet}.26",
+ "sec_group" : "{{ .Values.openStackSecurityGroup }}",
'scaling_vdns_name_0':'demovlbscaling${hostid}',
"vlb_private_net_cidr" : "192.168.10.0/24"
},
GLOBAL_SMS_SERVER_NAME = "aaf-sms.{{include "common.namespace" .}}"
GLOBAL_SMS_SERVER_PORT = "10443"
# vid info - everything is from the private oam network (also called onap private network)
-GLOBAL_VID_SERVER_PROTOCOL = "https"
-GLOBAL_VID_SERVER_PORT = "8443"
+GLOBAL_VID_SERVER_PROTOCOL = "{{ .Values.vidServerProtocol }}"
+GLOBAL_VID_SERVER_PORT = "{{ .Values.vidServerPort }}"
GLOBAL_VID_USERNAME = "{{ .Values.vidUsername }}"
GLOBAL_VID_PASSWORD = "{{ .Values.vidPassword}}"
GLOBAL_VID_HEALTH_USERNAME = "{{ .Values.vidHealthUsername }}"
export SERVICE_IP=$(kubectl get svc --namespace {{ include "common.namespace" . }} {{ include "common.name" . }} -o jsonpath='{.status.loadBalancer.ingress[0].ip}')
echo http://$SERVICE_IP:{{ .Values.service.externalPort }}
{{- else if contains "ClusterIP" .Values.service.type }}
- export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "so.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
+ export POD_NAME=$(kubectl get pods --namespace {{ include "common.namespace" . }} -l "app={{ template "common.name" . }},release={{ .Release.Name }}" -o jsonpath="{.items[0].metadata.name}")
echo "Visit http://127.0.0.1:8080 to use your application"
kubectl port-forward $POD_NAME 8080:{{ .Values.service.internalPort }}
{{- end }}
scriptVersion: "1.2.0-SNAPSHOT"
# Openstack network to which VNFs will bind their primary (first) interface. Maps to GLOBAL_INJECTED_NETWORK
openStackPrivateNetId: "e8f51956-00dd-4425-af36-045716781ffc"
-
+# Openstack security group for instantiating VNFs
+openStackSecurityGroup: "onap_sg"
# SDNC Preload configuration
# Openstack subnet UUID for the network defined by openStackPrivateNetId. Maps to onap_private_subnet_id
openStackPrivateSubnetId: "e8f51956-00dd-4425-af36-045716781ffc"
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/sdc-backend:1.3-STAGING-latest
-backendInitImage: onap/sdc-backend-init:1.3-STAGING-latest
+image: onap/sdc-backend:1.3.0
+backendInitImage: onap/sdc-backend-init:1.3.0
pullPolicy: Always
# flag to enable debugging - application support required
portName: sdc-be
nodePort: "04"
internalPort: 8443
+ externalPort: 8443
nodePort2: "05"
internalPort2: 8080
+ externalPort2: 8080
ingress:
enabled: false
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/sdc-cassandra:1.3-STAGING-latest
-cassandraInitImage: onap/sdc-cassandra-init:1.3-STAGING-latest
+image: onap/sdc-cassandra:1.3.0
+cassandraInitImage: onap/sdc-cassandra-init:1.3.0
pullPolicy: Always
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/sdc-elasticsearch:1.3-STAGING-latest
-elasticInitImage: onap/sdc-init-elasticsearch:1.3-STAGING-latest
+image: onap/sdc-elasticsearch:1.3.0
+elasticInitImage: onap/sdc-init-elasticsearch:1.3.0
pullPolicy: Always
# flag to enable debugging - application support required
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/sdc-frontend:1.3-STAGING-latest
+image: onap/sdc-frontend:1.3.0
pullPolicy: Always
config:
portName: sdc-fe
nodePort: "06"
internalPort: 8181
+ externalPort: 8181
nodePort2: "07"
internalPort2: 9443
+ externalPort2: 9443
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/sdc-kibana:1.3-STAGING-latest
+image: onap/sdc-kibana:1.3.0
pullPolicy: Always
config:
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/sdc-onboard-backend:1.3-STAGING-latest
-onboardingInitImage: onap/sdc-onboard-cassandra-init:1.3-STAGING-latest
+image: onap/sdc-onboard-backend:1.3.0
+onboardingInitImage: onap/sdc-onboard-cassandra-init:1.3.0
pullPolicy: Always
# flag to enable debugging - application support required
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/workflow-backend:1.3.0-SNAPSHOT
-configInitImage: onap/workflow-init:1.3.0-SNAPSHOT
+image: onap/workflow-backend:1.3.0
+configInitImage: onap/workflow-init:1.3.0
pullPolicy: Always
# flag to enable debugging - application support required
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/workflow-frontend:1.3.0-SNAPSHOT
+image: onap/workflow-frontend:1.3.0
pullPolicy: Always
# flag to enable debugging - application support required
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/sdnc-dmaap-listener-image:1.4-STAGING-latest
+image: onap/sdnc-dmaap-listener-image:1.4.1
pullPolicy: Always
# flag to enable debugging - application support required
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/sdnc-ansible-server-image:1.4-STAGING-latest
+image: onap/sdnc-ansible-server-image:1.4.1
pullPolicy: Always
# flag to enable debugging - application support required
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/admportal-sdnc-image:1.4-STAGING-latest
+image: onap/admportal-sdnc-image:1.3.4
pullPolicy: Always
# flag to enable debugging - application support required
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/sdnc-ueb-listener-image:1.4-STAGING-latest
+image: onap/sdnc-ueb-listener-image:1.4.1
pullPolicy: Always
# flag to enable debugging - application support required
# application images
repository: nexus3.onap.org:10001
pullPolicy: Always
-image: onap/sdnc-image:1.4-STAGING-latest
+image: onap/sdnc-image:1.4.1
# flag to enable debugging - application support required
debugEnabled: false
agnostic:
sniro:
endpoint: /sniro/api/v2/placement
- host: http://c1.vm1.mso.simpledemo.onap.org:30253
+ host: http://sniro-emulator:80
site-name: CamundaEngine
sniro:
auth: test:testpwd
secret: 64AG2hF4pYeG2pq7CT6XwUOT
service: ueb
poolMembers: message-router.{{ include "common.namespace" . }}:3904,message-router.{{ include "common.namespace" . }}:3904
+sniro:
+ conductor:
+ enabled: true
+ host: http://sniro-emulator:80
+ uri: /v1/release-orders
+ headers.auth: Basic dGVzdDp0ZXN0cHdk
+ manager:
+ timeout: PT30M
+ host: http://sniro-emulator:80
+ uri.v1: /sniro/api/v2/placement
+ uri.v2: /sniro/api/placement/v2
+ headers.auth: Basic dGVzdDp0ZXN0cHdk
+ headers.patchVersion: 1
+ headers.minorVersion: 1
+ headers.latestVersion: 2
server:
port: {{ index .Values.containerPort }}
tomcat:
-
username: mso_admin
password: '$2a$10$Fh9ffgPw2vnmsghsRD3ZauBL1aKXebigbq3BB1RPWtE62UDILsjke'
- role: ACTUATOR
+ role: ACTUATOR
\ No newline at end of file
# Application configuration defaults.
#################################################################
repository: nexus3.onap.org:10001
-image: onap/so/bpmn-infra:1.3.0-STAGING-latest
+image: onap/so/bpmn-infra:1.3.1
pullPolicy: Always
replicaCount: 1
# Application configuration defaults.
#################################################################
repository: nexus3.onap.org:10001
-image: onap/so/catalog-db-adapter:1.3.0-STAGING-latest
+image: onap/so/catalog-db-adapter:1.3.1
pullPolicy: Always
replicaCount: 1
type: {{ .Values.service.type }}
ports:
- name: {{ .Values.service.portName }}
- {{- if eq .Values.service.type "NodePort" }}
+ {{ if eq .Values.service.type "NodePort" }}
port: {{ .Values.service.internalPort }}
nodePort: {{ .Values.global.nodePortPrefix | default "302" }}{{ .Values.service.nodePort }}
- {{- else -}}
+ {{ else }}
port: {{ .Values.service.externalPort }}
targetPort: {{ .Values.service.internalPort }}
- {{- end}}
+ {{ end }}
selector:
app: {{ include "common.name" . }}
release: {{ .Release.Name }}
\ No newline at end of file
type: NodePort
portName: mariadb
internalPort: 3306
+ externalPort: 3306
nodePort: 52
ingress:
enabled: false
camunda:
rest:
api:
- url: http://bpmn-infra.{{ include "common.namespace" . }}:8081/engine-rest/engine/
+ url: http://so-bpmn-infra.{{ include "common.namespace" . }}:8081/sobpmnengine/engine/
engine: default
- auth: Basic YWRtaW46S3A4Yko0U1hzek0wV1hsaGFrM2VIbGNzZTJnQXc4NHZhb0dHbUp2VXkyVQ==
+ auth: Basic YXBpaEJwbW46cGFzc3dvcmQxJA==
mso:
database:
rest:
# Application configuration defaults.
#################################################################
repository: nexus3.onap.org:10001
-image: onap/so/so-monitoring:latest
+image: onap/so/so-monitoring:1.3.1
pullPolicy: Always
replicaCount: 1
logPath: app/logs/
app: so-monitoring
service:
- type: ClusterIP
+ type: NodePort
+ nodePort: 24
internalPort: 9091
externalPort: 9091
portName: so-monitor-port
#################################################################
# Application configuration defaults.
#################################################################
-image: onap/so/openstack-adapter:1.3.0-STAGING-latest
+image: onap/so/openstack-adapter:1.3.1
pullPolicy: Always
repository: nexus3.onap.org:10001
# Application configuration defaults.
#################################################################
repository: nexus3.onap.org:10001
-image: onap/so/request-db-adapter:1.3.0-STAGING-latest
+image: onap/so/request-db-adapter:1.3.1
pullPolicy: Always
replicaCount: 1
# Application configuration defaults.
#################################################################
repository: nexus3.onap.org:10001
-image: onap/so/sdc-controller:1.3.0-STAGING-latest
+image: onap/so/sdc-controller:1.3.1
pullPolicy: Always
replicaCount: 1
# Application configuration defaults.
#################################################################
repository: nexus3.onap.org:10001
-image: onap/so/sdnc-adapter:1.3.0-STAGING-latest
+image: onap/so/sdnc-adapter:1.3.1
pullPolicy: Always
replicaCount: 1
# Application configuration defaults.
#################################################################
repository: nexus3.onap.org:10001
-image: onap/so/vfc-adapter:1.3.0-STAGING-latest
+image: onap/so/vfc-adapter:1.3.1
pullPolicy: Always
replicaCount: 1
# Application configuration defaults.
#################################################################
repository: nexus3.onap.org:10001
-image: onap/so/api-handler-infra:1.3.0-STAGING-latest
+image: onap/so/api-handler-infra:1.3.1
pullPolicy: Always
replicaCount: 1
type: NodePort
nodePort: 77
internalPort: 8080
+ externalPort: 8080
portName: so-apih-port
updateStrategy:
type: RollingUpdate
# application image
repository: nexus3.onap.org:10001
-image: onap/usecase-ui:1.1.1
+image: onap/usecase-ui:1.2.0
pullPolicy: Always
# application configuration
flavor: small
repository: nexus3.onap.org:10001
-image: onap/vfc/catalog:1.1.0
+image: onap/vfc/catalog:1.2.0
pullPolicy: Always
#Istio sidecar injection policy
flavor: small
repository: nexus3.onap.org:10001
-image: onap/vfc/db:latest
+image: onap/vfc/db:1.2.0
pullPolicy: Always
# flag to enable debugging - application support required
flavor: small
repository: nexus3.onap.org:10001
-image: onap/vfc/emsdriver:1.1.0
+image: onap/vfc/emsdriver:1.2.0
pullPolicy: Always
#Istio sidecar injection policy
flavor: small
repository: nexus3.onap.org:10001
-image: onap/vfc/gvnfmdriver:1.1.0
+image: onap/vfc/gvnfmdriver:1.2.0
pullPolicy: Always
#Istio sidecar injection policy
flavor: small
repository: nexus3.onap.org:10001
-image: onap/vfc/nfvo/svnfm/huawei:1.1.0
+image: onap/vfc/nfvo/svnfm/huawei:1.2.0
pullPolicy: Always
#Istio sidecar injection policy
flavor: small
repository: nexus3.onap.org:10001
-image: onap/vfc/jujudriver:1.1.0
+image: onap/vfc/jujudriver:1.2.0
pullPolicy: Always
#Istio sidecar injection policy
flavor: small
repository: nexus3.onap.org:10001
-image: onap/vfc/multivimproxy:1.0.0
+image: onap/vfc/multivimproxy:1.2.0
pullPolicy: Always
#Istio sidecar injection policy
initialDelaySeconds: {{ .Values.readiness.initialDelaySeconds }}
periodSeconds: {{ .Values.readiness.periodSeconds }}
env:
+ - name: EXTERNAL_PORT
+ value: "8089"
- name: MSB_ADDR
value: "{{ .Values.global.config.msbServiceName }}:{{ .Values.global.config.msbPort }}"
resources:
flavor: small
repository: nexus3.onap.org:10001
-image: onap/vfc/nfvo/svnfm/nokiav2:1.1.0
+image: onap/vfc/nfvo/svnfm/nokiav2:1.3.0
pullPolicy: Always
#Istio sidecar injection policy
flavor: small
repository: nexus3.onap.org:10001
-image: onap/vfc/nslcm:1.1.0
+image: onap/vfc/nslcm:1.2.0
pullPolicy: Always
#Istio sidecar injection policy
flavor: small
repository: nexus3.onap.org:10001
-image: onap/vfc/resmanagement:1.1.0
+image: onap/vfc/resmanagement:1.2.0
pullPolicy: Always
#Istio sidecar injection policy
flavor: small
repository: nexus3.onap.org:10001
-image: onap/vfc/vnflcm:1.1.0
+image: onap/vfc/vnflcm:1.2.0
pullPolicy: Always
#Istio sidecar injection policy
flavor: small
repository: nexus3.onap.org:10001
-image: onap/vfc/vnfmgr:1.1.0
+image: onap/vfc/vnfmgr:1.2.0
pullPolicy: Always
#Istio sidecar injection policy
flavor: small
repository: nexus3.onap.org:10001
-image: onap/vfc/vnfres:1.1.0
+image: onap/vfc/vnfres:1.2.0
pullPolicy: Always
#Istio sidecar injection policy
flavor: small
repository: nexus3.onap.org:10001
-image: onap/vfc/wfengine-activiti:1.1.0
+image: onap/vfc/wfengine-activiti:1.2.0
pullPolicy: Always
#Istio sidecar injection policy
flavor: small
repository: nexus3.onap.org:10001
-image: onap/vfc/wfengine-mgrservice:1.1.0
+image: onap/vfc/wfengine-mgrservice:1.2.0
pullPolicy: Always
#Istio sidecar injection policy
flavor: small
repository: nexus3.onap.org:10001
-image: onap/vfc/ztesdncdriver:1.1.0
+image: onap/vfc/ztesdncdriver:1.2.0
pullPolicy: Always
#Istio sidecar injection policy
flavor: small
repository: nexus3.onap.org:10001
-image: onap/vfc/ztevnfmdriver:1.1.0
+image: onap/vfc/ztevnfmdriver:1.2.0
pullPolicy: Always
#Istio sidecar injection policy
- /root/job_complete.py
args:
- --job-name
- - vid-config-galera
+ - {{ include "common.fullname" . }}-galera-config
env:
- name: NAMESPACE
valueFrom:
- name: MSO_DME2_ENABLED
value: {{ .Values.global.debugEnabled | default .Values.debugEnabled | quote }}
- name: VID_ECOMP_REDIRECT_URL
- value: http://portal-app.{{ include "common.namespace" . }}:{{ .Values.config.onapport }}/ONAPPORTAL/login.htm
+ value: https://{{ .Values.config.portalhost }}:{{ .Values.config.onapport }}/ONAPPORTAL/login.htm
- name: VID_ECOMP_REST_URL
- value: http://portal-app.{{ include "common.namespace" . }}:{{ .Values.config.onapport }}/ONAPPORTAL/auxapi
+ value: https://portal-app.{{ include "common.namespace" . }}:{{ .Values.config.onapportrest }}/ONAPPORTAL/auxapi
+ - name: VID_ROLE_ACCESS_CENTRALIZED
+ value: "{{ .Values.config.roleaccesscentralized }}"
- name: VID_CONTACT_US_LINK
value: "{{ .Values.config.vidcontactuslink }}"
- name: VID_UEB_URL_LIST
apiVersion: batch/v1
kind: Job
metadata:
- name: vid-config-galera
+ name: {{ include "common.fullname" . }}-galera-config
namespace: {{ include "common.namespace" . }}
labels:
- app: vid-config-galera
+ app: {{ include "common.name" . }}-job
release: {{ .Release.Name }}
spec:
template:
metadata:
labels:
+ app: {{ include "common.name" . }}-job
release: {{ .Release.Name }}
- name: vid-galera-init
spec:
initContainers:
#This container checks that all galera instances are up before initializing it.
- - name: vid-init-galera-readiness
+ - name: {{ include "common.name" . }}-readiness
image: "{{ .Values.global.readinessRepository }}/{{ .Values.global.readinessImage }}"
imagePullPolicy: {{ .Values.global.pullPolicy | default .Values.pullPolicy }}
command:
apiVersion: v1
fieldPath: metadata.namespace
containers:
- - name: vid-config-galeradb
+ - name: {{ include "common.name" . }}-job
image: {{ .Values.mariadb_image }}
imagePullPolicy: "{{ .Values.pullPolicy }}"
volumeMounts:
# application image
repository: nexus3.onap.org:10001
-image: onap/vid:3.0-STAGING-latest
+image: onap/vid:3.0.0
pullPolicy: Always
# mariadb image for initializing
asdcclientrestauth: "Basic dmlkOktwOGJKNFNYc3pNMFdYbGhhazNlSGxjc2UyZ0F3ODR2YW9HR21KdlV5MlU="
asdcclientrestport: "8080"
vidaaiport: "8443"
- onapport: "8989"
+ onapport: "30225"
+ onapportrest: "8443"
+ portalhost: "portal.api.simpledemo.onap.org"
msoport: "8080"
vidmsopass: OBF:1ih71i271vny1yf41ymf1ylz1yf21vn41hzj1icz
msodme2serverurl: http://localhost:8081
vidmysqlmaxconnections: "5"
logstashServiceName: log-ls
logstashPort: 5044
+ roleaccesscentralized: remote
# subchart configuration
vid_mariadb_galera:
#################################################################
# application image
repository: nexus3.onap.org:10001
-image: onap/vnfsdk/refrepo:1.1.1
+image: onap/vnfsdk/refrepo:1.2.0
postgresRepository: crunchydata
postgresImage: crunchy-postgres:centos7-10.3-1.8.2
pullPolicy: Always