--- /dev/null
+apiVersion: v1
+kind: PersistentVolume
+metadata:
+ name: appc-db
+ namespace: "{{ .Values.nsPrefix }}-appc"
+ labels:
+ name: appc-db
+spec:
+ capacity:
+ storage: 2Gi
+ accessModes:
+ - ReadWriteMany
+ persistentVolumeReclaimPolicy: Retain
+ hostPath:
+ path: /dockerdata-nfs/{{ .Values.nsPrefix }}/appc/data
+---
+kind: PersistentVolumeClaim
+apiVersion: v1
+metadata:
+ name: appc-db
+ namespace: "{{ .Values.nsPrefix }}-appc"
+spec:
+ accessModes:
+ - ReadWriteMany
+ resources:
+ requests:
+ storage: 2Gi
+ selector:
+ matchLabels:
+ name: appc-db
restartPolicy: Always
volumes:
- name: appc-data
- hostPath:
- path: /dockerdata-nfs/{{ .Values.nsPrefix }}/appc/data
+ persistentVolumeClaim:
+ claimName: appc-db
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
from ubuntu:16.04
RUN mkdir -p /opt/config/src/
-VOLUME /config-init/
COPY onap-cfg.tar.gz /tmp/
RUN tar -zxvf /tmp/onap-cfg.tar.gz -C /opt/config/src/
mkdir -p /config-init/$NAMESPACE/aai/elasticsearch/es-data/
mkdir -p /config-init/$NAMESPACE/aai/search-data-service/logs/
mkdir -p /config-init/$NAMESPACE/aai/data-router/logs/
+mkdir -p /config-init/$NAMESPACE/mso/mariadb/data
chmod -R 777 /config-init/$NAMESPACE/sdc/logs/
chmod -R 777 /config-init/$NAMESPACE/portal/logs/
chmod -R 777 /config-init/$NAMESPACE/aai/elasticsearch/es-data/
chmod -R 777 /config-init/$NAMESPACE/aai/search-data-service/logs/
chmod -R 777 /config-init/$NAMESPACE/aai/data-router/logs/
+chmod -R 777 /config-init/$NAMESPACE/policy/mariadb/
# replace the default 'onap' namespace qualification of K8s hostnames within the config files
--- /dev/null
+0
+25
+ECOMP-PORTAL-OUTBOX-VID1 0 0
+PDPD-CONFIGURATION 0 2
+msgrtr.apinode.metrics.dmaap 1 26
+unauthenticated.SEC_MEASUREMENT_OUTPUT 1 1
+APPC-TEST2 0 0
+unauthenticated.TCA_EVENT_OUTPUT 1 1
+APPC-TEST1 0 0
+APPC-CL 0 2
+ECOMP-PORTAL-INBOX 0 0
+APPC-CL 1 0
+APPC-TEST2 1 1
+unauthenticated.TCA_EVENT_OUTPUT 0 1
+unauthenticated.SEC_MEASUREMENT_OUTPUT 0 1
+SDC-DISTR-NOTIF-TOPIC-SDC-OPENSOURCE-ENV1 0 0
+POLICY-CL-MGT 1 1
+PDPD-CONFIGURATION 1 0
+DCAE-CL-EVENT 1 1
+msgrtr.apinode.metrics.dmaap 0 4
+ECOMP-PORTAL-OUTBOX-APP1 0 0
+ECOMP-PORTAL-OUTBOX-SDC1 0 0
+POLICY-CL-MGT 0 1
+SDC-DISTR-STATUS-TOPIC-SDC-OPENSOURCE-ENV1 0 0
+DCAE-CL-EVENT 0 1
+ECOMP-PORTAL-OUTBOX-DBC1 0 0
+ECOMP-PORTAL-OUTBOX-POL1 0 0
--- /dev/null
+0
+25
+ECOMP-PORTAL-OUTBOX-VID1 0 0
+PDPD-CONFIGURATION 0 2
+msgrtr.apinode.metrics.dmaap 1 26
+unauthenticated.SEC_MEASUREMENT_OUTPUT 1 1
+APPC-TEST2 0 0
+unauthenticated.TCA_EVENT_OUTPUT 1 1
+APPC-TEST1 0 0
+APPC-CL 0 2
+ECOMP-PORTAL-INBOX 0 0
+APPC-CL 1 0
+APPC-TEST2 1 1
+unauthenticated.TCA_EVENT_OUTPUT 0 1
+unauthenticated.SEC_MEASUREMENT_OUTPUT 0 1
+SDC-DISTR-NOTIF-TOPIC-SDC-OPENSOURCE-ENV1 0 0
+POLICY-CL-MGT 1 1
+PDPD-CONFIGURATION 1 0
+DCAE-CL-EVENT 1 1
+msgrtr.apinode.metrics.dmaap 0 4
+ECOMP-PORTAL-OUTBOX-APP1 0 0
+ECOMP-PORTAL-OUTBOX-SDC1 0 0
+POLICY-CL-MGT 0 1
+SDC-DISTR-STATUS-TOPIC-SDC-OPENSOURCE-ENV1 0 0
+DCAE-CL-EVENT 0 1
+ECOMP-PORTAL-OUTBOX-DBC1 0 0
+ECOMP-PORTAL-OUTBOX-POL1 0 0
--- /dev/null
+#!/bin/bash
+
+if [[ -z "$KAFKA_PORT" ]]; then
+ export KAFKA_PORT=9092
+fi
+if [[ -z "$KAFKA_ADVERTISED_PORT" ]]; then
+ export KAFKA_ADVERTISED_PORT=$(docker port `hostname` $KAFKA_PORT | sed -r "s/.*:(.*)/\1/g")
+fi
+if [[ -z "$KAFKA_BROKER_ID" ]]; then
+ # By default auto allocate broker ID
+ #export KAFKA_BROKER_ID=-1
+ export KAFKA_BROKER_ID=1
+fi
+#if [[ -z "$KAFKA_LOG_DIRS" ]]; then
+ #export KAFKA_LOG_DIRS="/kafka/kafka-logs-$HOSTNAME"
+ export KAFKA_LOG_DIRS="/kafka/kafka-logs"
+#fi
+if [[ -z "$KAFKA_ZOOKEEPER_CONNECT" ]]; then
+ export KAFKA_ZOOKEEPER_CONNECT=$(env | grep ZK.*PORT_2181_TCP= | sed -e 's|.*tcp://||' | paste -sd ,)
+fi
+
+if [[ -n "$KAFKA_HEAP_OPTS" ]]; then
+ sed -r -i "s/(export KAFKA_HEAP_OPTS)=\"(.*)\"/\1=\"$KAFKA_HEAP_OPTS\"/g" $KAFKA_HOME/bin/kafka-server-start.sh
+ unset KAFKA_HEAP_OPTS
+fi
+
+if [[ -z "$KAFKA_ADVERTISED_HOST_NAME" && -n "$HOSTNAME_COMMAND" ]]; then
+ export KAFKA_ADVERTISED_HOST_NAME=$(eval $HOSTNAME_COMMAND)
+fi
+
+for VAR in `env`
+do
+ if [[ $VAR =~ ^KAFKA_ && ! $VAR =~ ^KAFKA_HOME ]]; then
+ kafka_name=`echo "$VAR" | sed -r "s/KAFKA_(.*)=.*/\1/g" | tr '[:upper:]' '[:lower:]' | tr _ .`
+ env_var=`echo "$VAR" | sed -r "s/(.*)=.*/\1/g"`
+ if egrep -q "(^|^#)$kafka_name=" $KAFKA_HOME/config/server.properties; then
+ sed -r -i "s@(^|^#)($kafka_name)=(.*)@\2=${!env_var}@g" $KAFKA_HOME/config/server.properties #note that no config values may contain an '@' char
+ else
+ echo "$kafka_name=${!env_var}" >> $KAFKA_HOME/config/server.properties
+ fi
+ fi
+done
+
+if [[ -n "$CUSTOM_INIT_SCRIPT" ]] ; then
+ eval $CUSTOM_INIT_SCRIPT
+fi
+
+
+KAFKA_PID=0
+
+# see https://medium.com/@gchudnov/trapping-signals-in-docker-containers-7a57fdda7d86#.bh35ir4u5
+term_handler() {
+ echo 'Stopping Kafka....'
+ if [ $KAFKA_PID -ne 0 ]; then
+ kill -s TERM "$KAFKA_PID"
+ wait "$KAFKA_PID"
+ fi
+ echo 'Kafka stopped.'
+ exit
+}
+
+
+# Capture kill requests to stop properly
+trap "term_handler" SIGHUP SIGINT SIGTERM
+create-topics.sh &
+$KAFKA_HOME/bin/kafka-server-start.sh $KAFKA_HOME/config/server.properties &
+KAFKA_PID=$!
+
+wait "$KAFKA_PID"
--- /dev/null
+###############################################################################
+##
+## Cambria API Server config
+##
+## - Default values are shown as commented settings.
+##
+
+###############################################################################
+##
+## HTTP service
+##
+## - 3904 is standard as of 7/29/14.
+#
+## Zookeeper Connection
+##
+## Both Cambria and Kafka make use of Zookeeper.
+##
+config.zk.servers=zookeeper.onap-dcae:2181
+#config.zk.servers=172.17.0.1:2181
+#dmaap.onap-dcae:2181
+#10.208.128.229:2181
+#config.zk.root=/fe3c/cambria/config
+
+
+###############################################################################
+##
+## Kafka Connection
+##
+## Items below are passed through to Kafka's producer and consumer
+## configurations (after removing "kafka.")
+## if you want to change request.required.acks it can take this one value
+#kafka.metadata.broker.list=localhost:9092,localhost:9093
+kafka.metadata.broker.list=kafka.onap-dcae:9092
+#kafka.metadata.broker.list=172.17.0.1:9092
+#dmaap.onap-dcae:9092
+#10.208.128.229:9092
+##kafka.request.required.acks=-1
+#kafka.client.zookeeper=${config.zk.servers}
+consumer.timeout.ms=100
+zookeeper.connection.timeout.ms=6000
+zookeeper.session.timeout.ms=6000
+zookeeper.sync.time.ms=2000
+auto.commit.interval.ms=1000
+fetch.message.max.bytes =1000000
+auto.commit.enable=false
+
+
+###############################################################################
+##
+## Secured Config
+##
+## Some data stored in the config system is sensitive -- API keys and secrets,
+## for example. to protect it, we use an encryption layer for this section
+## of the config.
+##
+## The key is a base64 encode AES key. This must be created/configured for
+## each installation.
+#cambria.secureConfig.key=
+##
+## The initialization vector is a 16 byte value specific to the secured store.
+## This must be created/configured for each installation.
+#cambria.secureConfig.iv=
+
+## Southfield Sandbox
+cambria.secureConfig.key=b/7ouTn9FfEw2PQwL0ov/Q==
+cambria.secureConfig.iv=wR9xP5k5vbz/xD0LmtqQLw==
+authentication.adminSecret=fe3cCompound
+#cambria.secureConfig.key[pc569h]=YT3XPyxEmKCTLI2NK+Sjbw==
+#cambria.secureConfig.iv[pc569h]=rMm2jhR3yVnU+u2V9Ugu3Q==
+
+
+###############################################################################
+##
+## Consumer Caching
+##
+## Kafka expects live connections from the consumer to the broker, which
+## obviously doesn't work over connectionless HTTP requests. The Cambria
+## server proxies HTTP requests into Kafka consumer sessions that are kept
+## around for later re-use. Not doing so is costly for setup per request,
+## which would substantially impact a high volume consumer's performance.
+##
+## This complicates Cambria server failover, because we often need server
+## A to close its connection before server B brings up the replacement.
+##
+
+## The consumer cache is normally enabled.
+#cambria.consumer.cache.enabled=true
+
+## Cached consumers are cleaned up after a period of disuse. The server inspects
+## consumers every sweepFreqSeconds and will clean up any connections that are
+## dormant for touchFreqMs.
+#cambria.consumer.cache.sweepFreqSeconds=15
+#cambria.consumer.cache.touchFreqMs=120000
+
+## The cache is managed through ZK. The default value for the ZK connection
+## string is the same as config.zk.servers.
+#cambria.consumer.cache.zkConnect=${config.zk.servers}
+
+##
+## Shared cache information is associated with this node's name. The default
+## name is the hostname plus the HTTP service port this host runs on. (The
+## hostname is determined via InetAddress.getLocalHost ().getCanonicalHostName(),
+## which is not always adequate.) You can set this value explicitly here.
+##
+#cambria.api.node.identifier=<use-something-unique-to-this-instance>
+
+###############################################################################
+##
+## Metrics Reporting
+##
+## This server can report its metrics periodically on a topic.
+##
+#metrics.send.cambria.enabled=true
+#metrics.send.cambria.topic=cambria.apinode.metrics #msgrtr.apinode.metrics.dmaap
+#metrics.send.cambria.sendEverySeconds=60
+
+cambria.consumer.cache.zkBasePath=/fe3c/cambria/consumerCache
+
+##############################################################################
+#100mb
+maxcontentlength=10000
+
+
+##############################################################################
+#AAF Properties
+msgRtr.namespace.aaf=org.openecomp.dcae.dmaap.mtnje2.mr.topic
+msgRtr.topicfactory.aaf=org.openecomp.dcae.dmaap.topicFactory|:org.openecomp.dcae.dmaap.mtnje2.mr.topic:
+enforced.topic.name.AAF=org.openecomp
+forceAAF=false
+transidUEBtopicreqd=false
+defaultNSforUEB=org.openecomp.dmaap.mr.ueb
+##############################################################################
+#Mirror Maker Agent
+msgRtr.mirrormakeradmin.aaf=org.openecomp.dmaap.mr.dev.mirrormaker|*|admin
+msgRtr.mirrormakeruser.aaf=org.openecomp.dmaap.mr.dev.mirrormaker|*|user
+msgRtr.mirrormakeruser.aaf.create=org.openecomp.dmaap.mr.dev.topicFactory|:org.openecomp.dmaap.mr.dev.topic:
+msgRtr.mirrormaker.timeout=15000
+msgRtr.mirrormaker.topic=org.openecomp.dmaap.mr.prod.mm.agent
+msgRtr.mirrormaker.consumergroup=mmagentserver
+msgRtr.mirrormaker.consumerid=1
--- /dev/null
+basic_realm=openecomp.org
+basic_warn=TRUE
+
+cadi_loglevel=DEBUG
+#cadi_keyfile=target/swm/package/nix/dist_files/appl/${artifactId}/etc/keyfile2
+cadi_keyfile=/appl/dmaapMR1/etc/keyfile
+# Configure AAF
+aaf_url=https://DME2RESOLVE/service=org.openecomp.authz.AuthorizationService/version=2.0/envContext=DEV/routeOffer=BAU_SE
+
+aaf_id=dgl@openecomp.org
+aaf_password=enc:f2u5br1mh29M02-
+aaf_timeout=5000
+aaf_clean_interval=1200000
+aaf_user_expires=60000
+aaf_high_count=1000000
+
+
+# The following properties are being set by the AJSC Container and should NOT need to be set here.
+AFT_LATITUDE=33.823589
+AFT_LONGITUDE=-84.366982
+AFT_ENVIRONMENT=AFTUAT
--- /dev/null
+_sNOLphPzrU7L0L3oWv0pYwgV_ddGF1XoBsQEIAp34jfP-fGJFPfFYaMpDEZ3gwH59rNw6qyMZHk
+k-4irklvVcWk36lC3twNvc0DueRCVrws1bkuhOLCXdxHJx-YG-1xM8EJfRmzh79WPlPkbAdyPmFF
+Ah44V0GjAnInPOFZA6MHP9rNx9B9qECHRfmvzU13vJCcgTsrmOr-CEiWfRsnzPjsICxpq9OaVT_D
+zn6rNaroGm1OiZNCrCgvRkCUHPOOCw3j9G1GeaImoZNYtozbz9u4sj13PU-MxIIAa64b1bMMMjpz
+Upc8lVPI4FnJKg6axMmEGn5zJ6JUq9mtOVyPj__2GEuDgpx5H4AwodXXVjFsVgR8UJwI_BvS2JVp
+JoQk0J1RqXmAXVamlsMAfzmmbARXgmrBfnuhveZnh9ymFVU-YZeujdANniXAwBGI7c6hG_BXkH7i
+Eyf4Fn41_SV78PskP6qgqJahr9r3bqdjNbKBztIKCOEVrE_w3IM5r02l-iStk_NBRkj6cq_7VCpG
+afxZ2CtZMwuZMiypO_wOgbdpCSKNzsL-NH2b4b08OlKiWb263gz634KJmV5WEfCl-6eH-JUFbWOS
+JwQfActLNT2ZQPl2MyZQNBzJEWoJRgS6k7tPRO-zqeUtYYHGHVMCxMuMHGQcoilNNHEFeBCG_fBh
+yAKb9g9F86Cbx9voMLiyTX2T3rwVHiSJFOzfNxGmfN5JWOthIun_c5hEY1tLQ15BomzkDwk7BAj7
+VbRCrVD45B6xrmSTMBSWYmLyr6mnQxQqeh9cMbD-0ZAncE3roxRnRvPKjFFa208ykYUp2V83r_PJ
+fV5I9ZPKSjk9DwFyrjkcQQEYDhdK6IFqcd6nEthjYVkmunu2fsX0bIOm9GGdIbKGqBnpdgBO5hyT
+rBr9HSlZrHcGdti1R823ckDF0Ekcl6kioDr5NLIpLtg9zUEDRm3QrbX2mv5Zs8W0pYnOqglxy3lz
+bJZTN7oR7VasHUtjmp0RT9nLZkUs5TZ6MHhlIq3ZsQ6w_Q9Rv1-ofxfwfCC4EBrWKbWAGCf6By4K
+Ew8321-2YnodhmsK5BrT4zQ1DZlmUvK8BmYjZe7wTljKjgYcsLTBfX4eMhJ7MIW1kpnl8AbiBfXh
+QzN56Mki51Q8PSQWHm0W9tnQ0z6wKdck6zBJ8JyNzewZahFKueDTn-9DOqIDfr3YHvQLLzeXyJ8e
+h4AgjW-hvlLzRGtkCknjLIgXVa3rMTycseAwbW-mgdCqqkw3SdEG8feAcyntmvE8j2jbtSDStQMB
+9JdvyNLuQdNG4pxpusgvVso0-8NQF0YVa9VFwg9U6IPSx5p8FcW68OAHt_fEgT4ZtiH7o9aur4o9
+oYqUh2lALCY-__9QLq1KkNjMKs33Jz9E8LbRerG9PLclkTrxCjYAeUWBjCwSI7OB7xkuaYDSjkjj
+a46NLpdBN1GNcsFFcZ79GFAK0_DsyxGLX8Tq6q0Bvhs8whD8wlSxpTGxYkyqNX-vcb7SDN_0WkCE
+XSdZWkqTHXcYbOvoCOb_e6SFAztuMenuHWY0utX0gBfx_X5lPDFyoYXErxFQHiA7t27keshXNa6R
+ukQRRS8kMjre1U74sc-fRNXkXpl57rG4rgxaEX0eBeowa53KAsVvUAoSac2aC_nfzXrDvoyf9Xi3
+JpEZNhUDLpFCEycV4I7jGQ9wo9qNaosvlsr6kbLDNdb_1xrGVgjT3xEvRNJNPqslSAu-yD-UFhC3
+AmCdYUnugw_eEFqXCHTARcRkdPPvl2XsmEKY2IqEeO5tz4DyXQFaL-5hEVh6lYEU1EOWHk3UGIXe
+Vc5_Ttp82qNLmlJPbZvgmNTJzYTHDQ_27KBcp7IVVZgPDjVKdWqQvZ18KhxvfF3Idgy82LBZniFV
+IbtxllXiPRxoPQriSXMnXjh3XkvSDI2pFxXfEvLRn1tvcFOwPNCz3QfPIzYg8uYXN5bRt3ZOrR_g
+ZhIlrc7HO0VbNbeqEVPKMZ-cjkqGj4VAuDKoQc0eQ6X_wCoAGO78nPpLeIvZPx1X3z5YoqNA
\ No newline at end of file
--- /dev/null
+default-character-set=latin1
+default-collation=latin1_swedish_ci
--- /dev/null
+ÿlocalhost root \ 1 VUÚXÿae9df72d0f92 root \ 1 VUÚX
\ No newline at end of file
--- /dev/null
+default-character-set=utf8
+default-collation=utf8_general_ci
--- /dev/null
+default-character-set=latin1
+default-collation=latin1_swedish_ci
--- /dev/null
+default-character-set=latin1
+default-collation=latin1_swedish_ci
COMPONENT_X_MX_MB=1024
COMPONENT_X_MS_MB=1024
-REST_PAP_URL=http://pap:9091/pap/
-REST_PDP_ID=http://pdp:8081/pdp/
+REST_PAP_URL=http://pap.onap-policy:9091/pap/
+REST_PDP_ID=http://pdp.onap-policy:8081/pdp/
PDP_HTTP_USER_ID=testpdp
PDP_HTTP_PASSWORD=alpha123
REST_CONFIG_HOME=${{POLICY_HOME}}/servers/pap/webapps/Config/
REST_ACTION_HOME=${{POLICY_HOME}}/servers/pap/webapps/Action/
-REST_CONFIG_URL=http://pap:9091/
+REST_CONFIG_URL=http://pap.onap-policy:9091/
REST_CONFIG_WEBAPPS=${{POLICY_HOME}}/servers/pap/webapps/
# PAP account information
+++ /dev/null
-192.168.141.230
# pap properties
PAP_PDPS=${{POLICY_HOME}}/servers/pap/bin/pdps
-PAP_URL=http://pap:9091/pap/
+PAP_URL=http://pap.onap-policy:9091/pap/
PAP_INITIATE_PDP=true
PAP_HEARTBEAT_INTERVAL=10000
# PDP related properties
-PAP_PDP_URL=http://pdp:8081/pdp/
+PAP_PDP_URL=http://pdp.onap-policy:8081/pdp/
PAP_PDP_HTTP_USER_ID=testpdp
PAP_PDP_HTTP_PASSWORD=alpha123
LOGPARSER_X_MS_MB=1024
LOGPARSER_X_MX_MB=1024
-SERVER=http://pap:9091/pap/
+SERVER=http://pap.onap-policy:9091/pap/
LOGPATH=${{POLICY_HOME}}/servers/pap/logs/pap-rest.log
PARSERLOGPATH=IntegrityMonitor.log
UEB_CLUSTER=dmaap.onap-message-router
-REST_PAP_URL=http://pap:9091/pap/
-REST_PDP_ID=http://pdp:8081/pdp/
+REST_PAP_URL=http://pap.onap-policy:9091/pap/
+REST_PDP_ID=http://pdp.onap-policy:8081/pdp/
REST_PDP_CONFIG=${{POLICY_HOME}}/servers/pdp/bin/config
REST_PDP_WEBAPPS=${{POLICY_HOME}}/servers/pdp/webapps
REST_PDP_REGISTER=true
LOGPARSER_X_MS_MB=1024
LOGPARSER_X_MX_MB=1024
-SERVER=http://pdp:8081/pdp/
+SERVER=http://pdp.onap-policy:8081/pdp/
LOGPATH=${{POLICY_HOME}}/servers/pdp/logs/pdp-rest.log
PARSERLOGPATH=IntegrityMonitor.log
+++ /dev/null
-#! /bin/bash
-
-
-echo "Pushing default policies"
-
-# Sometimes brmsgw gets an error when trying to retrieve the policies on initial push,
-# so for the BRMS policies we will do a push, then delete from the pdp group, then push again.
-# Second push should be successful.
-
-curl -X PUT --header 'Content-Type: application/json' --header 'Accept: text/plain' --header 'ClientAuth: cHl0aG9uOnRlc3Q=' --header 'Authorization: Basic dGVzdHJlc3Q6M2MwbXBVI2gwMUBOMWMz' --header 'Environment: TEST' -d '{
- "pdpGroup": "default",
- "policyName": "vFirewall",
- "policyScope": "com",
- "policyType": "MicroService"
-}' 'http://pypdp:8480/PyPDPServer/pushPolicy'
-
-sleep 2
-
-curl -X PUT --header 'Content-Type: application/json' --header 'Accept: text/plain' --header 'ClientAuth: cHl0aG9uOnRlc3Q=' --header 'Authorization: Basic dGVzdHJlc3Q6M2MwbXBVI2gwMUBOMWMz' --header 'Environment: TEST' -d '{
- "pdpGroup": "default",
- "policyName": "vLoadBalancer",
- "policyScope": "com",
- "policyType": "MicroService"
-}' 'http://pypdp:8480/PyPDPServer/pushPolicy'
-
-sleep 2
-curl -X PUT --header 'Content-Type: application/json' --header 'Accept: text/plain' --header 'ClientAuth: cHl0aG9uOnRlc3Q=' --header 'Authorization: Basic dGVzdHJlc3Q6M2MwbXBVI2gwMUBOMWMz' --header 'Environment: TEST' -d '{
- "pdpGroup": "default",
- "policyName": "BRMSParamvLBDemoPolicy",
- "policyScope": "com",
- "policyType": "BRMS_Param"
-}' 'http://pypdp:8480/PyPDPServer/pushPolicy'
-
-sleep 2
-
-curl -X PUT --header 'Content-Type: application/json' --header 'Accept: text/plain' --header 'ClientAuth: cHl0aG9uOnRlc3Q=' --header 'Authorization: Basic dGVzdHJlc3Q6M2MwbXBVI2gwMUBOMWMz' --header 'Environment: TEST' -d '{
- "pdpGroup": "default",
- "policyName": "BRMSParamvFWDemoPolicy",
- "policyScope": "com",
- "policyType": "BRMS_Param"
-}' 'http://pypdp:8480/PyPDPServer/pushPolicy'
-
-sleep 2
-
-curl -X DELETE --header 'Content-Type: application/json' --header 'Accept: text/plain' --header 'ClientAuth: cHl0aG9uOnRlc3Q=' --header 'Authorization: Basic dGVzdHJlc3Q6M2MwbXBVI2gwMUBOMWMz' --header 'Environment: TEST' -d '{
-"pdpGroup": "default",
-"policyComponent": "PDP",
-"policyName": "com.Config_BRMS_Param_BRMSParamvFWDemoPolicy.1.xml"
-}' 'http://pypdp:8480/PyPDPServer/deletePolicy'
-
-
-
-curl -X DELETE --header 'Content-Type: application/json' --header 'Accept: text/plain' --header 'ClientAuth: cHl0aG9uOnRlc3Q=' --header 'Authorization: Basic dGVzdHJlc3Q6M2MwbXBVI2gwMUBOMWMz' --header 'Environment: TEST' -d '{
-"pdpGroup": "default",
-"policyComponent": "PDP",
-"policyName": "com.Config_BRMS_Param_BRMSParamvLBDemoPolicy.1.xml"
-}' 'http://pypdp:8480/PyPDPServer/deletePolicy'
-
-sleep 2
-curl -X PUT --header 'Content-Type: application/json' --header 'Accept: text/plain' --header 'ClientAuth: cHl0aG9uOnRlc3Q=' --header 'Authorization: Basic dGVzdHJlc3Q6M2MwbXBVI2gwMUBOMWMz' --header 'Environment: TEST' -d '{
- "pdpGroup": "default",
- "policyName": "BRMSParamvLBDemoPolicy",
- "policyScope": "com",
- "policyType": "BRMS_Param"
-}' 'http://pypdp:8480/PyPDPServer/pushPolicy'
-
-sleep 2
-
-curl -X PUT --header 'Content-Type: application/json' --header 'Accept: text/plain' --header 'ClientAuth: cHl0aG9uOnRlc3Q=' --header 'Authorization: Basic dGVzdHJlc3Q6M2MwbXBVI2gwMUBOMWMz' --header 'Environment: TEST' -d '{
- "pdpGroup": "default",
- "policyName": "BRMSParamvFWDemoPolicy",
- "policyScope": "com",
- "policyType": "BRMS_Param"
-}' 'http://pypdp:8480/PyPDPServer/pushPolicy'
-
+++ /dev/null
-# pypdp component installation configuration parameters
-
-# tomcat specific parameters
-
-TOMCAT_JMX_PORT=9994
-TOMCAT_SHUTDOWN_PORT=8405
-SSL_HTTP_CONNECTOR_PORT=8480
-SSL_AJP_CONNECTOR_PORT=8384
-SSL_AJP_CONNECTOR_REDIRECT_PORT=8443
-
-TOMCAT_X_MS_MB=1024
-TOMCAT_X_MX_MB=1024
-
-# pypdp parameters
-
-PDP_URL=http://pdp:8081/pdp/,testpdp,alpha123
-PAP_URL=http://pap:9091/pap/,testpap,alpha123
-PYPDP_ID=testrest
-PYPDP_PASSWORD=3c0mpU#h01@N1c3
-
-node_type=pypdp
-# the java property is RESOURCE_NAME (uppercase), but the conf parameter is lowercase
-resource_name=pypdp_1
-
-CLIENT_FILE=client.properties
hostPath:
path: /var/run/docker.sock
- name: kafka-data
- hostPath:
- path: /dockerdata-nfs/{{ .Values.nsPrefix }}/message-router/dcae-startup-vm-message-router/docker_files/data-kafka/
+ persistentVolumeClaim:
+ claimName: message-router-kafka
- name: start-kafka
hostPath:
path: /dockerdata-nfs/{{ .Values.nsPrefix }}/message-router/dcae-startup-vm-message-router/docker_files/start-kafka.sh
--- /dev/null
+apiVersion: v1
+kind: PersistentVolume
+metadata:
+ name: message-router-kafka
+ namespace: "{{ .Values.nsPrefix }}-message-router"
+ labels:
+ name: message-router-kafka
+spec:
+ capacity:
+ storage: 2Gi
+ accessModes:
+ - ReadWriteMany
+ persistentVolumeReclaimPolicy: Retain
+ hostPath:
+ path: /dockerdata-nfs/{{ .Values.nsPrefix }}/message-router/dcae-startup-vm-message-router/docker_files/data-kafka/
+---
+kind: PersistentVolumeClaim
+apiVersion: v1
+metadata:
+ name: message-router-kafka
+ namespace: "{{ .Values.nsPrefix }}-message-router"
+spec:
+ accessModes:
+ - ReadWriteMany
+ resources:
+ requests:
+ storage: 2Gi
+ selector:
+ matchLabels:
+ name: message-router-kafka
+---
+apiVersion: v1
+kind: PersistentVolume
+metadata:
+ name: message-router-zookeeper
+ namespace: "{{ .Values.nsPrefix }}-message-router"
+ labels:
+ name: message-router-zookeeper
+spec:
+ capacity:
+ storage: 2Gi
+ accessModes:
+ - ReadWriteMany
+ persistentVolumeReclaimPolicy: Retain
+ hostPath:
+ path: /dockerdata-nfs/{{ .Values.nsPrefix }}/message-router/dcae-startup-vm-message-router/docker_files/data-zookeeper
+---
+kind: PersistentVolumeClaim
+apiVersion: v1
+metadata:
+ name: message-router-zookeeper
+ namespace: "{{ .Values.nsPrefix }}-message-router"
+spec:
+ accessModes:
+ - ReadWriteMany
+ resources:
+ requests:
+ storage: 2Gi
+ selector:
+ matchLabels:
+ name: message-router-zookeeper
restartPolicy: Always
volumes:
- name: zookeeper-data
- hostPath:
- path: /dockerdata-nfs/{{ .Values.nsPrefix }}/message-router/dcae-startup-vm-message-router/docker_files/data-zookeeper
+ persistentVolumeClaim:
+ claimName: message-router-zookeeper
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
name: mso-mariadb-conf
- mountPath: /docker-entrypoint-initdb.d
name: mso-mariadb-docker-entrypoint-initdb
+ - mountPath: /var/lib/mysql
+ name: mso-mariadb-data
ports:
- containerPort: 3306
name: mariadb
- name: mso-mariadb-docker-entrypoint-initdb
hostPath:
path: /dockerdata-nfs/{{ .Values.nsPrefix }}/mso/mariadb/docker-entrypoint-initdb.d
+ - name: mso-mariadb-data
+ persistentVolumeClaim:
+ claimName: mso-db
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
--- /dev/null
+apiVersion: v1
+kind: PersistentVolume
+metadata:
+ name: mso-db
+ namespace: "{{ .Values.nsPrefix }}-mso"
+ labels:
+ name: mso-db
+spec:
+ capacity:
+ storage: 2Gi
+ accessModes:
+ - ReadWriteMany
+ persistentVolumeReclaimPolicy: Retain
+ hostPath:
+ path: /dockerdata-nfs/{{ .Values.nsPrefix }}/mso/mariadb/data
+---
+kind: PersistentVolumeClaim
+apiVersion: v1
+metadata:
+ name: mso-db
+ namespace: "{{ .Values.nsPrefix }}-mso"
+spec:
+ accessModes:
+ - ReadWriteMany
+ resources:
+ requests:
+ storage: 2Gi
+ selector:
+ matchLabels:
+ name: mso-db
}
create_onap_helm() {
- helm install ../$2/ --name $1-$2 --namespace $1 --set nsPrefix=$1 --set nodePortPrefix=$3
+ helm install ../$2/ --name $1-$2 --namespace $1 --set nsPrefix=$1,nodePortPrefix=$3
}
"--container-name",
"pap",
"--container-name",
- "pdp",
- "--container-name",
- "pypdp"
+ "pdp"
],
"command": [
"/root/ready.py"
}
}
],
- "image": "{{ .Values.image.readiness }}",
+ "image": "{{ .Values.image.readiness }}:{{ .Values.image.readinessVersion }}",
"imagePullPolicy": "{{ .Values.pullPolicy }}",
"name": "brmsgw-readiness"
}
- /bin/bash
- ./do-start.sh
- brmsgw
- image: {{ .Values.image.policyPe }}
+ image: "{{ .Values.image.policyPe }}:{{ .Values.image.policyPeVersion}}"
imagePullPolicy: {{ .Values.pullPolicy }}
name: brmsgw
volumeMounts:
"--container-name",
"pdp",
"--container-name",
- "pypdp",
- "--container-name",
"brmsgw"
],
"command": [
}
}
],
- "image": "{{ .Values.image.readiness }}",
+ "image": "{{ .Values.image.readiness }}:{{ .Values.image.readinessVersion }}",
"imagePullPolicy": "{{ .Values.pullPolicy }}",
"name": "drools-readiness"
}
- /bin/bash
- -c
- ./do-start.sh
- image: {{ .Values.image.policyDrools }}
+ image: "{{ .Values.image.policyDrools }}:{{ .Values.image.policyDroolsVersion }}"
imagePullPolicy: {{ .Values.pullPolicy }}
name: drools
ports:
- /bin/bash
- -c
- exec bash /tmp/do-start.sh
- image: {{ .Values.image.policyDb }}
+ image: "{{ .Values.image.policyDb }}:{{ .Values.image.policyDbVersion }}"
imagePullPolicy: {{ .Values.pullPolicy }}
name: mariadb
ports:
- containerPort: 3306
+ volumeMounts:
+ - mountPath: /var/lib/mysql
+ name: policy-mariadb-data
readinessProbe:
tcpSocket:
port: 3306
initialDelaySeconds: 5
periodSeconds: 10
+ volumes:
+ - name: policy-mariadb-data
+ persistentVolumeClaim:
+ claimName: policy-db
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
}
}
],
- "image": "{{ .Values.image.readiness }}",
+ "image": "{{ .Values.image.readiness }}:{{ .Values.image.readinessVersion }}",
"imagePullPolicy": "{{ .Values.pullPolicy }}",
"name": "nexus-readiness"
}
- /bin/bash
- -c
- bash -c "/opt/nexus/nexus-2.14.2-01/bin/nexus start && sleep 1000d"
- image: {{ .Values.image.policyNexus }}
+ image: "{{ .Values.image.policyNexus }}:{{ .Values.image.policyNexusVersion }}"
imagePullPolicy: {{ .Values.pullPolicy }}
name: nexus
imagePullSecrets:
}
}
],
- "image": "{{ .Values.image.readiness }}",
+ "image": "{{ .Values.image.readiness }}:{{ .Values.image.readinessVersion }}",
"imagePullPolicy": "{{ .Values.pullPolicy }}",
"name": "pap-readiness"
},
- /bin/bash
- ./do-start.sh
- pap
- image: {{ .Values.image.policyPe }}
+ image: "{{ .Values.image.policyPe }}:{{ .Values.image.policyPeVersion }}"
imagePullPolicy: {{ .Values.pullPolicy }}
name: pap
ports:
}
}
],
- "image": "{{ .Values.image.readiness }}",
+ "image": "{{ .Values.image.readiness }}:{{ .Values.image.readinessVersion }}",
"imagePullPolicy": "{{ .Values.pullPolicy }}",
"name": "pdp-readiness"
}
- /bin/bash
- ./do-start.sh
- pdp
- image: {{ .Values.image.policyPe }}
+ image: "{{ .Values.image.policyPe }}:{{ .Values.image.policyPeVersion }}"
imagePullPolicy: {{ .Values.pullPolicy }}
name: pdp
ports:
+++ /dev/null
-apiVersion: extensions/v1beta1
-kind: Deployment
-metadata:
- name: pypdp
- namespace: "{{ .Values.nsPrefix }}-policy"
-spec:
- replicas: 1
- selector:
- matchLabels:
- app: pypdp
- template:
- metadata:
- labels:
- app: pypdp
- name: pypdp
- annotations:
- pod.beta.kubernetes.io/init-containers: '[
- {
- "args": [
- "--container-name",
- "mariadb",
- "--container-name",
- "nexus",
- "--container-name",
- "pap",
- "--container-name",
- "pdp"
- ],
- "command": [
- "/root/ready.py"
- ],
- "env": [
- {
- "name": "NAMESPACE",
- "valueFrom": {
- "fieldRef": {
- "apiVersion": "v1",
- "fieldPath": "metadata.namespace"
- }
- }
- }
- ],
- "image": "{{ .Values.image.readiness }}",
- "imagePullPolicy": "{{ .Values.pullPolicy }}",
- "name": "pypdp-readiness"
- }
- ]'
- spec:
- containers:
- - command:
- - /bin/bash
- - ./do-start.sh
- - pypdp
- image: {{ .Values.image.policyPe }}
- imagePullPolicy: {{ .Values.pullPolicy }}
- name: pypdp
- ports:
- - containerPort: 8480
- readinessProbe:
- tcpSocket:
- port: 8480
- initialDelaySeconds: 5
- periodSeconds: 10
- volumeMounts:
- - mountPath: /tmp/policy-install/config
- name: pe
- volumes:
- - name: pe
- hostPath:
- path: /dockerdata-nfs/{{ .Values.nsPrefix }}/policy/opt/policy/config/pe/
- imagePullSecrets:
- - name: "{{ .Values.nsPrefix }}-docker-registry-key"
--- /dev/null
+apiVersion: v1
+kind: PersistentVolume
+metadata:
+ name: policy-db
+ namespace: "{{ .Values.nsPrefix }}-policy"
+ labels:
+ name: policy-db
+spec:
+ capacity:
+ storage: 2Gi
+ accessModes:
+ - ReadWriteMany
+ persistentVolumeReclaimPolicy: Retain
+ hostPath:
+ path: /dockerdata-nfs/{{ .Values.nsPrefix }}/policy/mariadb/data/
+---
+kind: PersistentVolumeClaim
+apiVersion: v1
+metadata:
+ name: policy-db
+ namespace: "{{ .Values.nsPrefix }}-policy"
+spec:
+ accessModes:
+ - ReadWriteMany
+ resources:
+ requests:
+ storage: 2Gi
+ selector:
+ matchLabels:
+ name: policy-db
pullPolicy: Always
nodePortPrefix: 302
image:
- readiness: oomk8s/readiness-check:1.0.0
- policyPe: nexus3.onap.org:10001/openecomp/policy/policy-pe:1.0-STAGING-latest
- policyDrools: nexus3.onap.org:10001/openecomp/policy/policy-drools:1.0-STAGING-latest
- policyDb: nexus3.onap.org:10001/openecomp/policy/policy-db:1.0-STAGING-latest
- policyNexus: nexus3.onap.org:10001/openecomp/policy/policy-nexus:1.0-STAGING-latest
+ readiness: oomk8s/readiness-check
+ readinessVersion: 1.0.0
+ policyPe: nexus3.onap.org:10001/openecomp/policy/policy-pe
+ policyPeVersion: 1.1-STAGING-latest
+ policyDrools: nexus3.onap.org:10001/openecomp/policy/policy-drools
+ policyDroolsVersion: 1.1-STAGING-latest
+ policyDb: nexus3.onap.org:10001/openecomp/policy/policy-db
+ policyDbVersion: 1.1-STAGING-latest
+ policyNexus: nexus3.onap.org:10001/openecomp/policy/policy-nexus
+ policyNexusVersion: 1.1-STAGING-latest
ubuntu: ubuntu:16.04
periodSeconds: 10
volumes:
- name: portal-mariadb-data
- hostPath:
- path: /dockerdata-nfs/{{ .Values.nsPrefix }}/portal/mariadb/data
+ persistentVolumeClaim:
+ claimName: portal-db
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
--- /dev/null
+apiVersion: v1
+kind: PersistentVolume
+metadata:
+ name: portal-db
+ namespace: "{{ .Values.nsPrefix }}-portal"
+ labels:
+ name: portal-db
+spec:
+ capacity:
+ storage: 2Gi
+ accessModes:
+ - ReadWriteMany
+ persistentVolumeReclaimPolicy: Retain
+ hostPath:
+ path: /dockerdata-nfs/{{ .Values.nsPrefix }}/portal/mariadb/data
+---
+kind: PersistentVolumeClaim
+apiVersion: v1
+metadata:
+ name: portal-db
+ namespace: "{{ .Values.nsPrefix }}-portal"
+spec:
+ accessModes:
+ - ReadWriteMany
+ resources:
+ requests:
+ storage: 2Gi
+ selector:
+ matchLabels:
+ name: portal-db
periodSeconds: 10
volumes:
- name: sdc-sdc-cs-cs
- hostPath:
- path: /dockerdata-nfs/{{ .Values.nsPrefix }}/sdc/sdc-cs/CS
+ persistentVolumeClaim:
+ claimName: sdc-cs-db
- name: sdc-environments
hostPath:
path: /dockerdata-nfs/{{ .Values.nsPrefix }}/sdc/environments
--- /dev/null
+apiVersion: v1
+kind: PersistentVolume
+metadata:
+ name: sdc-cs-db
+ namespace: "{{ .Values.nsPrefix }}-sdc"
+ labels:
+ name: sdc-cs-db
+spec:
+ capacity:
+ storage: 2Gi
+ accessModes:
+ - ReadWriteMany
+ persistentVolumeReclaimPolicy: Retain
+ hostPath:
+ path: /dockerdata-nfs/{{ .Values.nsPrefix }}/sdc/sdc-cs/CS
+---
+kind: PersistentVolumeClaim
+apiVersion: v1
+metadata:
+ name: sdc-cs-db
+ namespace: "{{ .Values.nsPrefix }}-sdc"
+spec:
+ accessModes:
+ - ReadWriteMany
+ resources:
+ requests:
+ storage: 2Gi
+ selector:
+ matchLabels:
+ name: sdc-cs-db
periodSeconds: 10
volumes:
- name: sdnc-data
- hostPath:
- path: /dockerdata-nfs/{{ .Values.nsPrefix }}/sdnc/data
+ persistentVolumeClaim:
+ claimName: sdnc-db
imagePullSecrets:
- name: "{{ .Values.nsPrefix }}-docker-registry-key"
--- /dev/null
+apiVersion: v1
+kind: PersistentVolume
+metadata:
+ name: sdnc-db
+ namespace: "{{ .Values.nsPrefix }}-sdnc"
+ labels:
+ name: sdnc-db
+spec:
+ capacity:
+ storage: 2Gi
+ accessModes:
+ - ReadWriteMany
+ persistentVolumeReclaimPolicy: Retain
+ hostPath:
+ path: /dockerdata-nfs/{{ .Values.nsPrefix }}/sdnc/data
+---
+kind: PersistentVolumeClaim
+apiVersion: v1
+metadata:
+ name: sdnc-db
+ namespace: "{{ .Values.nsPrefix }}-sdnc"
+spec:
+ accessModes:
+ - ReadWriteMany
+ resources:
+ requests:
+ storage: 2Gi
+ selector:
+ matchLabels:
+ name: sdnc-db
periodSeconds: 10
volumes:
- name: vid-mariadb-data
- hostPath:
- path: /dockerdata-nfs/{{ .Values.nsPrefix }}/vid/mariadb/data
+ persistentVolumeClaim:
+ claimName: vid-db
- name: vid-pre-init
hostPath:
path: /dockerdata-nfs/{{ .Values.nsPrefix }}/vid/vid/lf_config/vid-pre-init.sql
--- /dev/null
+apiVersion: v1
+kind: PersistentVolume
+metadata:
+ name: vid-db
+ namespace: "{{ .Values.nsPrefix }}-vid"
+ labels:
+ name: vid-db
+spec:
+ capacity:
+ storage: 2Gi
+ accessModes:
+ - ReadWriteMany
+ persistentVolumeReclaimPolicy: Retain
+ hostPath:
+ path: /dockerdata-nfs/{{ .Values.nsPrefix }}/vid/mariadb/data
+---
+kind: PersistentVolumeClaim
+apiVersion: v1
+metadata:
+ name: vid-db
+ namespace: "{{ .Values.nsPrefix }}-vid"
+spec:
+ accessModes:
+ - ReadWriteMany
+ resources:
+ requests:
+ storage: 2Gi
+ selector:
+ matchLabels:
+ name: vid-db