Merge "limit cpu of logstash pods"
authorBorislav Glozman <Borislav.Glozman@amdocs.com>
Sun, 3 Jun 2018 15:22:43 +0000 (15:22 +0000)
committerGerrit Code Review <gerrit@onap.org>
Sun, 3 Jun 2018 15:22:43 +0000 (15:22 +0000)
62 files changed:
docs/Access_Rancher_server_via_web_browser.jpeg [new file with mode: 0644]
docs/Add_Kubernetes_Environment_to_Rancher.png [new file with mode: 0644]
docs/Apply_customization_script_for_the_Rancher_VM.jpeg [new file with mode: 0644]
docs/Click_create.jpeg [new file with mode: 0644]
docs/Click_on_CLI_and_then_click_on_Generate_Config.jpeg [new file with mode: 0644]
docs/Click_on_Close_button.jpeg [new file with mode: 0644]
docs/Click_on_Copy_to_Clipboard-wait_until_you_see_a_token-do_not_copy_user+password-the_server_is_not_ready_at_that_point.jpeg [new file with mode: 0644]
docs/K8s-Assign_Floating_IP_for_external_access.jpeg [new file with mode: 0644]
docs/K8s-Key_Pair.jpeg [new file with mode: 0644]
docs/K8s-Launch_Instance.jpeg [new file with mode: 0644]
docs/K8s-Launch_new_VM_instance_to_create_a_Kubernetes_single_host_or_cluster.jpeg [new file with mode: 0644]
docs/K8s-Manage_Floating_IP_Associations.jpeg [new file with mode: 0644]
docs/K8s-Networking.jpeg [new file with mode: 0644]
docs/K8s-Security_Group.jpeg [new file with mode: 0644]
docs/K8s-Select_Flavor.jpeg [new file with mode: 0644]
docs/K8s-Select_Ubuntu_16.04_as_base_image.jpeg [new file with mode: 0644]
docs/Rancher-Allocate_Floating_IP.jpeg [new file with mode: 0644]
docs/Rancher-Key_Pair.jpeg [new file with mode: 0644]
docs/Rancher-Launch_Instance.jpeg [new file with mode: 0644]
docs/Rancher-Launch_new_VM_instance_to_host_the_Rancher_Server.jpeg [new file with mode: 0644]
docs/Rancher-Manage_Floating_IP_Associations.jpeg [new file with mode: 0644]
docs/Rancher-Networking.jpeg [new file with mode: 0644]
docs/Rancher-Security_Groups.jpeg [new file with mode: 0644]
docs/Rancher-Select_Flavor.jpeg [new file with mode: 0644]
docs/Rancher-Select_Ubuntu_16.04_as_base_image.jpeg [new file with mode: 0644]
docs/Return_to_Rancher_environment_eg_SB4_and_wait_for_services_to_complete_10-15_mins.jpeg [new file with mode: 0644]
docs/Select_Add_Environment.png [new file with mode: 0644]
docs/and_click_on_Save_accept_defaults.jpeg [new file with mode: 0644]
docs/index.rst
docs/k8s-topology.jpg [new file with mode: 0644]
docs/master_nfs_node.sh [new file with mode: 0644]
docs/oom_cloud_setup_guide.rst
docs/oom_setup_kubernetes_rancher.rst [new file with mode: 0644]
docs/openstack-k8s-node.sh [new file with mode: 0644]
docs/openstack-rancher.sh [new file with mode: 0644]
docs/otherwise_select_INFRASTRUCTURE_Hosts_and_click_on_Add_Host.jpg [new file with mode: 0644]
docs/slave_nfs_node.sh [new file with mode: 0644]
kubernetes/aai/charts/aai-babel/values.yaml
kubernetes/aai/charts/aai-champ/values.yaml
kubernetes/aai/charts/aai-data-router/values.yaml
kubernetes/aai/charts/aai-gizmo/resources/config/log/logback.xml
kubernetes/aai/charts/aai-gizmo/values.yaml
kubernetes/aai/charts/aai-modelloader/values.yaml
kubernetes/aai/charts/aai-resources/values.yaml
kubernetes/aai/charts/aai-search-data/values.yaml
kubernetes/aai/charts/aai-sparky-be/values.yaml
kubernetes/aai/charts/aai-traversal/values.yaml
kubernetes/appc/resources/config/appc/opt/onap/appc/bin/startODL.sh
kubernetes/appc/templates/service.yaml
kubernetes/appc/values.yaml
kubernetes/oof/charts/oof-has/charts/oof-has-data/templates/deployment.yaml
kubernetes/oof/charts/oof-has/resources/config/AAF_RootCA.cer [new file with mode: 0755]
kubernetes/oof/charts/oof-has/resources/config/conductor.conf
kubernetes/oof/charts/oof-has/values.yaml
kubernetes/oof/resources/config/osdf_config.yaml
kubernetes/oof/values.yaml
kubernetes/policy/charts/brmsgw/values.yaml
kubernetes/policy/charts/drools/values.yaml
kubernetes/policy/charts/pdp/values.yaml
kubernetes/policy/values.yaml
kubernetes/sdnc/resources/config/bin/installSdncDb.sh
kubernetes/so/resources/config/mso/mso-docker.json

diff --git a/docs/Access_Rancher_server_via_web_browser.jpeg b/docs/Access_Rancher_server_via_web_browser.jpeg
new file mode 100644 (file)
index 0000000..103fc27
Binary files /dev/null and b/docs/Access_Rancher_server_via_web_browser.jpeg differ
diff --git a/docs/Add_Kubernetes_Environment_to_Rancher.png b/docs/Add_Kubernetes_Environment_to_Rancher.png
new file mode 100644 (file)
index 0000000..0d60f15
Binary files /dev/null and b/docs/Add_Kubernetes_Environment_to_Rancher.png differ
diff --git a/docs/Apply_customization_script_for_the_Rancher_VM.jpeg b/docs/Apply_customization_script_for_the_Rancher_VM.jpeg
new file mode 100644 (file)
index 0000000..832dda3
Binary files /dev/null and b/docs/Apply_customization_script_for_the_Rancher_VM.jpeg differ
diff --git a/docs/Click_create.jpeg b/docs/Click_create.jpeg
new file mode 100644 (file)
index 0000000..fad1856
Binary files /dev/null and b/docs/Click_create.jpeg differ
diff --git a/docs/Click_on_CLI_and_then_click_on_Generate_Config.jpeg b/docs/Click_on_CLI_and_then_click_on_Generate_Config.jpeg
new file mode 100644 (file)
index 0000000..fec6a35
Binary files /dev/null and b/docs/Click_on_CLI_and_then_click_on_Generate_Config.jpeg differ
diff --git a/docs/Click_on_Close_button.jpeg b/docs/Click_on_Close_button.jpeg
new file mode 100644 (file)
index 0000000..cb02480
Binary files /dev/null and b/docs/Click_on_Close_button.jpeg differ
diff --git a/docs/Click_on_Copy_to_Clipboard-wait_until_you_see_a_token-do_not_copy_user+password-the_server_is_not_ready_at_that_point.jpeg b/docs/Click_on_Copy_to_Clipboard-wait_until_you_see_a_token-do_not_copy_user+password-the_server_is_not_ready_at_that_point.jpeg
new file mode 100644 (file)
index 0000000..bb0f60c
Binary files /dev/null and b/docs/Click_on_Copy_to_Clipboard-wait_until_you_see_a_token-do_not_copy_user+password-the_server_is_not_ready_at_that_point.jpeg differ
diff --git a/docs/K8s-Assign_Floating_IP_for_external_access.jpeg b/docs/K8s-Assign_Floating_IP_for_external_access.jpeg
new file mode 100644 (file)
index 0000000..fe84d11
Binary files /dev/null and b/docs/K8s-Assign_Floating_IP_for_external_access.jpeg differ
diff --git a/docs/K8s-Key_Pair.jpeg b/docs/K8s-Key_Pair.jpeg
new file mode 100644 (file)
index 0000000..1e335d3
Binary files /dev/null and b/docs/K8s-Key_Pair.jpeg differ
diff --git a/docs/K8s-Launch_Instance.jpeg b/docs/K8s-Launch_Instance.jpeg
new file mode 100644 (file)
index 0000000..7e2122c
Binary files /dev/null and b/docs/K8s-Launch_Instance.jpeg differ
diff --git a/docs/K8s-Launch_new_VM_instance_to_create_a_Kubernetes_single_host_or_cluster.jpeg b/docs/K8s-Launch_new_VM_instance_to_create_a_Kubernetes_single_host_or_cluster.jpeg
new file mode 100644 (file)
index 0000000..8d154d4
Binary files /dev/null and b/docs/K8s-Launch_new_VM_instance_to_create_a_Kubernetes_single_host_or_cluster.jpeg differ
diff --git a/docs/K8s-Manage_Floating_IP_Associations.jpeg b/docs/K8s-Manage_Floating_IP_Associations.jpeg
new file mode 100644 (file)
index 0000000..f15e5e3
Binary files /dev/null and b/docs/K8s-Manage_Floating_IP_Associations.jpeg differ
diff --git a/docs/K8s-Networking.jpeg b/docs/K8s-Networking.jpeg
new file mode 100644 (file)
index 0000000..ee4cafc
Binary files /dev/null and b/docs/K8s-Networking.jpeg differ
diff --git a/docs/K8s-Security_Group.jpeg b/docs/K8s-Security_Group.jpeg
new file mode 100644 (file)
index 0000000..5e85218
Binary files /dev/null and b/docs/K8s-Security_Group.jpeg differ
diff --git a/docs/K8s-Select_Flavor.jpeg b/docs/K8s-Select_Flavor.jpeg
new file mode 100644 (file)
index 0000000..c3cf4b2
Binary files /dev/null and b/docs/K8s-Select_Flavor.jpeg differ
diff --git a/docs/K8s-Select_Ubuntu_16.04_as_base_image.jpeg b/docs/K8s-Select_Ubuntu_16.04_as_base_image.jpeg
new file mode 100644 (file)
index 0000000..e9a03e7
Binary files /dev/null and b/docs/K8s-Select_Ubuntu_16.04_as_base_image.jpeg differ
diff --git a/docs/Rancher-Allocate_Floating_IP.jpeg b/docs/Rancher-Allocate_Floating_IP.jpeg
new file mode 100644 (file)
index 0000000..9d6ff3f
Binary files /dev/null and b/docs/Rancher-Allocate_Floating_IP.jpeg differ
diff --git a/docs/Rancher-Key_Pair.jpeg b/docs/Rancher-Key_Pair.jpeg
new file mode 100644 (file)
index 0000000..1bcb6b6
Binary files /dev/null and b/docs/Rancher-Key_Pair.jpeg differ
diff --git a/docs/Rancher-Launch_Instance.jpeg b/docs/Rancher-Launch_Instance.jpeg
new file mode 100644 (file)
index 0000000..3894369
Binary files /dev/null and b/docs/Rancher-Launch_Instance.jpeg differ
diff --git a/docs/Rancher-Launch_new_VM_instance_to_host_the_Rancher_Server.jpeg b/docs/Rancher-Launch_new_VM_instance_to_host_the_Rancher_Server.jpeg
new file mode 100644 (file)
index 0000000..2b3fde3
Binary files /dev/null and b/docs/Rancher-Launch_new_VM_instance_to_host_the_Rancher_Server.jpeg differ
diff --git a/docs/Rancher-Manage_Floating_IP_Associations.jpeg b/docs/Rancher-Manage_Floating_IP_Associations.jpeg
new file mode 100644 (file)
index 0000000..ae062c1
Binary files /dev/null and b/docs/Rancher-Manage_Floating_IP_Associations.jpeg differ
diff --git a/docs/Rancher-Networking.jpeg b/docs/Rancher-Networking.jpeg
new file mode 100644 (file)
index 0000000..16ef380
Binary files /dev/null and b/docs/Rancher-Networking.jpeg differ
diff --git a/docs/Rancher-Security_Groups.jpeg b/docs/Rancher-Security_Groups.jpeg
new file mode 100644 (file)
index 0000000..c8ff2e8
Binary files /dev/null and b/docs/Rancher-Security_Groups.jpeg differ
diff --git a/docs/Rancher-Select_Flavor.jpeg b/docs/Rancher-Select_Flavor.jpeg
new file mode 100644 (file)
index 0000000..a3f35b5
Binary files /dev/null and b/docs/Rancher-Select_Flavor.jpeg differ
diff --git a/docs/Rancher-Select_Ubuntu_16.04_as_base_image.jpeg b/docs/Rancher-Select_Ubuntu_16.04_as_base_image.jpeg
new file mode 100644 (file)
index 0000000..b80fae4
Binary files /dev/null and b/docs/Rancher-Select_Ubuntu_16.04_as_base_image.jpeg differ
diff --git a/docs/Return_to_Rancher_environment_eg_SB4_and_wait_for_services_to_complete_10-15_mins.jpeg b/docs/Return_to_Rancher_environment_eg_SB4_and_wait_for_services_to_complete_10-15_mins.jpeg
new file mode 100644 (file)
index 0000000..6d62b99
Binary files /dev/null and b/docs/Return_to_Rancher_environment_eg_SB4_and_wait_for_services_to_complete_10-15_mins.jpeg differ
diff --git a/docs/Select_Add_Environment.png b/docs/Select_Add_Environment.png
new file mode 100644 (file)
index 0000000..48fffa4
Binary files /dev/null and b/docs/Select_Add_Environment.png differ
diff --git a/docs/and_click_on_Save_accept_defaults.jpeg b/docs/and_click_on_Save_accept_defaults.jpeg
new file mode 100644 (file)
index 0000000..9b57927
Binary files /dev/null and b/docs/and_click_on_Save_accept_defaults.jpeg differ
index 712802d..340b43b 100644 (file)
@@ -14,3 +14,4 @@ OOM Documentation Repository
   oom_developer_guide.rst
   oom_cloud_setup_guide.rst
   release-notes.rst
+  oom_setup_kubernetes_rancher.rst
diff --git a/docs/k8s-topology.jpg b/docs/k8s-topology.jpg
new file mode 100644 (file)
index 0000000..eba8e72
Binary files /dev/null and b/docs/k8s-topology.jpg differ
diff --git a/docs/master_nfs_node.sh b/docs/master_nfs_node.sh
new file mode 100644 (file)
index 0000000..4a7a8db
--- /dev/null
@@ -0,0 +1,32 @@
+#!/bin/bash
+
+usage () {
+  echo "Usage:"
+  echo "   ./$(basename $0) node1_ip node2_ip ... nodeN_ip"
+  exit 1
+}
+
+if [ "$#" -lt 1 ]; then
+  echo "Missing NFS slave nodes"
+  usage
+fi
+
+#Install NFS kernel
+sudo apt-get update
+sudo apt-get install -y nfs-kernel-server
+
+#Create /dockerdata-nfs and set permissions
+sudo mkdir -p /dockerdata-nfs
+sudo chmod 777 -R /dockerdata-nfs
+sudo chown nobody:nogroup /dockerdata-nfs/
+
+#Update the /etc/exports
+NFS_EXP=""
+for i in $@; do
+  NFS_EXP+="$i(rw,sync,no_root_squash,no_subtree_check) "
+done
+echo "/dockerdata-nfs "$NFS_EXP | sudo tee -a /etc/exports
+
+#Restart the NFS service
+sudo exportfs -a
+sudo systemctl restart nfs-kernel-server
index 7582a4a..7350966 100644 (file)
@@ -12,6 +12,7 @@
 .. _Setting Up Kubernetes with Rancher: https://wiki.onap.org/display/DW/ONAP+on+Kubernetes+on+Rancher
 .. _Setting Up Kubernetes with Kubeadm: https://wiki.onap.org/display/DW/Deploying+Kubernetes+Cluster+with+kubeadm
 .. _Setting Up Kubernetes with Cloudify: https://wiki.onap.org/display/DW/ONAP+on+Kubernetes+on+Cloudify
+.. _ONAP on Kubernetes Wiki: https://wiki.onap.org/display/DW/ONAP+on+Kubernetes
 
 .. figure:: oomLogoV2-medium.png
    :align: right
@@ -63,26 +64,38 @@ least three if not six nodes to ensure there is no single point of failure.
   =====  =====  ======  ====================
   RAM    HD     vCores  Ports
   =====  =====  ======  ====================
-  128GB  160GB  16      0.0.0.0/0 (all open)
+  128GB  160GB  32      0.0.0.0/0 (all open)
   =====  =====  ======  ====================
 
+.. note::
+  Kubernetes supports a maximum of 110 pods per node which forces one to use at
+  least two nodes to deploy all of ONAP although at least three are recommended
+  (for example 4x32GB - 8 vCores each). Subsets of ONAP may still be deployed
+  on a single node.
+
 Cloud Installation
 ==================
 
-#. OOM supports deployment on major public clouds. The following guides
-   provide instructions on how to deploy ONAP on these clouds:
-
-   - `Microsoft Azure`_,
-   - `Amazon AWS`_,
-   - `Google GCE`_,
-   - `VMware VIO`_,
-   - IBM, and
-   - `Openstack`_.
-
-#. Alternatively, OOM can be deployed on a private set of physical hosts or VMs
-   (or even a combination of the two). The following guides describe how to
-   create a Kubernetes cluster with popular tools:
-
-   - `Setting up Kubernetes with Rancher`_ (recommended)
-   - `Setting up Kubernetes with Kubeadm`_
-   - `Setting up Kubernetes with Cloudify`_
+.. #. OOM supports deployment on major public clouds. The following guides
+..    provide instructions on how to deploy ONAP on these clouds:
+..
+..    - `Microsoft Azure`_,
+..    - `Amazon AWS`_,
+..    - `Google GCE`_,
+..    - `VMware VIO`_,
+..    - IBM, and
+..    - `Openstack`_.
+..
+.. #. Alternatively, OOM can be deployed on a private set of physical hosts or VMs
+..    (or even a combination of the two). The following guides describe how to
+..    create a Kubernetes cluster with popular tools:
+..
+..    - `Setting up Kubernetes with Rancher`_ (recommended)
+..    - `Setting up Kubernetes with Kubeadm`_
+..    - `Setting up Kubernetes with Cloudify`_
+
+OOM can be deployed on a private set of physical hosts or VMs (or even a
+combination of the two). The following guide describe the recommended method to
+setup a Kubernetes cluster: :ref:`onap-on-kubernetes-with-rancher`.
+
+There are alternative deployment methods described on the `ONAP on Kubernetes Wiki`_
diff --git a/docs/oom_setup_kubernetes_rancher.rst b/docs/oom_setup_kubernetes_rancher.rst
new file mode 100644 (file)
index 0000000..8e55bf1
--- /dev/null
@@ -0,0 +1,478 @@
+.. This work is licensed under a Creative Commons Attribution 4.0 International License.
+.. http://creativecommons.org/licenses/by/4.0
+.. Copyright 2018 Amdocs, Bell Canada
+
+.. Links
+.. _HELM Best Practices Guide: https://docs.helm.sh/chart_best_practices/#requirements
+.. _kubectl Cheat Sheet: https://kubernetes.io/docs/reference/kubectl/cheatsheet/
+.. _Kubernetes documentation for emptyDir: https://kubernetes.io/docs/concepts/storage/volumes/#emptydir
+.. _Docker DevOps: https://wiki.onap.org/display/DW/Docker+DevOps#DockerDevOps-DockerBuild
+.. _http://cd.onap.info:30223/mso/logging/debug: http://cd.onap.info:30223/mso/logging/debug
+.. _Onboarding and Distributing a Vendor Software Product: https://wiki.onap.org/pages/viewpage.action?pageId=1018474
+.. _README.md: https://gerrit.onap.org/r/gitweb?p=oom.git;a=blob;f=kubernetes/README.md
+
+.. figure:: oomLogoV2-medium.png
+   :align: right
+
+.. _onap-on-kubernetes-with-rancher:
+
+ONAP on Kubernetes with Rancher
+###############################
+
+The following instructions will step you through the installation of Kubernetes
+on an OpenStack environment with Rancher.  The development lab used for this
+installation is the ONAP Windriver lab.
+
+This guide does not cover all of the steps required to setup your OpenStack
+environment: e.g. OAM networks and security groups but there is a wealth of
+OpenStack information on the web.
+
+Rancher Installation
+====================
+
+The following instructions describe how to create an Openstack VM running
+Rancher. This node will not be used to host ONAP itself, it will be used
+exclusively by Rancher.
+
+Launch new VM instance to host the Rancher Server
+-------------------------------------------------
+
+.. image:: Rancher-Launch_new_VM_instance_to_host_the_Rancher_Server.jpeg
+
+Select Ubuntu 16.04 as base image
+---------------------------------
+Select "No" on "Create New Volume"
+
+.. image:: Rancher-Select_Ubuntu_16.04_as_base_image.jpeg
+
+Select Flavor
+-------------
+Known issues exist if flavor is too small for Rancher. Please select a flavor
+with at least 4 vCPU and 8GB ram.
+
+.. image:: Rancher-Select_Flavor.jpeg
+
+Networking
+----------
+
+.. image:: Rancher-Networking.jpeg
+
+Security Groups
+---------------
+
+.. image:: Rancher-Security_Groups.jpeg
+
+Key Pair
+--------
+Use an existing key pair (e.g. onap_key), import an existing one or create a
+new one to assign.
+
+.. image:: Rancher-Key_Pair.jpeg
+
+Apply customization script for the Rancher VM
+---------------------------------------------
+
+Click :download:`openstack-rancher.sh <openstack-rancher.sh>` to download the script.
+
+.. literalinclude:: openstack-rancher.sh
+   :language: bash
+
+This customization script will:
+
+* setup root access to the VM (comment out if you wish to disable this
+  capability and restrict access to ssh access only)
+* install docker
+* install rancher
+* install kubectl
+* install helm
+* install nfs server
+
+.. note::
+  The Beijing release of OOM only supports Helm 2.8.2 not the 2.7.2 shown in
+  the screen capture below. The supported versions of all the software components
+  are listed in the :ref:`cloud-setup-guide-label`.
+
+.. image:: Apply_customization_script_for_the_Rancher_VM.jpeg
+
+Launch Instance
+---------------
+
+.. image:: Rancher-Launch_Instance.jpeg
+
+Assign Floating IP for external access
+--------------------------------------
+
+.. image:: Rancher-Allocate_Floating_IP.jpeg
+
+.. image:: Rancher-Manage_Floating_IP_Associations.jpeg
+
+.. image:: Rancher-Launch_Instance.jpeg
+
+Kubernetes Installation
+=======================
+
+Launch new VM instance(s) to create a Kubernetes single host or cluster
+-----------------------------------------------------------------------
+
+To create a cluster:
+
+.. note::
+  #. do not append a '-1' suffix (e.g. sb4-k8s)
+  #. increase count to the # of of kubernetes worker nodes you want (eg. 3)
+
+.. image:: K8s-Launch_new_VM_instance_to_create_a_Kubernetes_single_host_or_cluster.jpeg
+
+Select Ubuntu 16.04 as base image
+---------------------------------
+Select "No" on "Create New Volume"
+
+.. image:: K8s-Select_Ubuntu_16.04_as_base_image.jpeg
+
+Select Flavor
+-------------
+The size of a Kubernetes host depends on the size of the ONAP deployment that
+will be installed.
+
+As of the Beijing release a minimum of 3 x 32GB hosts will be needed to run a
+full ONAP deployment (all components).
+
+If a small subset of ONAP components are being deployed for testing purposes,
+then a single 16GB or 32GB host should suffice.
+
+.. image:: K8s-Select_Flavor.jpeg
+
+Networking
+-----------
+
+.. image:: K8s-Networking.jpeg
+
+Security Group
+---------------
+
+.. image:: K8s-Security_Group.jpeg
+
+Key Pair
+--------
+Use an existing key pair (e.g. onap_key), import an existing one or create a
+new one to assign.
+
+.. image:: K8s-Key_Pair.jpeg
+
+Apply customization script for Kubernetes VM(s)
+-----------------------------------------------
+
+Click :download:`openstack-k8s-node.sh <openstack-k8s-node.sh>` to
+download the script.
+
+.. literalinclude:: openstack-k8s-node.sh
+   :language: bash
+
+This customization script will:
+
+* setup root access to the VM (comment out if you wish to disable this
+  capability and restrict access to ssh access only)
+* install docker
+* install kubectl
+* install helm
+* install nfs common (see configuration step here)
+
+.. note::
+  Ensure you are using the correct versions as described in the
+  :ref:`cloud-setup-guide-label`
+
+Launch Instance
+---------------
+
+.. image:: K8s-Launch_Instance.jpeg
+
+Assign Floating IP for external access
+--------------------------------------
+
+.. image:: K8s-Assign_Floating_IP_for_external_access.jpeg
+
+.. image:: K8s-Manage_Floating_IP_Associations.jpeg
+
+.. image:: K8s-Launch_Instance.jpeg
+
+Setting up an NFS share for Multinode Kubernetes Clusters
+=========================================================
+The figure below illustrates a possible topology of a multinode Kubernetes
+cluster.
+
+.. image:: k8s-topology.jpg
+
+One node, the Master Node, runs Rancher and Helm clients and connects to all
+the Kubernetes nodes in the cluster. Kubernetes nodes, in turn, run Rancher,
+Kubernetes and Tiller (Helm) agents, which receive, execute, and respond to
+commands issued by the Master Node (e.g. kubectl or helm operations). Note that
+the Master Node can be either a remote machine that the user can log in to or a
+local machine (e.g. laptop, desktop) that has access to the Kubernetes cluster.
+
+Deploying applications to a Kubernetes cluster requires Kubernetes nodes to
+share a common, distributed filesystem. One node in the cluster plays the role
+of NFS Master (not to confuse with the Master Node that runs Rancher and Helm
+clients, which is located outside the cluster), while all the other cluster
+nodes play the role of NFS slaves. In the figure above, the left-most cluster
+node plays the role of NFS Master (indicated by the crown symbol). To properly
+set up an NFS share on Master and Slave nodes, the user can run the scripts
+below.
+
+Click :download:`master_nfs_node.sh <master_nfs_node.sh>` to download the script.
+
+.. literalinclude:: master_nfs_node.sh
+   :language: bash
+
+Click :download:`slave_nfs_node.sh <slave_nfs_node.sh>` to download the script.
+
+.. literalinclude:: slave_nfs_node.sh
+   :language: bash
+
+The master_nfs_node.sh script runs in the NFS Master node and needs the list of
+NFS Slave nodes as input, e.g.::
+
+    > sudo ./master_nfs_node.sh node1_ip node2_ip ... nodeN_ip
+
+The slave_nfs_node.sh script runs in each NFS Slave node and needs the IP of
+the NFS Master node as input, e.g.::
+
+    > sudo ./slave_nfs_node.sh master_node_ip
+
+Configuration (Rancher and Kubernetes)
+======================================
+
+Access Rancher server via web browser
+-------------------------------------
+(e.g.  http://10.12.6.16:8080/env/1a5/apps/stacks)
+
+.. image:: Access_Rancher_server_via_web_browser.jpeg
+
+Add Kubernetes Environment to Rancher
+-------------------------------------
+
+1. Select “Manage Environments”
+
+.. image:: Add_Kubernetes_Environment_to_Rancher.png
+
+2. Select “Add Environment”
+
+.. image:: Select_Add_Environment.png
+
+3. Add unique name for your new Rancher environment
+
+4. Select the Kubernetes template
+
+5. Click "create"
+
+.. image:: Click_create.jpeg
+
+6. Select the new named environment (ie. SB4) from the dropdown list (top left).
+
+Rancher is now waiting for a Kubernetes Host to be added.
+
+.. image:: K8s-Assign_Floating_IP_for_external_access.jpeg
+
+Add Kubernetes Host
+-------------------
+
+1.  If this is the first (or only) host being added - click on the "Add a host" link
+
+.. image:: K8s-Assign_Floating_IP_for_external_access.jpeg
+
+and click on "Save" (accept defaults).
+
+.. image:: and_click_on_Save_accept_defaults.jpeg
+
+otherwise select INFRASTRUCTURE→ Hosts and click on "Add Host"
+
+.. image:: otherwise_select_INFRASTRUCTURE_Hosts_and_click_on_Add_Host.jpg
+
+2. Enter the management IP for the k8s VM (e.g. 10.0.0.4) that was just created.
+
+3. Click on “Copy to Clipboard” button
+
+4. Click on “Close” button
+
+.. image:: Click_on_Close_button.jpeg
+
+Without the 10.0.0.4 IP - the CATTLE_AGENT will be derived on the host - but it
+may not be a routable IP.
+
+Configure Kubernetes Host
+-------------------------
+
+1. Login to the new Kubernetes Host::
+
+    > ssh -i ~/oom-key.pem ubuntu@10.12.5.1
+    The authenticity of host '10.12.5.172 (10.12.5.172)' can't be established.
+    ECDSA key fingerprint is SHA256:tqxayN58nCJKOJcWrEZzImkc0qKQHDDfUTHqk4WMcEI.
+    Are you sure you want to continue connecting (yes/no)? yes
+    Warning: Permanently added '10.12.5.172' (ECDSA) to the list of known hosts.
+    Welcome to Ubuntu 16.04.2 LTS (GNU/Linux 4.4.0-64-generic x86_64)
+
+     * Documentation: https://help.ubuntu.com
+     * Management: https://landscape.canonical.com
+     * Support: https://ubuntu.com/advantage
+
+     Get cloud support with Ubuntu Advantage Cloud Guest:
+       http://www.ubuntu.com/business/services/cloud
+
+    180 packages can be updated.
+    100 updates are security updates.
+
+    The programs included with the Ubuntu system are free software;
+    the exact distribution terms for each program are described in the
+    individual files in /usr/share/doc/*/copyright.
+
+    Ubuntu comes with ABSOLUTELY NO WARRANTY, to the extent permitted by
+    applicable law.
+
+    To run a command as administrator (user "root"), use "sudo <command>".
+    See "man sudo_root" for details.
+
+    ubuntu@sb4-k8s-1:~$
+
+
+2. Paste Clipboard content and hit enter to install Rancher Agent::
+
+    ubuntu@sb4-k8s-1:~$ sudo docker run -e CATTLE_AGENT_IP="10.0.0.4“ --rm --privileged -v /var/run/docker.sock:/var/run/docker.sock -v /var/lib/rancher:/var/lib/rancher rancher/agent:v1.2.9 http://10.12.6.16:8080/v1/scripts/5D757C68BD0A2125602A:1514678400000:yKW9xHGJDLvq6drz2eDzR2mjato
+    Unable to find image 'rancher/agent:v1.2.9' locally
+    v1.2.9: Pulling From rancher/agent
+    b3e1c725a85f: Pull complete
+    6071086409fc: Pull complete
+    d0ac3b234321: Pull complete
+    87f567b5cf58: Pull complete
+    a63e24b217c4: Pull complete
+    d0a3f58caef0: Pull complete
+    16914729cfd3: Pull complete
+    dc5c21984c5b: Pull complete
+    d7e8f9784b20: Pull complete
+    Digest: sha256:c21255ac4d94ffbc7b523F870F20ea5189b68Fa3d642800adb4774aab4748e66
+    Status: Downloaded newer image for rancher/agent:v1.2.9
+
+    INFO: Running Agent Registration Process, CATTLE_URL=http://10.12.6.16:8080/v1
+    INFO: Attempting to connect to: http://10.12.6.16:8080/v1
+    INFO: http://10.12.6.16:8080/v1 is accessible
+    INFO: Inspecting host capabilities
+    INFO: Boot2Docker: false
+    INFO: Host writable: true
+    INFO: Token: xxxxxxxx
+    INFO: Running registration
+    INFO: Printing Environment
+    INFO: ENV: CATTLE_ACCESS_KEY=98B35AC484FBF820E0AD
+    INFO: ENV: CATTLE_AGENT_IP=10.0.9.4
+    INFO: ENV: CATTLE_HOME=/var/lib/cattle
+    INFO: ENV: CATTLE_REGISTRATION_ACCESS_KEY=registrationToken
+    INFO: ENV: CATTLE_REGISTRATION_SECRET_KEY=xxxxxxx
+    INFO: ENV: CATTLE_SECRET_KEY=xxxxxxx
+    INFO: ENV: CATTLE_URL=http://10.12.6.16:8080/v1
+    INFO: ENV: DETECTED_CATTLE_AGENT_IP=10.12.5.172
+    INFO: ENV: RANCHER_AGENT_IMAGE=rancher/agent:v1.2.9
+    INFO: Launched Rancher Agent: c27ee0f3dc4c783b0db647ea1f73c35b3843a4b8d60b96375b1a05aa77d83136
+    ubuntu@sb4-k8s-1:~$
+
+3. Return to Rancher environment (e.g. SB4) and wait for services to complete
+   (~ 10-15 mins)
+
+.. image:: Return_to_Rancher_environment_eg_SB4_and_wait_for_services_to_complete_10-15_mins.jpeg
+
+Configure kubectl and helm
+==========================
+In this example we are configuring kubectl and helm that have been installed
+(as a convenience) onto the rancher and kubernetes hosts.  Typically you would
+install them both on your PC and remotely connect to the cluster. The following
+procedure would remain the same.
+
+1. Click on CLI and then click on “Generate Config”
+
+.. image:: Click_on_CLI_and_then_click_on_Generate_Config.jpeg
+
+2. Click on “Copy to Clipboard” - wait until you see a "token" - do not copy
+   user+password - the server is not ready at that point
+
+.. image:: Click_on_Copy_to_Clipboard-wait_until_you_see_a_token-do_not_copy_user+password-the_server_is_not_ready_at_that_point.jpeg
+
+3. Create a .kube directory in user directory (if one does not exist)::
+
+    ubuntu@sb4-kSs-1:~$ mkdir .kube
+    ubuntu@sb4-kSs-1:~$ vi .kube/config
+
+4. Paste contents of Clipboard into a file called “config” and save the file::
+
+    apiVersion: v1
+    kind : Config
+    clusters:
+    - cluster:
+        api-version: v1
+        insecure-skip-tls-verify: true
+        server: "https://10.12.6.16:8080/r/projects/1a7/kubernetes:6443"
+      name: "SB4"
+    contexts:
+    - context:
+        cluster: "SB4"
+        user: "SB4"
+      name: "SB4"
+    current-context: "SB4"
+    users:
+    - name: "SB4"
+      user:
+        token: "QmFzaWMgTlRBd01qZzBOemc)TkRrMk1UWkNOMFpDTlVFNlExcHdSa1JhVZreE5XSm1TRGhWU2t0Vk1sQjVhalZaY0dWaFVtZGFVMHQzWW1WWVJtVmpSQT09"
+    ~
+    ~
+    ~
+    - INSERT --
+
+5. Validate that kubectl is able to connect to the kubernetes cluster::
+
+    ubuntu@sb4-k8s-1:~$ kubectl config get-contexts
+    CURRENT   NAME   CLUSTER   AUTHINFO   NAMESPACE
+    *         SB4    SB4       SB4
+    ubuntu@sb4-kSs-1:~$
+
+and show running pods::
+
+    ubuntu@sb4-k8s-1:~$ kubectl get pods --all-namespaces -o=wide
+    NAMESPACE    NAME                                  READY   STATUS    RESTARTS   AGE   IP             NODE
+    kube-system  heapster—7Gb8cd7b5 -q7p42             1/1     Running   0          13m   10.42.213.49   sb4-k8s-1
+    kube-system  kube-dns-5d7bM87c9-c6f67              3/3     Running   0          13m   10.42.181.110  sb4-k8s-1
+    kube-system  kubernetes-dashboard-f9577fffd-kswjg  1/1     Running   0          13m   10.42.105.113  sb4-k8s-1
+    kube-system  monitoring-grafana-997796fcf-vg9h9    1/1     Running   0          13m   10.42,141.58   sb4-k8s-1
+    kube-system  monitoring-influxdb-56chd96b-hk66b    1/1     Running   0          13m   10.4Z.246.90   sb4-k8s-1
+    kube-system  tiller-deploy-cc96d4f6b-v29k9         1/1     Running   0          13m   10.42.147.248  sb4-k8s-1
+    ubuntu@sb4-k8s-1:~$
+
+6. Validate helm is running at the right version. If not, an error like this
+   will be displayed::
+
+    ubuntu@sb4-k8s-1:~$ helm list
+    Error: incompatible versions c1ient[v2.8.2] server[v2.6.1]
+    ubuntu@sb4-k8s-1:~$
+
+7. Upgrade the server-side component of helm (tiller) via `helm init --upgrade`::
+
+    ubuntu@sb4-k8s-1:~$ helm init --upgrade
+    Creating /home/ubuntu/.helm
+    Creating /home/ubuntu/.helm/repository
+    Creating /home/ubuntu/.helm/repository/cache
+    Creating /home/ubuntu/.helm/repository/local
+    Creating /home/ubuntu/.helm/plugins
+    Creating /home/ubuntu/.helm/starters
+    Creating /home/ubuntu/.helm/cache/archive
+    Creating /home/ubuntu/.helm/repository/repositories.yaml
+    Adding stable repo with URL: https://kubernetes-charts.storage.googleapis.com
+    Adding local repo with URL: http://127.0.0.1:8879/charts
+    $HELM_HOME has been configured at /home/ubuntu/.helm.
+
+    Tiller (the Helm server-side component) has been upgraded to the current version.
+    Happy Helming!
+    ubuntu@sb4-k8s-1:~$
+
+ONAP Deployment via OOM
+=======================
+Now that kubernetes and Helm are installed and configured you can prepare to
+deploy ONAP. Follow the instructions in the README.md_ or look at the official
+documentation to get started:
+
+- :ref:`quick-start-label` - deploy ONAP on an existing cloud
+- :ref:`user-guide-label` - a guide for operators of an ONAP instance
+
+
diff --git a/docs/openstack-k8s-node.sh b/docs/openstack-k8s-node.sh
new file mode 100644 (file)
index 0000000..7d49ad5
--- /dev/null
@@ -0,0 +1,46 @@
+#!/bin/bash
+
+DOCKER_VERSION=17.03
+KUBECTL_VERSION=1.8.10
+HELM_VERSION=2.8.2
+
+# setup root access - default login: oom/oom - comment out to restrict access too ssh key only
+sed -i 's/PermitRootLogin.*/PermitRootLogin yes/' /etc/ssh/sshd_config
+sed -i 's/PasswordAuthentication.*/PasswordAuthentication yes/' /etc/ssh/sshd_config
+service sshd restart
+echo -e "oom\noom" | passwd root
+
+apt-get update
+curl https://releases.rancher.com/install-docker/$DOCKER_VERSION.sh | sh
+mkdir -p /etc/systemd/system/docker.service.d/
+cat > /etc/systemd/system/docker.service.d/docker.conf << EOF
+[Service]
+ExecStart=
+ExecStart=/usr/bin/dockerd -H fd:// --insecure-registry=nexus3.onap.org:10001
+EOF
+systemctl daemon-reload
+systemctl restart docker
+apt-mark hold docker-ce
+
+IP_ADDY=`ip address |grep ens|grep inet|awk '{print $2}'| awk -F / '{print $1}'`
+HOSTNAME=`hostname`
+
+echo "$IP_ADDY $HOSTNAME" >> /etc/hosts
+
+docker login -u docker -p docker nexus3.onap.org:10001
+
+sudo apt-get install make -y
+
+sudo curl -LO https://storage.googleapis.com/kubernetes-release/release/v$KUBECTL_VERSION/bin/linux/amd64/kubectl
+sudo chmod +x ./kubectl
+sudo mv ./kubectl /usr/local/bin/kubectl
+sudo mkdir ~/.kube
+wget http://storage.googleapis.com/kubernetes-helm/helm-v${HELM_VERSION}-linux-amd64.tar.gz
+sudo tar -zxvf helm-v${HELM_VERSION}-linux-amd64.tar.gz
+sudo mv linux-amd64/helm /usr/local/bin/helm
+
+# install nfs
+sudo apt-get install nfs-common -y
+
+
+exit 0
diff --git a/docs/openstack-rancher.sh b/docs/openstack-rancher.sh
new file mode 100644 (file)
index 0000000..d51abe4
--- /dev/null
@@ -0,0 +1,51 @@
+#!/bin/bash
+
+DOCKER_VERSION=17.03
+RANCHER_VERSION=1.6.14
+KUBECTL_VERSION=1.8.10
+HELM_VERSION=2.8.2
+
+# setup root access - default login: oom/oom - comment out to restrict access too ssh key only
+sed -i 's/PermitRootLogin.*/PermitRootLogin yes/' /etc/ssh/sshd_config
+sed -i 's/PasswordAuthentication.*/PasswordAuthentication yes/' /etc/ssh/sshd_config
+service sshd restart
+echo -e "oom\noom" | passwd root
+
+apt-get update
+curl https://releases.rancher.com/install-docker/$DOCKER_VERSION.sh | sh
+mkdir -p /etc/systemd/system/docker.service.d/
+cat > /etc/systemd/system/docker.service.d/docker.conf << EOF
+[Service]
+ExecStart=
+ExecStart=/usr/bin/dockerd -H fd:// --insecure-registry=nexus3.onap.org:10001
+EOF
+systemctl daemon-reload
+systemctl restart docker
+apt-mark hold docker-ce
+
+IP_ADDY=`ip address |grep ens|grep inet|awk '{print $2}'| awk -F / '{print $1}'`
+HOSTNAME=`hostname`
+
+echo "$IP_ADDY $HOSTNAME" >> /etc/hosts
+
+docker login -u docker -p docker nexus3.onap.org:10001
+
+sudo apt-get install make -y
+
+sudo docker run -d --restart=unless-stopped -p 8080:8080 --name rancher_server rancher/server:v$RANCHER_VERSION
+sudo curl -LO https://storage.googleapis.com/kubernetes-release/release/v$KUBECTL_VERSION/bin/linux/amd64/kubectl
+sudo chmod +x ./kubectl
+sudo mv ./kubectl /usr/local/bin/kubectl
+sudo mkdir ~/.kube
+wget http://storage.googleapis.com/kubernetes-helm/helm-v${HELM_VERSION}-linux-amd64.tar.gz
+sudo tar -zxvf helm-v${HELM_VERSION}-linux-amd64.tar.gz
+sudo mv linux-amd64/helm /usr/local/bin/helm
+
+# nfs server
+sudo apt-get install nfs-kernel-server -y
+
+sudo mkdir -p /nfs_share
+sudo chown nobody:nogroup /nfs_share/
+
+
+exit 0
diff --git a/docs/otherwise_select_INFRASTRUCTURE_Hosts_and_click_on_Add_Host.jpg b/docs/otherwise_select_INFRASTRUCTURE_Hosts_and_click_on_Add_Host.jpg
new file mode 100644 (file)
index 0000000..18af758
Binary files /dev/null and b/docs/otherwise_select_INFRASTRUCTURE_Hosts_and_click_on_Add_Host.jpg differ
diff --git a/docs/slave_nfs_node.sh b/docs/slave_nfs_node.sh
new file mode 100644 (file)
index 0000000..5cb164c
--- /dev/null
@@ -0,0 +1,25 @@
+#!/bin/bash
+
+usage () {
+  echo "Usage:"
+  echo "   ./$(basename $0) nfs_master_ip"
+  exit 1
+}
+
+if [ "$#" -ne 1 ]; then
+  echo "Missing NFS mater node"
+  usage
+fi
+
+MASTER_IP=$1
+
+#Install NFS common
+sudo apt-get update
+sudo apt-get install -y nfs-common
+
+#Create NFS directory
+sudo mkdir -p /dockerdata-nfs
+
+#Mount the remote NFS directory to the local one
+sudo mount $MASTER_IP:/dockerdata-nfs /dockerdata-nfs/
+echo "$MASTER_IP:/dockerdata-nfs /dockerdata-nfs  nfs auto,nofail,noatime,nolock,intr,tcp,actimeo=1800 0 0" | sudo tee -a /etc/fstab
index 7118d16..2239c9a 100644 (file)
@@ -24,7 +24,7 @@ global:
 #################################################################
 
 # application image
-image: onap/babel:1.2-STAGING-latest
+image: onap/babel:1.2.0
 
 # application configuration
 config:
index 13cc9e7..78afd41 100644 (file)
@@ -24,7 +24,7 @@ global:
 #################################################################
 
 # application image
-image: onap/champ:1.2-STAGING-latest
+image: onap/champ:1.2.3
 
 # application configuration
 config:
index 5ce7689..51fc65a 100644 (file)
@@ -7,7 +7,7 @@ global: # global defaults
 
 # application image
 repository: nexus3.onap.org:10001
-image: onap/data-router:1.2-STAGING-latest
+image: onap/data-router:1.2.2
 pullPolicy: Always
 restartPolicy: Always
 
index f63afd3..a04d44c 100644 (file)
-<?xml version="1.0" encoding="UTF-8"?>
-<configuration scan="true" scanPeriod="3 seconds" debug="true">
-    <!--<jmxConfigurator /> -->
-    <!-- directory path for all other type logs -->
-    <property name="logDir" value="/var/log/onap" />
-    <property name="componentName" value="aai" />
-    <property name="subComponentName" value="aai-CRUD" />
-    <property name="logDirectory" value="${logDir}/${componentName}/${subComponentName}" />
-    <!--  default eelf log file names -->
-    <property name="generalLogName" value="application" />
-    <property name="errorLogName" value="error" />
-    <property name="metricsLogName" value="metrics" />
-    <property name="auditLogName" value="audit" />
-    <property name="debugLogName" value="debug" />
-    <property name="queueSize" value="256" />
-    <property name="maxFileSize" value="50MB" />
-    <property name="maxHistory" value="30" />
-    <property name="totalSizeCap" value="10GB" />
-    <property name="pattern" value="%d{&quot;yyyy-MM-dd'T'HH:mm:ss.SSSXXX&quot;, UTC}\t[%thread]\t%-5level\t%logger\t%replace(%replace(%replace(%mdc){'\t','\\\\t'}){', ','\t'}){'\n', '\\\\n'}\t%replace(%replace(%msg){'\n', '\\\\n'}){'\t','\\\\t'}%n" />
-    <!-- Example evaluator filter applied against console appender -->
-    <appender name="STDOUT" class="ch.qos.logback.core.ConsoleAppender">
-        <encoder>
-            <pattern>${pattern}</pattern>
-        </encoder>
-    </appender>
-    <!-- ============================================================================ -->
-    <!-- EELF Appenders -->
-    <!-- ============================================================================ -->
-    <!-- The EELFAppender is used to record events to the general application
+<configuration scan="true" scanPeriod="3 seconds" debug="false">
+  <!--<jmxConfigurator /> -->
+  <!-- directory path for all other type logs -->
+  
+  <property name="logDir"  value="/var/log/onap" />
+  
+  
+  <!--  specify the component name 
+       <ECOMP-component-name>::= "MSO" | "DCAE" | "ASDC " | "AAI" |"Policy" | "SDNC" | "AC"  -->
+  <property name="componentName" value="AAI-CRUD" />
+  
+  <!--  default eelf log file names -->
+  <property name="generalLogName" value="error" />
+  <property name="metricsLogName" value="metrics" />
+  <property name="auditLogName" value="audit" />
+  <property name="debugLogName" value="debug" />
+  
+  <property name="errorLogPattern" value="%d{yyyy-MM-dd'T'HH:mm:ss.SSSXXX}|%mdc{RequestId}|%thread|GIZMO|%mdc{PartnerName}|%logger||%.-5level|%msg%n" />
+  <property name="auditMetricPattern" value="%m%n" />
+
+  <property name="logDirectory" value="${logDir}/${componentName}" />
+  
+  <!-- Example evaluator filter applied against console appender -->
+  <appender name="STDOUT" class="ch.qos.logback.core.ConsoleAppender">
+    <encoder>
+      <pattern>${errorLogPattern}</pattern>
+    </encoder>
+  </appender>
+
+  <!-- ============================================================================ -->
+  <!-- EELF Appenders -->
+  <!-- ============================================================================ -->
+  
+  <!-- The EELFAppender is used to record events to the general application 
        log -->
-    <appender name="EELF" class="ch.qos.logback.core.rolling.RollingFileAppender">
-        <file>${logDirectory}/${generalLogName}.log</file>
-        <rollingPolicy class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy">
-            <fileNamePattern>${logDirectory}/${generalLogName}.%d{yyyy-MM-dd}.%i.log</fileNamePattern>
-            <timeBasedFileNamingAndTriggeringPolicy class="ch.qos.logback.core.rolling.SizeAndTimeBasedFNATP">
-                <maxFileSize>${maxFileSize}</maxFileSize>
-            </timeBasedFileNamingAndTriggeringPolicy>
-            <maxHistory>${maxHistory}</maxHistory>
-            <totalSizeCap>${totalSizeCap}</totalSizeCap>
-        </rollingPolicy>
-        <encoder>
-            <pattern>${pattern}</pattern>
-        </encoder>
-    </appender>
-    <appender name="asyncEELF" class="ch.qos.logback.classic.AsyncAppender">
-        <!-- deny all events with a level below INFO, that is TRACE and DEBUG -->
-        <filter class="ch.qos.logback.classic.filter.ThresholdFilter">
-            <level>INFO</level>
-        </filter>
-        <queueSize>${queueSize}</queueSize>
-        <appender-ref ref="EELF" />
-    </appender>
-    <!-- EELF Audit Appender. This appender is used to record audit engine
-       related logging events. The audit logger and appender are specializations
-       of the EELF application root logger and appender. This can be used to segregate
-       Policy engine events from other components, or it can be eliminated to record
+  
+  <appender name="EELF"
+            class="ch.qos.logback.core.rolling.RollingFileAppender">
+    <file>${logDirectory}/${generalLogName}.log</file>
+    <rollingPolicy
+        class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy">
+      <fileNamePattern>${logDirectory}/${generalLogName}.%d{yyyy-MM-dd}.log.zip
+      </fileNamePattern>
+      <maxHistory>60</maxHistory>
+    </rollingPolicy>
+    <encoder>
+      <pattern>${errorLogPattern}</pattern>
+    </encoder>
+  </appender>
+  <appender name="asyncEELF" class="ch.qos.logback.classic.AsyncAppender">
+    <!-- deny all events with a level below INFO, that is TRACE and DEBUG -->
+    <filter class="ch.qos.logback.classic.filter.ThresholdFilter">
+      <level>INFO</level>
+    </filter>
+    <queueSize>256</queueSize>
+    <appender-ref ref="EELF" />
+  </appender>
+
+  
+  <!-- EELF Audit Appender. This appender is used to record audit engine 
+       related logging events. The audit logger and appender are specializations 
+       of the EELF application root logger and appender. This can be used to segregate 
+       Policy engine events from other components, or it can be eliminated to record 
        these events as part of the application root log. -->
-    <appender name="EELFAudit" class="ch.qos.logback.core.rolling.RollingFileAppender">
-        <file>${logDirectory}/${auditLogName}.log</file>
-        <rollingPolicy class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy">
-            <fileNamePattern>${logDirectory}/${auditLogName}.%d{yyyy-MM-dd}.%i.log</fileNamePattern>
-            <timeBasedFileNamingAndTriggeringPolicy class="ch.qos.logback.core.rolling.SizeAndTimeBasedFNATP">
-                <maxFileSize>${maxFileSize}</maxFileSize>
-            </timeBasedFileNamingAndTriggeringPolicy>
-            <maxHistory>${maxHistory}</maxHistory>
-            <totalSizeCap>${totalSizeCap}</totalSizeCap>
-        </rollingPolicy>
-        <encoder>
-            <pattern>${pattern}</pattern>
-        </encoder>
-    </appender>
-    <appender name="asyncEELFAudit" class="ch.qos.logback.classic.AsyncAppender">
-        <queueSize>${queueSize}</queueSize>
-        <appender-ref ref="EELFAudit" />
-    </appender>
-    <appender name="EELFMetrics" class="ch.qos.logback.core.rolling.RollingFileAppender">
-        <file>${logDirectory}/${metricsLogName}.log</file>
-        <rollingPolicy class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy">
-            <fileNamePattern>${logDirectory}/${metricsLogName}.%d{yyyy-MM-dd}.%i.log</fileNamePattern>
-            <timeBasedFileNamingAndTriggeringPolicy class="ch.qos.logback.core.rolling.SizeAndTimeBasedFNATP">
-                <maxFileSize>${maxFileSize}</maxFileSize>
-            </timeBasedFileNamingAndTriggeringPolicy>
-            <maxHistory>${maxHistory}</maxHistory>
-            <totalSizeCap>${totalSizeCap}</totalSizeCap>
-        </rollingPolicy>
-        <encoder>
-            <pattern>${pattern}</pattern>
-        </encoder>
-    </appender>
-    <appender name="asyncEELFMetrics" class="ch.qos.logback.classic.AsyncAppender">
-        <queueSize>${queueSize}</queueSize>
-        <appender-ref ref="EELFMetrics" />
-    </appender>
-    <appender name="EELFError" class="ch.qos.logback.core.rolling.RollingFileAppender">
-        <file>${logDirectory}/${errorLogName}.log</file>
-        <rollingPolicy class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy">
-            <fileNamePattern>${logDirectory}/${errorLogName}.%d{yyyy-MM-dd}.%i.log</fileNamePattern>
-            <timeBasedFileNamingAndTriggeringPolicy class="ch.qos.logback.core.rolling.SizeAndTimeBasedFNATP">
-                <maxFileSize>${maxFileSize}</maxFileSize>
-            </timeBasedFileNamingAndTriggeringPolicy>
-            <maxHistory>${maxHistory}</maxHistory>
-            <totalSizeCap>${totalSizeCap}</totalSizeCap>
-        </rollingPolicy>
-        <encoder>
-            <pattern>${pattern}</pattern>
-        </encoder>
-        <filter class="ch.qos.logback.classic.filter.ThresholdFilter">
-            <level>INFO</level>
-        </filter>
-    </appender>
-    <appender name="asyncEELFError" class="ch.qos.logback.classic.AsyncAppender">
-        <queueSize>${queueSize}</queueSize>
-        <appender-ref ref="EELFError" />
-    </appender>
-    <appender name="EELFDebug" class="ch.qos.logback.core.rolling.RollingFileAppender">
-        <file>${logDirectory}/${debugLogName}.log</file>
-        <rollingPolicy class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy">
-            <fileNamePattern>${logDirectory}/${debugLogName}.%d{yyyy-MM-dd}.%i.log</fileNamePattern>
-            <timeBasedFileNamingAndTriggeringPolicy class="ch.qos.logback.core.rolling.SizeAndTimeBasedFNATP">
-                <maxFileSize>${maxFileSize}</maxFileSize>
-            </timeBasedFileNamingAndTriggeringPolicy>
-            <maxHistory>${maxHistory}</maxHistory>
-            <totalSizeCap>${totalSizeCap}</totalSizeCap>
-        </rollingPolicy>
-        <encoder>
-            <pattern>${pattern}</pattern>
-        </encoder>
-    </appender>
-    <appender name="asyncEELFDebug" class="ch.qos.logback.classic.AsyncAppender">
-        <queueSize>${queueSize}</queueSize>
-        <appender-ref ref="EELFDebug" />
-        <includeCallerData>true</includeCallerData>
-    </appender>
-    <!-- ============================================================================ -->
-    <!--  EELF loggers -->
-    <!-- ============================================================================ -->
-    <logger name="com.att.eelf" level="info" additivity="false">
-        <appender-ref ref="asyncEELF" />
-    </logger>
-    <logger name="com.att.eelf.security" level="info" additivity="false">
-        <appender-ref ref="asyncEELFSecurity" />
-    </logger>
-    <logger name="com.att.eelf.perf" level="info" additivity="false">
-        <appender-ref ref="asyncEELFPerformance" />
-    </logger>
-    <logger name="com.att.eelf.server" level="info" additivity="false">
-        <appender-ref ref="asyncEELFServer" />
-    </logger>
-    <logger name="com.att.eelf.policy" level="info" additivity="false">
-        <appender-ref ref="asyncEELFPolicy" />
-    </logger>
-    <logger name="com.att.eelf.debug" level="debug" additivity="false">
-        <appender-ref ref="asyncEELFDebug" />
-    </logger>
-    <logger name="com.att.eelf.audit" level="info" additivity="false">
-        <appender-ref ref="asyncEELFAudit" />
-    </logger>
-    <logger name="com.att.eelf.metrics" level="info" additivity="false">
-        <appender-ref ref="asyncEELFMetrics" />
-    </logger>
-    <logger name="com.att.eelf.error" level="info" additivity="false">
-        <appender-ref ref="asyncEELFError" />
-    </logger>
-    <!-- Spring related loggers -->
-    <logger name="org.springframework" level="WARN" />
-    <logger name="org.springframework.beans" level="WARN" />
-    <logger name="org.springframework.web" level="WARN" />
-    <logger name="com.blog.spring.jms" level="WARN" />
-    <!-- AJSC Services (bootstrap services) -->
-    <logger name="ajsc" level="WARN" />
-    <logger name="ajsc.RouteMgmtService" level="WARN" />
-    <logger name="ajsc.ComputeService" level="WARN" />
-    <logger name="ajsc.VandelayService" level="WARN" />
-    <logger name="ajsc.FilePersistenceService" level="WARN" />
-    <logger name="ajsc.UserDefinedJarService" level="WARN" />
-    <logger name="ajsc.UserDefinedBeansDefService" level="WARN" />
-    <logger name="ajsc.LoggingConfigurationService" level="WARN" />
-    <!-- AJSC related loggers (DME2 Registration, csi logging, restlet, servlet
-       logging) -->
-    <logger name="ajsc.utils" level="WARN" />
-    <logger name="ajsc.utils.DME2Helper" level="WARN" />
-    <logger name="ajsc.filters" level="WARN" />
-    <logger name="ajsc.beans.interceptors" level="WARN" />
-    <logger name="ajsc.restlet" level="WARN" />
-    <logger name="ajsc.servlet" level="WARN" />
-    <logger name="com.att" level="INFO" />
-    <logger name="com.att.ajsc.csi.logging" level="WARN" />
-    <logger name="com.att.ajsc.filemonitor" level="WARN" />
-    <!-- CRUD Service loggers -->
-    <logger name="org.openecomp.crud" level="INFO" />
-    <!-- Other Loggers that may help troubleshoot -->
-    <logger name="net.sf" level="WARN" />
-    <logger name="org.apache" level="WARN" />
-    <logger name="org.apache.commons.httpclient" level="WARN" />
-    <logger name="org.apache.commons" level="WARN" />
-    <logger name="org.apache.coyote" level="WARN" />
-    <logger name="org.apache.jasper" level="WARN" />
-    <!-- Camel Related Loggers (including restlet/servlet/jaxrs/cxf logging.
+  
+  <appender name="EELFAudit"
+            class="ch.qos.logback.core.rolling.RollingFileAppender">
+    <file>${logDirectory}/${auditLogName}.log</file>
+    <rollingPolicy
+        class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy">
+      <fileNamePattern>${logDirectory}/${auditLogName}.%d{yyyy-MM-dd}.log.zip
+      </fileNamePattern>
+      <maxHistory>60</maxHistory>
+    </rollingPolicy>
+    <encoder>
+      <pattern>${auditMetricPattern}</pattern>
+    </encoder>
+  </appender>
+  <appender name="asyncEELFAudit" class="ch.qos.logback.classic.AsyncAppender">
+    <queueSize>256</queueSize>
+    <appender-ref ref="EELFAudit" />
+  </appender>
+  
+  <appender name="EELFMetrics"
+            class="ch.qos.logback.core.rolling.RollingFileAppender">
+    <file>${logDirectory}/${metricsLogName}.log</file>
+    <rollingPolicy
+        class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy">
+      <fileNamePattern>${logDirectory}/${metricsLogName}.%d{yyyy-MM-dd}.log.zip
+      </fileNamePattern>
+      <maxHistory>60</maxHistory>
+    </rollingPolicy>
+    <encoder>
+      <!-- <pattern>"%d{HH:mm:ss.SSS} [%thread] %-5level %logger{1024} - 
+           %msg%n"</pattern> -->
+      <pattern>${auditMetricPattern}</pattern>
+    </encoder>
+  </appender>
+  
+  
+  <appender name="asyncEELFMetrics" class="ch.qos.logback.classic.AsyncAppender">
+    <queueSize>256</queueSize>
+    <appender-ref ref="EELFMetrics"/>
+  </appender>
+  
+  <appender name="EELFDebug"
+            class="ch.qos.logback.core.rolling.RollingFileAppender">
+    <file>${logDirectory}/${debugLogName}.log</file>
+    <rollingPolicy
+        class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy">
+      <fileNamePattern>${logDirectory}/${debugLogName}.%d{yyyy-MM-dd}.log.zip
+      </fileNamePattern>
+      <maxHistory>60</maxHistory>
+    </rollingPolicy>
+    <encoder>
+      <pattern>${errorLogPattern}</pattern>
+    </encoder>
+  </appender>
+  
+  <appender name="asyncEELFDebug" class="ch.qos.logback.classic.AsyncAppender">
+    <queueSize>256</queueSize>
+    <appender-ref ref="EELFDebug" />
+    <includeCallerData>false</includeCallerData>
+  </appender>
+  
+  
+  <!-- ============================================================================ -->
+  <!--  EELF loggers -->
+  <!-- ============================================================================ -->
+  <logger name="com.att.eelf" level="info" additivity="false">
+    <appender-ref ref="asyncEELF" />
+    <appender-ref ref="asyncEELFDebug" />
+  </logger>
+
+  <logger name="com.att.eelf.audit" level="info" additivity="false">
+    <appender-ref ref="asyncEELFAudit" />
+  </logger>
+  <logger name="com.att.eelf.metrics" level="info" additivity="false">
+    <appender-ref ref="asyncEELFMetrics" />
+  </logger>
+  
+  <!-- Spring related loggers -->
+  <logger name="org.springframework" level="WARN" />
+  <logger name="org.springframework.beans" level="WARN" />
+  <logger name="org.springframework.web" level="WARN" />
+  <logger name="com.blog.spring.jms" level="WARN" />
+
+  <!-- ARMAdapter Service loggers -->
+  <logger name="com.amdocs.aai.armadapter" level="INFO" />
+
+  <!-- Other Loggers that may help troubleshoot -->
+  <logger name="net.sf" level="WARN" />
+  <logger name="org.apache" level="WARN" />
+  <logger name="org.apache.commons.httpclient" level="WARN" />
+  <logger name="org.apache.commons" level="WARN" />
+  <logger name="org.apache.coyote" level="WARN" />
+  <logger name="org.apache.jasper" level="WARN" />
+
+  <!-- Camel Related Loggers (including restlet/servlet/jaxrs/cxf logging. 
        May aid in troubleshooting) -->
-    <logger name="org.apache.camel" level="WARN" />
-    <logger name="org.apache.cxf" level="WARN" />
-    <logger name="org.apache.camel.processor.interceptor" level="WARN" />
-    <logger name="org.apache.cxf.jaxrs.interceptor" level="WARN" />
-    <logger name="org.apache.cxf.service" level="WARN" />
-    <logger name="org.restlet" level="WARN" />
-    <logger name="org.apache.camel.component.restlet" level="WARN" />
-    <!-- logback internals logging -->
-    <logger name="ch.qos.logback.classic" level="WARN" />
-    <logger name="ch.qos.logback.core" level="WARN" />
-    <root level="INFO">
-        <appender-ref ref="asyncEELFDebug" />
-    </root>
+  <logger name="org.apache.camel" level="WARN" />
+  <logger name="org.apache.cxf" level="WARN" />
+  <logger name="org.apache.camel.processor.interceptor" level="WARN" />
+  <logger name="org.apache.cxf.jaxrs.interceptor" level="WARN" />
+  <logger name="org.apache.cxf.service" level="WARN" />
+  <logger name="org.restlet" level="WARN" />
+  <logger name="org.apache.camel.component.restlet" level="WARN" />
+  
+
+  <!-- logback internals logging -->
+  <logger name="ch.qos.logback.classic" level="WARN" />
+  <logger name="ch.qos.logback.core" level="WARN" />
+
+  <root>
+    <appender-ref ref="asyncEELF" /> 
+    <!-- <appender-ref ref="asyncEELFDebug" /> -->
+  </root>
+  
 </configuration>
index 1ea924f..acf6417 100644 (file)
@@ -22,7 +22,7 @@ global:
 #################################################################
 
 # application image
-image: onap/gizmo:1.1-STAGING-latest
+image: onap/gizmo:1.2.1
 
 # application configuration
 config:
index 676fc6f..8acb28f 100644 (file)
@@ -7,7 +7,7 @@ global: # global defaults
 
 # application image
 repository: nexus3.onap.org:10001
-image: onap/model-loader:1.2-STAGING-latest
+image: onap/model-loader:1.2.1
 pullPolicy: Always
 restartPolicy: Always
 
index 7b55aa7..50719e4 100644 (file)
@@ -9,7 +9,7 @@ global: # global defaults
 
 # application image
 repository: nexus3.onap.org:10001
-image: onap/aai-resources:1.2-STAGING-latest
+image: onap/aai-resources:1.2.1
 pullPolicy: Always
 restartPolicy: Always
 
index c6f42bd..f781662 100644 (file)
@@ -9,7 +9,7 @@ global: # global defaults
 
 # application image
 repository: nexus3.onap.org:10001
-image: onap/search-data-service:1.2-STAGING-latest
+image: onap/search-data-service:1.2.1
 pullPolicy: Always
 restartPolicy: Always
 
index b9ef4ca..c87ec5b 100644 (file)
@@ -14,7 +14,7 @@ global: # global defaults
 
 # application image
 repository: nexus3.onap.org:10001
-image: onap/sparky-be:1.2-STAGING-latest
+image: onap/sparky-be:1.2.1
 pullPolicy: Always
 restartPolicy: Always
 
index c22ae75..943d34a 100644 (file)
@@ -9,7 +9,7 @@ global: # global defaults
 
 # application image
 repository: nexus3.onap.org:10001
-image: onap/aai-traversal:1.2-STAGING-latest
+image: onap/aai-traversal:1.2.1
 pullPolicy: Always
 restartPolicy: Always
 
index e13193f..a990739 100755 (executable)
@@ -173,5 +173,8 @@ fi
         appcInstallEndTime=$(date +%s)
         echo "Total Appc install took $(expr $appcInstallEndTime - $appcInstallStartTime) seconds"
 
+echo "Starting cdt-proxy-service jar, logging to ${APPC_HOME}/cdt-proxy-service/jar.log"
+java -jar ${APPC_HOME}/cdt-proxy-service/cdt-proxy-service.jar > ${APPC_HOME}/cdt-proxy-service/jar.log &
+
 exec ${ODL_HOME}/bin/karaf
 
index 6c85985..0ba73e3 100644 (file)
@@ -31,6 +31,9 @@ spec:
     - port: {{ .Values.service.externalPort2 }}
       nodePort: {{ .Values.global.nodePortPrefix | default .Values.nodePortPrefix }}{{ .Values.service.nodePort2 }}
       name: "{{ .Values.service.portName }}-1830"
+    - port: {{ .Values.service.externalPort3 }}
+      nodePort: {{ .Values.global.nodePortPrefix | default .Values.nodePortPrefix }}{{ .Values.service.nodePort3 }}
+      name: "{{ .Values.service.portName }}-9090"
     {{- else -}}
     - port: {{ .Values.service.externalPort }}
       targetPort: {{ .Values.service.internalPort }}
@@ -38,6 +41,9 @@ spec:
     - port: {{ .Values.service.externalPort2 }}
       targetPort: {{ .Values.service.internalPort2 }}
       name: {{ .Values.service.portName }}-1830
+    - port: {{ .Values.service.externalPort3 }}
+      targetPort: {{ .Values.service.internalPort3 }}
+      name: {{ .Values.service.portName }}-9090
     {{- end}}
   selector:
     app: {{ include "common.name" . }}
index 79991f5..1bb7e23 100644 (file)
@@ -103,6 +103,10 @@ service:
   nodePort2: 31
   clusterPort: 2550
 
+  internalPort3: 9191
+  externalPort3: 9090
+  nodePort3: 11
+
 ## Persist data to a persitent volume
 persistence:
   enabled: true
index 4effba8..fde094f 100755 (executable)
@@ -92,9 +92,9 @@ spec:
           - mountPath: /usr/local/bin/aai_key.key
             name: {{ .Values.global.commonConfigPrefix }}-config
             subPath: aai_key.key
-          - mountPath: /usr/local/bin/bundle.pem
+          - mountPath: /usr/local/bin/AAF_RootCA.cer
             name: {{ .Values.global.commonConfigPrefix }}-config
-            subPath: bundle.pem
+            subPath: AAF_RootCA.cer
           resources:
 {{ toYaml .Values.resources | indent 12 }}
         {{- if .Values.nodeSelector }}
@@ -123,7 +123,7 @@ spec:
               path: aai_cert.cer
             - key: aai_key.key
               path: aai_key.key
-            - key: bundle.pem
-              path: bundle.pem
+            - key: AAF_RootCA.cer
+              path: AAF_RootCA.cer
       imagePullSecrets:
       - name: "{{ include "common.namespace" . }}-docker-registry-key"
diff --git a/kubernetes/oof/charts/oof-has/resources/config/AAF_RootCA.cer b/kubernetes/oof/charts/oof-has/resources/config/AAF_RootCA.cer
new file mode 100755 (executable)
index 0000000..e9a50d7
--- /dev/null
@@ -0,0 +1,31 @@
+-----BEGIN CERTIFICATE-----
+MIIFPjCCAyagAwIBAgIJAJ6u7cCnzrWdMA0GCSqGSIb3DQEBCwUAMCwxDjAMBgNV
+BAsMBU9TQUFGMQ0wCwYDVQQKDARPTkFQMQswCQYDVQQGEwJVUzAeFw0xODA0MDUx
+NDE1MjhaFw0zODAzMzExNDE1MjhaMCwxDjAMBgNVBAsMBU9TQUFGMQ0wCwYDVQQK
+DARPTkFQMQswCQYDVQQGEwJVUzCCAiIwDQYJKoZIhvcNAQEBBQADggIPADCCAgoC
+ggIBAMA5pkgRs7NhGG4ew5JouhyYakgYUyFaG121+/h8qbSdt0hVQv56+EA41Yq7
+XGie7RYDQK9NmAFF3gruE+6X7wvJiChp+Cyd7sFMnb65uWhxEdxWTM2BJFrgfzUn
+H8ZCxgaCo3XH4PzlKRy2LQQJEJECwl/RZmRCXijMt5e9h8XoZY/fKkKcZZUsWNCM
+pTo266wjvA9MXLmdgReRj0+vrCjrNqy+htwJDztoiHWiYPqT6o8EvGcgjNqjlZx7
+NUNf8MfLDByqKF6+wRbHv1GKjn3/Vijd45Fv8riyRYROiFanvbV6jIfBkv8PZbXg
+2VDWsYsgp8NAvMxK+iV8cO+Ck3lBI2GOPZbCEqpPVTYbLUz6sczAlCXwQoPzDIZY
+wYa3eR/gYLY1gP2iEVHORag3bLPap9ZX5E8DZkzTNTjovvLk8KaCmfcaUMJsBtDd
+ApcUitz10cnRyZc1sX3gE1f3DpzQM6t9C5sOVyRhDcSrKqqwb9m0Ss04XAS9FsqM
+P3UWYQyqDXSxlUAYaX892u8mV1hxnt2gjb22RloXMM6TovM3sSrJS0wH+l1nznd6
+aFXftS/G4ZVIVZ/LfT1is4StoyPWZCwwwly1z8qJQ/zhip5NgZTxQw4mi7ww35DY
+PdAQOCoajfSvFjqslQ/cPRi/MRCu079heVb5fQnnzVtnpFQRAgMBAAGjYzBhMB0G
+A1UdDgQWBBRTVTPyS+vQUbHBeJrBKDF77+rtSTAfBgNVHSMEGDAWgBRTVTPyS+vQ
+UbHBeJrBKDF77+rtSTAPBgNVHRMBAf8EBTADAQH/MA4GA1UdDwEB/wQEAwIBhjAN
+BgkqhkiG9w0BAQsFAAOCAgEAPx/IaK94n02wPxpnYTy+LVLIxwdq/kawNd6IbiMz
+L87zmNMDmHcGbfoRCj8OkhuggX9Lx1/CkhpXimuYsZOFQi5blr/u+v4mIbsgbmi9
+7j+cUHDP0zLycvSvxKHty51LwmaX9a4wkJl5zBU4O1sd/H9tWcEmwJ39ltKoBKBx
+c94Zc3iMm5ytRWGj+0rKzLDAXEWpoZ5bE5PLJauA6UDCxDLfs3FwhbS7uDggxYvf
+jySF5FCNET94oJ+m8s7VeHvoa8iPGKvXrIqdd7XDHnqJJlVKr7m9S0fMbyEB8ci2
+RtOXDt93ifY1uhoEtEykn4dqBSp8ezvNMnwoXdYPDvTd9uCAFeWFLVreBAWxd25h
+PsBTkZA5hpa/rA+mKv6Af4VBViYr8cz4dZCsFChuioVebe9ighrfjB//qKepFjPF
+CyjzKN1u0JKm/2x/ORqxkTONG8p3uDwoIOyimUcTtTMv42bfYD88RKakqSFXE9G+
+Z0LlaKABqfjK49o/tsAp+c5LoNlYllKhnetO3QAdraHwdmC36BhoghzR1jpX751A
+cZn2VH3Q4XKyp01cJNCJIrua+A+bx6zh3RyW6zIIkbRCbET+UD+4mr8WIcSE3mtR
+ZVlnhUDO4z9//WKMVzwS9Rh8/kuszrGFI1KQozXCHLrce3YP6RYZfOed79LXaRwX
+dYY=
+-----END CERTIFICATE-----
index 4d32dee..588b936 100755 (executable)
@@ -149,22 +149,29 @@ complex_cache_refresh_interval = 60
 
 # Base URL for A&AI, up to and not including the version, and without a
 # trailing slash. (string value)
-server_url = https://oof.api.simpledemo.onap.org:8443/aai
+server_url = https://{{.Values.config.aai.serviceName}}.{{ include "common.namespace" . }}:{{.Values.config.aai.port}}/aai
 
 # The version of A&AI in v# format. (string value)
-server_url_version = v11
+server_url_version = v13
 
 # SSL/TLS certificate file in pem format. This certificate must be registered
 # with the A&AI endpoint. (string value)
-certificate_file = /usr/local/bin/aai_cert.cer
+certificate_file =
 
 # Private Certificate Key file in pem format. (string value)
-certificate_key_file = /usr/local/bin/aai_key.key
+certificate_key_file =
 
 # Certificate Authority Bundle file in pem format. Must contain the appropriate
 # trust chain for the Certificate file. (string value)
 #certificate_authority_bundle_file = /opt/app/conductor/etc/certs/ca_bundle.pem
-certificate_authority_bundle_file =/usr/local/bin/bundle.pem
+certificate_authority_bundle_file =/usr/local/bin/AAF_RootCA.cer
+
+# Basic Authentication Username (string value)
+username = OOF
+
+# Basic Authentication Password (string value)
+password = OOF
+
 
 [api]
 
@@ -355,3 +362,31 @@ concurrent = true
 [reservation]
 concurrent = true
 
+
+[multicloud]
+
+#
+# From conductor
+#
+
+# Base URL for Multicloud without a trailing slash. (string value)
+server_url = http://{{.Values.config.msb.serviceName}}:{{.Values.config.msb.port}}/api/multicloud
+
+# Timeout for Multicloud Rest Call (string value)
+multicloud_rest_timeout = 30
+
+# Number of retry for Multicloud Rest Call (string value)
+multicloud_retries = 3
+
+# The version of Multicloud API. (string value)
+server_url_version = v0
+
+
+[vim_controller]
+#
+# From conductor
+#
+
+# Extensions list to use (list value)
+extensions = multicloud
+
index 25a076b..00f59a6 100755 (executable)
@@ -30,3 +30,11 @@ global:
 pullPolicy: Always
 nodePortPrefix: 302
 dataRootDir: /dockerdata-nfs
+config:
+  aai:
+    serviceName: aai
+    port: 8443
+  msb:
+    serviceName: msb-iag
+    port: 80
+
index 7fe63a5..f286be4 100755 (executable)
@@ -1,37 +1,18 @@
-# Copyright © 2017 Amdocs, Bell Canada
-#
-# Licensed under the Apache License, Version 2.0 (the "License");
-# you may not use this file except in compliance with the License.
-# You may obtain a copy of the License at
-#
-#       http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-
-
-# Credentials for the OOF placement service – SO. Newly added.
 osdfPlacementSOUsername: {{ .Values.config.osdfPlacementSOUsername }}
 osdfPlacementSOPassword: {{ .Values.config.osdfPlacementSOPassword }}
 
-# Credentials for the OOF CM scheduling service – Generic. Newly added
 osdfCMSchedulerUsername: {{ .Values.config.osdfCMSchedulerUsername }}
 osdfCMSchedulerPassword: {{ .Values.config.osdfCMSchedulerPassword }}
 
-# msoUrl: ""   # The SO url for call back. This will be part of the request, so no need
 soUsername: {{ .Values.config.soUsername }}
 soPassword: {{ .Values.config.soPassword }}
 
-conductorUrl: "{{.Release.Name}}-has-api.{{include "common.namespace" .}}:TODO-PORT-FOR-HAS-API"
+conductorUrl: {{ .Values.config.conductorUrl }}
 conductorUsername: {{ .Values.config.conductorUsername }}
 conductorPassword: {{ .Values.config.conductorPassword }}
 conductorPingWaitTime: {{ .Values.config.conductorPingWaitTime }}
 conductorMaxRetries: {{ .Values.config.conductorMaxRetries }}
 
-# Policy Platform -- requires ClientAuth, Authorization, and Environment
 policyPlatformUrl: {{ .Values.config.policyPlatformUrl }}
 policyPlatformEnv: {{ .Values.config.policyPlatformEnv }}
 policyPlatformUsername: {{ .Values.config.policyPlatformUsername }}
index 9f74271..51114a6 100644 (file)
@@ -34,7 +34,7 @@ debugEnabled: false
 
 # application configuration
 config:
-  msbgateway: msb-iag.{{ include "common.namespace" . }}
+  msbgateway: msb-iag
   msbPort: 80
   # Credentials for the OOF placement service – SO.
   osdfPlacementSOUsername: so_test
@@ -46,7 +46,7 @@ config:
   soUsername: ""   # SO username for call back.
   soPassword: ""   # SO password for call back.
   # Url and credentials for Conductor.
-  conductorUrl: https://oof-has-api.{{ include "common.namespace" . }}:8091/v1/plans/
+  conductorUrl: http://oof-has-api:8091/v1/plans/
   conductorUsername: admin1
   conductorPassword: plan.15
   conductorPingWaitTime: 60
index efc3976..d8cbda4 100644 (file)
@@ -25,7 +25,7 @@ global:
 #################################################################
 # application image
 repository: nexus3.onap.org:10001
-image: onap/policy-pe:1.2.2
+image: onap/policy-pe:1.2.3
 pullPolicy: Always
 
 # flag to enable debugging - application support required
index ca62f2b..1bc09a7 100644 (file)
@@ -27,7 +27,7 @@ global:
 #################################################################
 # application image
 repository: nexus3.onap.org:10001
-image: onap/policy-drools:1.2.2
+image: onap/policy-drools:1.2.3
 pullPolicy: Always
 
 # flag to enable debugging - application support required
index 811830e..707b975 100644 (file)
@@ -27,7 +27,7 @@ global:
 #################################################################
 # application image
 repository: nexus3.onap.org:10001
-image: onap/policy-pe:1.2.2
+image: onap/policy-pe:1.2.3
 pullPolicy: Always
 
 # flag to enable debugging - application support required
index 42d5847..6aa8c7e 100644 (file)
@@ -40,7 +40,7 @@ global:
 #################################################################
 # application image
 repository: nexus3.onap.org:10001
-image: onap/policy-pe:1.2.2
+image: onap/policy-pe:1.2.3
 pullPolicy: Always
 
 subChartsOnly:
index e574bd7..ab0bc35 100644 (file)
@@ -53,3 +53,6 @@ done
 
 # Create VNIs 100-199
 ${SDNC_HOME}/bin/addVnis.sh 100 199
+
+# Drop FK_NETWORK_MODEL foreign key as workaround for SDNC-291.
+${SDNC_HOME}/bin/rmForeignKey.sh NETWORK_MODEL FK_NETWORK_MODEL
\ No newline at end of file
index 8276d25..167a8ed 100755 (executable)
       "sdncTimeoutFirewall": "20",
       "callbackRetryAttempts": "30",
       "callbackRetrySleepTime": "1000",
-      "appcClientTopicRead": "APPC-LCM-READ",
-      "appcClientTopicWrite": "APPC-LCM-WRITE",
-      "appcClientTopicSdncRead": "SDNC-LCM-READ",
-      "appcClientTopicSdncWrite": "SDNC-LCM-WRITE",
+      "appcClientTopicRead": "APPC-LCM-WRITE",
+      "appcClientTopicWrite": "APPC-LCM-READ",
+      "appcClientTopicSdncRead": "SDNC-LCM-WRITE",
+      "appcClientTopicSdncWrite": "SDNC-LCM-READ",
       "appcClientTopicReadTimeout": "360000",
       "appcClientResponseTime": "360000",
       "appcClientPoolMembers": "{{.Values.config.messagerouter.serviceName}}:{{.Values.config.messagerouter.port}}",