X-Git-Url: https://gerrit.onap.org/r/gitweb?a=blobdiff_plain;f=docs%2Foom_user_guide.rst;h=74f8c57f6e124819c1fcfe7f2b54254a8014a523;hb=4501e19d3e57b38fa02c866abc812ec42ba63a63;hp=02f5c483b5cba7de74317bb8e800775dbb3c7e6e;hpb=94b5be41218a2d6be673dffd5fef938bfd995970;p=oom.git diff --git a/docs/oom_user_guide.rst b/docs/oom_user_guide.rst index 02f5c483b5..74f8c57f6e 100644 --- a/docs/oom_user_guide.rst +++ b/docs/oom_user_guide.rst @@ -15,8 +15,8 @@ .. _Kubernetes LoadBalancer: https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer .. _user-guide-label: -OOM User Guide helm3 (experimental) -################################### +OOM User Guide +############## The ONAP Operations Manager (OOM) provide the ability to manage the entire life-cycle of an ONAP installation, from the initial deployment to final @@ -55,8 +55,8 @@ ONAP with a few simple commands. Pre-requisites -------------- -Your environment must have both the Kubernetes `kubectl` and Helm setup as a -one time activity. +Your environment must have the Kubernetes `kubectl` with Cert-Manager +and Helm setup as a one time activity. Install Kubectl ~~~~~~~~~~~~~~~ @@ -64,7 +64,7 @@ Enter the following to install kubectl (on Ubuntu, there are slight differences on other O/Ss), the Kubernetes command line interface used to manage a Kubernetes cluster:: - > curl -LO https://storage.googleapis.com/kubernetes-release/release/v1.15.11/bin/linux/amd64/kubectl + > curl -LO https://storage.googleapis.com/kubernetes-release/release/v1.19.11/bin/linux/amd64/kubectl > chmod +x ./kubectl > sudo mv ./kubectl /usr/local/bin/kubectl > mkdir ~/.kube @@ -78,13 +78,18 @@ Verify that the Kubernetes config is correct:: At this point you should see Kubernetes pods running. +Install Cert-Manager +~~~~~~~~~~~~~~~~~~~~ +Details on how to install Cert-Manager can be found +:doc:`here `. + Install Helm ~~~~~~~~~~~~ Helm is used by OOM for package and configuration management. To install Helm, enter the following:: - > wget https://get.helm.sh/helm-v3.5.2-linux-amd64.tar.gz - > tar -zxvf helm-v3.5.2-linux-amd64.tar.gz + > wget https://get.helm.sh/helm-v3.6.3-linux-amd64.tar.gz + > tar -zxvf helm-v3.6.3-linux-amd64.tar.gz > sudo mv linux-amd64/helm /usr/local/bin/helm Verify the Helm version with:: @@ -199,7 +204,7 @@ deployment:: To install a specific version of a single ONAP component (`so` in this example) with the given release name enter:: - > helm deploy so onap/so --version 8.0.0 --set global.masterPassword=password --set global.flavor=unlimited --namespace onap + > helm deploy so onap/so --version 9.0.0 --set global.masterPassword=password --set global.flavor=unlimited --namespace onap .. note:: The dependent components should be installed for component being installed @@ -327,7 +332,7 @@ file: dependencies: <...> - name: so - version: ~8.0.0 + version: ~9.0.0 repository: '@local' condition: so.enabled <...> @@ -439,23 +444,24 @@ the portal and then simply access now the new ssl-encrypted URL: | Alternatives Considered: - - Kubernetes port forwarding was considered but discarded as it would require - the end user to run a script that opens up port forwarding tunnels to each of - the pods that provides a portal application widget. + - Kubernetes port forwarding was considered but discarded as it would + require the end user to run a script that opens up port forwarding tunnels + to each of the pods that provides a portal application widget. - Reverting to a VNC server similar to what was deployed in the Amsterdam - release was also considered but there were many issues with resolution, lack - of volume mount, /etc/hosts dynamic update, file upload that were a tall order - to solve in time for the Beijing release. + release was also considered but there were many issues with resolution, + lack of volume mount, /etc/hosts dynamic update, file upload that were + a tall order to solve in time for the Beijing release. Observations: - - If you are not using floating IPs in your Kubernetes deployment and directly attaching - a public IP address (i.e. by using your public provider network) to your K8S Node - VMs' network interface, then the output of 'kubectl -n onap get services | grep "portal-app"' + - If you are not using floating IPs in your Kubernetes deployment and + directly attaching a public IP address (i.e. by using your public provider + network) to your K8S Node VMs' network interface, then the output of + 'kubectl -n onap get services | grep "portal-app"' will show your public IP instead of the private network's IP. Therefore, - you can grab this public IP directly (as compared to trying to find the floating - IP first) and map this IP in /etc/hosts. + you can grab this public IP directly (as compared to trying to find the + floating IP first) and map this IP in /etc/hosts. .. figure:: oomLogoV2-Monitor.png :align: right @@ -556,15 +562,15 @@ Below is the example for the same:: > helm list NAME REVISION UPDATED STATUS CHART APP VERSION NAMESPACE - dev 1 Wed Oct 14 13:49:52 2020 DEPLOYED onap-8.0.0 Honolulu onap - dev-cassandra 5 Thu Oct 15 14:45:34 2020 DEPLOYED cassandra-8.0.0 onap - dev-contrib 1 Wed Oct 14 13:52:53 2020 DEPLOYED contrib-8.0.0 onap - dev-mariadb-galera 1 Wed Oct 14 13:55:56 2020 DEPLOYED mariadb-galera-8.0.0 onap + dev 1 Wed Oct 14 13:49:52 2020 DEPLOYED onap-9.0.0 Istanbul onap + dev-cassandra 5 Thu Oct 15 14:45:34 2020 DEPLOYED cassandra-9.0.0 onap + dev-contrib 1 Wed Oct 14 13:52:53 2020 DEPLOYED contrib-9.0.0 onap + dev-mariadb-galera 1 Wed Oct 14 13:55:56 2020 DEPLOYED mariadb-galera-9.0.0 onap Here the Name column shows the RELEASE NAME, In our case we want to try the scale operation on cassandra, thus the RELEASE NAME would be dev-cassandra. -Now we need to obtain the chart name for casssandra. Use the below +Now we need to obtain the chart name for cassandra. Use the below command to get the chart name:: > helm search cassandra @@ -573,21 +579,21 @@ Below is the example for the same:: > helm search cassandra NAME CHART VERSION APP VERSION DESCRIPTION - local/cassandra 8.0.0 ONAP cassandra - local/portal-cassandra 8.0.0 Portal cassandra - local/aaf-cass 8.0.0 ONAP AAF cassandra - local/sdc-cs 8.0.0 ONAP Service Design and Creation Cassandra + local/cassandra 9.0.0 ONAP cassandra + local/portal-cassandra 9.0.0 Portal cassandra + local/aaf-cass 9.0.0 ONAP AAF cassandra + local/sdc-cs 9.0.0 ONAP Service Design and Creation Cassandra Here the Name column shows the chart name. As we want to try the scale -operation for cassandra, thus the correponding chart name is local/cassandra +operation for cassandra, thus the corresponding chart name is local/cassandra Now we have both the command's arguments, thus we can perform the -scale opeartion for cassandra as follows:: +scale operation for cassandra as follows:: > helm upgrade dev-cassandra local/cassandra --set replicaCount=3 -Using this command we can scale up or scale down the cassadra db instances. +Using this command we can scale up or scale down the cassandra db instances. The ONAP components use Kubernetes provided facilities to build clustered, @@ -638,7 +644,7 @@ Prior to doing an upgrade, determine of the status of the deployed charts:: > helm list NAME REVISION UPDATED STATUS CHART NAMESPACE - so 1 Mon Feb 5 10:05:22 2020 DEPLOYED so-8.0.0 onap + so 1 Mon Feb 5 10:05:22 2020 DEPLOYED so-9.0.0 onap When upgrading a cluster a parameter controls the minimum size of the cluster during the upgrade while another parameter controls the maximum number of nodes @@ -675,7 +681,7 @@ To fetch release history enter:: > helm history so REVISION UPDATED STATUS CHART DESCRIPTION 1 Mon Feb 5 10:05:22 2020 SUPERSEDED so-8.0.0 Install complete - 2 Mon Feb 5 10:10:55 2020 DEPLOYED so-8.0.1 Upgrade complete + 2 Mon Feb 5 10:10:55 2020 DEPLOYED so-9.0.0 Upgrade complete Unfortunately, not all upgrades are successful. In recognition of this the lineup of pods within an ONAP deployment is tagged such that an administrator @@ -698,7 +704,7 @@ For example, to roll-back back to previous system revision enter:: > helm history so REVISION UPDATED STATUS CHART DESCRIPTION 1 Mon Feb 5 10:05:22 2020 SUPERSEDED so-8.0.0 Install complete - 2 Mon Feb 5 10:10:55 2020 SUPERSEDED so-8.0.1 Upgrade complete + 2 Mon Feb 5 10:10:55 2020 SUPERSEDED so-9.0.0 Upgrade complete 3 Mon Feb 5 10:14:32 2020 DEPLOYED so-8.0.0 Rollback to 1 .. note::