2 ###################################
3 # Resources configuration entries #
4 ###################################
6 # Resource host information
8 # Directory on resource host where tars with resources are present
11 # tarfile name within resources_dir directory with offline infrastructure binaries.
12 # Content of APP_BINARY_RESOURCES_DIR (defined in package.conf) packaged by package.sh to single tar file.
15 # tarfile name within resources_dir directory with auxiliary resources.
16 # Content of APP_AUX_BINARIES (defined in package.conf) packaged by package.sh to single tar file.
17 # the purpose of auxiliary resources is to provide user an interface
18 # to distribute to infra node tar file with application specific files.
19 # Docker images in tar format are currently the only supported content of aux_resources package.
20 aux_resources_filename:
22 # resources can be exported via nfs
23 # default is no - client will use ssh
24 # if set yes but nfs-utils is missing then fallback to ssh
27 # Infra node specific information
29 # Offline solution source data binaries (resources_filename tar) will be
30 # decompressed in this directory on target infra server.
31 # e.g. app_data_path: /opt/onap
34 # Path for auxiliary data in target infra server.
35 # Data from resource host defined by aux_resources_filename variable is placed to this directory.
36 # Currently docker images in tar format are supported (see runtime_images parameter).
37 # Could be used for other kind of application specific data also.
38 # e.g. aux_data_path: /opt/onap/my_extra_pods_docker_images
39 aux_data_path: "{{ app_data_path }}/runtime_images_source_dir"
42 ##########################################
43 # Offline Infrastructure specific params #
44 ##########################################
46 # information from which rootCA is created
48 # organization_name: Samsung
49 # state_or_province_name: Poland
51 # locality_name: Krakow
54 state_or_province_name:
58 # Force k8s cluster redeploy if it exists already
59 # Default value is to allow redeploy
62 # Distribute offline rpm repository
63 # Default value is to distribute rpm
64 deploy_rpm_repository: yes
66 # Offline solution is deploying app specific rpm repository and requires some name
67 # also for k8s cluster
71 # runtime_images provides an way to insert docker images
72 # into nexus during infrastructure playbook execution (populated to nexus at runtime).
73 # images specified must be available inside aux_resources_filename
74 # tar file that is extracted by installer into aux_data_path directory in infra server.
75 # Source format of an image is .tar file in aux_data_path directory and all .tar
76 # files in that dir are checked to match runtime_images definition.
77 # if runtime_images are not specified nothing is inserted on top of existing
78 # prebuilt nexus blob in installation time.
79 # Component name must match with tar filename!
81 # aaa-component-0.0.1.tar is expected in aux_data_path for aaa-component image
83 # aaa-component-0.0.1:
84 # registry: "nexus3.onap.org:10001"
85 # path: "/onap/components/aaa-component"
89 ###############################
90 # Application specific params #
91 ###############################
93 # Project name to utilize same codebase. Just helper variable inside ansible
94 # configuration files (like this file) to avoid writing own project name multiple
95 # times for paths, namespaces, Helm release, derived variables..
96 # e.g. project_configuration: onap-casablanca
97 project_configuration:
99 # App Helm charts directory location in installation package.
100 # The path is absolute path (even locates relative inside of this sw package
101 # installation folder) because it must be visible for ansible docker/chroot
102 # process to find directory and to transfer it into machine (infra node) running
104 # Content of the folder must be Helm chart directories of the app with Makefile.
105 # In case of ONAP OOM it would be <oom_repo>/kubernetes folder content.
106 # NOTE: This default value should not be changed if not really needed and it
107 # must match with the variable "HELM_CHARTS_DIR_IN_PACKAGE" value in package.sh
109 app_helm_charts_install_directory: "/ansible/application/helm_charts"
111 # Specify target dir where helm charts are copied into on infra node.
112 # (same as content of "app_helm_charts_install_directory" copied by installer to this dir.)
113 # This must be directory with all charts and Makefile.
114 # e.g. app_helm_charts_infra_directory: "{{ app_data_path }}/helm_charts"
115 app_helm_charts_infra_directory: "{{ app_data_path }}/helm_charts"
117 # Main Helm chart to install
118 # e.g. app_helm_chart_name: onap
121 # Targets for helm charts repository build
122 # e.g. for ONAP Casablanca
123 # app_helm_build_targets:
126 app_helm_build_targets:
128 # Directory with helm plugins
129 # It's an optional parameter used e.g. in OOM Casablanca
130 # app_helm_plugins_directory: "{{ app_helm_charts_install_directory}}/helm/plugins/"
131 app_helm_plugins_directory:
133 # Helm release name (visible in POD names) used by Helm
134 # e.g. app_helm_release_name: "{{ project_configuration }}"
135 app_helm_release_name: "{{ project_configuration }}"
137 # Kubernetes namespace where application is installed
138 # e.g. app_kubernetes_namespace: onap
139 app_kubernetes_namespace:
141 # Optional application custom Ansible roles name for pre and post install logic.
142 # Location of additional custom roles is defined in ansible.cfg with roles_path.
143 # e.g. application_pre_install_role: "{{ project_configuration }}-patch-role"
144 application_pre_install_role:
145 application_post_install_role:
147 # any other application specific params can be specified in this file
150 # openStackKeyStoneUrl: "http://1.2.3.4:5000"
151 # openStackServiceTenantName: "services"
152 # openStackDomain: "Default"
153 # openStackUserName: "admin"
154 # openStackEncryptedPassword: "f7920677e15e2678b0f33736189e8965"