from cloudify.decorators import operation
from cloudify.exceptions import NonRecoverableError, RecoverableError
import dockering as doc
+from dcaepolicy import Policies, POLICIES, POLICY_MESSAGE_TYPE
from dockerplugin import discovery as dis
from dockerplugin.decorators import monkeypatch_loggers, wrap_error_handling_start, \
merge_inputs_for_start
SERVICE_COMPONENT_NAME = "service_component_name"
SELECTED_CONTAINER_DESTINATION = "selected_container_destination"
CONTAINER_ID = "container_id"
+APPLICATION_CONFIG = "application_config"
+
+
+# Utility methods
+
+def _get_docker_logins(consul_host=CONSUL_HOST):
+ """Get Docker logins
+
+ The assumption is that all Docker logins to be used will be available in
+ Consul's key-value store under "docker_plugin/docker_logins" as a list of
+ json objects where each object is a single login:
+
+ [{ "username": "dcae_dev_ro", "password": "att123ro",
+ "registry": "nexus01.research.att.com:18443" }]
+ """
+ # REVIEW: The error handling may have to be re-examined. The current thought is
+ # that the logins *must* be setup even with an empty list otherwise the task
+ # will fail (fail fast). One alterative is to pass back empty list upon any
+ # issues but this would push potential issues to a later point of the
+ # deployment.
+ kv_conn = dis.create_kv_conn(consul_host)
+ return dis.get_kv_value(kv_conn, "docker_plugin/docker_logins")
+
# Lifecycle interface calls for dcae.nodes.DockerContainer
"""Setup for config discovery"""
try:
name = kwargs['name']
- application_config = kwargs['application_config']
+ application_config = kwargs[APPLICATION_CONFIG]
# NOTE: application_config is no longer a json string and is inputed as a
# YAML map which translates to a dict. We don't have to do any
ctx.logger.info("Done setting up: {0}".format(name))
return kwargs
+def _merge_policy_updates(**kwargs):
+ app_config = kwargs[APPLICATION_CONFIG]
+ kwargs[APPLICATION_CONFIG] = Policies.shallow_merge_policies_into(app_config)
+ return kwargs
+
@monkeypatch_loggers
+@Policies.gather_policies_to_node
@operation
def create_for_components(**kwargs):
"""Create step for Docker containers that are components
"""
_done_for_create(
**_setup_for_discovery(
- **_generate_component_name(
- **ctx.node.properties)))
+ **_merge_policy_updates(
+ **_generate_component_name(
+ **ctx.node.properties))))
def _parse_streams(**kwargs):
@monkeypatch_loggers
+@Policies.gather_policies_to_node
@operation
def create_for_components_with_streams(**kwargs):
"""Create step for Docker containers that are components that use DMaaP
_done_for_create(
**_setup_for_discovery(
**_setup_for_discovery_streams(
- **_parse_streams(
- **_generate_component_name(
- **ctx.node.properties)))))
+ **_merge_policy_updates(
+ **_parse_streams(
+ **_generate_component_name(
+ **ctx.node.properties))))))
@monkeypatch_loggers
docker_host_ip = _lookup_service(docker_host, consul_host=consul_host)
- client = doc.create_client(docker_host_ip, DOCKER_PORT)
+ logins = _get_docker_logins(consul_host=consul_host)
+ client = doc.create_client(docker_host_ip, DOCKER_PORT, logins=logins)
hcp = doc.add_host_config_params_volumes(volumes=kwargs.get("volumes",
None))
docker_host_ip = _lookup_service(docker_host)
- client = doc.create_client(docker_host_ip, DOCKER_PORT)
+ logins = _get_docker_logins()
+ client = doc.create_client(docker_host_ip, DOCKER_PORT, logins=logins)
container_id = ctx.instance.runtime_properties[CONTAINER_ID]
doc.stop_then_remove_container(client, container_id)
raise RecoverableError(e)
+def _notify_container(**kwargs):
+ """Notify container using the policy section in the docker_config"""
+ dc = kwargs["docker_config"]
+
+ if "policy" in dc:
+ if dc["policy"]["trigger_type"] == "docker":
+ # REVIEW: Need to finalize on the docker config policy data structure
+ script_path = dc["policy"]["script_path"]
+ app_config = kwargs["application_config"]
+ updated_policies = kwargs["updated_policies"]
+ cmd = doc.build_policy_update_cmd(script_path, use_sh=False,
+ updated_policies=updated_policies,
+ application_config=app_config)
+
+ docker_host = kwargs[SELECTED_CONTAINER_DESTINATION]
+ docker_host_ip = _lookup_service(docker_host)
+ logins = _get_docker_logins()
+ client = doc.create_client(docker_host_ip, DOCKER_PORT, logins=logins)
+
+ container_id = kwargs["container_id"]
+
+ doc.notify_for_policy_update(client, container_id, cmd)
+ # else the default is no trigger
+
+ return kwargs
+
+def _done_for_policy_update(**kwargs):
+ name = kwargs['name']
+ ctx.instance.runtime_properties.update(kwargs)
+ ctx.logger.info("Done updating for policy: {0}".format(name))
+ return kwargs
+
+@monkeypatch_loggers
+@Policies.update_policies_on_node(configs_only=True)
+@operation
+def policy_update(updated_policies, **kwargs):
+ """Policy update task
+
+ This method is responsible for updating the application configuration and
+ notifying the applications that the change has occurred. This is to be used
+ for the dcae.interfaces.policy.policy_update operation.
+
+ :updated_policies: contains the list of changed policy-configs when configs_only=True
+ (default) Use configs_only=False to bring the full policy objects in :updated_policies:.
+ """
+ update_inputs = copy.deepcopy(ctx.instance.runtime_properties)
+ update_inputs["updated_policies"] = updated_policies
+
+ # Merge in policy updates into application config and make available
+ _done_for_policy_update(
+ **_notify_container(
+ **_setup_for_discovery(
+ **_merge_policy_updates(**update_inputs))))
+
+
# Lifecycle interface calls for dcae.nodes.DockerHost
--- /dev/null
+tosca_definitions_version: cloudify_dsl_1_3
+
+description: >
+ This Blueprint installs a chain of two laika instances on a Docker cluster
+
+imports:
+ - http://www.getcloudify.org/spec/cloudify/3.4/types.yaml
+ - {{ ONAPTEMPLATE_RAWREPOURL_org_onap_dcaegen2 }}/type_files/docker/2.3.0/node-type.yaml
+ - {{ ONAPTEMPLATE_RAWREPOURL_org_onap_dcaegen2 }}/type_files/relationship/1.0.0/node-type.yaml
+ - {{ ONAPTEMPLATE_RAWREPOURL_org_onap_dcaegen2 }}/type_files/dcaepolicy/1.0.0/node-type.yaml
+
+inputs:
+ laika_image:
+ type: string
+
+ host_capacity_policy_id:
+ type: string
+ default: DCAE_alex.Config_host_capacity_policy_id_value
+
+ host_location_policy_id:
+ type: string
+ default: DCAE_alex.Config_host_location_policy_id_value
+
+ db_server_policy_id:
+ type: string
+ default: DCAE_alex.Config_db_server_policy_id_value
+
+node_templates:
+
+ host_capacity_policy:
+ type: dcae.nodes.policy
+ properties:
+ policy_id: { get_input: host_capacity_policy_id }
+
+ host_location_policy:
+ type: dcae.nodes.policy
+ properties:
+ policy_id: { get_input: host_location_policy_id }
+
+ db_server_policy:
+ type: dcae.nodes.policy
+ properties:
+ policy_id: { get_input: db_server_policy_id }
+
+ laika-zero:
+ type: dcae.nodes.DockerContainerForComponents
+ properties:
+ service_component_type:
+ 'laika'
+ location_id:
+ 'rework-central'
+ service_id:
+ 'foo-service'
+ application_config:
+ some-param: "Lorem ipsum dolor sit amet"
+ downstream-laika: "{{ laika }}"
+ image: { get_input : laika_image }
+ docker_config:
+ healthcheck:
+ type: "http"
+ endpoint: "/health"
+ policy:
+ trigger_type: "docker"
+ script_path: "/bin/echo"
+ relationships:
+ # Link to downstream laika
+ - type: dcae.relationships.component_connected_to
+ target: laika-one
+ - type: dcae.relationships.component_contained_in
+ target: docker_host
+ - type: cloudify.relationships.depends_on
+ target: host_capacity_policy
+ - type: cloudify.relationships.depends_on
+ target: host_location_policy
+ interfaces:
+ cloudify.interfaces.lifecycle:
+ start:
+ inputs:
+ ports:
+ - "8080:5432"
+ envs:
+ SOME-ENV: "BAM"
+ max_wait:
+ 120
+ stop:
+ inputs:
+ cleanup_image:
+ False
+
+ laika-one:
+ type: dcae.nodes.DockerContainerForComponents
+ properties:
+ service_component_type:
+ 'laika'
+ application_config:
+ some-param: "Lorem ipsum dolor sit amet"
+ image: { get_input : laika_image }
+ # Trying without health check
+ relationships:
+ - type: dcae.relationships.component_contained_in
+ target: docker_host
+ - type: cloudify.relationships.depends_on
+ target: db_server_policy
+ interfaces:
+ cloudify.interfaces.lifecycle:
+ stop:
+ inputs:
+ cleanup_image:
+ False
+
+ docker_host:
+ type: dcae.nodes.SelectedDockerHost
+ properties:
+ location_id:
+ 'rework-central'
+ docker_host_override:
+ 'component_dockerhost'