- Created by Alexis de Talhouët, last modified by Eliezio Oliveira on Aug 23, 2019
You are viewing an old version of this page. View the current version.
Compare with Current View Page History
« Previous Version 116 Next »
CDS is a framework to automate the resolution of resources for instantiation and any config provisioning operation, such as day0, day1 or day2 configuration.
CDS has a both design time and run time activities; during design time, Designer can define what actions are required for a given service, along with anything comprising the action. The design produce a CBA Package . Its content is driven from a catalog of reusable data dictionary and component, delivering a reusable and simplified self service experience.
CDS modelling is mainly based on TOSCA standard, using JSON as reprensentation.
Most of the TOSCA modeled entity presented in the bellow documentation can be found here.
Controller Blueprint Archive (.cba)
The Controller Blueprint Archive is the overall service design, fully model-driven, intent based package needed for provisioning and configuration management automation.
The CBA is .zip file, comprised of the following folder structure, the files may vary:
├── Definitions │ ├── blueprint.json Overall TOSCA service template (worfklow + node_template) │ ├── artifact_types.json (generated by enrichment) │ ├── data_types.json (generated by enrichment) │ ├── node_types.json (generated by enrichment) │ ├── relationship_types.json (generated by enrichment) │ └── resources_definition_types.json (generated by enrichment) ├── Environments Contains *.properties files as required by the service ├── Plans Contains Directed Graph ├── Scripts Contains scripts │ ├── python Python scripts │ └── kotlin Kotlin scripts ├── TOSCA-Metadata │ └── TOSCA.meta Meta-data of overall package └── Templates Contains combination of mapping and template
Dynamic payload
One of the most important API provided by the run time is to execute a CBA Package.
The nature of this API request and response is model driven and dynamic.
Here is how the a generic request and response look like.
{ "commonHeader": { "originatorId": "", "requestId": "", "subRequestId": "" }, "actionIdentifiers": { "blueprintName": "", "blueprintVersion": "", "actionName": "", "mode": "" }, "payload": { "$actionName-request": { "$actionName-properties": { } } } }
{ "commonHeader": { "originatorId": "", "requestId": "", "subRequestId": "" }, "actionIdentifiers": { "blueprintName": "", "blueprintVersion": "", "actionName": "", "mode": "" }, "payload": { "$actionName-response": { } } }
The actionName
, under the actionIdentifiers
refers to the name of a Workflow (see workflow)
The content of the payload
is what is fully dynamic / model driven.
The first top level element will always be either $actionName-request
for a request or $actionName-response
for a response.
Then the content within this element is fully based on the workflow inputs and outputs.
During the Enrichment, CDS will aggregate all the resources defined to be resolved as input, within mapping definition files, as data-type, that will then be use as type of an input called $actionName-properties
.
Enrichment
The enrichment process will complete the package by providing all the definition of types used:
- gather all the node-type used and put them into a
node_types.json
file - gather all the data-type used and put them into a
data_types.json
file - gather all the artifact-type used and put them into a
artifact_types.json
file - gather all the resource definitions used and put them into a
resources_definition_types.json
file
The enrichment can be run using REST API, and required the .zip file as input. It will return an enriched-cba.zip file.
curl -X POST \ 'http://{{ip}}:{{cds-designtime}}/api/v1/blueprint-model/enrich' \ -H 'content-type: multipart/form-data' \ -F file=@cba.zip
The enrichment process will also, for all resources to be resolved as input and default:
- dynamically gather them under a data-type, named
dt-${actionName}-properties
- will add it as a input of the workflow, as follow using this name:
${actionName}-properties
Example for workflow named resource-assignment:
"resource-assignment-properties": { "required": true, "type": "dt-resource-assignment-properties" }
External Systems support
Interaction with external systems is made dynamic, removing development cycle to support new endpoint.
In order to define the external system information, TOSCA provides dsl_definitions
. Link to TOSCA spec info 1, info 2.
Use cases:
- Resource resolution using REST or SQL external systems
- gRPC is supported for remote execution
- Any REST endpoint can be dynamically injected as part of the scripting framework.
Here are some examples on how to populate the system information within the package:
{ . . . "dsl_definitions": { "ipam-1": { "type": "token-auth", "url": "http://netbox-nginx.netprog:8080", "token": "Token 0123456789abcdef0123456789abcdef01234567" } } . . . }
{ . . . "dsl_definitions": { "ipam-1": { "type": "basic-auth", "url": "http://localhost:8080", "username": "bob", "password": "marley" } } . . . }
{ . . . "dsl_definitions": { "ipam-1": { "type" : "ssl-basic-auth", "url" : "http://localhost:32778", "keyStoreInstance": "JKS or PKCS12", "sslTrust": "trusture", "sslTrustPassword": "trustore password", "sslKey": "keystore", "sslKeyPassword: "keystore password" } } . . . }
{ . . . "dsl_definitions": { "remote-executor": { "type": "token-auth", "host": "cds-command-executor.netprog", "port": "50051", "token": "Basic Y2NzZGthcHBzOmNjc2RrYXBwcw==" }, } . . . }
{ . . . "dsl_definitions": { "netprog-db": { "type": "maria-db", "url": "jdbc:mysql://10.195.196.123:32050/netprog", "username": "netprog", "password": "netprog" } } . . . }
Expression
TOSCA provides for a set of functions to reference elements within the template or to retrieve runtime values.
Below is a list of supported expressions
get_input
The get_input function is used to retrieve the values of properties declared within the inputs section of a TOSCA Service Template.
Within CDS, this is mainly Workflow inputs.
Example:
"resolution-key": { "get_input": "resolution-key" }
get_property
The get_property function is used to retrieve property values between modelable entities defined in the same service template.
Example
TBD
get_attribute
The get_attribute function is used to retrieve the values of named attributes declared by the referenced node or relationship template name.
Example:
"get_attribute": [ "resource-assignment", "assignment-params" ]
get_operation_output
The get_operation_output function is used to retrieve the values of variables exposed / exported from an interface operation.
Example
TBD
get_artifact
The get_artifact function is used to retrieve artifact location between modelable entities defined in the same service template.
Example
TBD
Data Dictionary
A data dictionary models the how a specific resource can be resolved.
A resource is a variable/parameter in the context of the service. It can be anything, but it should not be confused with SDC or Openstack resources.
A data dictionary can have multiple sources to handle resolution in different ways.
The main goal of data dictionary is to define re-usable entity that could be shared.
Creation of data dictionaries is a standalone activity, separated from the blueprint design.
As part of modelling a data dictionary entry, the following generic information should be provided:
Property | Description | Scope |
---|---|---|
name | Data dictionary name | Mandatory |
tags | Information related | Mandatory |
updated-by | The creator | Mandatory |
property | Defines type and description, as nested JSON | Mandatory |
sources | List of resource source instance (see resource source) | Mandatory |
Bellow are properties that all the resource source have will have
The modeling does allow for data translation between external capability and CDS for both input and output key mapping.
Property | Description | Scope |
input-key-mapping | map of resources required to perform the request/query. The left hand-side is what is used within the query/request, the right hand side refer to a data dictionary instance. | Optional |
output-key-mapping | name of the resource to be resolved mapped to the value resolved by the request/query. | Optional |
key-dependencies | list of data dictionary instances to be resolved prior the resolution of this specific resource. during run time execution the key dependencies are recursively sorted and resolved in batch processing using the acyclic graph algorithm. |
Optional |
Example:vf-module-model-customization-uuid
and vf-module-label
are two data dictionaries. A SQL table, VF_MODULE_MODEL, exist to correlate them.
Here is how input-key-mapping, output-key-mapping and key-dependencies can be used:
{ "name" : "vf-module-label", "tags" : "vf-module-label", "updated-by" : "adetalhouet", "property" : { "description" : "vf-module-label", "type" : "string" }, "sources" : { "primary-db" : { "type" : "source-primary-db", "properties" : { "type" : "SQL", "query" : "select sdnctl.VF_MODULE_MODEL.vf_module_label as vf_module_label from sdnctl.VF_MODULE_MODEL where sdnctl.VF_MODULE_MODEL.customization_uuid=:customizationid", "input-key-mapping" : { "customizationid" : "vf-module-model-customization-uuid" }, "output-key-mapping" : { "vf-module-label" : "vf_module_label" }, "key-dependencies" : [ "vf-module-model-customization-uuid" ] } } } }
Data type
Represents the schema of a specific type of data.
Supports both primitive and complex data types:
Primitive | Complex |
---|---|
|
|
For complex data type, an entry schema is required, defining the type of value contained within the complex type, if list or array.
Users can create as many data type as needed.
Data type are useful to manipulate data during resource resolution. They can be used to format the JSON output as needed.
List of existing data type: https://github.com/onap/ccsdk-cds/tree/master/components/model-catalog/definition-type/starter-type/data_type
Bellow is a list of existing data types
datatype-resource-assignment
Used to define entries within artifact-mapping-resource
That datatype represent a resource to be resolved. We also refer this as an instance of a data dictionary as it's directly linked to its definition.
Property | Description |
---|---|
property | Defines how the resource looks like (see datatype-property) |
input-param | Whether the resource can be provided as input. |
dictionary-name | Reference to the name of the data dictionary (see Data Dictionary). |
dictionary-source | Reference the source to use to resolve the resource (see Resource source). |
dependencies | List of dependencies required to resolve this resource. |
updated-date | Date when mapping was upload. |
updated-by | Name of the person that updated the mapping. |
{ "version": "1.0.0", "description": "This is Resource Assignment Data Type", "properties": { "property": { "required": true, "type": "datatype-property" }, "input-param": { "required": true, "type": "boolean" }, "dictionary-name": { "required": false, "type": "string" }, "dictionary-source": { "required": false, "type": "string" }, "dependencies": { "required": true, "type": "list", "entry_schema": { "type": "string" } }, "updated-date": { "required": false, "type": "string" }, "updated-by": { "required": false, "type": "string" } }, "derived_from": "tosca.datatypes.Root" }
datatype-property
Used to defined the property entry of a resource assignment.
Property | Description |
---|---|
type | Whether it's a primitive type, or a defined data-type |
description | Description of for the property |
required | Whether it's required or not |
default | If there is a default value to provide |
entry_schema | If the type is a complex one, such as list, define what is the type of element within the list. |
{ "version": "1.0.0", "description": "This is Entry point Input Data Type, which is dynamic datatype, The parameter names will be populated during the Design time for each inputs", "properties": { "type": { "required": true, "type": "string" }, "description": { "required": false, "type": "string" }, "required": { "required": false, "type": "boolean" }, "default": { "required": false, "type": "string" }, "entry_schema": { "required": false, "type": "string" } }, "derived_from": "tosca.datatypes.Root" }
Artifact Type
Represents the type of a artifact, used to identify the implementation of the functionality supporting this type of artifact.
This node was created, derived from tosca.artifacts.Root
to be the root TOSCA node for all artifact.
{ "description": "TOSCA base type for implementation artifacts", "version": "1.0.0", "derived_from": "tosca.artifacts.Root" }
Bellow is a list of supported artifact types
artifact-directed-graph
Represents an Directed Graph
Here is the TOSCA artifact type:
{ "description": "Directed Graph File", "version": "1.0.0", "file_ext": [ "json", "xml" ], "derived_from": "tosca.artifacts.Implementation" }
artifact-template-velocity
Represents an Apache Velocity template.
Apache Velocity allow to insert logic (if / else / loops / etc) when processing the output of a template/text.
File must have .vtl extension.
The template can represent anything, such as device config, payload to interact with 3rd party systems, resource-accumulator template, etc...
Often a template will be parameterized, and each parameter must be defined within an mapping file.
Here is the TOSCA artifact type:
{ "description": " Velocity Template used for Configuration", "version": "1.0.0", "file_ext": [ "vtl" ], "derived_from": "tosca.artifacts.Implementation" }
artifact-template-jinja
Represents an Jinja template.
Jinja template allow to insert logic (if / else / loops / etc) when processing the output of a template/text.
File must have .jinja extension.
The template can represent anything, such as device config, payload to interact with 3rd party systems, resource-accumulator template, etc...
Often a template will be parameterized, and each parameter must be defined within an mapping file.
Here is the TOSCA artifact type:
{ "description": " Jinja Template used for Configuration", "version": "1.0.0", "file_ext": [ "jinja" ], "derived_from": "tosca.artifacts.Implementation" }
artifact-mapping-resource
This type is meant to represent mapping files defining the contract of each resource to be resolved.
Each parameter in a template must have a corresponding mapping definition, modeled using datatype-resource-assignment.
Hence the mapping file is meant to be a list of entries defined using datatype-resource-assignment.
File must have .json extension.
Here is the TOSCA artifact type:
{ "description": "Resource Mapping File used along with Configuration template", "version": "1.0.0", "file_ext": [ "json" ], "derived_from": "tosca.artifacts.Implementation" }
The mapping file basically contains a reference to the data dictionary to use to resolve a particular resource.
The data dictionary defines the HOW and the mapping defines the WHAT.
Relation between data dictionary, mapping and template.
Below are two examples using color coding to help understand the relationships.
In orange is the information regarding the template. As mentioned before, template is part of the blueprint itself, and for the blueprint to know what template to use, the name has to match.
In green is the relationship between the value resolved within the template, and how it's mapped coming from the blueprint.
In blue is the relationship between a resource mapping to a data dictionary.
In red is the relationship between the resource name to be resolved and the HEAT environment variables.
The key takeaway here is that whatever the value is for each color, it has to match all across. This means both right and left hand side are equivalent; it's all on the designer to express the modeling for the service. That said, best practice is example 1.
artifact-directed-graph
Represents a directed graph.
This is to represent a workflow.
File must have .xml extension.
Here is the list of executors currently supported (see here for explanation and full potential list: Service Logic Interpreter Nodes)
- execute
- block
- return
- break
- exit
Here is the TOSCA artifact type:
{ "description": "Directed Graph File", "version": "1.0.0", "file_ext": [ "json", "xml" ], "derived_from": "tosca.artifacts.Implementation" }
Node type
In CDS, we have mainly two distinct types: components and source. We have some other type as well, listed in the other section.
Component
Used to represent a functionality along with its contract, such as inputs, ouputs, and attributes
Here is the root component TOSCA node type from which other node type will derive:
{ "description": "This is default Component Node", "version": "1.0.0", "derived_from": "tosca.nodes.Root" }
Bellow is a list of supported components
component-resource-resolution
Used to perform resolution of resources.
Requires as many as artifact-mapping-resource AND artifact-template-velocity or artifact-template-jinja as needed.
Ouput result
Will put the resolution result as an attribute in the workflow context called assignment-params.
Using the get_attribute expression, this attribute can be retrieve to be provided as workflow output (see Workflow).
Specify which template to resolve
Currently, resolution is bounded to a template. To specify which template to use, you need to fill in the `artifact-prefix-names` field.
See Template to understand what the artifact prefix name is.
Storing the result
To store each resource being resolved, along with their status, and the resolved template, `store-result` should be set to `true`.
Also, when storing the data, it must be in the context of either a `resource-id` and resoure-type`, or based on a given `resolution-key`
Run the resolution multiple time
If you need to run the same resolution component multiple times, use the field `occurence`. This will add the notion of occurrence to the resolution, and if storing the results, resources and templates, they will be accessible for each occurrence.
Occurrence is a number between 1 and N; when retrieving information for a given occurrence, the first iteration starts at 1.
This feature is usefull when you need to apply the same configuration accross network elements.
Here is the definition:
{ "description": "This is Resource Assignment Component API", "version": "1.0.0", "attributes": { "assignment-params": { "required": true, "type": "string" } }, "capabilities": { "component-node": { "type": "tosca.capabilities.Node" } }, "interfaces": { "ResourceResolutionComponent": { "operations": { "process": { "inputs": { "resolution-key": { "description": "Key for service instance related correlation.", "required": false, "type": "string" }, "occurrence": { "description": "Number of time to perform the resolution.", "required": false, "default": 1, "type": "integer" }, "store-result": { "description": "Whether or not to store the output.", "required": false, "type": "boolean" }, "resource-type": { "description": "Request type.", "required": false, "type": "string" }, "artifact-prefix-names": { "required": true, "description": "Template , Resource Assignment Artifact Prefix names", "type": "list", "entry_schema": { "type": "string" } }, "request-id": { "description": "Request Id, Unique Id for the request.", "required": true, "type": "string" }, "resource-id": { "description": "Resource Id.", "required": false, "type": "string" }, "action-name": { "description": "Action Name of the process", "required": false, "type": "string" }, "dynamic-properties": { "description": "Dynamic Json Content or DSL Json reference.", "required": false, "type": "json" } }, "outputs": { "resource-assignment-params": { "required": true, "type": "string" }, "status": { "required": true, "type": "string" } } } } } }, "derived_from": "tosca.nodes.Component" }
component-script-executor
Used to execute a script to perform NETCONF, RESTCONF, SSH commands from within the runtime container of CDS.
Two type of scripts are supported:
- Kotlin: offer a way more integrated scripting framework, along with a way faster processing capability. See more about Kotlin script: https://github.com/Kotlin/KEEP/blob/master/proposals/scripting-support.md
- Python: uses Jython which is bound to Python 2.7, end of life Januray 2020. See more about Jython: https://www.jython.org/
The `script-class-reference` field need to reference
- for kotlin: the package name up to the class. e.g. com.example.Bob
- for python: it has to be the path from the Scripts folder, e.g. Scripts/python/Bob.py
Here is the definition
{ "description": "This is Netconf Transaction Configuration Component API", "version": "1.0.0", "interfaces": { "ComponentScriptExecutor": { "operations": { "process": { "inputs": { "script-type": { "description": "Script type, kotlin type is supported", "required": true, "type": "string", "default": "internal", "constraints": [ { "valid_values": [ "kotlin", "jython", "internal" ] } ] }, "script-class-reference": { "description": "Kotlin Script class name with full package or jython script name.", "required": true, "type": "string" }, "dynamic-properties": { "description": "Dynamic Json Content or DSL Json reference.", "required": false, "type": "json" } }, "outputs": { "response-data": { "description": "Execution Response Data in JSON format.", "required": false, "type": "string" }, "status": { "description": "Status of the Component Execution ( success or failure )", "required": true, "type": "string" } } } } } }, "derived_from": "tosca.nodes.Component" }
component-remote-script-executor
Used to execute a python script in a dedicated micro-service, providing a Python 3.6 environment.
Ouput result
prepare-environment-logs: will contain the logs for all the pip install of ansible_galaxy setup
execute-command-logs: will contain the execution logs of the script, that were printed into sdout
Using the get_attribute expression, this attribute can be retrieve to be provided as workflow output (see Workflow).
Params
The `command` field need to reference the path from the Scripts folder of the scripts to execute, e.g. Scripts/python/Bob.py
The `packages` field allow to provide a list of PIP package to install in the target environment, or a requirements.txt file. Also, it supports Ansible role.
If requirements.txt is specified, then it should be provided as part of the Environment folder of the CBA.
"packages": [ { "type": "pip", "package": [ "requirements.txt" ] }, { "type": "ansible_galaxy", "package": [ "juniper.junos" ] } ]
The `argument-properties` allows to specified input argument to the script to execute. They should be expressed in a DSL, and they will be ordered as specified.
"ansible-argument-properties": { "arg0": "-i", "arg1": "Scripts/ansible/inventory.yaml", "arg2": "--extra-vars", "arg3": { "get_attribute": [ "resolve-ansible-vars", "", "assignment-params", "ansible-vars" ] } } }
The `dynamic-properties` can be anything that needs to be passed to the script that couldn't be passed as an argument, such as JSON object, etc... If used, they will be passed in as the last argument of the Python script.
Here is the definition
{ "description": "This is Remote Python Execution Component.", "version": "1.0.0", "attributes": { "prepare-environment-logs": { "required": false, "type": "string" }, "execute-command-logs": { "required": false, "type": "list", "entry_schema": { "type": "string" } }, "response-data": { "required": false, "type": "json" } }, "capabilities": { "component-node": { "type": "tosca.capabilities.Node" } }, "interfaces": { "ComponentRemotePythonExecutor": { "operations": { "process": { "inputs": { "endpoint-selector": { "description": "Remote Container or Server selector name.", "required": false, "type": "string", "default": "remote-python" }, "dynamic-properties": { "description": "Dynamic Json Content or DSL Json reference.", "required": false, "type": "json" }, "argument-properties": { "description": "Argument Json Content or DSL Json reference.", "required": false, "type": "json" }, "command": { "description": "Command to execute.", "required": true, "type": "string" }, "packages": { "description": "Packages to install based on type.", "required": false, "type" : "list", "entry_schema" : { "type" : "dt-system-packages" } } } } } } }, "derived_from": "tosca.nodes.Component" }
component-remote-ansible-executor
Used to execute an ansible playbook hosted in AWX/Anisble Tower.
Ouput result
ansible-command-status: status of the command
ansible-command-logs: will contain the execution logs of the playbook
Using the get_attribute expression, this attribute can be retrieve to be provided as workflow output (see Workflow).
Param
TBD
Here is the definition
{ "description": "This is Remote Ansible Playbook (AWX) Execution Component.", "version": "1.0.0", "attributes": { "ansible-command-status": { "required": true, "type": "string" }, "ansible-command-logs": { "required": true, "type": "string" } }, "capabilities": { "component-node": { "type": "tosca.capabilities.Node" } }, "interfaces": { "ComponentRemoteAnsibleExecutor": { "operations": { "process": { "inputs": { "job-template-name": { "description": "Primary key or name of the job template to launch new job.", "required": true, "type": "string" }, "limit": { "description": "Specify host limit for job template to run.", "required": false, "type": "string" }, "inventory": { "description": "Specify inventory for job template to run.", "required": false, "type": "string" }, "extra-vars" : { "required" : false, "type" : "json", "description": "json formatted text that contains extra variables to pass on." }, "tags": { "description": "Specify tagged actions in the playbook to run.", "required": false, "type": "string" }, "skip-tags": { "description": "Specify tagged actions in the playbook to omit.", "required": false, "type": "string" }, "endpoint-selector": { "description": "Remote AWX Server selector name.", "required": true, "type": "string" } } } } } }, "derived_from": "tosca.nodes.Component" }
Source
Used to represent a type of source to resolve a resource, along with the expected properties
Defines the contract to resolve a resource.
Here is the root component TOSCA node type from which other node type will derive:
{ "description": "TOSCA base type for Resource Sources", "version": "1.0.0", "derived_from": "tosca.nodes.Root" }
Bellow is a list of supported sources
Input
Expects the value to be provided as input to the request.
Here is the definition:
{ "description": "This is Input Resource Source Node Type", "version": "1.0.0", "properties": {}, "derived_from": "tosca.nodes.ResourceSource" }
Default
Expects the value to be defaulted in the model itself.
Here is the definition:
{ "description": "This is Default Resource Source Node Type", "version": "1.0.0", "properties": {}, "derived_from": "tosca.nodes.ResourceSource" }
REST
Expects the URI along with the VERB and the payload, if needed.
CDS is currently deployed along the side of SDNC, hence the default rest connection provided by the framework is to SDNC MDSAL.
Property | Description | Scope |
---|---|---|
type | Expected output value, only JSON supported for now | Optional |
verb | HTTP verb for the request - default value is GET | Optional |
payload | Payload to sent | Optional |
endpoint-selector | Specific REST system to interact with to (see Dynamic Endpoint) | Optional |
url-path |
URI | Mandatory |
path | JSON path to the value to fetch from the response | Mandatory |
expression-type |
Path expression type - default value is JSON_PATH | Optional |
Here is the definition:
{ "description": "This is Rest Resource Source Node Type", "version": "1.0.0", "properties": { "type": { "required": false, "type": "string", "default": "JSON", "constraints": [ { "valid_values": [ "JSON" ] } ] }, "verb": { "required": false, "type": "string", "default": "GET", "constraints": [ { "valid_values": [ "GET", "POST", "DELETE", "PUT" ] } ] }, "payload": { "required": false, "type": "string", "default": "" }, "endpoint-selector": { "required": false, "type": "string" }, "url-path": { "required": true, "type": "string" }, "path": { "required": true, "type": "string" }, "expression-type": { "required": false, "type": "string", "default": "JSON_PATH", "constraints": [ { "valid_values": [ "JSON_PATH", "JSON_POINTER" ] } ] }, "input-key-mapping": { "required": false, "type": "map", "entry_schema": { "type": "string" } }, "output-key-mapping": { "required": false, "type": "map", "entry_schema": { "type": "string" } }, "key-dependencies": { "required": true, "type": "list", "entry_schema": { "type": "string" } } }, "derived_from": "tosca.nodes.ResourceSource" }
SQL
Expects the SQL query to be modeled; that SQL query can be parameterized, and the parameters be other resources resolved through other means. If that's the case, this data dictionary definition will have to define key-dependencies
along with input-key-mapping
.
CDS is currently deployed along the side of SDNC, hence the primary database connection provided by the framework is to SDNC database.
Property | Description | Scope |
---|---|---|
type | Database type, only SQL supported for now | Mandatory |
endpoint-selector | Specific Database system to interact with to (see Dynamic Endpoint) | Optional |
query | Statement to execute | Mandatory |
Here is the definition:
{ "description": "This is Database Resource Source Node Type", "version": "1.0.0", "properties": { "type": { "required": true, "type": "string", "constraints": [ { "valid_values": [ "SQL" ] } ] }, "endpoint-selector": { "required": false, "type": "string" }, "query": { "required": true, "type": "string" }, "input-key-mapping": { "required": false, "type": "map", "entry_schema": { "type": "string" } }, "output-key-mapping": { "required": false, "type": "map", "entry_schema": { "type": "string" } }, "key-dependencies": { "required": true, "type": "list", "entry_schema": { "type": "string" } } }, "derived_from": "tosca.nodes.ResourceSource" }
Capability
Expects a script to be provided.
Property | Description | Scope |
---|---|---|
script-type | The type of the script - default value is Koltin | Optional |
script-class-reference |
The name of the class to use to create an instance of the script | Mandatory |
Here is the definition:
{ "description": "This is Component Resource Source Node Type", "version": "1.0.0", "properties": { "script-type": { "required": true, "type": "string", "default": "kotlin", "constraints": [ { "valid_values": [ "internal", "kotlin", "jython" ] } ] }, "script-class-reference": { "description": "Capability reference name for internal and kotlin, for jython script file path", "required": true, "type": "string" }, "key-dependencies": { "description": "Resource Resolution dependency dictionary names.", "required": true, "type": "list", "entry_schema": { "type": "string" } } }, "derived_from": "tosca.nodes.ResourceSource" }
Other
dg-generic
Identifies a Directed Graph used as imperative workflow.
Property | Description | Scope |
---|---|---|
dependency-node-templates | The node template the workflow depends on | Required |
Here is the definition:
{ "description": "This is Generic Directed Graph Type", "version": "1.0.0", "properties": { "content": { "required": true, "type": "string" }, "dependency-node-templates": { "required": true, "description": "Dependent Step Components NodeTemplate name.", "type": "list", "entry_schema": { "type": "string" } } }, "derived_from": "tosca.nodes.DG" }
A node_template of this type always provide one artifact, of type artifact-directed-graph
, which will be located under the Plans/ folder within the CBA.
"config-deploy-process" : { "type" : "dg-generic", "properties" : { "content" : { "get_artifact" : [ "SELF", "dg-config-deploy-process" ] }, "dependency-node-templates" : [ "nf-account-collection", "execute" ] }, "artifacts" : { "dg-config-deploy-process" : { "type" : "artifact-directed-graph", "file" : "Plans/CONFIG_ConfigDeploy.xml" } } }
In the DG bellow, the execute node refers to the node_template.
<service-logic xmlns='http://www.onap.org/sdnc/svclogic' xmlns:xsi='http://www.w3.org/2001/XMLSchema-instance' xsi:schemaLocation='http://www.onap.org/sdnc/svclogic ./svclogic.xsd' module='CONFIG' version='1.0.0'> <method rpc='ConfigDeploy' mode='sync'> <block atomic="true"> <execute plugin="nf-account-collection" method="process"> <outcome value='failure'> <return status="failure"> </return> </outcome> <outcome value='success'> <execute plugin="execute" method="process"> <outcome value='failure'> <return status="failure"> </return> </outcome> <outcome value='success'> <return status='success'> </return> </outcome> </execute> </outcome> </execute> </block> </method> </service-logic>
tosca.nodes.VNF
Identifies a VNF, can be used to correlate any type of VNF related information.
{ "description": "This is VNF Node Type", "version": "1.0.0", "derived_from": "tosca.nodes.Root" }
vnf-netconf-device
Represents the VNF information to establish a NETCONF communication.
{ "description": "This is VNF Device with Netconf Capability", "version": "1.0.0", "capabilities": { "netconf": { "type": "tosca.capabilities.Netconf", "properties": { "login-key": { "required": true, "type": "string", "default": "sdnc" }, "login-account": { "required": true, "type": "string", "default": "sdnc-tacacs" }, "source": { "required": false, "type": "string", "default": "npm" }, "target-ip-address": { "required": true, "type": "string" }, "port-number": { "required": true, "type": "integer", "default": 830 }, "connection-time-out": { "required": false, "type": "integer", "default": 30 } } } }, "derived_from": "tosca.nodes.Vnf" }
Workflow
Workflow Scope within CDS Framework
The workflow is within the scope of the micro provisioning and configuration management in controller domain and does NOT account for the MACRO service orchestration workflow which is covered by the SO Project.
A workflow defines an overall action to be taken on the service, hence is an entry-point for the run-time execution of the CBA package.
A workflow also defines inputs and outputs that will defined the payload contract of the request and response (see Dynamic API)
A workflow can be composed of one or multiple sub-actions to execute.
A CBA package can have as many workflows as needed.
Single action
The workflow is directly backed by a Component
In the example bellow, the target of the workflow's steps resource-assignment
is `resource-assignment` which actually is the name of the node_template
defined after, of type component-resource-resolution
.
Link to the example.
. . . "topology_template": { "workflows": { "resource-assignment": { "steps": { "resource-assignment": { "description": "Resource Assign Workflow", "target": "resource-assignment" ] } }, "inputs": { "resource-assignment-properties": { "description": "Dynamic PropertyDefinition for workflow(resource-assignment).", "required": true, "type": "dt-resource-assignment-properties" } }, "outputs": { "meshed-template": { "type": "json", "value": { "get_attribute": [ "resource-assignment", "assignment-params" ] } } } }, "node_templates": { "resource-assignment": { "type": "component-resource-resolution", "interfaces": { "ResourceResolutionComponent": { "operations": { "process": { "inputs": { "artifact-prefix-names": [ "vf-module-1" ] } } } } }, "artifacts": { "vf-module-1-template": { "type": "artifact-template-velocity", "file": "Templates/vf-module-1-template.vtl" }, "vf-module-1-mapping": { "type": "artifact-mapping-resource", "file": "Templates/vf-module-1-mapping.json" } } } } . . .
Multiple sub-actions
The workflow is backed by a Directed Graph engine, dg-generic
, and is an imperative workflow.
A DG used as workflow for CDS is composed of multiple execute nodes; each individual execute node refers to an modelled Component
instance.
In the example above, you can see the target of the workflow's steps execute-script
is `execute-remote-ansible-process`, which is a node_template
of type dg_generic
. . . "topology_template": { "workflows": { "execute-remote-ansible": { "steps": { "execute-script": { "description": "Execute Remote Ansible Script", "target": "execute-remote-ansible-process" ] } }, "inputs": { "ip": { "required": false, "type": "string" }, "username": { "required": false, "type": "string" }, "password": { "required": false, "type": "string" }, "execute-remote-ansible-properties": { "description": "Dynamic PropertyDefinition for workflow(execute-remote-ansible).", "required": true, "type": "dt-execute-remote-ansible-properties" } }, "outputs": { "ansible-variable-resolution": { "type": "json", "value": { "get_attribute": [ "resolve-ansible-vars", "assignment-params" ] } }, "prepare-environment-logs": { "type": "string", "value": { "get_attribute": [ "execute-remote-ansible", "prepare-environment-logs" ] } }, "execute-command-logs": { "type": "string", "value": { "get_attribute": [ "execute-remote-ansible", "execute-command-logs" ] } } } } }, "node_templates": { "execute-remote-ansible-process": { "type": "dg-generic", "properties": { "content": { "get_artifact": [ "SELF", "dg-execute-remote-ansible-process" ] }, "dependency-node-templates": [ "resolve-ansible-vars", "execute-remote-ansible" ] }, "artifacts": { "dg-execute-remote-ansible-process": { "type": "artifact-directed-graph", "file": "Plans/CONFIG_ExecAnsiblePlaybook.xml" } } }
Properties of a workflow
Property | Description | |||||||
---|---|---|---|---|---|---|---|---|
workflow-name | Defines the name of the action that can be triggered by external system | |||||||
inputs |
They are two types of inputs, the dynamic ones, and the static one. Specified at workflow level
These will end up under Represent the resources defined as input within mapping definition files. The enrichment process will (see Enrichment)
Example for workflow named resource-assignment: dynamic input
"resource-assignment-properties": { "required": true, "type": "dt-resource-assignment-properties" } |
|||||||
outputs | Defines the outputs of the execution; there can be as many output as needed. Depending on the
|
|||||||
steps | Defines the actual step to execute as part of the workflow
|
Example:
{ "workflow": { "resource-assignment": { <- workflow-name "inputs": { "vnf-id": { <- static inputs "required": true, "type": "string" }, "resource-assignment-properties": { <- dynamic inputs "required": true, "type": "dt-resource-assignment-properties" } }, "steps": { "call-resource-assignment": { <- step-name "description": "Resource Assignment Workflow", "target": "resource-assignment-process" <- node_template targeted by the step } }, "outputs": { "template-properties": { <- output "type": "json", <- complex type "value": { "get_attribute": [ <- uses expression to retrieve attribute from context "resource-assignment", "assignment-params" ] } } } } } }
Template
A template is an artifact, and uses artifact-mapping-resource and artifact-template-velocity.
A template is parameterized and each parameter must be defined in a corresponding mapping file.
In order to know which mapping correlates to which template, the file name must start with an artifact-prefix
, serving as identifier to the overall template + mapping.
The requirement is as follows:
${artifact-prefix}-template
${artifact-prefix}-mapping
Scripts
Library
NetconfClient
In order to facilitate NETCONF interaction within scripts, a python NetconfClient binded to our Kotlin implementation is made available. This NetconfClient can be used when using the component-netconf-executor
.
The client can be find here: https://github.com/onap/ccsdk-cds/blob/master/components/scripts/python/ccsdk_netconf/netconfclient.py
ResolutionHelper
When executing a component executor script, designer might want to perform resource resolution along with template meshing directly from the script itself.
The helper can be find here: https://github.com/onap/ccsdk-cds/blob/master/components/scripts/python/ccsdk_netconf/common.py
- No labels