Installation
Deployment Prerequisite/dependencies
VES-Mapper can be deployed individually though it will throw errors if it can’t reach to DMaaP instance’s APIs. To test it functionally, DMaaP is the only required prerequisite outside DCAE. As VES-Mapper is integrated with Consul / CBS, it fetches the initial configuration from Consul.
Blueprint/model/image
VES-Mapper blueprint is available @ https://git.onap.org/dcaegen2/platform/blueprints/tree/blueprints/k8s-ves-mapper.yaml?h=guilin
VES-Mapper docker image is available in Nexus repo @ nexus3.onap.org:10001/onap/org.onap.dcaegen2.services.mapper.vesadapter.universalvesadaptor:latest
1.To Run via blueprint
a. Verify DMaaP configurations in the blueprint as per setup
Dmaap Configuration consists of subscribe url to fetch notifications from the respective collector and publish url to publish ves event.
streams_publishes
and streams_subscribes
point to the publishing topic and subscribe topic respectively. Update these urls
as per your DMaaP configurations in the blueprint.
*b. Verify the Smooks mapping configuration in the blueprint as per the usecase. Blueprint contains default mapping for each supported collector ( SNMP Collector and RESTConf collector currently) which may serve the purpose for the usecase. The mapping-files
in collectors
contains the contents of the mapping file.
c. Upload the blueprint in the DCAE’s Cloudify instance
For this step, DCAE’s Cloudify instance should be in running state. Transfer blueprint file in DCAE bootstrap POD under /blueprints directory. Log-in to the DCAE bootstrap POD’s main container.
Note
For doing this, we should run the below commands
To get the bootstrap pod name, run this: kubectl get pods -n onap | grep bootstrap
To transfer blueprint file in bootstrap pod, run this: kubectl cp <source file path> <bootstrap pod>:/blueprints -n onap
To login to bootstrap pod name, run this: kubectl exec -it <bootstrap pod> bash -n onap
Note
Verify the below versions before validate blueprint
The version of the plugin used is different from “cfy plugins list”, update the blueprint import to match.
If the tag_version under inputs is old, update with the latest
Validate blueprint
cfy blueprints validate /blueprints/k8s-ves-mapper.yaml
Use following command for validated blueprint to upload:
cfy blueprints upload -b ves-mapper /blueprints/k8s-ves-mapper.yaml
d. Create the Deployment After VES-Mapper’s validated blueprint is uploaded, create Cloudify Deployment by following command
cfy deployments create -b ves-mapper ves-mapper
e. Deploy the component by using following command
cfy executions start -d ves-mapper install
To undeploy running ves-mapper, follow the below steps
a. cfy uninstall ves-mapper -f
Note
The deployment uninstall will also delete the blueprint. In some case you might notice 400 error reported indicating active deployment exist such as below.
Ex: An error occurred on the server: 400: Can’t delete deployment ves-mapper - There are running or queued executions for this deployment. Running executions ids: d89fdd0c-8e12-4dfa-ba39-a6187fcf2f18
b. In that case, cancel the execution ID then run uninstall as below
cfy executions cancel <Running executions ID>
cfy uninstall ves-mapper
2.To run on standalone mode
Though this is not a preferred way, to run VES-Mapper container on standalone mode using local configuration file carried in the docker image, following docker run command can be used.
docker run -d nexus3.onap.org:10003/onap/org.onap.dcaegen2.services.mapper.vesadapter.universalvesadaptor:1.1.0