OOM Quick Start Guide


Once a kubernetes environment is available (follow the instructions in OOM Cloud Setup Guide if you don’t have a cloud environment available), follow the following instructions to deploy ONAP.

Step 1. Clone the OOM repository from ONAP gerrit:

> git clone -b <BRANCH> http://gerrit.onap.org/r/oom --recurse-submodules
> cd oom/kubernetes

where <BRANCH> can be an offical release tag, such as 4.0.0-ONAP for Dublin 5.0.1-ONAP for El Alto

Step 2. Install Helm Plugins required to deploy ONAP:

> sudo cp -R ~/oom/kubernetes/helm/plugins/ ~/.helm

Step 3. Customize the helm charts like oom/kubernetes/onap/values.yaml or an override file like onap-all.yaml, onap-vfw.yaml or openstack.yaml file to suit your deployment with items like the OpenStack tenant information.


Standard and example override files (e.g. onap-all.yaml, openstack.yaml) can be found in the oom/kubernetes/onap/resources/overrides/ directory.
  1. You may want to selectively enable or disable ONAP components by changing the enabled: true/false flags.
  2. Encyrpt the OpenStack password using the shell tool for robot and put it in the robot helm charts or robot section of openstack.yaml
  3. Encrypt the OpenStack password using the java based script for SO helm charts or SO section of openstack.yaml.
  4. Update the OpenStack parameters that will be used by robot, SO and APPC helm charts or use an override file to replace them.

a. Enabling/Disabling Components: Here is an example of the nominal entries that need to be provided. We have different values file available for different contexts.

b. Generating ROBOT Encrypted Password: The ROBOT encrypted Password uses the same encryption.key as SO but an openssl algorithm that works with the python based Robot Framework.


To generate ROBOT openStackEncryptedPasswordHere :

cd so/resources/config/mso/

/oom/kubernetes/so/resources/config/mso# echo -n "<openstack tenant password>" | openssl aes-128-ecb -e -K `cat encryption.key` -nosalt | xxd -c 256 -p

c. Generating SO Encrypted Password: The SO Encrypted Password uses a java based encryption utility since the Java encryption library is not easy to integrate with openssl/python that ROBOT uses in Dublin.


To generate SO openStackEncryptedPasswordHere and openStackSoEncryptedPassword:

SO_ENCRYPTION_KEY=`cat ~/oom/kubernetes/so/resources/config/mso/encryption.key`


git clone http://gerrit.onap.org/r/integration

cd integration/deployment/heat/onap-rke/scripts

javac Crypto.java

[ if javac is not installed ‘apt-get update ; apt-get install default-jdk’ ]


  1. Update the OpenStack parameters:

There are assumptions in the demonstration VNF heat templates about the networking available in the environment. To get the most value out of these templates and the automation that can help confirm the setup is correct, please observe the following constraints.


This network should allow heat templates to add interfaces. This need not be an external network, floating IPs can be assigned to the ports on the VMs that are created by the heat template but its important that neutron allow ports to be created on them.

openStackPrivateNetCidr: “”

This ip address block is used to assign OA&M addresses on VNFs to allow ONAP connectivity. The demonstration heat templates assume that 10.0 prefix can be used by the VNFs and the demonstration ip addressing plan embodied in the preload template prevent conflicts when instantiating the various VNFs. If you need to change this, you will need to modify the preload data in the robot helm chart like integration_preload_parametes.py and the demo/heat/preload_data in the robot container. The size of the CIDR should be sufficient for ONAP and the VMs you expect to create.

openStackOamNetworkCidrPrefix: “10.0”

This ip prefix mush match the openStackPrivateNetCidr and is a helper variable to some of the robot scripts for demonstration. A production deployment need not worry about this setting but for the demonstration VNFs the ip asssignment strategy assumes 10.0 ip prefix.

Example Keystone v2.0 .. literalinclude:: example-integration-override.yaml


Example Keystone v3 (required for Rocky and later releases) .. literalinclude:: example-integration-override-v3.yaml


Step 4. To setup a local Helm server to server up the ONAP charts:

> helm serve &

Note the port number that is listed and use it in the Helm repo add as follows:

> helm repo add local

Step 5. Verify your Helm repository setup with:

> helm repo list

Step 6. Build a local Helm repository (from the kubernetes directory):

> make all; make onap

Step 7. Display the onap charts that available to be deployed:

> helm search onap -l
NAME                	CHART VERSION	APP VERSION	DESCRIPTION                                 
local/onap                	5.0.0        	Dublin  Open Network Automation Platform (ONAP)
local/aaf                 	5.0.0        	        ONAP Application Authorization Framework
local/aai                 	5.0.0        	        ONAP Active and Available Inventory
local/appc                	5.0.0        	        Application Controller
local/cassandra           	5.0.0        	        ONAP cassandra
local/cds                 	5.0.0        	        ONAP Controller Design Studio (CDS)
local/clamp               	5.0.0        	        ONAP Clamp
local/cli                 	5.0.0        	        ONAP Command Line Interface
local/common              	5.0.0        	        Common templates for inclusion in other charts
local/consul              	5.0.0        	        ONAP Consul Agent
local/contrib             	5.0.0        	        ONAP optional tools
local/dcaegen2            	5.0.0        	        ONAP DCAE Gen2
local/dgbuilder           	5.0.0        	        D.G. Builder application
local/dmaap               	5.0.0        	        ONAP DMaaP components
local/esr                 	5.0.0        	        ONAP External System Register
local/log                 	5.0.0        	        ONAP Logging ElasticStack
local/mariadb-galera      	5.0.0        	        Chart for MariaDB Galera cluster
local/mongo               	5.0.0        	        MongoDB Server
local/msb                 	5.0.0        	        ONAP MicroServices Bus
local/multicloud          	5.0.0        	        ONAP multicloud broker
local/music               	5.0.0        	        MUSIC - Multi-site State Coordination Service
local/mysql               	5.0.0        	        MySQL Server
local/nbi                 	5.0.0        	        ONAP Northbound Interface
local/network-name-gen    	5.0.0        	        Name Generation Micro Service
local/nfs-provisioner     	5.0.0        	        NFS provisioner
local/oof                 	5.0.0        	        ONAP Optimization Framework
local/pnda                	5.0.0        	        ONAP DCAE PNDA
local/policy              	5.0.0        	        ONAP Policy Administration Point
local/pomba               	5.0.0        	        ONAP Post Orchestration Model Based Audit
local/portal              	5.0.0        	        ONAP Web Portal
local/postgres            	5.0.0        	        ONAP Postgres Server
local/robot               	5.0.0        	        A helm Chart for kubernetes-ONAP Robot
local/sdc                 	5.0.0        	        Service Design and Creation Umbrella Helm charts
local/sdnc                	5.0.0        	        SDN Controller
local/sdnc-prom           	5.0.0        	        ONAP SDNC Policy Driven Ownership Management
local/sniro-emulator      	5.0.0        	        ONAP Mock Sniro Emulator
local/so                  	5.0.0        	        ONAP Service Orchestrator
local/uui                 	5.0.0        	        ONAP uui
local/vfc                 	5.0.0        	        ONAP Virtual Function Controller (VF-C)
local/vid                 	5.0.0        	        ONAP Virtual Infrastructure Deployment
local/vnfsdk              	5.0.0        	        ONAP VNF SDK


The setup of the Helm repository is a one time activity. If you make changes to your deployment charts or values be sure to use make to update your local Helm repository.

Step 8. Once the repo is setup, installation of ONAP can be done with a single command


The –timeout 900 is currently required in Dublin to address long running initialization tasks for DMaaP and SO. Without this timeout value both applications may fail to deploy.

To deploy all ONAP applications use this command:

> cd oom/kubernetes
> helm deploy dev local/onap --namespace onap -f onap/resources/overrides/onap-all.yaml -f onap/resources/overrides/environment.yaml -f onap/resources/overrides/openstack.yaml --timeout 900

All override files may be customized (or replaced by other overrides) as per needs.


Enables the modules in the ONAP deployment. As ONAP is very modular, it is possible to customize ONAP and disable some components through this configuration file.


Includes configuration values specific to the deployment environment.

Example: adapt readiness and liveness timers to the level of performance of your infrastructure


Includes all the Openstack related information for the default target tenant you want to use to deploy VNFs from ONAP and/or additional parameters for the embedded tests.

Step 9. Verify ONAP installation

Use the following to monitor your deployment and determine when ONAP is ready for use:

> kubectl get pods -n onap -o=wide


While all pods may be in a Running state, it is not a guarantee that all components are running fine.

Launch the healthcheck tests using Robot to verify that the components are healthy.

> ~/oom/kubernetes/robot/ete-k8s.sh onap health

Step 10. Undeploy ONAP

> helm undeploy dev –purge

More examples of using the deploy and undeploy plugins can be found here: https://wiki.onap.org/display/DW/OOM+Helm+%28un%29Deploy+plugins