The new Beijing release scalability, resiliency, and manageablity are described here. These capabilities apply to the OOM/Kubernetes installation.
Follow the OOM installation instructions at
TLab specific Upgrade of Portal
Below are the notes specific for the TLab environment. If a change is made in OOM charts then the following needs to be completed to upgrade the running charts in
- Find Rancher IP in R_Control-Plane tenant - we use “onap_dev” key and “Ubuntu” user to SSH. For example: “ssh -i onap_dev firstname.lastname@example.org”.
- Login as ubuntu, then run "sudo -i" to login as root. The “oom” git repo is in the rancher vm's root directory, under “/root/oom”.
- Edit portal files at /root/oom/kubernetes/portal....
When complete run the following from /root/oom/kubernetes dir
In case, if it is a fresh install, try below command:
Stop command to stop any old namespace:
- Rancher gui is at 192.168.xx.xxx:8080
There is an interactive cli as well in the rancher gui where you can run kubectl commands.
Below command will show the list of portal services.
Accessing the ONAP Portal using OOM and a Kubernetes Cluster
If everything is successful, then to access Portal - http://onap.readthedocs.io/en/latest/submodules/oom.git/docs/oom_user_guide.html#accessing-the-onap-portal-using-oom-and-a-kubernetes-cluster
Overview of the running system
Verify that the portal healtcheck passes by the robot framework:
To scale a new portal-app, set the replica count appropriately.
In our tests below, we are going to work with the OOM portal component in isolation. In this exercise, we scale the portal-app with 2 new replicas.
The below code needs to be added to the integration-override.yaml file.
Run below command from "/root” folder to do helm upgrade
helm upgrade -i dev local/onap -f integration-override.yaml
A portal-app container failure can be simulated by stopping the portal-app container. The kubernetes liveness operation will detect that the ports are down, inferring there's a problem with the service, and in turn, will restart the container.
Here is where one of the running instances was deleted. And another is starting.
Here the new instance has started and the old one is still terminating
After the deleted instance has terminated the 3 instances are all running normally.
During this time there was no issues with the Portal Website.
TODO: In the following releases, the portal-db, portal-widget needs to be tested with similar resiliency and scaling tests. The portal-cassandra and portal-zookeeper requires coordination with MUSIC team to test the resiliency as there are some plans to provide MUSIC as a service.
Check Portal UI and perform sanity tests.
- After 3 instances of Portal are up, edit IP in /etc/hosts file, and logon as demo user on http://portal.api.simpledemo.onap.org:30215/ONAPPORTAL/login.htm
- Then killed 1 instance, I am able to continue on Portal page seamlessly
- Another test on failover timing, when killed all 3 instances, the new Portal processes are coming up within 30 seconds
In case of failures, below commands may get handy for accessing the pods/logs and troubleshooting the issue.
To find all portal pods:
From above list, now to check DB logs:
To check portal-app logs:
To get inside the portal-app docker and access the application logs:
Portal’s Rocket Chat channel - http://onap-integration.eastus.cloudapp.azure.com:3000/channel/portal