You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 106 Next »

vCPE Use Case - Customer Service Instantiation - 171025.pptx

The latest SNIRO Emulator tool jar:

11/1/2017

  1. vCpeResCust custom workflow: 
    1. Michael Lando found a solution for  SDC-564 - Getting issue details... STATUS . SDNC should be able to parse service template properly after the fix is committed.

    2. Brian manually configured SDNC to bypass service template parsing and checking as a temp hacking so that SDNC could pass "service assign".
    3. SO flow continued to create TunnelXConn. An exception occurred. It is determined that the cause is homing. The problem is tracked by INT-317 - Getting issue details... STATUS .
    4. The team determined that vCpeResCust needs to include a BRG allotted resource. We were able to modify the service model in SDC but distribution to SO ended in error. Waiting to be solved.

10/31/2017

  1. vCpeResCust custom workflow: 
    1. Service assign request from SO to SDNC has a format error. Fixed it by modifying the following line in mso docker containter: /etc/mso/config.d/mso.sdnc.properties. Note that the above is lost once the container is restarted. A JIRA ticket is opened to request a permanent fix: SO-295 - Getting issue details... STATUS .

      org.openecomp.mso.adapters.sdnc..service-topology-operation.assign=POST|270000|sdncurl8|sdnc-request-header|org:onap:sdnc:northbound:generic-resource
    2. In SDNC ueb listener docker container, /opt/onap/sdnc/data/properties/ueb-listener.properties does not have the correct sdc address, fixed. But the user name and password to authenticate with SDC are incorrect. So it's not able to get service model distributed from SDC. This is now fixed, see the file content below. Dan has committed the new config file to the repo.

      ueb-listener.properties
      org.onap.ccsdk.sli.northbound.uebclient.asdc-address=c2.vm1.sdc.simpledemo.openecomp.org:8443
      org.onap.ccsdk.sli.northbound.uebclient.consumer-group=sdc-OpenSource-Env1-sdnc-dockero
      org.onap.ccsdk.sli.northbound.uebclient.consumer-id=sdc-COpenSource-Env11-sdnc-dockero
      org.onap.ccsdk.sli.northbound.uebclient.environment-name=AUTO
      org.onap.ccsdk.sli.northbound.uebclient.password=Kp8bJ4SXszM0WXlhak3eHlcse2gAw84vaoGGmJvUy2U
      org.onap.ccsdk.sli.northbound.uebclient.user=sdnc
      org.onap.ccsdk.sli.northbound.uebclient.sdnc-user=admin
      org.onap.ccsdk.sli.northbound.uebclient.sdnc-passwd=Kp8bJ4SXszM0WXlhak3eHlcse2gAw84vaoGGmJvUy2U
      org.onap.ccsdk.sli.northbound.uebclient.asdc-api-base-url=http://sdnchost:8282/restconf/operations/
      org.onap.ccsdk.sli.northbound.uebclient.asdc-api-namespace=org:onap:ccsdk
      org.onap.ccsdk.sli.northbound.uebclient.spool.incoming=/opt/onap/sdnc/ueb-listener/spool/incoming
      org.onap.ccsdk.sli.northbound.uebclient.spool.archive=/opt/onap/sdnc/ueb-listener/spool/archive
      org.onap.ccsdk.sli.northbound.uebclient.polling-interval=30
      org.onap.ccsdk.sli.northbound.uebclient.polling-timeout=15
      org.onap.ccsdk.sli.northbound.uebclient.client-startup-timeout=900
      org.onap.ccsdk.sli.northbound.uebclient.relevant-artifact-types=YANG_XML,VF_LICENSE,TOSCA_TEMPLATE,TOSCA_CSAR,UCPE_LAYER_2_CONFIGURATION
      org.onap.ccsdk.sli.northbound.uebclient.activate-server-tls-auth=false
      org.onap.ccsdk.sli.northbound.uebclient.keystore-path=
      org.onap.ccsdk.sli.northbound.uebclient.keystore-password=
      org.onap.ccsdk.sli.northbound.uebclient.xslt-path-list=
      org.onap.ccsdk.sli.northbound.uebclient.artifact-map=/opt/onap/sdnc/data/properties/artifact.map
    3. Now SDNC is able to pick up service template distributed from SDC. But reports an error when Cambria parses yml template. Waiting to be fixed:  SDC-564 - Getting issue details... STATUS
    4. service-Vcperescust-template.yml
    5. service-Vcperescust-template.yml

10/30/2017

  1. vCpeResCust custom workflow: 
    1. Identified configuration mistake in mso.bpmn.urn.properties: both hostname and url path needs to be corrected to provide correct callback url to SNIRO. The updated file is here: mso.bpmn.urn.properties
    2. Identified a bug. When SO service level flow requests SDNC for service instance assignment, the request does not have correct format and got rejected. Dan provides a sample and Yang model in  SDNC-153 - Getting issue details... STATUS . A new ticket is created.  SO-289 - Getting issue details... STATUS .
  2. Eric and team have the following update on VNFs:
    1. There should be vnf snapshot images in the Integration environment now.  Note, that the vbng image Brian made today needs to be replaced with one Matt is in process of building. 
    2. Per Kang’s question below, the last line of each vnf yaml file runs an install script – this is what sets up the interface addresses, per the env file.  The snapshot images skip over need for time consuming compilation of the vpp/honeycomb code.
    3. We have tested manually configuring vxlan tunnels, etc. in the VCPE project and verified that the data path is working.
    4. we would like to remove and rebuild the vnfs in the vcpe project tomorrow morning.
      1. one item we’re in process of testing is control plane connectivity from sdnc to the vbrg
    5. A couple areas where there may be some unkowns that could come up during integration:
      1. interaction with the onap vdhcp vnf (different than what was tested in our local lab)
      2. interaction with the aaa vnf  (did not have this in our local lab setup)

10/27/2017

  1. vCpeResCust custom workflow: 

    1. Passed: Get and decompose service template from catalog.
    2. Passed: Query SNIRO emulator to get homing information. The current config files for SO after manual changes are:  mso.bpmn.urn.propertiesmso-docker.json. They are supposed to be updated by SO so that no more manual changes are needed in the future. The callback URL provided by SO is incorrect, needs to be fixed ( SO-278 - Getting issue details... STATUS ). SNIRO emulator needs to be modified to use the callback URL from SO request ( INT-311 - Getting issue details... STATUS ). Currently we use the following hacking to send the required info to SO.

      Manually send SNIRO results to SO
      curl -i -X POST -d @sniro.json -H Content-Type=application/json --user 'MSOClient:password1$' http://so:8080/workflows/messages/message/SNIROResponse/c39fe43a-8f31-4d70-957d-b1c11e161636
      
      
      sniro.json is preloaded into SNIRO emulator and will be used to feed SO request.
      {
         "solutionInfo" : {
            "placement" : [
               {
                  "serviceResourceId" : "61d563e8-e714-4393-8f99-cc480144a05e",
                  "resourceModuleName" : "TunnelXConn",
                  "serviceInstanceId" : "GMuxInfra-UUID",
                  "cloudRegionId" : "RegionOne",
                  "inventoryType" : "service"
               },
               {
                  "resourceModuleName" : "vG",
                  "serviceResourceId" : "91d563e8-e714-4393-8f99-cc480144a05e",
                  "cloudRegionId" : "RegionOne",
                  "serviceInstanceId" : "vG-service-instance-id",
                  "inventoryType" : "cloud"
               },
               {
                  "inventoryType" : "service",
                  "serviceInstanceId" : "BRG_EMU_UUID",
                  "cloudRegionId" : "RegionOne",
                  "resourceModuleName" : "BRG",
                  "serviceResourceId" : "71d563e8-e714-4393-8f99-cc480144a05e"
               }
            ]
         },
         "requestId" : "111-111-1111",
         "statusMessage" : "",
         "transactionId" : "111-111-1111",
         "requestState" : "complete"
      }
      
      
      
    3. Passed: SO queries AAI to get service and other info include globalCustomerID. We preload AAI with the following info. Note that the ASDC_TOSCA_UUID part is questionable. It seems not necessary. It is tracked by SO-279 - Getting issue details... STATUS .

      Preload AAI for vCpeResCust flow
      PUT /aai/v11 /business/customers/customer/SDN-ETHERNET-INTERNET
      
      {
          "global-customer-id": "SDN-ETHERNET-INTERNET",
          "subscriber-name": "SDN-ETHERNET-INTERNET",
          "subscriber-type": "INFRA",
          "service-subscriptions": {
              "service-subscription": [
                  {
                      "service-type": "123456789",
                      "service-instances": {
                          "service-instance": [
                              {
                                  "service-instance-id": "fbe9ad27-7ddd-49a6-ab2f-8c08e31fe12c",
                                  "service-instance-name": "fbe9ad27-7ddd-49a6-ab2f-8c08e31fe12c",
                                  "service-type": "vcpe"
                              }
                          ]
                      }
                  }
              ]
          }
      }
      
      
      After the sdc models were loaded, I added this model into AAI:
      
      PUT /aai/v11 /service-design-and-creation/models/model/1963dd8b-9375-4cab-aa59-0ee06e8333fa/model-vers/model-ver/ASDC_TOSCA_UUID
      {
          "model-version-id": "ASDC_TOSCA_UUID",
          "model-name": "vCpeResCust",
          "model-version": "1.0",
          "model-description": "Some ASDC Tosca Model"
      }
    4. Passed: SO creates a service instance UUID and  put it in AAI. 
    5. Blocked: SO calls SDNC assign service (type=vCpeResCust, UUID), see  SDNC-153 - Getting issue details... STATUS .
  2. General Infrastructure
    1. Brian has successfully instantiated general infrastructure, hahahaha~~~~~~~~~~
  3. Notes for upcoming test
    1. SO allows only one workflow to executive at a time. To clear the current one: 

      delete from mso_requests.infra_active_requests;
    2. To manually send event to DMaaP to invoke SDNC to create BRG record in AAI (this emulates the event from DHCP), do the following

      http://{{mr}}:3904/events/VCPE-DHCP-EVENT/group1/C1?timeout=5000
      [
          "{\"msg_name\":\"DHCPACK\",\"macaddr\":\"e2:91:8c:7a:1e:9d\",\"yiaddr\":\"10.3.0.2\"}"
      ]


10/26/2017

  1. vCpeResCust custom workflow: 
    1. The config changes for sniro end point should be visible at : /shared/mso-docker.json. We should look at "sniroEndpoint". Also Jim Hahn found out that the following line needs to be added to the same file. A ticket is opened:  SO-274 - Getting issue details... STATUS . Important: It seems that the mso container needs to be restarted to pick up the change in this file.

      "adaptersWorkflowMessageEndpoint": "http://mso:8080/workflows/messages/Resteasy",
    2. SNIRO emulator config tested. The related wiki page is here: SNIRO Emulator
    3. Add the following to /etc/mso/config.d/mso.bpmn.urn.properties to enable debug. Important: The changes are lost each time the container is restarted.

      Debug mode
      log.debug.CreateVcpeResCustService=true
      log.debug.DeleteVcpeResCustService=true
      log.debug.DoCreateAllottedResourceBRG=true
      log.debug.DoCreateAllottedResourceBRGRollback=true
      log.debug.DoCreateAllottedResourceTXC=true
      log.debug.DoCreateAllottedResourceTXCRollback=true
      log.debug.DoDeleteAllottedResourceBRG=true
      log.debug.DoDeleteAllottedResourceTXC=true
      log.debug.Homing=true
      
    4. A bug was reported when SO tried to create a SNIRO request. It is tracked by  SO-273 - Getting issue details... STATUS .
  2. General Infrastructure:
    1. Brian created preload as following

      VNF-API Preload
      POST http://10.12.25.64:8282/restconf/operations/VNF-API:preload-vnf-topology-operation
      Content-Type: application/json
      
      {
         "VNF-API:input": {
           "VNF-API:request-information": {
             "VNF-API:request-id": "robot12",
             "VNF-API:notification-url": "https://so.onap.org",
             "VNF-API:order-number": "robot12",
             "VNF-API:request-sub-action": "SUPP",
             "VNF-API:request-action": "PreloadVNFRequest",
             "VNF-API:source": "VID",
             "VNF-API:order-version": "1.0"
           },
           "VNF-API:sdnc-request-header": {
             "VNF-API:svc-action": "reserve",
             "VNF-API:svc-notification-url": "https://son.onap.org",
             "VNF-API:svc-request-id": "robot12"
           },
           "VNF-API:vnf-topology-information": {
                   "vnf-topology-identifier": {
                     "service-type": "3c8da74-9225-4b9d-8915-214f8ab273a3",
                     "vnf-type": "VcpeGenInfra..base_vcpe_infra..module-0",
                     "generic-vnf-name": "vcpe-gen-infra 0",
                     "generic-vnf-type": "VcpeGenInfra..base_vcpe_infra..module-0",
                     "vnf-name": "vCPEInfraVF102604"
             },
             "VNF-API:vnf-parameters": [
                    {
                    "vnf-parameter-name": "vcpe_image_name",
                    "vnf-parameter-value": "ubuntu-14-04-cloud-amd64"
                    },
                    {
                    "vnf-parameter-name": "vcpe_flavor_name",
                    "vnf-parameter-value": "m1.medium"
                    },
                    {
                    "vnf-parameter-name": "public_net_id",
                    "vnf-parameter-value": "external"
                    },
                    {
                    "vnf-parameter-name": "cpe_signal_net_id",
                    "vnf-parameter-value": "vCPEInfraCPESIGNAL102604"
                    },
                    {
                    "vnf-parameter-name": "cpe_signal_subnet_id",
                    "vnf-parameter-value": "vCPEInfraCPESIGNALSUB102604"
                    },
                    {
                    "vnf-parameter-name": "cpe_public_net_id",
                    "vnf-parameter-value": "vCPEInfraCPEPUBLIC102604"
                    },
                    {
                    "vnf-parameter-name": "cpe_public_subnet_id",
                    "vnf-parameter-value": "vCPEInfraCPEPUBLICSUB102604"
                    },
                    {
                    "vnf-parameter-name": "onap_private_net_id",
                    "vnf-parameter-value": "oam_onap_hUnI"
                    },
                    {
                    "vnf-parameter-name": "onap_private_subnet_id",
                    "vnf-parameter-value": "oam_onap_hUnI"
                    },
                    {
                    "vnf-parameter-name": "onap_private_net_cidr",
                    "vnf-parameter-value": "10.0.0.0/16"
                    },
                    {
                    "vnf-parameter-name": "cpe_signal_net_cidr",
                    "vnf-parameter-value": "10.4.0.0/24"
                    },
                    {
                    "vnf-parameter-name": "cpe_public_net_cidr",
                    "vnf-parameter-value": "10.2.0.0/24"
                    },
                    {
                    "vnf-parameter-name": "vdhcp_private_ip_0",
                    "vnf-parameter-value": "10.4.0.1"
                    },
                    {
                    "vnf-parameter-name": "vdhcp_private_ip_1",
                    "vnf-parameter-value": "10.0.101.1"
                    },
                    {
                    "vnf-parameter-name": "vaaa_private_ip_0",
                    "vnf-parameter-value": "10.4.0.4"
                    },
                    {
                    "vnf-parameter-name": "vaaa_private_ip_1",
                    "vnf-parameter-value": "10.0.101.2"
                    },
                    {
                    "vnf-parameter-name": "vdns_private_ip_0",
                    "vnf-parameter-value": "10.2.0.1"
                    },
                    {
                    "vnf-parameter-name": "vdns_private_ip_1",
                    "vnf-parameter-value": "10.0.101.3"
                    },
                    {
                    "vnf-parameter-name": "vweb_private_ip_0",
                    "vnf-parameter-value": "10.2.0.10"
                    },
                    {
                    "vnf-parameter-name": "vweb_private_ip_1",
                    "vnf-parameter-value": "10.0.101.40"
                    },
                    {
                    "vnf-parameter-name": "mr_ip_addr",
                    "vnf-parameter-value": "10.0.11.1"
                    },
                    {
                    "vnf-parameter-name": "vaaa_name_0",
                    "vnf-parameter-value": "zdcpe1cpe01aaa01_102604"
                    },
                    {
                    "vnf-parameter-name": "vdns_name_0",
                    "vnf-parameter-value": "zdcpe1cpe01dns01_102604"
                    },
                    {
                    "vnf-parameter-name": "vdhcp_name_0",
                    "vnf-parameter-value": "zdcpe1cpe01dhcp01_102604"
                    },
                    {
                    "vnf-parameter-name": "vweb_name_0",
                    "vnf-parameter-value": "zdcpe1cpe01web01_102604"
                    },
                    {
                    "vnf-parameter-name": "vnf_id",
                    "vnf-parameter-value": "vCPE_Infrastructure_demo_app_102604"
                    },
                    {
                    "vnf-parameter-name": "vf_module_id",
                    "vnf-parameter-value": "vCPE_Intrastructure_102604"
                    },
                    {
                    "vnf-parameter-name": "dcae_collector_ip",
                    "vnf-parameter-value": "10.0.4.102"
                    },
                    {
                    "vnf-parameter-name": "dcae_collector_port",
                    "vnf-parameter-value": "8080"
                    },
                    {
                    "vnf-parameter-name": "repo_url_blob",
                    "vnf-parameter-value": "https://nexus.onap.org/content/sites/raw"
                    },
                    {
                    "vnf-parameter-name": "repo_url_artifacts",
                    "vnf-parameter-value": "https://nexus.onap.org/content/groups/staging"
                    },
                    {
                    "vnf-parameter-name": "demo_artifacts_version",
                    "vnf-parameter-value": "1.1.0"
                    },
                    {
                    "vnf-parameter-name": "install_script_version",
                    "vnf-parameter-value": "1.1.0-SNAPSHOT"
                    },
                    {
                    "vnf-parameter-name": "key_name",
                    "vnf-parameter-value": "vaaa_key"
                    },
                    {
                    "vnf-parameter-name": "pub_key",
                    "vnf-parameter-value": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQDQXYJYYi3/OUZXUiCYWdtc7K0m5C0dJKVxPG0eI8EWZrEHYdfYe6WoTSDJCww+1qlBSpA5ac/Ba4Wn9vh+lR1vtUKkyIC/nrYb90ReUd385Glkgzrfh5HdR5y5S2cL/Frh86lAn9r6b3iWTJD8wBwXFyoe1S2nMTOIuG4RPNvfmyCTYVh8XTCCE8HPvh3xv2r4egawG1P4Q4UDwk+hDBXThY2KS8M5/8EMyxHV0ImpLbpYCTBA6KYDIRtqmgS6iKyy8v2D1aSY5mc9J0T5t9S2Gv+VZQNWQDDKNFnxqYaAo1uEoq/i1q63XC5AD3ckXb2VT6dp23BQMdDfbHyUWfJN"
                    },
                    {
                    "vnf-parameter-name": "cloud_env",
                    "vnf-parameter-value": "Openstack"
                    }
             ],
             "VNF-API:vnf-assignments": {
               }
             }
           }
      }
      
      
      
      

10/25/2017

  1. vCpeResCust custom workflow: 
    1. Re-created SO maria DB to reflect the latest change to the DB tables. This done by adding a line to /opt/test_lab/deploy.sh: "$DOCKER_COMPOSE_CMD up -d --force-recreate mariadb".

    2. The workflow successfully obtained service template from the catalog and decomposed it.

    3. A problem was observed when the workflow tried to call the homing service. It is tracked by SO-268 - Getting issue details... STATUS .
  2. General Infrastructure:
    1. Brian modified the neutron heat template and parameters in SO DB, pre-loaded SDNC, and successfully created virtual networks. (The scripts and template on 10/24 notes have been updated accordingly.) SDNC preload data for CPE_SIGNAL is below.

      SDNC Preload for CPE_SIGNAL
      {
        "VNF-API:input": {
          "VNF-API:request-information": {
            "VNF-API:request-id": "robot0012",
            "VNF-API:notification-url": "http://so.onap.org",
            "VNF-API:order-number": "robot0012",
            "VNF-API:request-sub-action": "SUPP",
            "VNF-API:request-action": "PreloadNetworkRequest",
            "VNF-API:source": "robot",
            "VNF-API:order-version": "1.0"
          },
          "VNF-API:network-topology-information": {
            "VNF-API:network-topology-identifier": {
              "VNF-API:network-role": "cpe-public",
              "VNF-API:network-technology": "neutron",
              "VNF-API:service-type": "vCPE",
              "VNF-API:network-name": "vCPEInfraCPEPUBLICL1025BDF01",
              "VNF-API:network-type": "Generic NeutronNet"
            },
            "VNF-API:provider-network-information": {
              "VNF-API:is-external-network": "true",
              "VNF-API:physical-network-name": "vCPEInfraCPEPUBLIC1025BDF01",
              "VNF-API:is-provider-network": "true",
              "VNF-API:is-shared-network": "true"
            },
            "VNF-API:subnets": [
              {
                "VNF-API:start-address": "10.2.0.2",
                "VNF-API:cidr-mask": "24",
                "VNF-API:subnet-name": "vCPEInfraCPEPUBLIC1025BDF01",
                "VNF-API:ip-version": "4",
                "VNF-API:dhcp-enabled": "N",
                "VNF-API:gateway-address": "10.2.0.1"
              }
                    ]
          },
          "VNF-API:sdnc-request-header": {
            "VNF-API:svc-action": "reserve",
            "VNF-API:svc-notification-url": "http://so.onap.org",
            "VNF-API:svc-request-id": "reobot0012"
          }
        }
      }
    2. During VNF instantiation, a bug was discovered. The content of DB table vnf_component_recipe is incorrect. It is tracked by  SO-267 - Getting issue details... STATUS .
    3. Brian found a few bugs related to SDNC: 

      Key Summary Status
      Loading...
      Refresh

10/24/2017

  1. vCpeResCust custom workflow: 
    1. Use the following to add the custom workflow into the recipe table:

      insert into recipe
      INSERT INTO `service_recipe` (`ACTION`, `VERSION_STR`, `DESCRIPTION`, `ORCHESTRATION_URI`, `SERVICE_PARAM_XSD`, `RECIPE_TIMEOUT`, `SERVICE_TIMEOUT_INTERIM`, `CREATION_TIMESTAMP`, `SERVICE_MODEL_UUID`) VALUES ('createInstance','1','vCpeRestCust','/mso/async/services/CreateResCustService',NULL,180,NULL,'2017-10-24 13:48:00','d8784dc7-4ec8-4609-87bb-5e5bfc183613');
    2. Successfully invoked the custom workflow from SO NBI using curl (note that VID currently only support a la carte so cannot invoke this flow). 

      invoke vCpeResCust workflow
      curl -X POST \
        http://so:8080/ecomp/mso/infra/serviceInstances/v5 \
        -H 'accept: application/json' \
        -H 'authorization: Basic SW5mcmFQb3J0YWxDbGllbnQ6cGFzc3dvcmQxJA==' \
        -H 'cache-control: no-cache' \
        -H 'content-type: application/json' \
        -H 'postman-token: 0c4f0ea7-736f-4999-3399-982de75ceecf' \
        -d '{
         "requestDetails" : {
            "requestInfo" : {
               "productFamilyId" : "a9a77d5a-123e-4ca2-9eb9-0b015d2ee0fb",
               "suppressRollback" : "false",
               "instanceName" : "vcperescust-102404",
               "requestorId": "demo",
               "source" : "VID"
            },
            "requestParameters" : {
               "subscriptionServiceType" : "123456789",
               "userParams" : [{
                  "name":"BRG_WAN_MAC_Address",
                  "value":"brgmac"
               }],
               "aLaCarte" : "false"
            },
            "subscriberInfo" : {
               "subscriberName" : "Kaneohe",
               "globalSubscriberId" : "SDN-ETHERNET-INTERNET"
            },
            "cloudConfiguration" : {
               "lcpCloudRegionId" : "RegionOne",
               "tenantId" : "466979b815b5415ba14ada713e6e1846"
            },
            "modelInfo" : {
               "modelType" : "service",
               "modelVersionId" : "ASDC_TOSCA_UUID",
               "modelName" : "vCpeResCust",
               "modelVersion" : "1.0",
               "modelInvariantId" : "1963dd8b-9375-4cab-aa59-0ee06e8333fa"
            }
         }
      }
      '
    3. A bug is discovered for the workflow and is being worked on (tracked by  SO-262 - Getting issue details... STATUS .
  2. Genera Infrastructure: With the following manual fix we are able to distribute the service to SO.
    1. The generic neutron network HEAT template is missing from the SO DB. Brian found a way to manually fix it. SO will include this in the repo. It is tracked by  SO-265 - Getting issue details... STATUS . The scipt and heat template are given below (updated on 10/25 based on Brian Freeman's comments to  SO-265 - Getting issue details... STATUS ).

      insert HEAT into SO DB
      INSERT INTO `heat_template_params` (`HEAT_TEMPLATE_ARTIFACT_UUID`,`PARAM_NAME`,`IS_REQUIRED`,`PARAM_TYPE`,`PARAM_ALIAS` ) VALUES ('efee1d84-b8ec-11e7-abc4-cec278b6b50a','network_name',1 ,'string', NULL);
      INSERT INTO `heat_template_params` (`HEAT_TEMPLATE_ARTIFACT_UUID`,`PARAM_NAME`,`IS_REQUIRED`,`PARAM_TYPE`,`PARAM_ALIAS` ) VALUES ('efee1d84-b8ec-11e7-abc4-cec278b6b50a','shared',0 ,'string', NULL);
      INSERT INTO `heat_template` (`ARTIFACT_UUID`, `NAME`,`VERSION` , `BODY`, `TIMEOUT_MINUTES`,`DESCRIPTION`, `CREATION_TIMESTAMP`, `ARTIFACT_CHECKSUM` ) VALUES ('efee1d84-b8ec-11e7-abc4-cec278b6b50a', 'GenericNeutronNet', '1', LOAD_FILE('/tmp/generic_neutron.yaml') , 10 ,'Generic Neutron Template',NOW(), 'MANUAL RECORD');
       
      heat_template_version: 2013-05-23
      description: A simple Neutron network
      parameters:
        network_name:
          type: string
          description: Name of the Neutron Network
          default: ONAP-NW1
        shared:
           type: boolean
           description: Shared amongst tenants
           default: True
      outputs:
        network_id:
          description: Openstack network identifier
          value: { get_resource: network }
      resources:
        network:
          type: OS::Neutron::Net
          properties:
            name: { get_param: network_name }
            shared: { get_param: shared }
      
      
      
      
    2. network_resource table model_invariant_uuid was too short (20 instead of at least 36). Manually increased to 120. It is tracked by  SO-266 - Getting issue details... STATUS
  3. Genera Infrastructure: Instantiation
    1. Need to run "/opt/demo.sh init" in the robot VM first.
    2. A service was created using VID.
    3. Tried to add a neutron network to it. SO received an error from SDNC. This is tracked by  SDNC-143 - Getting issue details... STATUS
    4. The 404 error in VID on deploy is due to missing zone data in AAI. Preload AAI with the following:

      Preload AAI with zone data
      PUT https://{{aai}}:8443/aai/v11/network/zones/zone/nova1
      {
          "zone-id": "nova1",
          "zone-name": "nova",
          "design-type": "integration",
          "zone-context": "labs",
          "status": "Active"
      }
  4. VNFs:
    1. Updated doc is available: ONAP vCPE VPP-based VNF Installation and Usage Information
  5. Closed loop: APPC-MultiCloud:
    1. The team is debugging the API request from APPC to MultiCloud. It is tracked by  MULTICLOUD-119 - Getting issue details... STATUS

10/23/2017

  1. MultiCloud found a bug and fixed it, tracked by  MULTICLOUD-118 - Getting issue details... STATUS .
  2. vGMUX: Eric has created a snapshot image for vGMUX and tested instantiation without ONAP. He and his team is working on the other VNFs.
  3. SDC: David Shadmi and Kang finished TunnelXConn VF design and vcperescust service design and distribution.
    1. To successfully distributed vcperescust, we need to first distribute vgmux service due to dependency.

10/19/2017

  1. APPC was able to reboot VM using legacy provider API. The team is currently debugging using the LCM framework where there is an issue with CDP PAL (Cloud Delivery Platform-Provider Abstraction Layer) interaction with MultiCloud.
  2. vGMUX: Eric and the team choose to build a image for each working VNF to avoid building and installation during instantiation. The team will also start to test the other VNFs including vBRG, vBNG, and vG.

10/18/2017

  1. APPC and MultiCloud located the problem to be API call request prefix. Bin Yang provided the following sample code to call MultiCloud API to do VM restart. APPC will fix the problem accordingly.

    export MULTICLOUD_PLUGIN_ENDPOINT=http://10.0.14.1:9005/api/multicloud-titanium_cloud/v0/pod25_RegionOne
    export TOKEN=$(curl -v -s -H "Content-Type: application/json" -X POST -d '{ }' $MULTICLOUD_PLUGIN_ENDPOINT/identity/v3/auth/tokens 2>&1 | grep X-Subject-Token | sed "s/^.*: //")
    export PROJECT_ID=466979b815b5415ba14ada713e6e1846
    curl -v -s -H "Content-Type: application/json" -H "X-Auth-Token: $TOKEN" -X GET $MULTICLOUD_PLUGIN_ENDPOINT/compute/v2.1/$PROJECT_ID/servers
    curl -v -s -H "Content-Type: application/json" -H "X-Auth-Token: $TOKEN" -X POST -d '{"os-stop":null}' $MULTICLOUD_PLUGIN_ENDPOINT/compute/v2.1/$PROJECT_ID/servers/0a06842a-4ec4-4918-b046-399f6b38f5f9/action
    curl -v -s -H "Content-Type: application/json" -H "X-Auth-Token: $TOKEN" -X POST -d '{"os-start":null}' $MULTICLOUD_PLUGIN_ENDPOINT/compute/v2.1/$PROJECT_ID/servers/0a06842a-4ec4-4918-b046-399f6b38f5f9/action

  2. vGMUX can be instantiated using the HEAT but VPP code build failed. Eric was able to manual build using the same script. He is looking into the problem.

10/17/2017

  1. APPC has made good progress to fix a few bugs. The execution has reached MultiCloud API. A call is set up on 10/18 to look into this problem with MultiCloud people.

10/16/2017

  1. DCAE to Policy: Vijay Venkatesh Kumar has fixed TCA policy to put in the correct closed loop name. However, multiple problems exist in the TCA output, tracked by  DCAEGEN2-164 - Getting issue details... STATUS .
    1. Fixed vCPE policies on the wiki to use VNF as the controlLoopSchemaType. 
    2. Fixed TCA jar to include "version" as is needed by Policy.
  2. Passed test: → DCAE → Policy → . Manually fed a VES data to DCAE, observed an event from Policy to APPC.
    1. Note that currently Policy does not check the time stamp in the event so for testing purpose there is no need to increase the time stamp in the VES data.
    2. Once Policy sends out a restart event to APPC, it will not send another restart event until the current operation is completed or time expires even if it receives subsequent ONSET events. To retest in case operation fails in APPC, Policy needs to be restarted.
  3. Eric Multanen has fixed the even data content problem to include the correct vnf_id passed from the HEAT. See details appended by Eric to to 10/13/2017 notes.

10/13/2017

  1. DCAE to Policy: Vijay Venkatesh Kumar updated TCA and VES data is now processed by collector and TCA and output from TCA is observed on unauthenticated.DCAE_CL_OUTPUT. Policy does observe that event from TCA. But the closed-loop name is incorrect. Vijay Venkatesh Kumar will load the correct TCA Policy listed on Policy R1 Amsterdam Functional Test Cases and the test will continue next week.
  2. APPC to MultiCloud: Problems from the previous days have been fixed. Scott Seabolt tested through Swagger UI and identified a new problem caused by prefix spaces in DG. He will fix it soon.
  3. vGMUX: Eric Multanen has identified how to configure VES collector: one has to delete the current VES collector setting before setting a new one. This applies to both command line and API configuration. A question raised is that the current VES event uses VM name and VM ID. What we use in AAI is actually the VNF ID. Need to figure out how to set that VNF ID in vGMUX.
    1. From Eric: The "sourceName" and "sourceId" in the event data are generated by 'libevel.so' which comes from source code in demo/vnfs/VES5.0/evel/evel-library/code/evel_library.  These elements are currently populated by openstack metadata corresponding to the vm name and vm id.  The vG-MUX server has a property vnf_id='vCPE_Infrastructure_vGMUX_demo_app' - so it is presumably feasible to modify the libevel.so code to use this value (the same one for both 'sourceName' and 'sourceId' ?).  The question is: should this be a quick and dirty hack for the vG-MUX VNF specifically?  or, a more general fix in the library code?  I'd be worried about changing the library code in the 'demo' repository and breaking some other use case or demo.
    2. From Kang
      1. A quick hack should be good enough. In this use case, we only need vGMUX to report events so the other VNFs do not need a similar hack.
      2. Let's give vnf_id to both sourceName and sourceId. In principle only sourceId should be used by DCAE. But my test shows that only sourceName is used instead. Since vnf_id is specified in the HEAT and there is no sourceName at all in the HEAT, the above approach should work well to avoid potential inconsistency.
      3. Update based on communication with John Choma: VNF UUID is the same as VNF ID used in the HEAT.
    3. From Eric:

      Below is output of the updated event data - check the bold items to verify this is what is desired (note:  "vCPE_Infrastructure_vGMUX_demo_app" is the value of the "vnf_id" property from the OpenStack metadata of the vG-MUX VM):

      {"event": {"commonEventHeader": {"domain": "measurementsForVfScaling", "eventId": "Generic_traffic", "eventName": "Measurement_vGMUX", "lastEpochMicrosec": 1508188493486856, "priority": "Normal", "reportingEntityName": "zdcpe1cpe01mux01", "sequence": 55, "sourceName": "vCPE_Infrastructure_vGMUX_demo_app", "startEpochMicrosec": 1508188483486856, "version": 1.2, "eventType": "HTTP request rate", "reportingEntityId": "No UUID available", "sourceId": "vCPE_Infrastructure_vGMUX_demo_app"}, "measurementsForVfScalingFields": {"measurementInterval": 10, "cpuUsageArray": [{"cpuIdentifier": "cpu1", "cpuIdle": 100.000000, "cpuUsageSystem": 0.000000, "cpuUsageUser": 0.000000, "percentUsage": 0.000000}], "requestRate": 9956, "vNicUsageArray": [{"receivedOctetsDelta": 0.000000, "receivedTotalPacketsDelta": 0.000000, "transmittedOctetsDelta": 0.000000, "transmittedTotalPacketsDelta": 0.000000, "valuesAreSuspect": "true", "vNicIdentifier": "eth0"}], "additionalMeasurements": [{"name": "ONAP-DCAE", "arrayOfFields": [{"name": "Packet-Loss-Rate", "value": "0.0"}]}], "measurementsForVfScalingVersion": 2.1}}}

  4. Meeting recordings: 
    1. GMT20171013-135015_Kang-Xi-s-_1920x1200.mp4
    2. GMT20171013-191357_Kang-Xi-s-_1920x1200.mp4

10/12/2017

  1. AAI has a problem to pre-load data using this script closed-loop-biny993.zip, tracked by  AAI-433 - Getting issue details... STATUS .
  2. Ethan Lynn has created a doc to show to register VMWare Openstack Fake Cloud into AAI and how to use it: https://gerrit.onap.org/r/#/c/18493/1/docs/Multicloud-Fake_Cloud-Guide.rst
  3. ESR installation guide: http://onap.readthedocs.io/en/latest/submodules/aai/esr-server.git/docs/platform/installation.html
  4. DCAE TCA is tracked by  DCAEGEN2-158 - Getting issue details... STATUS .
  5. Eric Multanen continues to fix vGMUX and made good progress. vGMUX in the ONAP-vCPE work space is now reporting VES data. Some fixes are needed. This is tracked by  INT-275 - Getting issue details... STATUS .

10/11/2017

  1. Test plan for CLAMP (Ron Shacham, Gervais-Martial Ngueko, Xue Gao, Lusheng Ji, Vijay Venkatesh Kumar):
    1. Integration team will install ONAP in a separate work space and let CLAMP do the testing. This is tracked in  INT-271 - Getting issue details... STATUS .
    2. CLAMP will start by test with Policy and SDC on the design part. The goal is to configure Policy and create Blueprint template for DCAE.
    3. Currently DCAEGEN2 cannot be installed in the open lab due to the lack of Openstack Designate support. This is expected to be fixed in a few days. In the meanwhile, CLAMP will do pairwise test with DCAE in the AT&T internal environment in parallel.
  2. MultiCloud (Bin YangEthan Lynn)
    1. MultiCloud sending query to AAI to obtain cloud info passed.
    2. MultiCloud performing VM restart (two commands: stop followed by start) passed. This was done by manually posting to the MultiCloud broker API.
    3. A VMWare VIM is simulated and passed VM restart test.
    4. Normally VIM is registered to ESR, and ESR will populate AAI with the cloud info. In this test, Bin Yang used a script to preload AAI with such info. MultiCloud queries AAI using VIM-ID to get the cloud info to do actual cloud operations.
  3. APPC→MultiCloud to restart VNF (Scott Seabolt, Ryan Young, Randa Maher)
    1. A bug is discovered inside APPC, tracked by APPC-268 - Getting issue details... STATUS .
  4. DCAE (Vijay Venkatesh Kumar)
    1. With the sampe VES provided by Eric Multanen on 10/10, the team has updated the TCA policy accordingly and will load it to TCA later today. It is tracked by  DCAEGEN2-135 - Getting issue details... STATUS .
    2. Note that in the final release, this TCA policy will be designed either using CLAMP or Policy GUI and distributed. No manual loading is needed.
  5. vGMUX (Eric Multanen)
    1. Eric is debugging vGMUX in his environment. Once it is done, he will see if it is feasible to create a VM image and use the image in the open lab instead.
  6. VxLAN config using SDNC
    1. We suggest that SDNC generates the VxLAN config request and send them to Eric Multanen to verify the correctness. It will help the upcoming pairwise test once the VNFs are ready in the open lab. This is tracked by  SDNC-120 - Getting issue details... STATUS .
  7. Customer flow testing in SO
    1. Update from Saryu Shah and BORALE, SHAILENDRA S <sb8915@att.com>: 

      The custom flow testing was performed using Junit testing.

      “mvn install” in so/bpmn/MSOInfrastructureBPMN will runs all the tests.

      Test is self-contained and all the required data files and simulator code was checked in. As of now, Jim is in the process of checking in bug-fixes.

  8. Meeting recording: GMT20171011-140734_Kang-Xi-s-_1920x1200.mp4

10/10/2017

  1. APPC: Scott Seabolt used Swagger API to invoke VNF restart. APPC passed AAI named query. The DG in APPC failed to query AAI because it did not point to the correct AAI. This is tracked by  APPC-267 - Getting issue details... STATUS .
  2. vGMUX: Eric has created a sample VES data as shown below. DCAE needs to be adjusted to process the data. This is tracked by DCAEGEN2-147 - Getting issue details... STATUS . Eric is also working to fix vGMUX.

    vCPE VES data
    {"event": {"commonEventHeader": {"domain": "measurementsForVfScaling", "eventId": "Generic_traffic", "eventName": "Measurement_vGMUX", "lastEpochMicrosec": 1507676920903343, "priority": "Normal", "reportingEntityName": "vg-mux", "sequence": 1, "sourceName": "Dummy VM name - No Metadata available", "startEpochMicrosec": 1507676910903343, "version": 1.2, "eventType": "HTTP request rate", "reportingEntityId": "No UUID available", "sourceId": "Dummy VM UUID - No Metadata available"}, "measurementsForVfScalingFields": {"measurementInterval": 10, "cpuUsageArray": [{"cpuIdentifier": "cpu1", "cpuIdle": 85.700000, "cpuUsageSystem": 14.300000, "cpuUsageUser": 0.000000, "percentUsage": 0.000000}], "requestRate": 9383, "vNicUsageArray": [{"receivedOctetsDelta": 0.000000, "receivedTotalPacketsDelta": 0.000000, "transmittedOctetsDelta": 0.000000, "transmittedTotalPacketsDelta": 0.000000, "valuesAreSuspect": "true", "vNicIdentifier": "eth0"}], "additionalMeasurements": [{"name": "ONAP-DCAE", "arrayOfFields": [{"name": "Packet-Loss-Rate", "value": "49.0"}]}], "measurementsForVfScalingVersion": 2.1}}}
  3. Meeting recording: GMT20171010-181134_Kang-Xi-s-_1920x1200.mp4

10/9/2017

Closed loop control test is decomposed into the following items.

  1. Design and load policy rules to policy engine. Jorge Hernandez created a closed loop for vCPE. Policy is tested by posting a DCAE TCA event on DMaaP. Policy reacted by sending an event to APPC to trigger VNF restart. The message format needs to be fixed, see POLICY-300 - Getting issue details... STATUS . And the topics need to be added to APPC properties, see  APPC-265 - Getting issue details... STATUS .

  2. Scripts are created to preload AAI. Scott Seabolt tested them in the integration environment. Both preloading and named query works.
    closed-loop.zip contains the necessary data (json) as well as shell scripts to load the Integration vGMUX.  Note the attached files load data specific the the "vCPE_Infrastructure_vGMUX_demo_app" VNF.  If you'd like to load a different VNF each of the json files owuld need to be modified to reflect the new VNF.
    1. put_closed_loop.sh

      • usage: put_closed_loop.sh A&AI-IP
    2. verify.sh
      • usage: verify.sh A&AI-IP
  3. VES collector is working. Waiting for sample VES data from Eric Multanen. See  DCAEGEN2-135 - Getting issue details... STATUS . Once the data is available, the TCA policy on Policy R1 Amsterdam Functional Test Cases may need to be adjusted and then TCA testing can be performed.
  4. vGMUX cannot be properly instantiated. Eric MultanenMarco Platania, and Kang Xi are working to solve it.
  5. Docker info shows that the multi-cloud services are running inside the open-o VM as multiple containers
    1. The VM is vm1-openo-server with IP 10.12.25.113
    2. Service ports are listed below. More details of each service are described on Setup MultiCloud Development Env 
      1. broker: 9001
      2. VIO plugin: 9004
      3. Ocata plugin: 9006
      4. WindRiver plugin: 9005


Meeting recording is GMT20171009-170725_Kang-Xi-s-_1920x1200.mp4

10/6/2017

Closed loop control test is decomposed into the following items.

  1. Design and load policy rules to policy engine. Ideally this should be done by CLAMP. Currently CLAMP is working on CLAMP-59 - Getting issue details... STATUS . The workaround is to use Policy GUI. Jorge Hernandez will work on this.

  2. Preload AAI with vCPE data. The purpose is to lookup vServer using vGMUX VNF ID. Scott Seabolt (APPC) will provide a sample data to Venkata Harish Kajur (AAI) to show the named query. Venkata Harish Kajur will then create a script to preload such data to AAI for the testing.

  3. vGMUX sends packet loss VES to DCAE VES collector.

    • Kang Xi created a vGMUX instance in the open lab but the VPP-based VNF inside the VM is not functioning. According to Eric Multanen, most likely it is caused by network issues. Eric is working on this. Kang will also inform Danny Zhou and Johnson Li about this.

    • The following VES data caused an error. The most possible reason is that the data itself is incorrect. Vijay Venkatesh Kumar is looking into that.

      VES Data
      curl -H "Accept: application/json" -H "Content-Type: application/json" -d '{"event":{"measurementsForVfScalingFields":{"measurementInterval":10,"measurementsForVfScalingVersion":1.1,"vNicUsageArray":[{"multicastPacketsIn":0,"bytesIn":4300,"unicastPacketsIn":0,"multicastPacketsOut":0,"broadcastPacketsOut":0,"packetsOut":0,"bytesOut":0,"broadcastPacketsIn":0,"packetsIn":101,"unicastPacketsOut":0,"vNicIdentifier":"eth1"}]},"commonEventHeader":{"reportingEntityName":"zdfw1fwl01fwl01","startEpochMicrosec":1500379662497999,"lastEpochMicrosec":1500379672497999,"eventId":"1","sourceName":"zdfw1fwl01fwl01","sequence":1,"priority":"Normal","functionalRole":"vFirewall","domain":"measurementsForVfScaling","reportingEntityId":"No UUID available","sourceId":"75ec15e4-1a9a-4ee3-bb3c-31556903558d","version":1.2}}}' 10.12.25.84:8080/eventListener/v5
      
      
  4. DCAE TCA sends an event to Policy through DMaaP. This is tested as described in  DCAEGEN2-124 - Getting issue details... STATUS

  5. Policy process the event from DCAE, and then sends a “restart” event to APPC through DMaaP. Jorge Hernandez will work on this.

  6. APPC captures the event from Policy, performs a named query to AAI to get the vServer ID based on the VNF ID, and then restarts the VM through MultiVIM. 

    • APPC is working on some changes and will be able to test next week. 
    • There should be a MultiVIM VM in the stack but we don't see any. Refer INT-258 - Getting issue details... STATUS  to track this problem.

In addition, the bottom of Policy R1 Amsterdam Functional Test Cases shows important sample data for vCPE test cases.

Update:

  • Scott Seabolt created the data pertaining to the vGMUX in POD25 which needs to be inventoried in POD25 A&AI. Venkata Harish Kajur then created a script to preload AAI with the following instructions: "Unzip this closed-loop.zip and go into closed-loop directory and run the ./put_closed_loop.sh ${AAI_VM1_IP_ADDRESS} And then you can verify that it works by doing a ./verify.sh ${AAI_VM1_IP_ADDRESS} and it should return you that named query response that you sent me."
  • Meeting recording

Related Documents:

  1. VNF Doc: ONAP vCPE VPP-based VNF Installation and Usage Information
  2. DCAE mS Deployment (Standalone instantiation)
  3. SNIRO Emulator


export MULTICLOUD_PLUGIN_ENDPOINT=http://10.0.14.1:9005/api/multicloud-titanium_cloud/v0/pod25_RegionOne

 

export TOKEN=$(curl -v -s -H "Content-Type: application/json" -X POST -d '{ }'  $MULTICLOUD_PLUGIN_ENDPOINT/identity/v3/auth/tokens 2>&1 | grep X-Subject-Token | sed "s/^.*: //")

 

export PROJECT_ID=466979b815b5415ba14ada713e6e1846

 

curl -v -s  -H "Content-Type: application/json" -H "X-Auth-Token: $TOKEN" -X GET $MULTICLOUD_PLUGIN_ENDPOINT/compute/v2.1/$PROJECT_ID/servers

 

curl -v -s -H "Content-Type: application/json" -H "X-Auth-Token: $TOKEN" -X POST -d '{"os-stop":null}'  $MULTICLOUD_PLUGIN_ENDPOINT/compute/v2.1/$PROJECT_ID/servers/0a06842a-4ec4-4918-b046-399f6b38f5f9/action

 

curl -v -s -H "Content-Type: application/json" -H "X-Auth-Token: $TOKEN" -X POST -d '{"os-start":null}'  $MULTICLOUD_PLUGIN_ENDPOINT/compute/v2.1/$PROJECT_ID/servers/0a06842a-4ec4-4918-b046-399f6b38f5f9/action

  • No labels