Skip to end of metadata
Go to start of metadata

You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 17 Next »

ETSI-DMS on O-Cloud, Tacker installation 


This document describes how to deploy an openstack-helm environment based on the INF with Ceph.
  1. Preparation

sysadmin@controller-1:~$ source /etc/platform/openrc
[sysadmin@controller-1 ~(keystone_admin)]$ system storage-backend-list
+--------------------------------------+------------+---------+------------+-------------------+----------+---------------+
| uuid                                 | name       | backend | state      | task              | services | capabilities  |
+--------------------------------------+------------+---------+------------+-------------------+----------+---------------+
| da70e0b7-34c8-488a-9e1f-08e057d6a4be | ceph-store | ceph    | configured | provision-storage | None     | replication:  |
|                                      |            |         |            |                   |          | 2 min_replica |
|                                      |            |         |            |                   |          | tion: 1       |
|                                      |            |         |            |                   |          |               |
+--------------------------------------+------------+---------+------------+-------------------+----------+---------------+


[sysadmin@controller-1 ~(keystone_admin)]$ ceph -s
  cluster:
    id:     c5663990-249a-4b71-988f-19b402784429
    health: HEALTH_OK

  services:
    mon: 1 daemons, quorum controller (age 8h)
    mgr: controller-1(active, since 8h), standbys: controller-0
    mds: kube-cephfs:1 {0=controller-1=up:active} 1 up:standby
    osd: 2 osds: 2 up (since 8h), 2 in (since 8h)

  data:
    pools:   3 pools, 192 pgs
    objects: 181 objects, 258 MiB
    usage:   2.4 GiB used, 269 GiB / 271 GiB avail
    pgs:     192 active+clean



    • Prepare a Linux environment in which the network is reachable to the INF AIO-SX with Bash CLI. And, install the required packages.


      $ sudo apt-get install git make patch jq
      # Make sure your CLI has the 'kubectl', 'openstack' and 'helm' installed

    • Clone openstack-helm and openstack-helm-infra code on your Linux.


      # The commit id is 82a6aa8ce96b1669af0b9e8da85b537d02fc5fd3 that used in this demo.
      $ git clone --depth 1 --branch master https://opendev.org/openstack/openstack-helm.git
      # The commit id is 07c735f632147378c4af8e7b4ce6f390d38e3d69 that used in this demo.

      $ git clone --depth 1 --branch master https://opendev.org/openstack/openstack-helm-infra.git

    • Copy the "/etc/kubernetes/admin.conf" from the INF controller node to your local Linux.

      $ scp <INF-controller-0>:/etc/kubernetes/admin.conf ~/.kube/config
      # Change the IP address in the ~/.kube/config
      # server: https://<INF-OAM-IP>:6443
      # You can get the OAM IP through this command on controller node
      # system addrpool-show `system addrpool-list | grep oam | awk '{print $2}'` | grep floating
    • Add labels to controller-0 node.

      $ kubectl label node controller-0 openstack-control-plane=enabled
    • Create namespaces.

      $ kubectl create namespace openstack
  1. Deploy Ingress Controller Original procedure: https://docs.openstack.org/openstack-helm/latest/install/developer/kubernetes-and-common-setup.html#deploy-the-ingress-controller

    • Modify openstack-helm/tools/deployment/component/common/ingress.sh file as follows:

      • diff --git a/tools/deployment/component/common/ingress.sh b/tools/deployment/component/common/ingress.sh
        index 9ae0371..3229dcb 100755
        --- a/tools/deployment/component/common/ingress.sh
        +++ b/tools/deployment/component/common/ingress.sh
        @@ -29,6 +29,23 @@ deployment:
           type: DaemonSet
         network:
           host_namespace: true
        +endpoints:
        +  ingress:
        +    port:
        +      http:
        +        default: 10080
        +      https:
        +        default: 10443
        +      healthz:
        +        default: 11254
        +      status:
        +        default: 11246
        +      stream:
        +        default: 11247
        +      profiler:
        +        default: 11245
        +      server:
        +        default: 18181
         EOF

         touch /tmp/ingress-component.yaml
        @@ -48,21 +65,21 @@ pod:
         EOF
         fi

        -helm upgrade --install ingress-kube-system ${HELM_CHART_ROOT_PATH}/ingress \
        -  --namespace=kube-system \
        -  --values=/tmp/ingress-kube-system.yaml \
        -  ${OSH_EXTRA_HELM_ARGS} \
        -  ${OSH_EXTRA_HELM_ARGS_INGRESS} \
        -  ${OSH_EXTRA_HELM_ARGS_INGRESS_KUBE_SYSTEM}
        -
        -#NOTE: Wait for deploy
        -./tools/deployment/common/wait-for-pods.sh kube-system
        +#helm upgrade --install ingress-kube-system ${HELM_CHART_ROOT_PATH}/ingress \
        +#  --namespace=kube-system \
        +#  --values=/tmp/ingress-kube-system.yaml \
        +#  ${OSH_EXTRA_HELM_ARGS} \
        +#  ${OSH_EXTRA_HELM_ARGS_INGRESS} \
        +#  ${OSH_EXTRA_HELM_ARGS_INGRESS_KUBE_SYSTEM}
        +#
        +##NOTE: Wait for deploy
        +#./tools/deployment/common/wait-for-pods.sh kube-system

         #NOTE: Deploy namespace ingress
         helm upgrade --install ingress-openstack ${HELM_CHART_ROOT_PATH}/ingress \
           --namespace=openstack \
           --values=/tmp/ingress-component.yaml \
        -  --set deployment.cluster.class=nginx \
        +  --set deployment.cluster.class=nginx-openstack \
           ${OSH_EXTRA_HELM_ARGS} \
           ${OSH_EXTRA_HELM_ARGS_INGRESS} \
           ${OSH_EXTRA_HELM_ARGS_INGRESS_OPENSTACK}
        @@ -70,13 +87,13 @@ helm upgrade --install ingress-openstack ${HELM_CHART_ROOT_PATH}/ingress \
         #NOTE: Wait for deploy
         ./tools/deployment/common/wait-for-pods.sh openstack

        -helm upgrade --install ingress-ceph ${HELM_CHART_ROOT_PATH}/ingress \
        -  --namespace=ceph \
        -  --values=/tmp/ingress-component.yaml \
        -  --set deployment.cluster.class=nginx-ceph \
        -  ${OSH_EXTRA_HELM_ARGS} \
        -  ${OSH_EXTRA_HELM_ARGS_INGRESS} \
        -  ${OSH_EXTRA_HELM_ARGS_INGRESS_CEPH}
        -
        -#NOTE: Wait for deploy
        -./tools/deployment/common/wait-for-pods.sh ceph
        +#helm upgrade --install ingress-ceph ${HELM_CHART_ROOT_PATH}/ingress \
        +#  --namespace=ceph \
        +#  --values=/tmp/ingress-component.yaml \
        +#  --set deployment.cluster.class=nginx-ceph \
        +#  ${OSH_EXTRA_HELM_ARGS} \
        +#  ${OSH_EXTRA_HELM_ARGS_INGRESS} \
        +#  ${OSH_EXTRA_HELM_ARGS_INGRESS_CEPH}
        +#
        +##NOTE: Wait for deploy
        +#./tools/deployment/common/wait-for-pods.sh ceph

    • Execute ingress.sh.

      $ cd $HOME/openstack-helm/
      $ ./tools/deployment/component/common/ingress.sh
  1. Deploy Other Component Original Procedure: https://docs.openstack.org/openstack-helm/latest/install/developer/deploy-with-ceph.html#deploy-mariadb ~ Install the following components to use Tacker: * MariaDB * RabbitMQ * Memcached * Keystone * Glance

    • Modify openstack-helm/tools/deployment/component/glance/glance.sh file as follows:

      diff --git a/tools/deployment/component/glance/glance.sh b/tools/deployment/component/glance/glance.sh
      index b388ec04..4d50c2c5 100755
      --- a/tools/deployment/component/glance/glance.sh
      +++ b/tools/deployment/component/glance/glance.sh
      @@ -27,7 +27,7 @@ make glance
       tee /tmp/glance.yaml <<EOF
       storage: ${GLANCE_BACKEND}
       volume:
      -  class_name: standard
      +  class_name: general
       bootstrap:
         structured:
           images:
    • Execute script files.

      $ ./tools/deployment/developer/ceph/050-mariadb.sh
      $ ./tools/deployment/developer/ceph/060-rabbitmq.sh
      $ ./tools/deployment/developer/ceph/070-memcached.sh
      $ ./tools/deployment/developer/ceph/080-keystone.sh
      $ ./tools/deployment/component/glance/glance.sh
  1. Deploy Barbican and Tacker

    • Modify openstack-helm/tacker/templates/pvc.yaml file as follows:

      diff --git a/tacker/templates/pvc.yaml b/tacker/templates/pvc.yaml
      index 8b1678b3..c0599b45 100644
      --- a/tacker/templates/pvc.yaml
      +++ b/tacker/templates/pvc.yaml
      @@ -23,7 +23,7 @@ metadata:
        name: {{ $name }}
      spec:
        accessModes:
      -   - "ReadWriteMany"
      +   - "ReadWriteOnce"
        resources:
          requests:
            storage: {{ $size }}
    • Modify openstack-helm/tacker/values.yaml file as follows:

      diff --git a/tacker/values.yaml b/tacker/values.yaml
      index 90702f95..3d2f2621 100644
      --- a/tacker/values.yaml
      +++ b/tacker/values.yaml
      @@ -105,12 +105,12 @@ pod:
         security_context:
           server:
             pod:
      -        runAsUser: 42424
      -        runAsNonRoot: true
      +        runAsUser: 0
      +        runAsNonRoot: false
           conductor:
             pod:
      -        runAsUser: 42424
      -        runAsNonRoot: true
      +        runAsUser: 0
      +        runAsNonRoot: false
         lifecycle:
           termination_grace_period:
             server:
    • Execute script files.

      $ ./tools/deployment/developer/common/085-barbican.sh
      $ ./tools/deployment/component/tacker/tacker.sh
  2. Verify successful deployment

    • The helm releases are deployed as follows:

      sysadmin@controller-0:~$ helm list -n openstack
      NAME                   NAMESPACE       REVISION       UPDATED                                 STATUS         CHART                           APP VERSION
      barbican               openstack       1               2023-10-04 14:11:54.122228604 +0000 UTC deployed       barbican-0.3.5                 v1.0.0    
      glance                 openstack       1               2023-10-05 01:14:44.18606719 +0000 UTC deployed       glance-0.4.13                   v1.0.0    
      ingress-openstack       openstack       1               2023-10-02 07:59:30.823441021 +0000 UTC deployed       ingress-0.2.17                 v1.5.1    
      keystone               openstack       2               2023-10-04 13:58:36.81624535 +0000 UTC deployed       keystone-0.3.4                 v1.0.0    
      mariadb                 openstack       1               2023-10-04 13:36:33.178219784 +0000 UTC deployed       mariadb-0.2.33                 v10.6.7    
      memcached               openstack       1               2023-10-04 13:44:40.7788406 +0000 UTC   deployed       memcached-0.1.13               v1.5.5    
      rabbitmq               openstack       1               2023-10-04 13:39:44.683045128 +0000 UTC deployed       rabbitmq-0.1.29                 v3.9.0    
      tacker                 openstack       1               2023-10-05 10:03:19.033603307 +0000 UTC deployed       tacker-0.1.1                   v1.0.0
    • The pods are read as follows (Check all pods are "Completed" or "Running" status):

      sysadmin@controller-0:~/openstack-helm$ kubectl get pod -n openstack
      NAME                                                       READY   STATUS     RESTARTS     AGE
      barbican-api-75fd4d79d7-ncz2c                               1/1     Running     0             46h
      barbican-db-init-mvhs4                                     0/1     Completed   0             46h
      barbican-db-sync-2hn96                                     0/1     Completed   0             46h
      barbican-ks-endpoints-57rm2                                 0/3     Completed   0             46h
      barbican-ks-service-x2jqn                                   0/1     Completed   0             46h
      barbican-ks-user-ds9h6                                     0/1     Completed   0             46h
      barbican-rabbit-init-gz647                                 0/1     Completed   0             46h
      barbican-test                                               0/1     Completed   0             46h
      glance-api-97df56ddb-pr598                                 1/1     Running     0             35h
      glance-bootstrap-fbmpq                                     0/1     Completed   0             35h
      glance-db-init-gtmdc                                       0/1     Completed   0             35h
      glance-db-sync-9jkb8                                       0/1     Completed   0             35h
      glance-ks-endpoints-dkb6m                                   0/3     Completed   0             35h
      glance-ks-service-xdhfk                                     0/1     Completed   0             35h
      glance-ks-user-9xhvf                                       0/1     Completed   0             35h
      glance-metadefs-load-rw2kc                                 0/1     Completed   0             35h
      glance-rabbit-init-c4wvr                                   0/1     Completed   0             35h
      glance-storage-init-lzn72                                   0/1     Completed   0             35h
      ingress-5448bbd7d-7rz99                                     1/1     Running     1 (47h ago)   4d4h
      ingress-error-pages-54c8fdfb4d-wgktt                       1/1     Running     1 (47h ago)   4d4h
      keystone-api-6cb7d765ff-srpwg                               1/1     Running     0             46h
      keystone-bootstrap-f9s5n                                   0/1     Completed   0             46h
      keystone-credential-setup-27qkx                             0/1     Completed   0             46h
      keystone-db-init-sr9dj                                     0/1     Completed   0             46h
      keystone-db-sync-7hnj8                                     0/1     Completed   0             46h
      keystone-domain-manage-2n6sf                               0/1     Completed   0             46h
      keystone-fernet-rotate-28275120-djbg7                       0/1     Completed   0             24h
      keystone-fernet-rotate-28275840-z2wnq                       0/1     Completed   0             12h
      keystone-fernet-rotate-28276560-z6rmr                       0/1     Completed   0             30m
      keystone-fernet-setup-x8px7                                 0/1     Completed   0             46h
      keystone-rabbit-init-w5h9q                                 0/1     Completed   0             46h
      mariadb-ingress-7f9bcfd79b-6flfw                           1/1     Running     0             46h
      mariadb-ingress-7f9bcfd79b-tlwkc                           1/1     Running     0             46h
      mariadb-ingress-error-pages-557b55c45f-tw8sw               1/1     Running     0             46h
      mariadb-server-0                                           1/1     Running     0             46h
      memcached-memcached-785bbdd4d8-zxh76                       1/1     Running     0             46h
      rabbitmq-cluster-wait-49khp                                 0/1     Completed   0             46h
      rabbitmq-rabbitmq-0                                         1/1     Running     0             46h
      rabbitmq-rabbitmq-1                                         1/1     Running     0             46h
      tacker-conductor-9f977f5b4-tx58c                           1/1     Running     0             26h
      tacker-db-init-4d7xz                                       0/1     Completed   0             26h
      tacker-db-sync-vwzg2                                       0/1     Completed   0             26h
      tacker-ks-endpoints-426wd                                   0/3     Completed   0             26h
      tacker-ks-service-lltsv                                     0/1     Completed   0             26h
      tacker-ks-user-5vpws                                       0/1     Completed   0             26h
      tacker-rabbit-init-2jkgb                                   0/1     Completed   0             26h
      tacker-server-76d9bbf6c8-skk8h                             1/1     Running     0             26h
    • Test if Tacker is working properly

      $ TACKER_SERVER_POD=tacker-server-76d9bbf6c8-skk8h
      $ TACKER_ENDPOINT=tacker-api.openstack.svc.cluster.local

      # Issue token from keystone
      $ kubectl exec -n openstack -it $TACKER_SERVER_POD \
      -- curl -i -X POST -H "Content-Type: application/json" \
      -d '{"auth":{"identity":{"methods":["password"],"password":{"user":{"domain":{"name":"default"},"name":"admin","password":"password"}}},"scope":{"project":{"domain":{"name":"default"},"name":"admin"}}}}' \
      http://keystone-api.openstack.svc.cluster.local:5000/v3/auth/tokens

      HTTP/1.1 201 CREATED
      Date: Fri, 06 Oct 2023 12:46:40 GMT
      Content-Type: application/json
      Content-Length: 3175
      Connection: keep-alive
      X-Subject-Token: gAAAAABlIAGv1RqxqMJ7rt_VyAtPTxF0XjMG19zp-0zaZmHdFkKmEjLfUus09GkPUdcbCeVuR8ZfmMjqg9C2kRCWWX4Llfdwld1lKM-beqQ7s127kjhpilf28e1oXh351CmBFy97PaZ9D5WBoe3fRrDkhhB_cEsB76Pyj6P2KQuNeMIhGmb1fKA
      Vary: X-Auth-Token
      x-openstack-request-id: req-408ef1f6-2b61-4a8d-89b0-0d987878cbbb

      # Set `X-Subject-Token` retrieved as TOKEN
      $ TOKEN=gAAAAABlIAGv1RqxqMJ7rt_VyAtPTxF0XjMG19zp-0zaZmHdFkKmEjLfUus09GkPUdcbCeVuR8ZfmMjqg9C2kRCWWX4Llfdwld1lKM-beqQ7s127kjhpilf28e1oXh351CmBFy97PaZ9D5WBoe3fRrDkhhB_cEsB76Pyj6P2KQuNeMIhGmb1fKA

      $ kubectl exec -n openstack -it $TACKER_SERVER_POD \
      -- curl -X GET http://${TACKER_ENDPOINT}:9890/vnflcm/v2/vnf_instances \
      -H "X-Auth-Token:$TOKEN" -H "Version: 2.0.0"
      []   *** Success if you can get an empty list ***
  • No labels