docs/doc/source/security/kubernetes/security-configure-container-backed-remote-clis-and-clients.rst
Keane Lim 3c5fa979a4 Security guide update
Re-organized topic hierarchy

Tiny edit to restart review workflow.

Squashed with Resolved index.rst conflict commit

Change-Id: I13472792cb19d1e9975ac76c6954d38054d606c5
Signed-off-by: Keane Lim <keane.lim@windriver.com>
Signed-off-by: MCamp859 <maryx.camp@intel.com>
2021-03-12 15:10:40 -05:00

13 KiB

Configure Container-backed Remote CLIs and Clients

The command line can be accessed from remote computers running Linux, OSX, and Windows.

This functionality is made available using a docker image for connecting to the remotely. This docker image is pulled as required by configuration scripts.

You must have Docker installed on the remote systems you connect from. For more information on installing Docker, see https://docs.docker.com/install/. For Windows remote clients, Docker is only supported on Windows 10.

For Windows remote clients, you must run the following commands from a Cygwin terminal. See https://www.cygwin.com/ for more information about the Cygwin project.

For Windows remote clients, you must also have winpty installed. Download the latest release tarball for Cygwin from https://github.com/rprichard/winpty/releases. After downloading the tarball, extract it to any location and change the Windows <PATH> variable to include its bin folder from the extracted winpty folder.

The following procedure shows how to configure the Container-backed Remote CLIs and Clients for an admin user with cluster-admin clusterrole. If using a non-admin user such as one with role privileges only within a private namespace, additional configuration is required in order to use helm.

  1. On the Controller, configure a Kubernetes service account for user on the remote client.

    You must configure a Kubernetes service account on the target system and generate a configuration file based on that service account.

    Run the following commands logged in as root on the local console of the controller.

    1. Set environment variables.

      You can customize the service account name and the output configuration file by changing the <USER> and <OUTPUT_FILE> variables shown in the following examples.

      % USER="admin-user"
      % OUTPUT_FILE="temp-kubeconfig"
    2. Create an account definition file.

      % cat <<EOF > admin-login.yaml
      apiVersion: v1
      kind: ServiceAccount
      metadata:
        name: ${USER}
        namespace: kube-system
      ---
      apiVersion: rbac.authorization.k8s.io/v1
      kind: ClusterRoleBinding
      metadata:
        name: admin-user
      roleRef:
        apiGroup: rbac.authorization.k8s.io
        kind: ClusterRole
        name: cluster-admin
      subjects:
      - kind: ServiceAccount
        name: ${USER}
        namespace: kube-system
      EOF
    3. Apply the definition.

      % kubectl apply -f admin-login.yaml
    4. Store the token value.

      % TOKEN_DATA=$(kubectl -n kube-system describe secret $(kubectl -n kube-system get secret | grep ${USER} | awk '{print $1}') | grep "token:" | awk '{print $2}')
    5. Source the platform environment.

      % source /etc/platform/openrc
    6. Store the OAM IP address.

      1. % OAM_IP=$(system oam-show awk '{print $4}')

      2. AIO-SX uses <oam_ip> instead of <oam_floating_ip>. The following shell code ensures that <OAM_IP> is assigned the correct IP address.

        % if [ -z "$OAM_IP" ]; then
            OAM_IP=$(system oam-show |grep oam_ip| awk '{print $4}')
        fi
      3. IPv6 addresses must be enclosed in square brackets. The following shell code does this for you.

        % if [[ $OAM_IP =~ .*:.* ]]; then
            OAM_IP="[${OAM_IP}]"
        fi
    7. Generate the temp-kubeconfig file.

      % sudo kubectl config --kubeconfig ${OUTPUT_FILE} set-cluster wrcp-cluster --server=https://${OAM_IP}:6443 --insecure-skip-tls-verify
      % sudo kubectl config --kubeconfig ${OUTPUT_FILE} set-credentials ${USER} --token=$TOKEN_DATA
      % sudo kubectl config --kubeconfig ${OUTPUT_FILE} set-context ${USER}@wrcp-cluster --cluster=wrcp-cluster --user ${USER} --namespace=default
      % sudo kubectl config --kubeconfig ${OUTPUT_FILE} use-context ${USER}@wrcp-cluster
  2. On the remote client, install the remote client tarball file from the StarlingX CENGEN build servers..

    • The tarball is available from the area on the StarlingX CENGEN build servers.
    • You can extract the tarball's contents anywhere on your client system.

    While it is not strictly required to extract the tarball before other steps, subsequent steps in this example copy files to the extracted directories as a convenience when running configuration scripts.

  3. Download the openrc file from the Horizon Web interface to the remote client.

    1. Log in to Horizon as the user and tenant that you want to configure remote access for.
    2. Navigate to Project > API Access > Download Openstack RC file.
    3. Select Openstack RC file.
  4. Copy the temp-kubeconfig file to the remote workstation.

    You can copy the file to any location on the remote workstation. For convenience, this example assumes that it is copied to the location of the extracted tarball.

    Note

    Ensure the temp-kubeconfig file has 666 permissions after copying the file to the remote workstation, otherwise, use the following command to change permissions, chmod 666 temp\_kubeconfig.

  5. On the remote client, configure the client access.

    1. Change to the location of the extracted tarball.

    2. Run the configure\_client.sh script to generate a client access file for the platform.

      Note

      For remote CLI commands that require local filesystem access, you can specify a working directory when running configure\_client.sh using the -w option. If no directory is specified, the location from which the configure\_client.sh was run is used for local file access by remote cli commands. This working directory is mounted at /wd in the docker container. If remote CLI commands need access to local files, copy the files to your configured work directory and then provide the command with the path to the mounted folder inside the container.

      $ mkdir -p $HOME/remote_wd
      $ ./configure_client.sh -t platform -r admin_openrc.sh -k temp-kubeconfig \
      -w HOME/remote_wd
      $ cd $HOME/remote_wd

      By default, configure_client.sh will use container images from docker.io for both platform clients and openstack clients. You can optionally use the -p and -a options to override the default docker with one or two others from a custom registry. For example, to use the container images from the WRS AWS ECR

      $ ./configure_client.sh -t platform -r admin_openrc.sh -k
      temp-kubeconfig -w HOME/remote_wd -p
      625619392498.dkr.ecr.us-west-2.amazonaws.com/starlingx/stx-platfo
      rmclients:stx.4.0-v1.3.0  -a
      625619392498.dkr.ecr.us-west-2.amazonaws.com/starlingx/stx-openst
      ackclients:master-centos-stable-latest

      If you are working with repositories that require authentication, you must first perform a docker login to that repository before using remote clients.

      A remote_client_platform.sh file is generated for use accessing the platform CLI.

  6. Test workstation access to the remote platform CLI.

    Enter your platform password when prompted.

    Note

    The first usage of a command will be slow as it requires that the docker image supporting the remote clients be pulled from the remote registry.

    root@myclient:/home/user/remote_wd# source remote_client_platform.sh
    Please enter your OpenStack Password for project admin as user admin:
    root@myclient:/home/user/remote_wd# system host-list
    +----+--------------+-------------+----------------+-------------+--------------+
    | id | hostname     | personality | administrative | operational | availability |
    +----+--------------+-------------+----------------+-------------+--------------+
    | 1  | controller-0 | controller  | unlocked       | enabled     | available    |
    | 2  | controller-1 | controller  | unlocked       | enabled     | available    |
    | 3  | compute-0    | worker      | unlocked       | enabled     | available    |
    | 4  | compute-1    | worker      | unlocked       | enabled     | available    |
    +----+--------------+-------------+----------------+-------------+--------------+
    root@myclient:/home/user/remote_wd# kubectl -n kube-system get pods
    NAME                                       READY   STATUS      RESTARTS   AGE
    calico-kube-controllers-767467f9cf-wtvmr   1/1     Running     1          3d2h
    calico-node-j544l                          1/1     Running     1          3d
    calico-node-ngmxt                          1/1     Running     1          3d1h
    calico-node-qtc99                          1/1     Running     1          3d
    calico-node-x7btl                          1/1     Running     4          3d2h
    ceph-pools-audit-1569848400-rrpjq          0/1     Completed   0          12m
    ceph-pools-audit-1569848700-jhv5n          0/1     Completed   0          7m26s
    ceph-pools-audit-1569849000-cb988          0/1     Completed   0          2m25s
    coredns-7cf476b5c8-5x724                   1/1     Running     1          3d2h
    ...
    root@myclient:/home/user/remote_wd#

    Note

    Some commands used by remote CLI are designed to leave you in a shell prompt, for example:

    root@myclient:/home/user/remote_wd# openstack

    or

    root@myclient:/home/user/remote_wd# kubectl exec -ti <pod_name> -- /bin/bash

    In some cases the mechanism for identifying commands that should leave you at a shell prompt does not identify them correctly. If you encounter such scenarios, you can force-enable or disable the shell options using the <FORCE_SHELL> or <FORCE_NO_SHELL> variables before the command.

    For example:

    root@myclient:/home/user/remote_wd# FORCE_SHELL=true kubectl exec -ti <pod_name> -- /bin/bash
    root@myclient:/home/user/remote_wd# FORCE_NO_SHELL=true kubectl exec <pod_name> -- ls

    You cannot use both variables at the same time.

  7. If you need to run a remote CLI command that references a local file, then that file must be copied to or created in the working directory specified on the ./config_client.sh command and referenced as under /wd/ in the actual command.

    For example:

    root@myclient:/home/user# cd $HOME/remote_wd
    root@myclient:/home/user/remote_wd# kubectl -n kube-system  create -f test/test.yml
    pod/test-pod created
    root@myclient:/home/user/remote_wd# kubectl -n kube-system  delete -f test/test.yml
    pod/test-pod deleted
  8. Do the following to use helm.

    Note

    For non-admin users, additional configuration is required first as discussed in Configure Remote Helm Client for Non-Admin Users.

    Note

    When using helm, any command that requires access to a helm repository (managed locally) will require that you be in the $HOME/remote_wd directory and use the --home ./helm option.

    1. Do the initial setup of the helm client.

      % helm init --client-only --home "./.helm"
    2. Run a helm command.

      $ helm list
      $ helm install --name wordpress stable/wordpress  --home "./helm"

Install Kubectl and Helm Clients Directly on a Host <security-install-kubectl-and-helm-clients-directly-on-a-host>

Configure Remote Helm Client for Non-Admin Users <configure-remote-helm-client-for-non-admin-users>