
Per recent updates, updating use of 'compute' node to be 'worker' node. StarlingX uses worker nodes, and a 'compute' node is a specialization of a worker node (OpenStack compute label applied). - Update narrative text to use worker node - Update shell scripts to use NODE instead of COMPUTE Updated non-versioned content and R3 installation guides. Change-Id: Ia3c5d354468f18efb79c823e5bfddf17e34998a9 Signed-off-by: Kristal Dale <kristal.dale@intel.com>
78 lines
2.4 KiB
ReStructuredText
78 lines
2.4 KiB
ReStructuredText
================================================
|
|
Kubernetes Intel GPU Device Plugin Configuration
|
|
================================================
|
|
|
|
This document describes how to enable the Intel GPU device plugin in StarlingX
|
|
and schedule pods on nodes with an Intel GPU.
|
|
|
|
------------------------------
|
|
Enable Intel GPU device plugin
|
|
------------------------------
|
|
|
|
You can pre-install the ``intel-gpu-plugin`` daemonset as follows:
|
|
|
|
#. Launch the ``intel-gpu-plugin`` daemonset.
|
|
|
|
Add the following lines to the ``localhost.yaml`` file before playing the
|
|
Ansible bootstrap playbook to configure the system.
|
|
|
|
::
|
|
|
|
k8s_plugins:
|
|
intel-gpu-plugin: intelgpu=enabled
|
|
|
|
#. Assign the ``intelgpu`` label to each node that should have the Intel GPU
|
|
plugin enabled. This will make any GPU devices on a given node available for
|
|
scheduling to containers. The following example assigns the ``intelgpu``
|
|
label to the worker-0 node.
|
|
|
|
::
|
|
|
|
$ NODE=worker-0
|
|
$ system host-lock $NODE
|
|
$ system host-label-assign $NODE intelgpu=enabled
|
|
$ system host-unlock $NODE
|
|
|
|
#. After the node becomes available, verify the GPU device plugin is registered
|
|
and that the available GPU devices on the node have been discovered and reported.
|
|
|
|
::
|
|
|
|
$ kubectl describe node $NODE | grep gpu.intel.com
|
|
gpu.intel.com/i915: 1
|
|
gpu.intel.com/i915: 1
|
|
|
|
-------------------------------------
|
|
Schedule pods on nodes with Intel GPU
|
|
-------------------------------------
|
|
|
|
Add a ``resources.limits.gpu.intel.com`` to your container specification in order
|
|
to request an available GPU device for your container.
|
|
|
|
::
|
|
|
|
...
|
|
spec:
|
|
containers:
|
|
- name: ...
|
|
...
|
|
resources:
|
|
limits:
|
|
gpu.intel.com/i915: 1
|
|
|
|
|
|
The pods will be scheduled to the nodes with available Intel GPU devices. A GPU
|
|
device will be allocated to the container and the available GPU devices will be
|
|
updated.
|
|
|
|
::
|
|
|
|
$ kubectl describe node $NODE | grep gpu.intel.com
|
|
gpu.intel.com/i915: 1
|
|
gpu.intel.com/i915: 0
|
|
|
|
For more details, refer to the following examples:
|
|
|
|
* `Kubernetes manifest file example <https://github.com/intel/intel-device-plugins-for-kubernetes/blob/master/demo/intelgpu-job.yaml>`_
|
|
* `Scheduling pods on nodes with Intel GPU example <https://github.com/intel/intel-device-plugins-for-kubernetes/blob/master/cmd/gpu_plugin/README.md#test-gpu-device-plugin>`_
|