Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
5 changes: 1 addition & 4 deletions confidential-containers/overview.rst
Original file line number Diff line number Diff line change
Expand Up @@ -94,7 +94,6 @@ The GPU Operator deploys the components needed to run Confidential Containers to
* NVIDIA Confidential Computing Manager (cc-manager) for Kubernetes - to set the confidential computing (CC) mode on the NVIDIA GPUs.
* NVIDIA Sandbox Device Plugin - to discover NVIDIA GPUs along with their capabilities, to advertise these to Kubernetes, and to allocate GPUs during pod deployment.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

https://github.com/kata-containers/kata-containers/pull/12651/changes has an updated version on the description of the sandbox device plugin

* NVIDIA VFIO Manager - to bind discovered NVIDIA GPUs to the vfio-pci driver for VFIO passthrough.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

see https://github.com/kata-containers/kata-containers/pull/12651/changes:
nvidia-vfio-manager: Binding discovered NVIDIA GPUs and nvswitches to
the vfio-pci driver for VFIO passthrough.

* NVIDIA Kata Manager for Kubernetes - to create host-side CDI specifications for GPU passthrough.

**Kata Deploy**

Expand Down Expand Up @@ -167,7 +166,6 @@ The following is the component stack to support the open Reference Architecture
| - NVIDIA VFIO Manager
| - NVIDIA Sandbox device plugin
| - NVIDIA Confidential Computing Manager for Kubernetes
| - NVIDIA Kata Manager for Kubernetes
- v25.10.0 and higher
* - CoCo release (EA)
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Is this intentional? I think for our latest stack we need a Kata 3.28 release.
I don't know what 'v0.18.0' is here and I ma not sure if we have the exact trustee/guest components in these versions. We are not using a concrete CoCo release. We are using a Kata release and this Kata release pull in CoCo components as dependencies

| - Kata 3.25 (w/ kata-deploy helm)
Expand All @@ -194,8 +192,7 @@ You can configure all the worker nodes in your cluster for running GPU workloads
* NVIDIA MIG Manager for Kubernetes
* Node Feature Discovery
* NVIDIA GPU Feature Discovery
- * NVIDIA Kata Manager for Kubernetes
* NVIDIA Confidential Computing Manager for Kubernetes
- * NVIDIA Confidential Computing Manager for Kubernetes
* NVIDIA Sandbox Device Plugin
* NVIDIA VFIO Manager
* Node Feature Discovery
Expand Down
19 changes: 2 additions & 17 deletions gpu-operator/confidential-containers-deploy.rst
Original file line number Diff line number Diff line change
Expand Up @@ -139,21 +139,7 @@ Perform the following steps to install the Operator for use with confidential co
-n gpu-operator --create-namespace \
nvidia/gpu-operator \
--set sandboxWorkloads.enabled=true \
--set kataManager.enabled=true \
--set kataManager.config.runtimeClasses=null \
--set kataManager.repository=nvcr.io/nvidia/cloud-native \
--set kataManager.image=k8s-kata-manager \
--set kataManager.version=v0.2.4 \
--set ccManager.enabled=true \
--set ccManager.defaultMode=on \
--set ccManager.repository=nvcr.io/nvidia/cloud-native \
--set ccManager.image=k8s-cc-manager \
--set ccManager.version=v0.2.0 \
--set sandboxDevicePlugin.repository=nvcr.io/nvidia/cloud-native \
--set sandboxDevicePlugin.image=nvidia-sandbox-device-plugin \
--set sandboxDevicePlugin.version=v0.0.1 \
--set 'sandboxDevicePlugin.env[0].name=P_GPU_ALIAS' \
--set 'sandboxDevicePlugin.env[0].value=pgpu' \
--set sandboxWorkloads.mode=kata \
--set nfd.enabled=true \
--set nfd.nodefeaturerules=true

Expand All @@ -172,7 +158,7 @@ Perform the following steps to install the Operator for use with confidential co
resource types (such as ``nvidia.com/GH100_H100L_94GB``) instead of the generic
``nvidia.com/pgpu``. For simplicity, this guide uses the generic alias.

3. Verify that all GPU Operator pods, especially the Kata Manager, Confidential Computing Manager, Sandbox Device Plugin and VFIO Manager operands, are running::
3. Verify that all GPU Operator pods, especially the Confidential Computing Manager, Sandbox Device Plugin and VFIO Manager operands, are running::

$ kubectl get pods -n gpu-operator

Expand All @@ -184,7 +170,6 @@ Perform the following steps to install the Operator for use with confidential co
gpu-operator-1766001809-node-feature-discovery-worker-mh4cv 1/1 Running 0 86s
gpu-operator-f48fd66b-vtfrl 1/1 Running 0 86s
nvidia-cc-manager-7z74t 1/1 Running 0 61s
nvidia-kata-manager-k8ctm 1/1 Running 0 62s
nvidia-sandbox-device-plugin-daemonset-d5rvg 1/1 Running 0 30s
nvidia-sandbox-validator-6xnzc 1/1 Running 1 30s
nvidia-vfio-manager-h229x 1/1 Running 0 62s
Expand Down
1 change: 1 addition & 0 deletions gpu-operator/index.rst
Original file line number Diff line number Diff line change
Expand Up @@ -56,6 +56,7 @@
:hidden:

KubeVirt <gpu-operator-kubevirt.rst>
Kata Containers <kata-containers-deploy.rst>
Confidential Containers <confidential-containers-deploy.rst>

.. toctree::
Expand Down
Loading