Skip to content

Add Kata docs#370

Open
a-mccarthy wants to merge 8 commits intoNVIDIA:mainfrom
a-mccarthy:add-kata-docs
Open

Add Kata docs#370
a-mccarthy wants to merge 8 commits intoNVIDIA:mainfrom
a-mccarthy:add-kata-docs

Conversation

@a-mccarthy
Copy link
Copy Markdown
Collaborator

@a-mccarthy a-mccarthy commented Mar 25, 2026

split from to #365 to just work on kata sandbox docs.

@github-actions
Copy link
Copy Markdown

Documentation preview

https://nvidia.github.io/cloud-native-docs/review/pr-370

@a-mccarthy
Copy link
Copy Markdown
Collaborator Author

@manuelh-dev if you have time to review :)

Signed-off-by: Abigail McCarthy <20771501+a-mccarthy@users.noreply.github.com>
Signed-off-by: Abigail McCarthy <20771501+a-mccarthy@users.noreply.github.com>
Signed-off-by: Abigail McCarthy <20771501+a-mccarthy@users.noreply.github.com>
Signed-off-by: Abigail McCarthy <20771501+a-mccarthy@users.noreply.github.com>
Signed-off-by: Abigail McCarthy <20771501+a-mccarthy@users.noreply.github.com>
* ``NVIDIA GPU Feature Discovery`` -- to detect NVIDIA GPUs and label worker nodes.
- * ``NVIDIA Confidential Computing Manager for Kubernetes`` -- to set the confidential computing (CC) mode on the NVIDIA GPUs.
This component is deployed to all nodes configured for Kata Containers, even if you are not planning to run Confidential Containers.
Refer to the :doc:`Confidential Containers <cc:index>` documentation for more details.
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

maybe we should additionally link to the section in the deployment guide which talks about mode changes? If a node is CC capable, the GPU will be in CC mode won't it?

Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I see this is aspect is being iterated on further below again. It might be good to make it clear at one place that a mode change may be required. Also, @jojimt is it possible to directly control this via GPU operator helm install flags?

$ kubectl delete -f cuda-vectoradd-kata.yaml


Troubleshooting Workloads
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Reminder for the future: Let's think about forking this out into a separate page. Please feel free to close.

:alt: Logical diagram of software components between Kubelet and containers when using Kata Containers.

flowchart LR
a[Kubelet] --> b[CRI] --> c[Kata\nRuntime] --> d[Lightweight\nQEMU VM] --> e[Lightweight\nGuest OS] --> f[Pod] --> g[Container]
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Potential to discuss: The Pod construct could be a box wrapping the VM, Guest OS, Container. Here, it looks like the runtime somewhat starts a VM, which somewhat starts a Guest OS which somewhat starts a Pod. Is this really correct? Maybe I'm over-thinking here, but would be good to run internally.

Signed-off-by: Abigail McCarthy <20771501+a-mccarthy@users.noreply.github.com>
Signed-off-by: Abigail McCarthy <20771501+a-mccarthy@users.noreply.github.com>
Signed-off-by: Abigail McCarthy <20771501+a-mccarthy@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants