Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

nvidia-device-plugin doesn't schedule plugin Pods to workers #20

Open
sfc-gh-lshi opened this issue Dec 12, 2024 · 1 comment
Open

nvidia-device-plugin doesn't schedule plugin Pods to workers #20

sfc-gh-lshi opened this issue Dec 12, 2024 · 1 comment

Comments

@sfc-gh-lshi
Copy link

With the currently documented steps, nvidia-device-plugin does not schedule the plugin DaemonSet to any of the cluster workers. This is because the worker nodes are missing required labels.

After installing the Helm chart, the worker(s) need to be labeled thusly:

kubectl label node "${KIND_CLUSTER_NAME}-worker" --overwrite nvidia.com/gpu.present=true

Only afterwards does kubectl --context=kind-${KIND_CLUSTER_NAME} get pod -n nvidia show any Pods.

@LogExE
Copy link

LogExE commented Dec 12, 2024

It looks like we have to pass --set gfd.enabled=true to helm install command

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants