Feature Request: Update GPU operator invocation #75

Closed
opened 2023-07-29 14:30:52 +00:00 by cloudymax · 5 comments
cloudymax commented 2023-07-29 14:30:52 +00:00 (Migrated from github.com)

Looks like Rancher is doing it like this: https://gist.github.com/bgulla/5ea0e7fd310b5db4f9b66036d1cdb3d3

helm repo add nvidia https://helm.ngc.nvidia.com/nvidia \
&& helm repo update
helm install --wait nvidiagpu \
     -n gpu-operator --create-namespace \
    --set toolkit.env[0].name=CONTAINERD_CONFIG \
    --set toolkit.env[0].value=/var/lib/rancher/k3s/agent/etc/containerd/config.toml \
    --set toolkit.env[1].name=CONTAINERD_SOCKET \
    --set toolkit.env[1].value=/run/k3s/containerd/containerd.sock \
    --set toolkit.env[2].name=CONTAINERD_RUNTIME_CLASS \
    --set toolkit.env[2].value=nvidia \
    --set toolkit.env[3].name=CONTAINERD_SET_AS_DEFAULT \
    --set-string toolkit.env[3].value=true \
     nvidia/gpu-operator

delete:
helm uninstall -n gpu-operator nvidiagpu

cluster-info:
kubectl get nodes -o wide

Looks like Rancher is doing it like this: https://gist.github.com/bgulla/5ea0e7fd310b5db4f9b66036d1cdb3d3 ```bash helm repo add nvidia https://helm.ngc.nvidia.com/nvidia \ && helm repo update ``` ```bash helm install --wait nvidiagpu \ -n gpu-operator --create-namespace \ --set toolkit.env[0].name=CONTAINERD_CONFIG \ --set toolkit.env[0].value=/var/lib/rancher/k3s/agent/etc/containerd/config.toml \ --set toolkit.env[1].name=CONTAINERD_SOCKET \ --set toolkit.env[1].value=/run/k3s/containerd/containerd.sock \ --set toolkit.env[2].name=CONTAINERD_RUNTIME_CLASS \ --set toolkit.env[2].value=nvidia \ --set toolkit.env[3].name=CONTAINERD_SET_AS_DEFAULT \ --set-string toolkit.env[3].value=true \ nvidia/gpu-operator ``` delete: helm uninstall -n gpu-operator nvidiagpu cluster-info: kubectl get nodes -o wide
jessebot commented 2023-08-03 09:28:27 +00:00 (Migrated from github.com)

I will take this as a feature if you can do the PR

I will take this as a feature if you can do the PR
jessebot commented 2023-08-12 15:37:48 +00:00 (Migrated from github.com)

@cloudymax this should first be an app in https://github.com/small-hack/argocd-apps/tree/main and then we can add it to the default config for smol k8s :3

@cloudymax this should first be an app in https://github.com/small-hack/argocd-apps/tree/main and then we can add it to the default config for smol k8s :3
jessebot commented 2023-09-10 08:10:51 +00:00 (Migrated from github.com)

Now that v1.0.0 is officially out, it's much easier to add this to the default applications. Some notes for that:

Now that `v1.0.0` is officially out, it's much easier to add this to the default applications. Some notes for that: - make sure the application is documented in [this section of the root README](https://github.com/small-hack/smol-k8s-lab#optionally-installed-applications), including adding a small icon, even if you have to make one up :) - make sure to set it to disabled by default in [default_config.yaml under apps](https://github.com/small-hack/smol-k8s-lab/blob/main/smol_k8s_lab/config/default_config.yaml#L64) (also make sure it's in alphabetical order) - Make sure it's well documented in [small-hack/argocd-apps](https://github.com/small-hack/argocd-apps/tree/main) both in the root README.md and in its application directory's README.
jessebot commented 2023-12-01 12:53:48 +00:00 (Migrated from github.com)

@cloudymax I'm marking this as blocked based on your work on this helm chart, but feel free to unblock it when you're ready

@cloudymax I'm marking this as blocked based on your work on this helm chart, but feel free to unblock it when you're ready
jessebot commented 2023-12-02 12:02:57 +00:00 (Migrated from github.com)
Closing based on https://github.com/small-hack/smol-k8s-lab/pull/58#issuecomment-1837129823
Sign in to join this conversation.
No milestone
No project
No assignees
1 participant
Notifications
Due date
The due date is invalid or out of range. Please use the format "yyyy-mm-dd".

No due date set.

Dependencies

No dependencies set.

Reference
small-hack/smol-k8s-lab#75
No description provided.