-
Notifications
You must be signed in to change notification settings - Fork 157
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add GPU edge node manege docs #654
base: master
Are you sure you want to change the base?
Conversation
[APPROVALNOTIFIER] This PR is NOT APPROVED This pull-request has been approved by: The full list of commands accepted by this bot can be found here.
Needs approval from an approver in each of these files:
Approvers can indicate their approval by writing |
399a860
to
d7fb3be
Compare
/lgtm |
New changes are detected. LGTM label has been removed. |
47188ae
to
0fa121e
Compare
Signed-off-by: wbc6080 <[email protected]> Co-authored-by: ming.tang <[email protected]>
PTAL @fisherxu @Shelley-BaoYue |
Signed-off-by: wbc6080 <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
gotta say this is very good information for user application at edge. thanks for taking care of this as documentation.
just got some minor comments, that probably good for maintenance i think.
|
||
## Abstract | ||
|
||
With the development of edge AI, the demand for deploying GPU applications on edge nodes is gradually increasing. Currently, KubeEdge can manage GPU nodes through some configurations, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
With the development of edge AI, the demand for deploying GPU applications on edge nodes is gradually increasing. Currently, KubeEdge can manage GPU nodes through some configurations, | |
With the development of edge AI, it is likely that the applications at edge demand and rely on the GPU acceleration. Currently, KubeEdge can manage GPU nodes through some configurations, |
## Abstract | ||
|
||
With the development of edge AI, the demand for deploying GPU applications on edge nodes is gradually increasing. Currently, KubeEdge can manage GPU nodes through some configurations, | ||
and allocate GPU resources to user edge applications through the k8s device-plugin component. If you need to use this feature, please refer to the steps below. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
and allocate GPU resources to user edge applications through the k8s device-plugin component. If you need to use this feature, please refer to the steps below. | |
and allocate GPU resources to user edge applications through the [k8s device-plugin](https://kubernetes.io/docs/concepts/extend-kubernetes/compute-storage-net/device-plugins/) component. If you need to use this feature, please refer to the steps below. |
|
||
1. Install GPU driver | ||
|
||
First you need to determine whether the edge node machine has GPU. You can use the `lspci | grep NVIDIA` command to check. Download the appropriate GPU driver according to the specific GPU model and complete the installation. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
so it turned out that this is only for Nvidia GPUs? if that is so, probably we could change the title into How to enable Nvidia GPUs on KubeEdge
or something else?
|
||
## Getting Started | ||
|
||
### GPU running environment construction |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Can we instead just point some URLs such as https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/install-guide.html to make sure Nvidia GPU is enabled and configured for container runtime?
This is obviously not the procedure KubeEdge provides, to avoid having the maintenance cost to catch up with Nvidia specific procedure, it would be easier for the doc maintainers?
|
||
Hosting edge GPU nodes mainly includes the following steps: | ||
|
||
1. Manage the node to the cluster |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
same here can we just point the reference to https://kubeedge.io/docs/setup/install-with-keadm/ to avoid the redundancy and possible maintenance problem?
Which issue(s) this PR fixes:
Fixes #
docs update
Added documentation for managing edge GPU nodes and introduce how to use GPU resources in edge applications.