-
Notifications
You must be signed in to change notification settings - Fork 23
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support to specify CUDA visible device id for model service #444
Comments
but experiment must be made to make sure this can work with k8s's gpu scheduling
Based on ray doc https://docs.ray.io/en/latest/ray-core/scheduling/accelerators.html#starting-ray-nodes-with-accelerators |
Furthermore if we allow user to specifiy This issue can be splitted to two tasks:
|
let me do a basic support like below:
|
I think we can support it now following the docs below: |
CUDA_DEVICE_ORDER=PCI_BUS_ID
CUDA_VISIBLE_DEVICES="0,3" # specify which GPU(s) to be used
For scenario that there are different type of GPUs on the same node, like T4 and A100 etc. We should support to deploy model with specified device id.
The text was updated successfully, but these errors were encountered: