Clarify Kubernetes version requirement and fallback plan in Key Features#380
Conversation
| - **Accelerator Fungibility**: llmaz supports serving the same LLM with various accelerators to optimize cost and performance. | ||
| - **Various Model Providers**: llmaz supports a wide range of model providers, such as [HuggingFace](https://huggingface.co/), [ModelScope](https://www.modelscope.cn), ObjectStores. llmaz will automatically handle the model loading, requiring no effort from users. | ||
| - **Multi-Host Support**: llmaz supports both single-host and multi-host scenarios with [LWS](https://github.com/kubernetes-sigs/lws) from day 0. | ||
| - **Multi-Host Support**: llmaz supports both single-host and multi-host scenarios with [LWS](https://github.com/kubernetes-sigs/lws) from day 0. **Important**: LWS requires Kubernetes version **v1.26 or higher**. If you are using a lower Kubernetes version and most of your workloads rely on single-node inference, we may consider replacing LWS with a deployment-based approach. This fallback plan would involve using Kubernetes Deployments to manage single-node inference workloads efficiently. See [#32](https://github.com/InftyAI/llmaz/issues/32) for more details and updates. |
There was a problem hiding this comment.
Could you update this in https://github.com/InftyAI/llmaz/blob/main/docs/installation.md#prerequisites? Thanks
There was a problem hiding this comment.
Sure. I will complete it and make a commit.
There was a problem hiding this comment.
Please revert this change, the note within installation.md is enough. Thanks
|
/kind documentation |
|
I've updated the Installation file. Can you please review it? |
| - **Accelerator Fungibility**: llmaz supports serving the same LLM with various accelerators to optimize cost and performance. | ||
| - **Various Model Providers**: llmaz supports a wide range of model providers, such as [HuggingFace](https://huggingface.co/), [ModelScope](https://www.modelscope.cn), ObjectStores. llmaz will automatically handle the model loading, requiring no effort from users. | ||
| - **Multi-Host Support**: llmaz supports both single-host and multi-host scenarios with [LWS](https://github.com/kubernetes-sigs/lws) from day 0. | ||
| - **Multi-Host Support**: llmaz supports both single-host and multi-host scenarios with [LWS](https://github.com/kubernetes-sigs/lws) from day 0. **Important**: LWS requires Kubernetes version **v1.26 or higher**. If you are using a lower Kubernetes version and most of your workloads rely on single-node inference, we may consider replacing LWS with a deployment-based approach. This fallback plan would involve using Kubernetes Deployments to manage single-node inference workloads efficiently. See [#32](https://github.com/InftyAI/llmaz/issues/32) for more details and updates. |
There was a problem hiding this comment.
Please revert this change, the note within installation.md is enough. Thanks
|
/lgtm Thanks for your patience. Welcome onboard! |
|
Thank you! I appreciate the review and approval. Excited to contribute further! |
|
/lgtm |
|
seems we need to rebase. |
Co-authored-by: Kante Yin <kerthcet@gmail.com>
971cf5e to
15fee24
Compare
|
Are there any final changes required? Please clarify, and I will make the necessary updates. |
|
/lgtm |
…res section
What this PR does / why we need it
This PR updates the README file to improve clarity on multiple host support and installation guidelines by specifying Kubernetes version requirements, helping to prevent setup issues.
Which issue(s) this PR fixes
Fixes #
#379
Special notes for your reviewer
Does this PR introduce a user-facing change?