-
Notifications
You must be signed in to change notification settings - Fork 50
Added Gateway API Inference Extension (GIE) installation for quickstart docs #332
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
Hi @nekomeowww thanks for taking a look! Can you checkout the ci-deps.sh script in chart-dependencies/ci-deps.sh that installs the latest GAIE v0.3.0 latest release CRDs with https://github.com/llm-d/llm-d-inference-scheduler/blob/main/deploy/components/crds-gie/kustomization.yaml. Not sure where the v0.8.0 GIE release in your patch came from? Im pasting the logs from a quick install from main. Can you share what you are getting? Thanks!
|
Signed-off-by: Neko <[email protected]>
Signed-off-by: Neko <[email protected]>
Ah sure, thanks for pointing that out, I got the script from Getting started - Kubernetes Gateway API Inference Extension section, and it appears that I pasted and modified the wrong version for it. Let me do a real quick test for what you suggested, thanks. |
I tried to install llm-d on my development use macOS over
Indeed showed that GAIE CRDs are installed correctly without any problem. But when I tried llm-d several days ago, since I was installing it into our existing cluster (with NVIDIA GPUs for testing and development purpose), with Istio already installed and configured by other team members (we got many other gateway CRs configured for other teams), after reading a small portion of To address this issue, would it be better if I could write another documentations to install it step by step? (A documentation version of |
Hi @nekomeowww apologies for the slow reply. We definitely have a more manual documenting opportunity coming up once we land #321 if interested. Cheers. |
Summary
Currently, if Gateway API Inference Extension (GIE) not installed in a fresh new testing used cluster, the following error will throw when following the guide of Quickstart, and for the
helm install
ed release, the createdllm-d-modelservice
will fail due to timeout on waiting CRD points toInferencePool.inference.networking.x-k8s.io
:While I do discovered that #321 is currently pending to be merged for adding the Inference Extension (GIE) as a sub chart, for others who currently trying out
llm-d
, it's obvious that missing pieces of GIE will result in errors.This Pull Reuqest temporarily add a new section for asking to install GIE before installation.
Related
Related to #312.
Workaround for #321