-
Notifications
You must be signed in to change notification settings - Fork 117
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
unable to run ramalama using --runtime vllm on macOS #801
Comments
https://docs.vllm.ai/en/latest/getting_started/installation/cpu/index.html it's not optimal but it can run on macOS
|
I think you found your answer in the above doc: "vLLM has experimental support for macOS with Apple silicon. For now, users shall build from the source vLLM to natively run on macOS." This is one for the vLLM folks |
A general issue around vLLM should be opened if not already, could do with some Containerfile work. llama.cpp is more suitable for macOS runtime today |
I don't see why it's being closed as fixed The user experience is bad. It tries to fetch an image that does not exists It should report a good error message |
some ramalama images are building llama.cpp from sources so it could also build vLLM for a later arm/macOS usage |
trying ramalama on macOS 15 using --runtime vllm
I got
it seems to fetch a cuda image while I am on Apple silicon.
The text was updated successfully, but these errors were encountered: