Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ollama api error #97

Open
Mrship12138 opened this issue Feb 18, 2025 · 2 comments
Open

ollama api error #97

Mrship12138 opened this issue Feb 18, 2025 · 2 comments

Comments

@Mrship12138
Copy link

Dear Sir:
LiteLLM completion() model= deepseek-coder-v2:16b; provider = ollama
2025-02-18 17:29:13,281 - LiteLLM - INFO -
LiteLLM completion() model= deepseek-coder-v2:16b; provider = ollama
2025-02-18 17:29:26,456 - httpx - INFO - HTTP Request: POST http://localhost:11434/api/generate "HTTP/1.1 500 Internal Server Error"

Give Feedback / Get Help: https://github.com/BerriAI/litellm/issues/new

Image

@Chenglong-MS
Copy link
Collaborator

Hello! I'm able to run deepseek-coder-v2 16b with the following configuration:

Image

I loaded the 16 model with Ollama using ollama run deepseek-coder-v2, so I use deepseek-coder-v2 in the model configuration.

@Chenglong-MS
Copy link
Collaborator

Actually I also encounter this issue every now and then. I debugged a little bit, and found it is an issue with ollama:

Error: an error was encountered while running the model: unexpected EOF

And this looks like an OOM issue during inference. I think the solution is to either use a smaller model, or use GPU for inference.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants