-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
List of Free Local LLM Models - Up and Running #967
Comments
@psymbio I faced the same problem when i was doing the hugging face AI agents course . So I figured out out this way: from smolagents import CodeAgent, LiteLLMModel
model = LiteLLMModel(
model_id="ollama_chat/qwen2.5:7b", #Can try diffrent model here I am using qwen2.5 7B model
api_base="http://127.0.0.1:11434",
num_ctx=8192,
)
agent = CodeAgent(tools=[], model=model, add_base_tools=True)
agent.run("Could you give me the 118th number in the Fibonacci sequence?") |
The solution described by @HKanoje is good! If you think it's worth it, please open a PR ion the docs wherever you faced the problem (in agents course?) to add this as a note like "if you run into credit limits, you can switch to local inference using ...". |
@aymeric-roucher Okay got it. Will do ASAP. Just a quick question should i make note everywhere HfApiModel() is used or just initial agent will do? |
Hey @aymeric-roucher , just letting you know that I’ve opened a PR huggingface/agents-course#327 with the note on using |
Thanks, @HKanoje, for the PR in agents-course. |
Is your feature request related to a problem? Please describe.
Can I get a list of free local LLM models I can run without the PRO subscription.
Describe the solution you'd like
A list of the args I can pass into HfApiModel() (most likely) so that I don't run into this error:
Is this not possible with the current options.
I've tried DeepSeek model, which I thought would be free - but turns out even that is connected to the Together API. How can I get a local free LLM model up and running with smolagents?
Maybe this code sample can be provided in the README section.
The text was updated successfully, but these errors were encountered: