Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[FEAT]: Multilingual Native Embedder #658

Open
timothycarambat opened this issue Jan 30, 2024 · 9 comments
Open

[FEAT]: Multilingual Native Embedder #658

timothycarambat opened this issue Jan 30, 2024 · 9 comments
Labels
enhancement New feature or request feature request

Comments

@timothycarambat
Copy link
Member

What would you like to see?

Currently, the built-in embedder uses the ONNX all-MiniLM-L6-v2 embedder, which does okay for most use cases and is much smaller to download.

There should be support for the larger multilingual-e5-large model (ONNX HERE) for multi-lingual support.

This should not be the default, but it should be something the user can opt to select. They may have to wait for the download to completely download for the embedder change to be saved as we cannot afford the latency to download the model at runtime.

@timothycarambat
Copy link
Member Author

Also, we don't want to pre-pack the docker image with models people may not use, so we will not be doing that in the future to keep the docker image portable enough for a reasonable size.

@vlbosch
Copy link

vlbosch commented Mar 5, 2024

I would also like the option to add another local embeddings model, like for example BGE-M3. I tried adding it in the models-folder myself, but couldn't get it to work yet, unfortunately. Hopefully this feature can be added on the short term, so that we don't need to really on OpenAI's models for multilingual documents. Thanks in advance! :-)

@oscar-7000
Copy link

bge-m3 would be nice

@sweco-nlmdek
Copy link

This would be a very welcome feature. i see in this thread : #645 someone tried multilingual-e5-large and it seems to help allot.

@Herz3h
Copy link

Herz3h commented Jan 29, 2025

Is there a manual way to use a multilingual embedding model in the mean time ? or one from sentence-transformers?

@timothycarambat
Copy link
Member Author

@Herz3h, yes, we support Ollama or LMStudio for embedder endpoints, where you can use any embedder you like to fit your use case.

@Herz3h
Copy link

Herz3h commented Jan 30, 2025

Thanks, however neither of ollama/lmstudio seem to support sentence-transformers, is there a way to still use them? since some of them rank very high in https://huggingface.co/spaces/mteb/leaderboard

@timothycarambat
Copy link
Member Author

timothycarambat commented Jan 30, 2025

Ollama and LMstudio both support embedders - you need to use the GGUF version since they cannot run tensorflow or pytorch models.

Ollamas select is certainly more limited since it relies on their registry, but you can import HF models directly into Ollama.
Example: https://huggingface.co/models?library=gguf&sort=downloads&search=embedding
Click on any of those and you will see the ability to pull it in Ollama or LMStudio.

Both of which support /embedding - so it works.

Image

@Herz3h
Copy link

Herz3h commented Jan 31, 2025

Oh didn't know there was an option from huggingface to pull anymodel in ollama. Thank you very much !

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request feature request
Projects
None yet
Development

No branches or pull requests

5 participants