Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[BUG]: In query mode, return answer citations with mach rates ranging from 0% to 3%. #3076

Closed
kaanvur opened this issue Jan 31, 2025 · 7 comments
Labels
possible bug Bug was reported but is not confirmed or is unable to be replicated.

Comments

@kaanvur
Copy link

kaanvur commented Jan 31, 2025

How are you running AnythingLLM?

AnythingLLM desktop app

What happened?

I added two files.
I selected query mode in chat settings.
Document similarity threshold
medium is set in vector database settings.
But the model still returns answers to irrelevant questions.
As citations, it shows a few examples with mach rates ranging from 0% to 3%.

Are there known steps to reproduce?

Image
Image
Image
Image
Image

@kaanvur kaanvur added the possible bug Bug was reported but is not confirmed or is unable to be replicated. label Jan 31, 2025
@kaanvur kaanvur changed the title [BUG]: [BUG]: In query mode, return answer citations with mach rates ranging from 0% to 3%. Jan 31, 2025
@timothycarambat
Copy link
Member

Turn off reranking when in query mode - that is the root cause of this (and also why it is not the default mode for a workspace)

@kaanvur
Copy link
Author

kaanvur commented Feb 1, 2025

Turn off reranking when in query mode - that is the root cause of this (and also why it is not the default mode for a workspace)

Thanks for your answer @timothycarambat but I couldn't find any instructions on where and how to do this setting. Can you help me?

@timothycarambat
Copy link
Member

Of course, in the UI we call reranking "Search Preference" since re-ranking is very ambiguous to non-technical people. If you click on a workspace's settings (gear icon) and navigate to "Vector Database" you will find a field "search preference" which you can change from "Accuracy Optimized" back to "default"

https://docs.anythingllm.com/llm-not-using-my-docs#vector-database-settings--search-preference

@kaanvur
Copy link
Author

kaanvur commented Feb 2, 2025

Thank you for your answer @timothycarambat . When I read your explanations, it gives the opposite impression :) I made the changes. This time it brought up more relevant information. However, none of the quotes that it said were 65% relevant to my question "egg bread recipe" have anything to do with this topic. How can I improve this?

Image
Image

@kaanvur
Copy link
Author

kaanvur commented Feb 2, 2025

Would changing the Embedding model here provide better results?

@timothycarambat
Copy link
Member

@kaanvur The default embedder is pretty bad at multi-lingual embedding (see #658). However we support using Ollama/LMStudio with any embedder you can find on HF, which goes a long way.

You would indeed get better responses with the non-default. The default we have is super small so that is why we use it. Some other models like Jina can be more comprehensive but also are 300MB

@kaanvur
Copy link
Author

kaanvur commented Feb 3, 2025

Thank you for all your answers and patience. I guess I need to try different combinations and see what the results are for the better results :) Thanks for your efforts!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
possible bug Bug was reported but is not confirmed or is unable to be replicated.
Projects
None yet
Development

No branches or pull requests

2 participants