-
Notifications
You must be signed in to change notification settings - Fork 43
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to access togethercomputer/m2-bert-80M-32k-retrieval
locally?
#42
Comments
togethercomputer/m2-bert-80M-32k-retrieval
Locally?togethercomputer/m2-bert-80M-32k-retrieval
locally?
Give this script a try: https://github.com/HazyResearch/m2/blob/main/bert/embed_text.py It prints out the values of the embedding, it should be the same between local and the API. The API is serving the togethercomputer model (“V0”). The V1 under hazyresearch has been fine tuned on more diverse data, so performance may vary between those two models. |
Thank you for the fast responses, @DanFu09!
Are Thank you, very interesting model. |
It’s the same architecture between v0 and v1, but slightly different data
and loss function.
I’m not aware of what “v2” means :)
The v1 in the together api URL refers to the overall API spec, not any
particular model.
…On Fri, Jan 3, 2025 at 11:11 AM Andreu Huguet ***@***.***> wrote:
Thank you for the fast responses, @DanFu09 <https://github.com/DanFu09>!
The API is serving the togethercomputer model (“V0”). The V1 under
hazyresearch has been fine tuned on more diverse data, so performance may
vary between those two models.
Are v0 and v1 the same architecture but trained on different data? Is the
private v2 from the GitHub docs
<https://github.com/HazyResearch/m2/blob/main/bert/EMBEDDINGS.md#generating-embeddings>
also the same architecture trained on different data?
Thank you, very interesting model.
—
Reply to this email directly, view it on GitHub
<#42 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/ABDDIITII6B3FZHWHZ5J64T2IZO5FAVCNFSM6AAAAABUQCXRB2VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDKNRYHE4DKNRQGA>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
With v2 I am referring at the Lines 69 to 73 in 7d359e8
Maybe it is a typo! |
Ah, thats a typo! At some point we internally switched “v2” to “v1” so
probably things are leftover from that.
…On Fri, Jan 3, 2025 at 11:23 AM Andreu Huguet ***@***.***> wrote:
With v2 I am referring at the
togethercomputer/m2-bert-80M-32k-retrieval-v2 version mentioned in the
EMBEDDINGS.md various times such in:
https://github.com/HazyResearch/m2/blob/7d359e8ce0d18294d840d1ed775ec920b3b48afd/bert/EMBEDDINGS.md?plain=1#L69-L73
Maybe it is a typo!
—
Reply to this email directly, view it on GitHub
<#42 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/ABDDIIXRTP4ID6W4ZCTWXFL2IZQLTAVCNFSM6AAAAABUQCXRB2VHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDKNRZGAYDCNZSHE>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
I am comparing the performance of three different approaches for the
togethercomputer/m2-bert-80M-32k-retrieval
model. Below are the results from each method. Based on the outcomes, the TogetherComputer API provides the best results. I need guidance on how to access this model locally.Performance Comparison
1. TogetherComputer API
Code:
Results:
2. HuggingFace
Code:
Results:
3. HuggingFace v1
Code:
Results:
Request for Help
The TogetherComputer API provides the best results in terms of similarity scores. However, I would like to access the same model locally for further experimentation and to avoid API dependency.
Question:
How can I download and use the
togethercomputer/m2-bert-80M-32k-retrieval
model locally, while achieving the same level of performance as with the TogetherComputer API?Thank you!
The text was updated successfully, but these errors were encountered: