R7B - Unsafe Results

#20
by Lewdiculous - opened
AetherArchitectural org
edited 7 days ago

@woofwolfy - Not sure if it was an issue with the quant process or a false positive, it was done as usual, I'll wait to hear back from HF and see if it's possible to solve, if not, well, it wasn't meant to be. The template was unmodified from the original, ugh.

@woofwolfy - Not sure if it was an issue with the quant process or a false positive, it was done as usual, I'll wait to hear back from HF and see if it's possible to solve, if not, well, it wasn't meant to be. The template was unmodified from the original, ugh.

As far as I remember when you uploaded the first few quants everything was fine, but then when the rest appeared something happened

AetherArchitectural org

I'm leaning on false positives because of their template. I'll wait to hear from someone.

This comment has been hidden
AetherArchitectural org
edited 7 days ago

Now this is very interesting, could b4416 be infected (?)

I downloaded it and it gave me a different result
изображение.png

llama-b4416-bin-win-cuda-cu11.7-x64.zip Same
llama-b4416-bin-win-cuda-cu12.4-x64.zip Same

AetherArchitectural org

Can you link your VT scans please?

AetherArchitectural org
edited 7 days ago

I did open an issue to ask about this:

https://github.com/ggerganov/llama.cpp/issues/11077

The size of this template though...:
https://github.com/ggerganov/llama.cpp/issues/11077#issuecomment-2571396549

I think that's causing the flagging, both for the builds as well as for the quants which retain the template, protectai also pointed.

AetherArchitectural org

Ah, yeah, bingo:
https://www.virustotal.com/gui/file/762f91efbfa8e278652b39c390354751e9a124ad1991cdabcadb4b32eabe5075

I've scanned other releases they all have the same triggers

AetherArchitectural org
edited 7 days ago

I'm kind of pointing at these template changes:
https://www.diffchecker.com/HJ5zHE2p/

(Then vs Now)

But there where made by the Model authors. I just make sure to use their latest repo:

https://huggingface.co/CohereForAI/c4ai-command-r7b-12-2024/commit/e8aabc6cdb30e851a2fdd21a27b4ba0fccff8624

This comment has been hidden
AetherArchitectural org
edited 6 days ago
AetherArchitectural org
edited 6 days ago

Welp, I reached out to let them know and see if this can be adjusted.

https://huggingface.co/CohereForAI/c4ai-command-r7b-12-2024/discussions/10

Did you close the repo? I think I want to download it. It doesn't seem to work on koboldcpp, but want to check in lm studio this imatrix version

AetherArchitectural org
edited 2 days ago

Yes since the warnings are there still.

There's a test version here:

https://huggingface.co/Lewdiculous/llama.cpp-11077-test-01/

Yes since the warnings are there still.

There's a test version here:

https://huggingface.co/Lewdiculous/llama.cpp-11077-test-01/

okay i'll make q4km imatrix from your txt

AetherArchitectural org

Yes since the warnings are there still.

There's a test version here:

https://huggingface.co/Lewdiculous/llama.cpp-11077-test-01/

okay i'll make q4km imatrix from your txt

If you want the Q4 I can open the old repo again.

AetherArchitectural org
edited 2 days ago

@woofwolfy

https://huggingface.co/Lewdiculous/c4ai-command-r7b-12-2024-GGUF-IQ-ARM-Imatrix

You can test and compare then, if it was solved.

oh... I already cooking in ggml-org/gguf-my-repo , okay i'll download it and test other one

AetherArchitectural org
edited 2 days ago

Uhum, good call, give it a go, hopefully it's good now. Not sure I want to reupload everything again.

mmm the same https://protectai.com/insights/models/woofwolfy/c4ai-command-r7b-12-2024-Q4_K_M-GGUF-Imatix/c9a54bebdb3fcd6474586cedc21b49e988acf12a/files?utm_source=huggingface

I think it's fine, but it seems that lm studio refuses to run it because of the unknown architecture.

AetherArchitectural org

#10900 needs to come for downstream projects. I imagine in the next release of each application.

Sign up or log in to comment