I loaded DeepSeek-V3-Q5_K_M up on my 10yrs old old Tesla M40 (Dell C4130)
1
#8 opened 1 day ago
by
gng2info
why use q5 for key cache?
1
#7 opened 1 day ago
by
CHNtentes
Advice on running llama-server with Q2_K_L quant
1
#6 opened 2 days ago
by
vmajor
What is the required GPU size to run Is a 4090 possible and does it support ollama
4
#5 opened 3 days ago
by
sminbb
I'm a newbie. How to use?
1
#4 opened 3 days ago
by
huangkk
Getting error with Q3-K-M
5
#2 opened 3 days ago
by
alain401
Are these imatrix GGUF quants?
4
#1 opened 4 days ago
by
Kearm