veg
ciprianv
AI & ML interests
None yet
Recent Activity
liked
a model about 23 hours ago
vpyn/Qwen3.5-397B-A17B-CARVE-v1-NVFP4 liked
a model 3 days ago
unsloth/Qwen3.5-35B-A3B-GGUF new activity
7 days ago
mratsim/MiniMax-M2.5-BF16-INT4-AWQ:Cant get it to work on 8x RTX3090 Organizations
None yet
Cant get it to work on 8x RTX3090
14
#1 opened 16 days ago
by
maglat
"w1_weight_scale_2 must match w3_weight_scale_2. Accuracy may be affected."
๐ 1
20
#2 opened 17 days ago
by
zenmagnets
accuracy
17
#4 opened 16 days ago
by
ktsaou
Fastest for my 3090x8
2
#1 opened 17 days ago
by
ciprianv
Hey i like the model could you maybe make a NVFP4 version or a version optimised for the dgx spark?
3
#1 opened about 2 months ago
by
Floris111
Please create also Minimax 2.1 REAP versions
2
#1 opened about 2 months ago
by
ciprianv
Report: getting 20 t/s with UD-Q4_K_XL and 72 VRAM
๐ฅ 1
10
#2 opened 2 months ago
by
SlavikF
Hot Damn This Model Cooks!
๐ 6
12
#5 opened 2 months ago
by
aaron-newsome
Please make 4 bit dwq mlx quant
2
#1 opened 2 months ago
by
Narutoouz
Please update llama.cpp to see improved performance!
๐ 4
4
#7 opened 3 months ago
by
danielhanchen
Updated Title: UDQ4_K_XL - Great Rust coder
๐ 3
5
#11 opened 7 months ago
by
wonderfuldestruction
download link creates Q5_K_M instead of UD-Q5_K_XL named files
1
#2 opened 7 months ago
by
ciprianv
Confused about the eval score
โค๏ธ 2
3
#15 opened 7 months ago
by
Denisssy
IQ3_KS metrics on mixed CUDA + CPU, pretty good model!
๐ฅ 2
34
#2 opened 8 months ago
by
Panchovix
What are the recommended settings?
1
#7 opened 8 months ago
by
ciprianv
Thanks for your work! Any chance for something between Q2_K_R and Q3_K_R?
๐ ๐ 5
19
#7 opened 9 months ago
by
Panchovix
Update - Tool Calling + Chat Template bug fixes
9
#20 opened 9 months ago
by
danielhanchen
Please share feedback here!
34
#6 opened 9 months ago
by
shimmyshimmer
Recommendation for 256 ram 48 vram
โค๏ธ 2
5
#2 opened 9 months ago
by
ciprianv
Q6 K
36
#1 opened 10 months ago
by
Autumnlight