Text Generation
Transformers
GGUF
minimax
unsloth
imatrix
conversational

Pls MXFP4

#4
by Kirara702 - opened

Please start including the MXFP4 alongside the other quantized versions.

Is MXFP4 really better than Q4_K_XL or Q4_K_M? Or it's just because OpenAI used it?

It's worse than the k quants. But it has native hardware support and is fast and energy efficient.

It's worse than the k quants. But it has native hardware support and is fast and energy efficient.

So if my gpu is older than blackwell it's not worth using? Only blackwell has native fp4 support right?

Good idea we are going to do that pretty soon hopefully for our future quants. The quality will degrade slightly

It's worse than the k quants. But it has native hardware support and is fast and energy efficient.

So if my gpu is older than blackwell it's not worth using? Only blackwell has native fp4 support right?

Also works on old GPUS

Good idea we are going to do that pretty soon hopefully for our future quants. The quality will degrade slightly

It's worse than the k quants. But it has native hardware support and is fast and energy efficient.

So if my gpu is older than blackwell it's not worth using? Only blackwell has native fp4 support right?

Also works on old GPUS

But is the speed much slower than blackwell gpus?

Sign up or log in to comment