File size: 1,693 Bytes
e39ff3a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
uv run python custom_mlx_lm/custom_convert.py --hf-path . --mlx-path MobileLLM-R1-950M-mixed-4bit-mlx --dynamic-quant --target-bpw 4.5 --group-size 64 --report-ppl
Loading model from ....
Loading calibration data...
Token indices sequence length is longer than the specified maximum sequence length for this model (110205 > 32768). Running this sequence through the model will result in indexing errors
Calculating perplexity of original model...
Original PPL: 50.262
Starting advanced mixed-precision quantization...
huggingface/tokenizers: The current process just got forked, after parallelism has already been used. Disabling parallelism to avoid deadlocks...
To disable this warning, you can either:
	- Avoid using `tokenizers` before the fork if possible
	- Explicitly set the environment variable TOKENIZERS_PARALLELISM=(true | false)
Estimating sensitivities: 100%|████████████████████████████████████| 54/54 [02:03<00:00,  2.28s/it]
Calculating perplexity of quantized model...
Quantized PPL: 59.059

✅ Model saved to MobileLLM-R1-950M-mixed-4bit-mlx

uv run python custom_mlx_lm/quant_summary.py --model-path MobileLLM-R1-950M-mixed-4bit-mlx --show 8 
Method: mixed_precision_dynamic
Group size: 64
Total linear layers: 154
4-bit layers: 153
8-bit layers: 1

Examples (8-bit):
- layers.0.attention.o_proj

Examples (4-bit):
- layers.0.attention.k_proj
- layers.0.attention.q_proj
- layers.0.attention.v_proj
- layers.0.feed_forward.down_proj
- layers.0.feed_forward.gate_proj
- layers.0.feed_forward.up_proj
- layers.1.attention.k_proj
- layers.1.attention.o_proj

weights.npz contains quantized tensors: True