inference-optimization/Meta-Llama-3.1-8B-Instruct-NVFP4-FP8-Dynamic_5.75-bits 6B • Updated Jan 26 • 3
inference-optimization/Meta-Llama-3.1-8B-Instruct-NVFP4-FP8-Dynamic_6.25-bits 6B • Updated Jan 26 • 3