LFM2-8B-A1B-GGUF - Quantized GGUF Model

This is a quantized GGUF model (Q4_K_M) compatible with Ollama.

Model Details

  • Base Model: LiquidAI/LFM2-8B-A1B-GGUF
  • Quantization: Q4_K_M
  • Framework: Ollama

Usage with Ollama

You can pull and run this model directly with Ollama:

ollama pull hf.co/Sadiah/ollama-q4_k_m-LFM2-8B-A1B-GGUF:Q4_K_M

Then run it:

ollama run hf.co/Sadiah/ollama-q4_k_m-LFM2-8B-A1B-GGUF:Q4_K_M "Write your prompt here"

Features

  • Efficient quantization (Q4_K_M) for faster inference
  • Compatible with Ollama's inference engine

License

Please refer to the original model card for licensing information.

Downloads last month
9
GGUF
Model size
8B params
Architecture
lfm2moe
Hardware compatibility
Log In to view the estimation

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Sadiah/ollama-q4_k_m-LFM2-8B-A1B-GGUF

Quantized
(1)
this model

Collection including Sadiah/ollama-q4_k_m-LFM2-8B-A1B-GGUF