Cygnis Alpha 2 i1

Imatrix Optimized v0.3 Reasoning

Enhanced Importance Matrix (imatrix) quantizations for Cygnis Alpha 2.
Higher intelligence floor for low-bit inference, preserving complex CoT patterns.

Open in Ollama
ollama run Simonc-44/Cygnis-Alpha-2-8B-v0.3-i1-GGUF

Available Imatrix Quants

Quantization Size Intelligence Level Download
Q6_K (i1) 6.7 GB REFERENCE GRADE Link
Q4_K_M (i1) 5.0 GB RECOMMENDED Link
IQ4_XS (i1) 4.5 GB HIGH LOGIC Link
IQ3_M (i1) 3.9 GB EFFICIENT Link
IQ2_M (i1) 3.0 GB EXPERIMENTAL Link

Why Imatrix?

Standard quantization treats all weights equally. Imatrix (Importance Matrix) uses a calibration dataset to identify which neurons are vital for the model's logic. By protecting these neurons, we achieve 3-bit or 4-bit quants that often match the performance of standard 6-bit files.

Perplexity vs Bitrate (Efficiency)

Visual representation of why Imatrix (i1) quants outperform standard static quants at lower bitrates.

Ollama Config

FROM ./Cygnis-Alpha-2-7B-v0.3.i1-Q4_K_M.gguf

TEMPLATE """<|im_start|>system {{ .System }}<|im_end|> <|im_start|>user {{ .Prompt }}<|im_end|> <|im_start|>assistant <|im_thought|> """

PARAMETER stop "<|im_end|>"

Quantized with ❤️ by Simonc-44 | Powered by Llama.cpp Imatrix

Downloads last month
1,778
GGUF
Model size
8B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

1-bit

2-bit

3-bit

4-bit

5-bit

6-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Simonc-44/Cygnis-Alpha-2-8B-v0.3-i1-GGUF

Datasets used to train Simonc-44/Cygnis-Alpha-2-8B-v0.3-i1-GGUF

Collection including Simonc-44/Cygnis-Alpha-2-8B-v0.3-i1-GGUF