GLM-4.7-Flash-REAP-50 Quantized GGUF (Q4_K_M)

This repository contains a Q4_K_M quantized GGUF version of the GLM-4.7-Flash-REAP-50 model, originally developed by Akicou.

Quantization Details

  • Original Format: F16 GGUF
  • Quantized Format: Q4_K_M
  • Converted using: llama.cpp
  • File: glm-4.7-flash-Q4_K_M.gguf

Feel free to use and experiment with this quantized version!

Downloads last month
248
GGUF
Model size
16B params
Architecture
deepseek2
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for gaionaus/GLM-4.7-Flash-REAP-50_Q4_K_M_GGUF

Quantized
(3)
this model