GLM-4.7-Flash-REAP-50 Quantized GGUF (Q4_K_M)
This repository contains a Q4_K_M quantized GGUF version of the GLM-4.7-Flash-REAP-50 model, originally developed by Akicou.
Quantization Details
- Original Format: F16 GGUF
- Quantized Format: Q4_K_M
- Converted using: llama.cpp
- File:
glm-4.7-flash-Q4_K_M.gguf
Feel free to use and experiment with this quantized version!
- Downloads last month
- 248
Hardware compatibility
Log In to add your hardware
4-bit
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support