A collection of my FP8 quants for models missing this.
Markus PRO
AI & ML interests
NLP
Recent Activity
updated
a collection
less than a minute ago
FP8 quants
reacted
to
danielhanchen's
post
with 🔥
about 1 hour ago
Run GLM-4.7-Flash locally on your device with 24GB RAM!🔥
It's the best performing 30B model on SWE-Bench and GPQA. With 200K context, it excels at coding, agents, chat & reasoning.
GGUF: https://huggingface.co/unsloth/GLM-4.7-Flash-GGUF
Guide: https://unsloth.ai/docs/models/glm-4.7-flash
reacted
to
danielhanchen's
post
with 🔥
about 1 hour ago
Run GLM-4.7-Flash locally on your device with 24GB RAM!🔥
It's the best performing 30B model on SWE-Bench and GPQA. With 200K context, it excels at coding, agents, chat & reasoning.
GGUF: https://huggingface.co/unsloth/GLM-4.7-Flash-GGUF
Guide: https://unsloth.ai/docs/models/glm-4.7-flash