Similar Models

GLM-4.7-Flash (Balanced Edge Agent)

THUDM/glm-4-7-flash-GGUF

OpenClaw-recommended general-purpose model. 7B params, 128K context, MIT license. Balanced speed/quality for daily assistant tasks, research, and multi-step reasoning.

How to Use

To get started, install the `transformers` library:

pip install transformers

Then, use the following snippet to load the model:

from transformers import AutoTokenizer, AutoModelForCausalLM

model_id = "THUDM/glm-4-7-flash-GGUF"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id)

# Your inference code here...

Available Versions

Tag / VariantSizeFormatDownload
THUDM/glm-4-7-flash-GGUF:Q4_K_M4.8GBGGUFLink
THUDM/glm-4-7-flash-GGUF:Q8_07.9GBGGUFLink

Model Details

Teacher Model

GLM-Edge-1.5B

Distillation Method

Knowledge Distillation (Logits)

Training Dataset

Flickr30k (Conceptual)

Primary Task

Multimodal Generation

Performance Metrics (Example)

MetricStudent ModelTeacher Model
Model Size4.8GB8.5GB
BLEU Score28.530.1