openai-community/gpt-oss-20b-GGUF
NewOpenClaw-compatible open-weight 20B model. 64K+ context, Apache 2.0. Balanced performance for tool-use, memory persistence, and multi-channel agentic workflows.
OpenClaw-compatible open-weight 20B model. 64K+ context, Apache 2.0. Balanced performance for tool-use, memory persistence, and multi-channel agentic workflows.
OpenClaw-recommended coding specialist. 7B params, 128K context, Apache 2.0. Optimized for tool-calling, shell commands, and multi-file edits in agentic workflows.
Alibaba Qwen3.5 2B edge-optimized model. Hybrid Gated DeltaNet+Attention architecture, 256K context, Apache 2.0. Built for tool-calling agents and multimodal workflows.
THUDM/glm-4-7-flash-GGUF
OpenClaw-recommended general-purpose model. 7B params, 128K context, MIT license. Balanced speed/quality for daily assistant tasks, research, and multi-step reasoning.
To get started, install the `transformers` library:
pip install transformersThen, use the following snippet to load the model:
from transformers import AutoTokenizer, AutoModelForCausalLM
model_id = "THUDM/glm-4-7-flash-GGUF"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id)
# Your inference code here...GLM-Edge-1.5B
Knowledge Distillation (Logits)
Flickr30k (Conceptual)
Multimodal Generation
| Metric | Student Model | Teacher Model |
|---|---|---|
| Model Size | 4.8GB | 8.5GB |
| BLEU Score | 28.5 | 30.1 |