Similar Models

Qwen3-4B-Instruct-2507 (Non-Thinking Specialist)

unsloth/Qwen3-4B-Instruct-2507-GGUF

Alibaba Qwen3 updated 4B instruct model. 256K native context, Apache 2.0. Optimized for instruction-following, tool-calling, and agentic workflows without CoT overhead.

How to Use

To get started, install the `transformers` library:

pip install transformers

Then, use the following snippet to load the model:

from transformers import AutoTokenizer, AutoModelForCausalLM

model_id = "unsloth/Qwen3-4B-Instruct-2507-GGUF"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id)

# Your inference code here...

Available Versions

Tag / VariantSizeFormatDownload
unsloth/Qwen3-4B-Instruct-2507-GGUF:UD-Q4_K_XL2.55GBGGUFLink
unsloth/Qwen3-4B-Instruct-2507-GGUF:Q4_K_M2.5GBGGUFLink

Model Details

Teacher Model

Qwen3-30B-A3B

Distillation Method

Knowledge Distillation (Logits)

Training Dataset

Flickr30k (Conceptual)

Primary Task

Multimodal Generation

Performance Metrics (Example)

MetricStudent ModelTeacher Model
Model Size2.5GB8.5GB
BLEU Score28.530.1