unsloth/Phi-4-mini-reasoning-GGUF
NewMicrosoft Phi-4-mini distilled for step-by-step reasoning. 3.8B params, 128K context, MIT license. Unsloth bug-fixed GGUF for reliable agentic tool-calling.
Microsoft Phi-4-mini distilled for step-by-step reasoning. 3.8B params, 128K context, MIT license. Unsloth bug-fixed GGUF for reliable agentic tool-calling.
Mistral AI edge-optimized 3.4B+0.4B vision model. Native function calling, JSON outputs, 256K context. Built for tool-using agentic pipelines.
A 2026-native 3B reasoning model from Hugging Face. Dual-mode `/think` and `/no_think` for agentic workflows with 64K-128K context. Fully open recipe.
unsloth/Phi-4-mini-instruct-GGUF
Microsoft Phi-4-mini distilled for edge reasoning. 3.8B params, 128K context, MIT license. Optimized for agentic tool-calling and multilingual tasks.
To get started, install the `transformers` library:
pip install transformersThen, use the following snippet to load the model:
from transformers import AutoTokenizer, AutoModelForCausalLM
model_id = "unsloth/Phi-4-mini-instruct-GGUF"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id)
# Your inference code here...Phi-4-14B
Knowledge Distillation (Logits)
Flickr30k (Conceptual)
Multimodal Generation
| Metric | Student Model | Teacher Model |
|---|---|---|
| Model Size | 2.8GB | 8.5GB |
| BLEU Score | 28.5 | 30.1 |