Similar Models

SmolLM2-135M-Instruct (Infrastructure Fallback)

HuggingFaceTB/SmolLM2-135M-Instruct-GGUF

Ultra-lightweight 135M instruct model from Hugging Face. Apache 2.0. Optimized for browser/mobile edge deployment, classification, and low-latency fallback tasks.

How to Use

To get started, install the `transformers` library:

pip install transformers

Then, use the following snippet to load the model:

from transformers import AutoTokenizer, AutoModelForCausalLM

model_id = "HuggingFaceTB/SmolLM2-135M-Instruct-GGUF"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id)

# Your inference code here...

Available Versions

Tag / VariantSizeFormatDownload
HuggingFaceTB/SmolLM2-135M-Instruct-GGUF:Q4_K_M145MBGGUFLink
HuggingFaceTB/SmolLM2-135M-Instruct-GGUF:Q8_0245MBGGUFLink

Model Details

Teacher Model

SmolLM2-Base

Distillation Method

Knowledge Distillation (Logits)

Training Dataset

Flickr30k (Conceptual)

Primary Task

Multimodal Generation

Performance Metrics (Example)

MetricStudent ModelTeacher Model
Model Size145MB8.5GB
BLEU Score28.530.1