LFM2.5-1.2B-Instruct-math-aggressive

MATH-optimized | Aggressive pruning | 35% weights pruned

This model is a aggressively pruned version of LiquidAI/LFM2.5-1.2B-Instruct.

Note: Minimal quality drop detected. The Wanda pruning algorithm effectively identifies and removes less important weights while preserving model capability.

Performance Comparison

Category Original Pruned Change
Python 5.0% 0.0% ↓ 5.0%
Html 15.0% 0.0% ↓ 15.0%
Trivia 90.0% 90.0% β†’
Math 55.0% 55.0% ⭐ β†’
Reasoning 45.0% 40.0% ↓ 5.0%
Medical 80.0% 80.0% β†’
Linux 50.0% 50.0% β†’
Writing 15.0% 15.0% β†’

Average: 44.4% -> 41.2% (-3.1%)

Math Retention: 100.0%

Comparison Graph

Quick Start

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("CompactAI/LFM2.5-1.2B-Instruct-math-aggressive")
tokenizer = AutoTokenizer.from_pretrained("CompactAI/LFM2.5-1.2B-Instruct-math-aggressive")

inputs = tokenizer("Your prompt here", return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))

Technical Details

Property Value
Base Model LiquidAI/LFM2.5-1.2B-Instruct
Specialization Math
Prune Mode Aggressive
Weight Reduction 35% weights pruned

License

This model inherits the license from the base model.

Downloads last month
10
Safetensors
Model size
1B params
Tensor type
F16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for CompactAI/LFM2.5-1.2B-Instruct-math-aggressive

Finetuned
(53)
this model

Collection including CompactAI/LFM2.5-1.2B-Instruct-math-aggressive