Qwen 2.5 0.5B โ€“ Coding Expert (FFN-Pruned)

This model is a coding-specialized expert derived from Qwen 2.5 0.5B using activation-based FFN neuron pruning on the HumanEval dataset.

Key features

  • ~25% FFN neuron pruning
  • No retraining
  • Stable Python code generation
  • Reduced compute per token

Method

  1. Collected per-neuron FFN activations during inference
  2. Pruned lowest-activation neurons per layer
  3. Preserved embeddings, attention, and early representations

Intended use

  • Research
  • Model compression experiments
  • Expert routing / MoE research

Limitations

  • Specialized for coding
  • General language performance may degrade

This is a research artifact, not a production model.

Downloads last month
24
Safetensors
Model size
0.5B params
Tensor type
F32
ยท
F16
ยท
I8
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Echoes123-3/qwen2.5-0.5b-coding-pruned

Base model

Qwen/Qwen2.5-0.5B
Quantized
(84)
this model