AIMER: Calibration-Free Task-Agnostic MoE Pruning

Zongfang Liu1,2, Shengkun Tang3, Yifan Shen3, Huan Wang2†, Xin Yuan2†
1Zhejiang University, 2Westlake University, 3Mohamed bin Zayed University of Artificial Intelligence
Corresponding authors: wanghuan@westlake.edu.cn, xyuan@westlake.edu.cn

Abstract

Mixture-of-Experts language models increase parameter capacity without proportional per-token compute, but deployment still requires storing all experts, making expert pruning important for reducing memory and serving overhead. Existing task-agnostic expert pruning methods are typically calibration-dependent: they estimate expert importance from routing or activation statistics on a calibration set, which makes pruning outcomes sensitive to the choice of calibration set and adds substantial preprocessing cost. We introduce AIMER (Absolute mean over root mean square IMportance for Expert Ranking), a simple calibration-free criterion that yields clear within-layer score separation and distinct expert stratification. Across 7B to 30B MoE language models at 25% and 50% pruning ratios over 16 benchmarks, AIMER consistently delivers competitive or stronger overall performance against state-of-the-art calibration-based expert pruning baselines with only 0.22-1.27 seconds for scoring the experts.

Overview of AIMER

AIMER is a calibration-free criterion for task-agnostic MoE expert pruning. For each expert, it combines the gate, up, and down projection matrices into a single weight vector and scores that expert by its mean absolute value normalized by its root mean square.

AIMER score

AIMER(w) = ||w||1 / (sqrt(N) ||w||2)

Comparison of statistics requirements for AIMER and calibration-based baselines

Method Calibration Data Expert Activations Router Weights
Frequency ×
SEER ×
EAN ×
REAP
AIMER (Ours) × × ×

Main Results

Layer-wise comparison between raw magnitude and AIMER score profiles for OLMoE-7B, ERNIE-21B, and Qwen3-30B.

Layer-wise Magnitude and AIMER score profiles across three MoE models. Compared with raw magnitude, AIMER yields a more separable within-layer distribution over experts, making pruning decisions less ambiguous near the pruning boundary.

Efficiency

OLMoE-7B

0.22 s

AIMER scoring time vs 0.75 h for REAP.

ERNIE-21B

0.51 s

AIMER scoring time vs 1.37 h for REAP.

Qwen3-30B

1.27 s

AIMER scoring time vs 2.96 h for REAP.

Model REAP Time AIMER Time REAP Peak Memory AIMER Peak Memory Loading Memory After / Before
OLMoE-7B 0.75 h 0.22 s 15.51 GB 13.00 GB 6.89 / 12.89 GB
ERNIE-21B 1.37 h 0.51 s 44.72 GB 40.85 GB 21.67 / 40.66 GB
Qwen3-30B 2.96 h 1.27 s 63.07 GB 57.01 GB 29.93 / 56.92 GB

BibTeX

@misc{liu2026aimercalibrationfreetaskagnosticmoe,
  title={AIMER: Calibration-Free Task-Agnostic MoE Pruning},
  author={Zongfang Liu and Shengkun Tang and Yifan Shen and Huan Wang and Xin Yuan},
  year={2026},
  eprint={2603.18492},
  archivePrefix={arXiv},
  primaryClass={cs.LG},
  url={https://arxiv.org/abs/2603.18492},
}