Open Access Green as soon as Postprint is submitted to ZB.
Investigating Structural Pruning and Recovery Techniques for Compressing Multimodal Large Language Models: An Empirical Study.
In: (Pattern Recognition). Berlin [u.a.]: Springer, 2026. 320 - 336 (Lect. Notes Comput. Sc. ; 16125 LNCS)
While Multimodal Large Language Models (MLLMs) demonstrate impressive capabilities, their substantial computational and memory requirements pose significant barriers to practical deployment. Current parameter reduction techniques primarily involve training MLLMs from Small Language Models (SLMs), but these methods offer limited flexibility and remain computationally intensive. To address this gap, we propose to directly compress existing MLLMs through structural pruning combined with efficient recovery training. Specifically, we investigate two structural pruning paradigms—layerwise and widthwise pruning—applied to the language model backbone of MLLMs, alongside supervised finetuning and knowledge distillation. Additionally, we assess the feasibility of conducting recovery training with only a small fraction of the available data. Our results show that widthwise pruning generally maintains better performance in low-resource scenarios with limited computational resources or insufficient finetuning data. As for the recovery training, finetuning only the multimodal projector is sufficient at small compression levels (<20%). Furthermore, a combination of supervised finetuning and hidden-state distillation yields optimal recovery across various pruning levels. Notably, effective recovery can be achieved with as little as 5% of the original training data, while retaining over 95% of original performance. Through empirical study on two representative MLLMs, i.e., LLaVA-v1.5-7B and Bunny-v1.0-3B, this study offers actionable insights for practitioners aiming to compress MLLMs effectively without extensive computation resources or sufficient data.
Altmetric
Additional Metrics?
Edit extra informations
Login
Publication type
Article: Conference contribution
Keywords
Model Compression ; Multimodal Llms ; Pruning
ISSN (print) / ISBN
0302-9743
e-ISSN
1611-3349
Conference Title
Pattern Recognition
Quellenangaben
Volume: 16125 LNCS,
Pages: 320 - 336
Publisher
Springer
Publishing Place
Berlin [u.a.]