Structurally Prune Anything (SPA)
Structurally Prune Anything (SPA)
A Universal Neural Network Pruning Framework
Overview
SPA is a groundbreaking structured pruning framework that addresses the limitations of existing pruning methods by providing universal compatibility across neural network architectures, deep learning frameworks, and training stages.
Key Features
- Universal Compatibility: Works with any neural network architecture (CNNs, Transformers, ResNets, etc.)
- Framework Agnostic: Compatible with PyTorch, TensorFlow, JAX, and other frameworks
- Flexible Timing: Supports pre-training, post-training, and fine-tuning scenarios
- ONNX Integration: Leverages standardized computational graphs for seamless operation
Technical Innovation
- Group-Level Importance Estimation: Novel algorithm for identifying and grouping dependent computational operators
- Optimal Brain SPA (OBSPA): State-of-the-art post-training pruning without requiring fine-tuning or calibration data
- Automated Pipeline: No manual intervention required for different architectures
Results
- Competitive performance across diverse architectures (ResNet, VGG, Transformer models)
- Significant computational savings (up to 80% parameter reduction)
- Maintained accuracy across different pruning ratios
Paper | Project Page |