Researchers from MIT and several European institutions have developed a new technique called CompreSSM that makes artificial intelligence models leaner and faster during the learning phase itself, according to a recent announcement.
The new method integrates compression directly into the AI training process, avoiding the typical post-training model shrinking that can degrade accuracy. This approach enables powerful AI systems with significantly lower computational and energy costs, streamlining the creation of sophisticated models for applications from natural language processing to robotics.
What We Know So Far
- A team of researchers from MIT's Computer Science and Artificial Intelligence Laboratory (CSAIL), the Max Planck Institute, the European Laboratory for Learning and Intelligent Systems, ETH, and Liquid AI developed the new method, named CompreSSM, as reported by MIT News.
- The technique compresses AI models during the training process, rather than as a separate step afterward, which helps avoid the typical performance loss associated with post-training compression.
- According to the researchers, CompreSSM uses mathematical tools from control theory to identify and remove redundant or unimportant components from the model early in its training.
- The process is effective because the relative importance of a model's different components tends to stabilize after just 10 percent of the total training process has been completed, the MIT report stated.
How MIT Makes AI Models Leaner and Faster
CompreSSM builds efficiency into the AI learning process from the start. Unlike conventional model optimization that prunes large, complex models after peak performance, CompreSSM integrates compression as part of learning. Senior author Daniela Rus, director of CSAIL, stated, "What's exciting about this work is that it turns compression from an afterthought into part of the learning process itself.”
The technique specifically targets a class of AI architectures known as state-space models (SSMs), which are increasingly used in complex sequential data tasks like language processing and audio generation. According to the research team, CompreSSM dynamically identifies which parts of the neural network are least critical to the learning task and removes them on the fly. "It's essentially a technique to make models grow smaller and faster as they are training," stated lead author Makram Chahine.
The performance gains reported by the researchers are notable. In tests on image classification benchmarks, models trained with CompreSSM were up to 1.5 times faster. In one specific experiment, a compressed model achieved higher accuracy on the CIFAR-10 dataset than a model that was designed to be small from the beginning. On Mamba, a prominent state-space model, the method reportedly achieved training speedups of approximately 4x by compressing a 128-dimensional model down to about 12 dimensions.
| Model Training Approach | Final State Dimension | CIFAR-10 Accuracy |
|---|---|---|
| Trained with CompreSSM | ~25% of original | 85.7% |
| Trained at small size from scratch | ~25% of original | 81.8% |
Benefits of Efficient AI Model Learning
CompreSSM reduces the time and computational resources needed to train high-performing AI models. By making models inherently leaner, the method lowers the barrier to entry for developing advanced generative AI and other sophisticated systems, which currently demand massive and expensive computing infrastructure.
CompreSSM's efficiency aligns with the AI industry's broader trend to mitigate the technology's significant energy footprint. As models grow, their power consumption during training and deployment is a major concern. According to an IEEE Spectrum report, decentralized training is a related effort, utilizing networks of underused hardware—like servers in different time zones or solar-powered homes—to create distributed AI training hubs.
Though distinct from CompreSSM, decentralized training strategies also aim to make AI development more sustainable. The IEEE Spectrum report suggests these approaches reduce carbon emissions by shifting computational loads to times and places with abundant renewable energy, underscoring the industry's focus on resource-efficient artificial intelligence.










