MCNC: Manifold Constrained Network Compression
Tech ID: 33976 / UC Case 2025-419-0
Abstract
Researchers at Vanderbilt University and the University of California,
Davis have developed MCNC software that significantly compresses large AI
models while maintaining their performance using a novel manifold-constrained
optimization approach.
Full Description
The Manifold-Constrained Model
Compression (MCNC) software introduces an approach to compress foundational AI
models like Vision Transformers and large language models such as GPT and
LLaMA. Unlike traditional methods, MCNC uses manifold-constrained optimization
to achieve over 100 times compression, enhancing storage, communication, and
customization efficiency without compromising model performance.
Applications
- Efficient deployment of complex AI models on smartphones,
IoT devices, and other hardware with constrained resources.
- Seamless model customization and fine-tuning for
personalized AI applications.
- Reduction in communication overhead for cloud-based AI
services and platforms
Features/Benefits
- Achieves high compression rates of over 100x without significantly compromising performance.
- Enhances efficiency in storage and communication, making it ideal for edge devices with limited memory.
- Enables high throughput, facilitating rapid model construction.
- Offers flexibility and compatibility with various AI models and other compression techniques.
- Overcomes challenges in storing and transmitting large-scale AI models due to their massive size.
- Addresses the limitations of edge devices in running sophisticated AI models by managing memory and processing constraints.
- Resolves memory bottlenecks encountered during model customization and fine-tuning.
- Reduces high data transfer costs and latency in model communication.
Patent Status
Patent Pending