Member-only story
Microsoft’s Open Sourced a New Library for Extreme Compression of Deep Learning Models
DeepSpeed Compression combines compression and system optimization techniques for building smaller and more efficient deep learning architectures.
I recently started an AI-focused educational newsletter, that already has over 125,000 subscribers. TheSequence is a no-BS (meaning no hype, no news etc) ML-oriented newsletter that takes 5 minutes to read. The goal is to keep you up to date with machine learning projects, research papers and concepts. Please give it a try by subscribing below:
Large neural networks have been dominating the deep learning space for the last few years. While the performance of large deep learning architectures is certainly impressive, its operational requirements remain prohibited for most organizations. Not surprisingly, there have been a lot of effort in areas like model compression that can help reduce the size of and inference computation deep learning models. Similarly, there have also…