Member-only story

OpenAI Helps Us Understand How Deep Learning Training Scales

Understanding the optimimal size of a training set remains one of the most interesting challenges of supervised learning models.

Jesus Rodriguez
3 min readNov 22, 2022
Source: https://datavizionsystems.com/product/ai-102-designing-and-implementing-a-microsoft-azure-ai-solution/

I recently started an AI-focused educational newsletter, that already has over 150,000 subscribers. TheSequence is a no-BS (meaning no hype, no news etc) ML-oriented newsletter that takes 5 minutes to read. The goal is to keep you up to date with machine learning projects, research papers and concepts. Please give it a try by subscribing below:

In the last few yew years, there have been an increasing interest in training parallelization methods that can be applicable to large deep learning models. Those training parallelism efforts have focused on both model-based and data-based approaches with the latter being more popular given their simplicity. Conceptually, data parallelism involves splitting a training dataset into batches of data, distributing those across multiple computing devices and aggregating the resulting gradients. One of the…

--

--

Jesus Rodriguez
Jesus Rodriguez

Written by Jesus Rodriguez

CEO of IntoTheBlock, President of Faktory, President of NeuralFabric and founder of The Sequence , Lecturer at Columbia University, Wharton, Angel Investor...

Responses (1)