Deep Learning on 33,000,000 data points using a few lines of YAML

May 4th, 2020 - Hamza Tahir


Over the last few years at maiot, we have regularly dealt with datasets that contain millions of data points. Today, I want to write about how the we use our machine learning platform, the Core Engine, to build production-ready distributed training pipelines. These pipelines are capable of dealing with millions of datapoints in a matter of hours. If you also want to build large-scale deep learning pipelines, sign up for the Core Engine for free here and follow along.