Site icon The Data Exchange

Why You Should Optimize Your Deep Learning Inference Platform

The Data Exchange Podcast: Yonatan Geifman and Ran El-Yaniv on the benefits that accrue from using an inference acceleration platform.


SubscribeApple • Android • Spotify • Stitcher • Google • RSS.

In this episode of the Data Exchange, I speak1 with Yonatan Geifman, CEO and co-founder of Deci, as well as with Ran El-Yaniv, Chief Scientist and co-founder of Deci and Professor of Computer Science at Technion. As companies deploy machine learning and deep learning to critical products and services, the number of predictions that models have to render can easily reach millions per day (even hundreds of trillions, in the case of Facebook).

These “prediction services” continue to grow in importance – 80% of content on Netflix is discovered through recommenders – and thus companies need to build platforms to serve predictions to an ever growing number of users and services. Deci builds tools to help companies accelerate and scale their inference platform to meet the requirements of their specific application and use case. They do so through an array of tools that looks at inference holistically and systematically:

[Image: Inference Acceleration Stack, from Deci.ai, used with permission.]

Download a complete transcript of this episode by filling out the form below:

Short excerpt:

 

Related content and resources:


[1] This post and episode are part of a collaboration between Gradient Flow and Deci. See our statement of editorial independence.

[Photo by Robynne Hu on Unsplash]

Exit mobile version