Convenient and flexible ML pipelines with Kubeflow

Logo
Presented by

Mattias Arro, Machine Learning Engineer, Subspace AI

About this talk

It is still early days for open source solutions for productionalising and deploying machine learning (ML) models, managing scalable data pipelines and data science experiments. Kubeflow is a collection of tools that are perfect for these use cases and is gaining popularity for a good reason. This talk describes a system built on top of Kubeflow which is generic enough to be used for managing ML pipelines of various shapes and sizes, yet flexible enough to allow entirely custom workflows. At its core, there is a set of conventions which determine where data is read from and written to, and expressing data preprocessing and models as a configuration of composable objects and functions. This approach makes it trivial to add new models, datasets, and training objectives to a production system, and enables training and deploying stacked models of arbitrary complexity.
Related topics:

More from this channel

Upcoming talks (0)
On-demand talks (265)
Subscribers (55718)
Dataiku is the world’s leading platform for Everyday AI, systemizing the use of data for exceptional business results. Organizations that use Dataiku elevate their people (whether technical and working in code or on the business side and low- or no-code) to extraordinary, arming them with the ability to make better day-to-day decisions with data. More than 450 companies worldwide use Dataiku to systemize their use of data and AI, driving diverse use cases from fraud detection to customer churn prevention, predictive maintenance to supply chain optimization, and everything in between.