βš™οΈKF Pipelines

What is Kubeflow Pipelines?

Kubeflow Pipelines (KFP) is a platform for building and deploying portable and scalable machine learning (ML) workflows using Docker containers.

With KFP you can author componentsarrow-up-right and pipelinesarrow-up-right using the KFP Python SDKarrow-up-right, compile pipelines to an intermediate representation YAMLarrow-up-right, and submit the pipeline to run on a KFP-conformant backend such as the open source KFP backendarrow-up-right.

The open source KFP backendarrow-up-right is available as a core component of Kubeflow or as a standalone installation. Follow the installationarrow-up-right instructions and Hello World Pipelinearrow-up-right example to quickly get started with KFP.

Why Kubeflow Pipelines?

KFP enables data scientists and machine learning engineers to:

  • Author end-to-end ML workflows natively in Python

  • Create fully custom ML components or leverage an ecosystem of existing components

  • Easily manage, track, and visualize pipeline definitions, runs, experiments, and ML artifacts

  • Efficiently use compute resources through parallel task execution and through caching to eliminating redundant executions

  • Maintain cross-platform pipeline portability through a platform-neutral IR YAML pipeline definitionarrow-up-right

What is a pipeline?

A pipelinearrow-up-right is a definition of a workflow that composes one or more componentsarrow-up-right together to form a computational directed acyclic graph (DAG). At runtime, each component execution corresponds to a single container execution, which may create ML artifacts. Pipelines may also feature control flowarrow-up-right.

Last updated