Back to glossaryDefinition

What is Data Pipeline?

An automated workflow that extracts, transforms, and loads data from sources to destinations.

A data pipeline is an automated series of processes that move data from one or more sources to one or more destinations, typically involving extraction (pulling data from source systems), transformation (cleaning, enriching, and restructuring), and loading (writing to the target system). Modern pipelines handle batch and real-time streaming, include data quality checks, and support observability and monitoring.

Put this into practice

Assess your maturity, discover initiatives, and build your transformation roadmap.

Start free assessment