dltHub

Build with dlt.

Ship with dltHub Pro.

dlt is the open-source Python library 10,000s of developers use to build data pipelines. dltHub Pro is the agentic platform that deploys, monitors, and scales them. One command. No manual environment setup. No silent failures.

OPEN SOURCE

dlt

The AI-native Python library for data movement. Write any pipeline, run anywhere, no backend needed.

$pip install dlt
Docs9,700+ sources · Agent-friendly docs

IN PUBLIC PREVIEW

Launch May 2026

dltHub Pro

Agents build your dlt pipelines from a prompt. Pro deploys them to production with scheduling, alerting, and observability — one command, zero manual setup.

DocsAgent-friendly docs
dltHub Pro

"What I didn't expect is how much it unblocks the team. A mid-level engineer can spin up a prototype, browse the raw data in dltHub Pro's local DuckDB workspace, validate the SQL schema - all without pulling in a senior. That loop of prototype, inspect, fix, re-run - that's the real unlock."

Marcello Victorino

Marcello Victorino

Staff Data Engineer, Tasman Analytics

THE AGENTIC DATA WORKFLOW

Describe it. The agent ships it. You verify it.

Available in dlt
Claude Code·~/crm-pipeline
>

Build a pipeline that loads CRM contacts and deals into my warehouse using dlt

Agentic Workflows

Complete agentic workflows for every phase of data engineering

Not autocomplete, not a chatbot on a dashboard. A guided sequence of skills, commands, rules, and MCP - with guardrails agents can't skip. Maintained by dltHub, controlling the infrastructure agents and pipelines operate on.

REST API Pipelinedlt
dltHub RuntimePRO
Data Explorationdlt
TransformationsPRO
Foundationdlt

Agentic Workflows in Detail

Discover individual skills per agentic workflow

See how each workflow guides your agent - step by step, from first prompt to production deployment.

REST API Pipeline1/7

Find a dlt source for a given API or data provider. Use when the user asks about a source, wants to find a connector, or asks to implement a pipeline for a specific data source.

Sonnet 4.6 · REST API Pipeline · ~/pipelines

? for shortcuts

dlt is the leading open-source Python library for building data pipelines using code and agents.

10M+

PyPI downloads per month

10,000+

Companies loading data into databases with dlt in production

800+

Companies loading into Snowflake with dlt in production

Agents now build 10x more dlt data pipelines than humans

In January 2025, the dlt community created 2,400 pipelines, almost entirely by hand. By January 2026, that number had grown to 81,000. The shift is not just the 34x year-over-year volume increase: agents now build roughly 10x more pipelines per month than human developers, accounting for ~91% of all new pipelines. The same inflection seen with databases such as Neon or Supabase is happening with data pipelines, only faster.

By agents
By humans
Early 2025
2026
10x more pipelines deployed by agents
Quotation mark icon{testimonial.author?.name}

The current machine learning revolution has been enabled by the Cambrian explosion of Python open-source tools that have become so accessible that a wide range of practitioners can use them. As a simple-to-use Python library, dlt is the first tool that this new wave of people can use. By leveraging this library, we can extend the machine learning revolution into enterprise data.

Quotation mark icon
Julien Chaumond
CTO/Co-Founder at Hugging Face
Quotation mark icon{testimonial.author?.name}

Python and machine learning under security constraints are key to our success. We found that our cloud ETL provider could not meet our needs. dlt is a lightweight yet powerful open source tool we can run together with Snowflake. Our event streaming and batch data loading performs at scale and low cost. Now anyone who knows Python can self-serve to fulfil their data needs.

Quotation mark icon
Maximilian Eber
CPTO & Co-Founder at Taktile

Learn agentic data engineering

Free, self-paced course. From first prompt to production deployment.

An image with the command "pip install "dlt[hub]" in the middle, and logos of REST API sources around it
DLTHUB CONTEXT

The context your agent needs to ship any pipeline

dltHub's agentic workflows come with a REST API toolkit that taps directly into dltHub Context - a hub of deeply researched, enriched context on REST APIs across SaaS sources, databases, and destinations. Your agent pulls exactly what it needs to code any dlt pipeline, in minutes.

We already cover more than 10,100 sources, with a clear path to hundreds of thousands. From prompt to pipeline to live reports in a notebook - all in one agentic flow, with outputs tailored to data users.

Frequently Asked Questions

How is dltHub Pro different from a Claude skill or tools like Replit?

Tools like Claude skills or Replit are great for writing and running code. But they are not built for data engineering workflows end to end.

dltHub Pro gives your team complete agentic workflows that cover every phase: coding, running, deploying, and debugging pipelines, on infrastructure you control. Not just a skill, not just an editor, but a guided workflow from first line to production.

How is dlt different from Fivetran or a Python script that uses the request library?

dlt is the perfect match between standardization and customization. You get the automation that matters: schema inference, incremental state, normalization, and loading, while keeping the full flexibility and portability of plain Python.

And with agentic dltHub workflows, your team can code, run, deploy, and debug pipelines faster, with the reliability you can trust at every step.

What is dltHub?

dltHub is the managed platform for deploying and operating data pipelines built with dlt. It provides a runtime, observability, data quality checks, and collaboration features so teams can go from development to production with one command.

What is dlt?

dlt (data load tool) is an open-source Python library for building data pipelines. It lets you write any connector, run anywhere, and requires no backend. dlt is Apache 2.0 licensed and always free to use.