Data integration, solved rewritten.

The enterprise-grade data integration platform — now with a real agent inside. CDC, ETL, ELT, APIs, files, at petabyte scale. Built for data teams. Faster with AI.

270+
Connectors
Petabyte
scale
SOC 2
+ HIPAA
Real-time
CDC
Etlworks Composer with Simba AI agent showing a Salesforce to Snowflake flow
How it works

From idea to production. Your way.

1

Design with intent

Architect data flows the way your business needs them. Templates, connectors, and best practices ready out of the box.

2

Build it your way

Click through the visual canvas. Write the code. Or hand it to the agent and watch it build. Same engine, same flows — pick the path that fits the task.

3

Run in production with confidence

Petabyte-scale execution. Real-time monitoring. SOC 2, HIPAA, SSO, RBAC, on-prem and hybrid deployments. The agent can watch alongside, alerting and debugging when things drift.

Trusted by data teams at

Platform

Everything you need. One platform.

A complete data integration platform — every technique, every direction, fully integrated.

Capabilities

Proof & depth

Composer

The new way to build flows in Etlworks. Drop connections on the canvas, wire them up, chat with Simba — no flow type to pick first.

  • Connect anything to anything
  • Drag-and-drop, code, or chat
  • Backed by the full platform underneath
Learn more about Composer

What’s underneath

3,979 pre-built templates

For known source-destination pairs. Discoverable by AI search; Simba uses them when building flows.

  • File-based pipelines2,478
  • Reverse ETL539
  • Data warehousing364
  • API integrations198
  • CDC & streaming165
  • Database replication65
200+ flow types · the engine primitives

Connectors

Pre-built integrations for the modern data stack — relational and NoSQL databases, SaaS apps, file formats, and APIs.

  • 65+ databases & warehouses
  • 115+ SaaS apps
  • 60+ file formats & protocols
  • Custom API connectors via configuration
Browse all connectors
270+ pre-built connectors
  • Databases & warehouses 65+
  • SaaS apps 115+
  • Files & protocols 60+
  • Custom (config-built) unlimited

ETL, ELT, reverse ETL

Move data any direction with the technique that fits the workload — visual transforms or full code, batch or streaming.

  • Drag-and-drop transforms with live preview
  • SQL, JavaScript, Python, Groovy code transforms
  • Reverse ETL — sync warehouse data back to SaaS
Learn more about ETL/ELT

Any direction · Any technique

Transform first Load first
Forward
ETL Classic
source → transform → warehouse

Pre-shape data before loading. Best for clean schemas.

ELT Pushdown
source → warehouse → transform

Land raw, transform in Snowflake or BigQuery.

Reverse
Reverse ETL Activation
warehouse → transform → SaaS

Push enriched data to Salesforce, HubSpot, etc.

Streaming ETL Continuous
source → transform mid–stream → sink

Continuous transforms over CDC or message streams.

Mix techniques in one workspace. Same engine, same connectors.

Real-time CDC

Sub-second change data capture from your databases to anywhere — warehouses, databases, lakes, queues, or files. Log-based streaming, no Kafka required.

  • MySQL, Postgres, SQL Server, Oracle, DB2, MongoDB sources
  • Stream to warehouses, databases, lakes, queues, or files
  • Schema evolution, exactly-once delivery, replay
  • Built-in — no Kafka or Debezium required
Learn more about CDC
In production
<1s
latency
10K+/sec
events peak
100%
exactly-once

log–based streaming · schema evolution · replay

Security & compliance

Compliance-grade security baked into every layer of the platform — same controls in cloud, hybrid, and on-prem deployments.

  • SOC 2 Type II, HIPAA, GDPR
  • SSO via SAML
  • Role-based access control with audit logs
  • End-to-end encryption + customer-managed keys
Read the security overview
Verified compliance

SOC 2 Type II

audited

Annual refresh · report on request

HIPAA

compliant

BAA available, not gated

GDPR

compliant

DPA published

Pen testing

annual

Third-party · summary on request

Same posture, every deployment

Cloud Hybrid On-prem

Scale & reliability

Petabyte engine, horizontal and vertical scaling, high availability, and fault tolerance. The same engine in every customer environment.

  • Petabyte-scale processing in production
  • Horizontal & vertical scaling
  • Active-passive HA with automatic failover
  • Self-recovery from failures and pressure
See real production scale
In production today
500B+ records replicated · single customer
1,500+ MySQL DBs synced via CDC
99.999% pipeline uptime
1,000+ autonomous agents · zero-touch

Verified by customers

Tallink · Intertek Alchemy · OpenGov · BioUrja

API integration

Any external API as source or destination. Publish your own APIs from any flow. Pagination, auth, and retry handled — so the integration actually works in production.

  • External APIs as source or destination — REST, SOAP, GraphQL, OData
  • Publish pull, push, CRUD APIs from your flows
  • Auto-pagination · 9 auth protocols · retry & throttle handled
  • Sync or async · webhooks both directions
Learn more about API integration

Any API · Any direction

External APIs — source or destination in any flow

GET api.salesforce.com/v59/query
POST api.hubspot.com/crm/v3/objects
POST api.stripe.com/graphql

Etlworks-published APIs — clients call your endpoints

GET your–domain.com/api/v1/orders pull
POST your–domain.com/api/v1/customers CRUD
POST your–domain.com/in/webhook push

What’s handled for you

auto-pagination auto-retry throttle handling OAuth 1/2 JWT HMAC NTLM AWS SigV4 +3 more

Developer experience

Built for builders. Full CLI, scripting in your language, version control on every flow — and an agent API so your AI stack can drive Etlworks programmatically.

  • Full CLI — wraps every platform action
  • JavaScript, Python, SQL transforms — host your own scripts
  • Agent API for LangChain, CrewAI, AutoGen
  • Built-in version control, export/import, patch, recycling bin
Read developer docs
Three developer surfaces
CLI Wraps every platform action
$ etlworks flow run salesforce_sync
  ↳ 11,902 rows · 4.2s · v34
$ etlworks deploy --env prod
Scripts Inline transforms in JS, Python, SQL
function transform(row) {
  row.email = row.email.toLowerCase();
  return row;
}
Agent API For LangChain, CrewAI, AutoGen
from etlworks import AgentClient
client = AgentClient(api_key=...)
client.create_flow(prompt="sync stripe to snowflake")

Platform plumbing

version control export/import patch recycling bin self-hosted scripts

Comparing data integration platforms? See how Etlworks stacks up against 30+ alternatives

Deployment

Run it where your data lives.

Same engine, every environment. Pick the deployment that fits your security posture and operational maturity.

Cloud SaaS

Fastest to start. Cloud-hosted platform, cloud data sources.

Hybrid agents

SaaS UX, on-prem data. Outbound 443 only — no inbound ports.

Full on-prem

Air-gappable. Defense, banking-isolated, classified networks.

See deployment options
Pricing

Pick a plan. Start free.

Unlimited users, connections, flows, and RBAC on every plan. 14-day free trial — no credit card, no sales call.

Starter

For small teams getting started

$300/ month
3M records · shared infrastructure
  • 5 scheduled flows, hourly
  • Unlimited users + connections
  • REST API access
  • $10/mo AI agent credits — vibe-build, tools
  • Custom REST APIs
  • Real-time CDC
  • Dedicated instance
Start free 14-day trial
Business

For growing teams

$600/ month
30M records · shared infrastructure
  • 20 scheduled flows, hourly
  • Unlimited users + connections
  • REST API + custom APIs
  • On-premise data via agent
  • $10/mo AI agent credits — vibe-build, tools
  • Real-time CDC
  • Dedicated instance
Start free 14-day trial
Enterprise

For dedicated deployments

$1,000+ / month
Unlimited records · dedicated instance
  • Unlimited flows, real-time
  • Real-time CDC + streaming
  • Dedicated, isolated instance
  • SSO + multitenancy
  • $20/mo AI agent credits — vibe-build, tools
  • On-prem and hybrid deployment
  • SOC 2, HIPAA, GDPR
Start free 14-day trial
Need on-prem licensing, perpetual licenses, or extended support? See full pricing →

14 days. No card. Real workloads.

Spin up a free trial, point it at your real data, see what production on Etlworks actually feels like.