S1 · HERO
v2.5 now available · 45 MCP tools
⬡ npm · datavor v2.5

The AI-Native Database
Sync & Pipeline MCP

Sync, transform, schedule, and monitor your data pipelines — with real-time CDC, per-record fault tolerance, proactive suggestions, and dependency-aware scheduling — all through natural language.

No SQL. No complex UIs. No data engineers. Just ask.

$ npm install -g datavor ⧉ copy
Read the Docs →
Free forever for Phase 1 Works with Claude Desktop 5-minute setup
Claude Desktop · Datavor
You
Sync my customers table from MySQL to PostgreSQL
Claude · Datavor
Connected to MySQL — 10,542 rows found
Connected to PostgreSQL
⟳ Syncing in batches of 100...
Done — 10,542 customers synced in 5.2s
You
Now sync only USA customers incrementally
Claude · Datavor
Incremental mode — checking updated_at...
Synced 47 changed rows in 0.3s
Live sync status
🐬
MySQL
✓ customers
✓ orders
✓ products
syncing
🐘
PostgreSQL
✓ customers
✓ orders
· products
10,542 rows · 5.2s · 0 errors
Ready to use with Claude Desktop, Claude Code, and any MCP client ⬇ Download on npm

S2 · HOW IT WORKS

Three steps.
Seconds, not hours.

1
🔌

Install & Connect

Add Datavor to Claude Desktop with one config block. Works on macOS in minutes.

~5 min setup
2
💬

Ask Claude

Tell Claude what you need in plain English. No SQL required — Datavor handles the rest.

natural language
3

Done

Datavor executes precise operations, handles errors, and gives you a full sync report.

instant results
claude_desktop_config.json — the complete setup
{
  "mcpServers": {
    "datavor": {
      "command": "datavor"
    }
  }
}
One config block. Restart Claude. Done.

S3 · FEATURES

Everything your databases need.
Nothing you don't.

Visual Schema Tools

See differences before you sync.

Ask Claude to compare any table between your MySQL and PostgreSQL databases. Get a clean visual diff showing missing columns, type mismatches, and index differences — all before a single row is touched.

  • Side-by-side column comparison
  • Type mismatch highlighting
  • Complete migration analysis across all tables
  • AI-powered recommendations on what to fix first
Schema comparison · customers table
🐬 MySQL
🐘 PostgreSQL
idINT
idINTEGER
nameVARCHAR(100)
nameVARCHAR(100)
emailVARCHAR(100)
emailmissing
created_atDATETIME
created_atTIMESTAMP
countryVARCHAR(50)
countryVARCHAR(50)
⚠ 2 differences found — sync may need type conversion
Incremental Sync

Stop re-syncing everything, every time.

After your first full sync, Datavor remembers the checkpoint. Every subsequent sync fetches only new and changed rows — making daily updates 50–98% faster as your tables grow.

  • Timestamp-based change detection
  • Automatic checkpoint tracking
  • Works with any table that has an updated_at column
  • Falls back to full sync gracefully
Incremental sync performance
SYNCING 10,000 ROWS — CUSTOMERS TABLE
Full sync (first run)
15.0s
all 10,000 rows
Incremental (next day)
0.3s
47 changed rows only
98%
faster than full sync
after first run
⭐ New in v2.0 — Multi-Connector

Five databases. One tool. Any direction.

Datavor v2.0 connects to MySQL, PostgreSQL, SQL Server, SQLite, and Snowflake — and syncs between any combination. One natural-language request, any source to any target.

  • All 5 engines fully supported — read, write, sync, transform
  • Universal type engine with O(n) complexity — no per-pair mappings
  • Cross-engine migrations: MySQL → Snowflake, SQLite → PostgreSQL and more
  • 34 MCP tools across all connectors
Also compatible with cloud databases
Aiven
PlanetScale
AWS RDS
Azure SQL
Supabase
Supported databases — v2.0
Sync between any combination
MySQL
v1.0+
PostgreSQL
v1.0+
SQL Server
v2.0 ⭐
SQLite
v2.0 ⭐
Snowflake
v2.0 ⭐
"Sync orders from MySQL to Snowflake daily"
✅ Connected MySQL · Snowflake
✅ DATETIME→TIMESTAMP_NTZ, JSON→VARIANT
✅ 45,231 rows synced in 18.2s
v1.5 — Scheduler

Set it and forget it.

Create automated sync jobs in plain English. The daemon runs independently — your pipelines keep moving while you sleep.

  • Create jobs in seconds — "sync orders every night at 2am"
  • Natural language to cron: daily, hourly, weekly, monthly
  • Pause, resume, or run any job manually on demand
  • pm2-compatible daemon for 24/7 uptime
v1.5 — Transform Pipeline

Shape data as it moves.

Rename columns, convert types, filter rows, remap values, and add computed columns — all inline during sync. Preview on real data before writing anything.

  • Rename: customer_id → cust_id
  • Cast: string score → float · Filter: active = 1 only
  • Computed: first + last → full_name
  • Preview on 5 sample rows before committing
v1.5 — Sync Dashboard

Every sync recorded. Always visible.

Datavor automatically logs every operation. Ask Claude for a health check — success rates, row counts, failure diagnostics, and table history.

  • Summary: success rate, rows moved, daily trend
  • Table history: every run with row count and duration
  • Failure diagnostics: error message + specific fix hints
  • Zero config — all data written automatically
v2.5 — Change Data Capture

Real-time replication. No polling.

Stream every INSERT, UPDATE, and DELETE from PostgreSQL (WAL) or MySQL (binlog) into your destination as it happens. Checkpoints survive restarts.

  • PostgreSQL: logical replication (wal_level=logical)
  • MySQL: binary log streaming (binlog-format=ROW)
  • Sub-100ms lag · LSN / binlog position persisted
  • Prerequisites checked automatically before starting
v2.5 — Per-Record Fault Tolerance

Bad rows no longer kill the pipeline.

One malformed row used to abort the whole batch. Now the bulk write tries first, falls back per-row on failure, and isolates offenders — the rest commits.

  • Reports Rows synced / Rows failed on every run
  • Failures captured by type: null_violation, duplicate_key…
  • Full failed-row payloads available via dashboard_failures
  • Error Learner tracks patterns and surfaces fix suggestions
"Create a job to sync orders every night at 2am"
✅ Job created: Nightly orders sync
Schedule: Daily 02:00 (0 2 * * *)
Mode: full · Table: orders

"Show me all my sync jobs"
✅ 4 active jobs
orders_sync Daily 02:00 ✓ last ran 6h ago
customers_sync Daily 02:10 ✓ last ran 6h ago
products_sync Hourly ✓ last ran 44min ago
events_rollup Weekly Sun ✓ last ran 2 days ago
"Preview: rename customer_id→cust_id, cast score to float"
── Preview (5 sample rows) ──────────────────────────
customer_id → cust_id [rename]
score: "8.5" → 8.5 [cast string→float]
✅ Preview looks correct — apply?

"Yes, sync with those transforms"
✅ 45,231 rows synced · 3 transforms applied · 22.1s
"Show me a dashboard summary for the last 7 days"
── Sync Summary (last 7 days) ───────────────────────
Success rate: 98.6%
Total syncs: 42 · Rows moved: 890,432
Most active: orders (18 runs, 0 failures)
Avg duration: 4.2s
❌ 2 failures → run dashboard_failures for details
"Stream orders from prod MySQL to analytics Postgres in real time"
✅ CDC stream started.
Source: production_mysql.orders
Target: analytics_postgres.orders
Method: MySQL binlog (ROW format)
Lag: ~50ms

"Show CDC status"
✅ 1 active stream · 12,847 events applied · lag 48ms
"Sync new_users from staging to production"
✅ Synced 9,847 rows
❌ Failed 3 rows
Failures by type:
null_violation: 2 (email is NOT NULL)
duplicate_key: 1

"Show dashboard_failures"
#3,047 email: NULL — integrity.null_violation
#7,891 duplicate key on user_id=88421
v2.0 — Context Engine

Datavor learns your data over time.

A persistent local knowledge brain silently accumulates everything it learns — schema changes, business rules, sync patterns, and relationships.

  • Remembers schema and detects changes automatically
  • Learns business rules — auto-applies them on future syncs
  • Tracks foreign key relationships across databases
  • "Explain my production database" → full plain-English briefing
v2.5 — Recipe Manager

Save transforms once. Reuse forever.

Stop re-typing the same transform configuration on every sync. Name it, save it, and apply it with one command across any environment.

  • Save any transform set as a named recipe
  • Apply with one command: "apply the cleanup_orders recipe"
  • Filter by connection, table, or tags with list_recipes
  • Frequently-used transforms auto-captured from sync runs
v2.5 — Proactive Suggestions

Datavor tells you what to fix before you notice.

Schema changes, recurring errors, and slow syncs are detected, scored, and surfaced as actionable cards. Accept or dismiss with one command.

  • Schema changes: new column detected on a synced table
  • Error patterns: recurring failures counted and explained
  • Optimisation: full-replace syncs that should be CDC or incremental
  • Dismissed suggestions never re-raised; accepted ones return action_config
v2.5 — Dependency-Aware Scheduling

Wire real pipelines, not timing guesses.

Express job dependencies as a DAG. Cycles are rejected at definition time. Failed parents block children; retries use exponential backoff.

  • scheduler_add_dependency: "B runs after A"
  • scheduler_show_graph: renders the full topology
  • Cycle detection at definition time — invalid DAGs rejected
  • Failed jobs retry 1m → 2m → 4m → 8m backoff
"Explain my production MySQL database"
📖 production_db (MySQL)
Tables: 23 known · Rules: 4 (2 auto-applied)
Syncs: 142 recorded · 97.2% success rate
Rows: 890,000 total · 2 schema changes detected
Since: 2026-01-15

💡 Auto-rule active: exclude_test_orders
email NOT LIKE '%@test.com%' · used 47×
"Save transform set as 'cleanup_orders': lowercase email, drop test rows"
✅ Recipe 'cleanup_orders' saved (recipe_id: 7)

30 days later:
"Apply the cleanup_orders recipe and sync orders to staging"
✅ Recipe loaded · transforms applied · 12,420 rows synced · 0 failed

"List my recipes"
#7 cleanup_orders · used 12× · orders
#12 users_sanitise · used 8× · users
"Show me what Datavor recommends I look at"
3 suggestions pending:
1. [schema_change · 0.96] New column 'discount_pct' on orders
syncs to analytics.orders — destination is missing it
2. [error_fix · 7×] Null violation on shipments since Apr 12
3. [optimisation · 12 runs] events full-replace avg 84s
Switch to CDC or incremental?

"Accept 1 and 3. Dismiss 2 — those NULLs are intentional"
✅ Done. Suggestion 2 will not be raised again.
"Make analytics_rollup wait for orders_sync AND customers_sync"
✅ 2 dependencies added. Cycle check passed.
orders_sync → analytics_rollup
customers_sync → analytics_rollup

"Show me the scheduler graph"
orders_sync [✓ done]
customers_sync [✓ done]
└─ analytics_rollup [⏳ running]
└─ email_digest [⌛ waiting]
v1.0 — Explore & Query

Browse your database like a conversation.

Ask Claude to show a tree of all tables, describe any schema, or run a query in plain English. No GUI needed.

  • Database tree view — all tables, columns, row counts
  • describe_table: types, constraints, foreign keys
  • Natural language queries — Claude writes and runs the SQL
  • compare_table_schemas — side-by-side diff before migration
Database tree view
"Show me a tree of my production database"

📂 production_mysql (23 tables)
├─ orders (142,380 rows)
├─ customers (89,241 rows)
├─ products (4,521 rows)
└─ ... 20 more tables

"Describe the orders table"
✅ 8 columns · 2 indexes · 2 foreign keys detected

S4 · By the numbers

Fast. Free.
Built to scale.

5s
Avg sync time for 10k rows
98%
Faster than full daily backup
$0
Free. No credit card. No time limit.
45
MCP tools available

S5 · Ecosystem

Works wherever
Claude works.

One line of config. Any MCP-compatible client. Claude Desktop, Claude Code, Cursor, Cline — Datavor works with all of them.

👤 You — natural language request
choose your client
Claude Desktop
direct · recommended
Claude Code
direct · CLI
Any MCP Client
Cline, Cursor & more
Datavor MCP
45 MCP tools
MySQL·PostgreSQL·SQL Server·SQLite·Snowflake

S6 · Use Cases

What teams actually
use Datavor for.

🔄Scheduled Sync

Nightly Production → Analytics Sync

Mirror production MySQL to analytics PostgreSQL every night. Zero manual work.

↓ saves ~3h/week
Real-Time CDC

Real-Time CDC Pipeline

Stream every INSERT, UPDATE, DELETE into analytics within 50ms. No polling, no stale reports.

~50ms lag
🔀Data Migration

Data Migration with Transforms

Move tables between engines with inline renames, casts, and value remaps. Save as a recipe for future runs.

reusable recipes
🧠AI Pipelines

AI-Powered Pipeline Management

Datavor learns your schema, auto-applies your rules, spots errors, and suggests optimisations before you notice.

self-managing pipelines

S8 · FAQ

Common questions.

Is Datavor really free?

Yes — completely free in Phase 1. No credit card, no time limit, no feature restrictions. All 45 MCP tools included.

What databases does Datavor support?

MySQL, PostgreSQL, SQL Server, SQLite, and Snowflake — plus cloud variants: Supabase, AWS RDS, Azure SQL, Aiven, and PlanetScale.

What is CDC and do I need it?

Change Data Capture streams every row-level change in real time using PostgreSQL WAL or MySQL binlog. You need it if analytics data needs to be current within seconds. Regular syncs are fine for hourly/daily freshness.

Does Datavor work without Claude Desktop open?

Yes — run node build/scheduler-daemon.js (or use pm2) and all scheduled jobs and CDC streams run 24/7 without Claude.

What MCP clients work with Datavor?

Claude Desktop, Claude Code, Cursor, and Cline — all confirmed. Any MCP-compatible client works with the same one-line config.

Is my data sent to any server?

No. Datavor runs entirely on your machine. Your actual row data never leaves your environment.

What is the Context Engine?

A persistent local knowledge brain that learns from every sync — saving business rules that auto-apply, tracking history, and proactively surfacing suggestions when something looks off.

How do I get started?

Run npm install -g datavor, add one line to your Claude Desktop config, restart Claude, and ask "connect to my database at..." — syncing in under 5 minutes.


One command.
Infinite possibilities.

Install Datavor, add one line to your Claude Desktop config, and start syncing databases through natural language in under 5 minutes.

bash
npm install -g datavor
Get Started — Free →
Free. No account. No credit card. 45 MCP tools.