Cloud Data Migration: Best Practices & Tools for 2026
Plan your cloud data migration with tools from AWS, Azure, and Google Cloud. Includes migration patterns, best practices, and real examples like Samsung 1.1B user migration.

Cloud data migration moves your databases, files, and applications from on-premises servers to AWS, Azure, or Google Cloud. Samsung migrated 1.1 billion users from Oracle to Amazon Aurora using AWS Database Migration Service. Here's how to plan yours.
Companies now allocate $95 million of their $329 million IT budget to cloud products and services. That number keeps growing—63% of IT leaders reported accelerated cloud migrations in 2024. Whether you're moving a single database or an entire data center, this guide covers the tools, patterns, and practices you need.
What Is Cloud Data Migration?
Cloud data migration transfers data from one location to another, typically from on-premises infrastructure to a cloud environment. This includes:
- Databases: Moving production databases to managed services like Amazon RDS, Azure SQL, or Cloud SQL
- Files and objects: Transferring file servers and blob storage to S3, Azure Blob, or Cloud Storage
- Applications: Migrating application data alongside the workloads that use it
- Data warehouses: Moving analytics platforms to Snowflake, BigQuery, or Redshift
The migration path you choose depends on your timeline, budget, technical debt tolerance, and long-term architecture goals.
Migration Patterns: The 6 Rs
Every migration fits into one of six patterns. Pick the wrong one and you'll waste months. Pick the right one and you'll save 30% on infrastructure costs like Western Sydney University did with Azure.
1. Lift and Shift (Rehosting)
Move applications to the cloud with minimal changes. You're essentially copying your on-premises setup to virtual machines in the cloud.
When to use it:
- Datacenter lease is expiring
- You need to migrate fast (weeks, not months)
- Applications work fine as-is
Trade-offs:
- Fastest migration path
- Carries technical debt into the cloud
- You won't get cloud-native cost benefits
- May over-provision resources
Example: A financial services firm rehosting their reporting database to minimize disruption during fiscal year-end.
2. Re-platforming (Lift, Tinker, and Shift)
Make minor modifications to leverage managed services. You're not rewriting the application, but you're swapping out components.
When to use it:
- Database can move to a managed service (RDS, Cloud SQL)
- ETL jobs can run on managed Spark or serverless functions
- You want moderate optimization without major rewrites
Trade-offs:
- Balances speed with optimization
- Requires some development time
- Better long-term cost profile than lift-and-shift
Example: A retail company refactoring ETL jobs from on-premises Informatica to AWS Glue or Azure Data Factory.
3. Re-architecting (Refactoring)
Redesign applications for cloud-native capabilities. This means microservices, serverless functions, and managed databases designed from scratch.
When to use it:
- Application needs major modernization anyway
- Scalability requirements will grow significantly
- You want maximum cloud benefits
Trade-offs:
- Longest timeline and highest effort
- Maximum cloud-native benefits
- Risk of scope creep
- Requires skilled cloud architects
Example: Rebuilding a monolithic e-commerce platform as microservices on Kubernetes with serverless data pipelines.
4. Rebuilding
Build applications from scratch when legacy code is too outdated or incompatible with cloud platforms.
When to use it:
- Legacy technology stack has no cloud equivalent
- Application provides competitive differentiation
- Clean-slate design would be faster than migration
Trade-offs:
- Most resource-intensive approach
- Full freedom to innovate
- Longest timeline
- Requires parallel systems during transition
5. Replacing (SaaS Adoption)
Replace custom applications with SaaS solutions. Why migrate your CRM when Salesforce exists?
When to use it:
- Application is a commodity (CRM, HRIS, ERP)
- Maintenance burden outweighs customization benefits
- SaaS solutions have matured for your use case
Trade-offs:
- Fastest time to value
- Limited customization
- Vendor lock-in risk
- Data migration to new format required
6. Hybrid Cloud
Keep some workloads on-premises while moving others to the cloud. This isn't a cop-out—it's often the right choice.
When to use it:
- Regulatory requirements mandate data residency
- Legacy systems can't migrate but modern apps need cloud
- Risk-averse organization needs gradual transition
Trade-offs:
- Added complexity in management
- Integration challenges between environments
- Flexibility to migrate at your own pace
- Maintains stability for regulated workloads
Cloud Provider Migration Tools
Each major cloud provider offers native migration tools. Here's what they do, what they cost, and when to use them.
AWS Migration Tools
AWS Database Migration Service (DMS)
AWS DMS handles database migrations with minimal downtime. Samsung used it to migrate 1.1 billion users across three continents from Oracle to Amazon Aurora.
What it does:
- Migrates databases to RDS, Aurora, Redshift, DynamoDB, or S3
- Supports Oracle, SQL Server, PostgreSQL, MySQL, MongoDB, and MariaDB as sources
- Provides continuous replication until cutover
- Includes AI-assisted schema conversion with up to 90% automated conversion rates
Pricing: Hourly based on instance type, with a serverless scaling option. Free tier available for getting started.
Key capabilities:
- Near-zero downtime migrations
- Multi-AZ redundancy for high availability
- SSL/TLS encryption in transit
- Data validation and re-synchronization if discrepancies occur
Best for: Any database migration to AWS managed databases, especially when downtime must be minimal.
AWS DataSync
DataSync automates data transfers between on-premises storage and AWS.
What it does:
- Transfers files from NFS, SMB, HDFS, or self-managed object storage to S3, EFS, or FSx
- Runs up to 10x faster than open-source tools
- Automatically encrypts and validates data
Pricing: Pay per GB transferred.
Best for: File server migrations, NAS consolidation, or ongoing data synchronization between on-premises and cloud.
AWS Migration Hub and Snowball: Status Changes
Two popular AWS tools have changed status:
AWS Migration Hub is no longer open to new customers as of November 7, 2025. AWS now directs customers to AWS Transform for enterprise migration planning.
AWS Snowball Edge is also no longer accepting new customers. If you need offline data transfer for petabyte-scale migrations, AWS recommends:
- AWS DataSync for online transfers
- AWS Data Transfer Terminal for physical transfers
- AWS Partner solutions for specialized needs
If you're already using these services, they continue to function. But new migrations should plan around these alternatives.
Microsoft Azure Migration Tools
Azure Migrate
Azure Migrate provides a unified platform for cloud migration with AI-assisted discovery and assessment.
What it does:
- Discovers and assesses servers, databases, and web applications
- Generates business cases with cost estimates
- Maps dependencies between systems
- Plans phased migrations with application awareness
Pricing: Free with your Azure subscription. AI assistance features are in early access at no additional cost.
Supported workloads:
- VMware and Hyper-V virtual machines
- Physical servers
- Databases (SQL Server, MySQL, PostgreSQL)
- Web applications
- VDI environments
Customer results:
- Western Sydney University: 30% cost savings
- Del Monte Foods: 99.99% uptime and 57% infrastructure savings
Best for: Enterprise-wide assessment and migration planning when moving multiple workloads to Azure.
Azure Database Migration Service
Azure DMS handles database migrations to Azure SQL Database, Azure SQL Managed Instance, and SQL Server on Azure VMs.
What it does:
- Migrates SQL Server, MySQL, PostgreSQL, and MongoDB databases
- Provides near-zero downtime online migrations
- Integrates with Azure Data Studio for schema assessment
- Automates through PowerShell cmdlets
Best for: Database migrations from on-premises or other clouds to Azure SQL services.
Google Cloud Migration Tools
Database Migration Service
Google's DMS is serverless and free for homogeneous migrations (same database engine on both sides).
What it does:
- Migrates MySQL, PostgreSQL, SQL Server, and Oracle databases
- Targets Cloud SQL and AlloyDB for PostgreSQL
- Uses Gemini AI for schema and code conversion
- Provides continuous replication with minimal downtime
Pricing:
- Free for homogeneous migrations (MySQL to Cloud SQL for MySQL)
- Per-byte pricing for heterogeneous migrations (Oracle to PostgreSQL)
Customer results:
- Accenture: Migrated to Cloud SQL with BigQuery federation
- Ryde: Moved production databases from Amazon RDS to Cloud SQL in less than a day
Best for: Migrations targeting Cloud SQL or AlloyDB, especially when you want AI-assisted schema conversion.
Cloud Provider Comparison
| Feature | AWS DMS | Azure DMS | Google DMS |
|---|---|---|---|
| Homogeneous pricing | Hourly instance | Included with Azure | Free |
| Heterogeneous pricing | Hourly instance | Included with Azure | Per-byte |
| AI schema conversion | Yes (90% rates) | Assessment only | Gemini-powered |
| Supported sources | 10+ engines | SQL Server, MySQL, PostgreSQL, MongoDB | MySQL, PostgreSQL, SQL Server, Oracle |
| Supported targets | RDS, Aurora, Redshift, DynamoDB, S3 | Azure SQL, SQL MI, SQL VM | Cloud SQL, AlloyDB |
| Serverless option | Yes | No | Yes |
Choose AWS DMS when: You're migrating to Amazon Aurora or need the broadest source database support.
Choose Azure DMS when: You're already on Azure or need tight integration with Azure SQL Managed Instance.
Choose Google DMS when: You want free homogeneous migrations or need Gemini AI for complex schema conversions.
Third-Party Migration Tools
Cloud provider tools work well for their own ecosystems. But what if you're moving data between clouds, or from dozens of SaaS applications to a data warehouse? Third-party tools fill that gap.
Fivetran
Fivetran is a fully managed data pipeline platform with 700+ pre-built connectors.
What it does:
- Extracts data from SaaS apps, databases, files, and events
- Loads into Snowflake, BigQuery, Redshift, Databricks, and other destinations
- Manages schema changes automatically
- Syncs every 15 minutes (Standard) or 1 minute (Enterprise)
Pricing tiers:
- Free: Up to 500K monthly active rows, 5K monthly model runs
- Standard: Starts around $500/month for 1M monthly active rows
- Enterprise: 1-minute syncs, custom roles, VPN tunnels
- Business Critical: Customer-managed encryption, PCI DSS Level 1, private networking
- Enterprise License Agreement: Fixed annual price with unlimited consumption
Best for: Teams that want fully managed ELT pipelines without building connectors. Strong choice when you need compliance certifications (SOC 2, HIPAA, PCI DSS).
Airbyte
Airbyte is an open-source data integration platform with 600+ connectors.
What it does:
- Moves data from databases, SaaS apps, files, and APIs
- Supports any destination (warehouses, lakes, operational systems)
- Offers self-hosted (free) or cloud-managed options
- Includes a connector builder for custom sources
Pricing tiers:
- Core (Self-hosted): Free, open-source
- Standard (Cloud): Volume-based pricing
- Plus: Capacity-based with annual billing
- Pro: Faster syncs, RBAC, row filtering
- Enterprise: SSO, SCIM, private networking
Best for: Teams that want open-source flexibility with the option for cloud management. Good choice when you need custom connectors or want to avoid vendor lock-in.
Stitch (Now Part of Qlik Talend Cloud)
Stitch was a simple cloud ETL tool with 130+ connectors. It's been acquired by Qlik and integrated into Qlik Talend Cloud.
What happened: Stitch's functionality now lives within Qlik Talend Cloud. If you're evaluating Stitch, you'll need to look at Qlik Talend Cloud pricing instead.
Best for: Existing Stitch customers or organizations already using Qlik Talend for data integration.
Third-Party Tool Comparison
| Feature | Fivetran | Airbyte | Stitch/Qlik Talend |
|---|---|---|---|
| Connectors | 700+ | 600+ | 130+ (in Talend) |
| Open source | No | Yes | No |
| Self-hosted option | No | Yes | No |
| Free tier | 500K rows/month | Unlimited (self-hosted) | Trial only |
| Fastest sync | 1 minute (Enterprise) | Real-time CDC | Varies |
| dbt integration | Built-in | Available | Available |
Choose Fivetran when: You want turnkey managed pipelines and don't mind the cost.
Choose Airbyte when: You want open-source flexibility or need custom connectors.
Choose Qlik Talend when: You're already in the Qlik ecosystem or need broader data integration beyond ELT.
Migration Best Practices
Migration tools only work when you use them correctly. Here's a step-by-step checklist.
Phase 1: Assessment and Planning
1. Discover and inventory all data assets
You can't migrate what you don't know exists. Catalog every database, file share, and data pipeline. Document:
- Data size and growth rate
- Data sensitivity and compliance requirements
- Current performance baselines
- Downstream dependencies
2. Map data dependencies
Identify how applications use data. A database migration fails if the application pointing to it doesn't get updated. Map:
- Applications that read from each data source
- ETL jobs and their schedules
- Reporting tools and dashboards
- APIs that expose data
3. Generate a business case
Calculate the total cost of ownership for both options:
- Current on-premises costs (hardware, licensing, personnel, facilities)
- Projected cloud costs (compute, storage, egress, managed services)
- Migration costs (tools, consulting, downtime)
- Expected savings and timeline to break even
Use Azure Migrate's business case generator or AWS Migration Evaluator for data-driven estimates.
4. Prioritize workloads
Not everything migrates at once. Create migration waves based on:
- Business criticality (start with low-risk workloads)
- Technical complexity (build team skills on simpler migrations)
- Dependencies (migrate data before applications that use it)
- Quick wins (migrations that prove value fast)
5. Choose migration patterns per workload
Match each application to the right pattern (lift-and-shift, re-platform, re-architect). Different workloads in the same organization often need different approaches.
Phase 2: Security and Compliance
1. Encrypt data in transit
Never transfer data over unencrypted connections. Use:
- SSL/TLS for database connections
- VPN or AWS Direct Connect / Azure ExpressRoute for network transfers
- Private networking options from your migration tools
2. Encrypt data at rest
Cloud providers offer encryption by default, but verify:
- Storage encryption is enabled (S3 default encryption, Azure Storage encryption)
- Database encryption is enabled (RDS encryption, Azure SQL TDE)
- Consider customer-managed keys for sensitive workloads
3. Implement access controls
Apply least-privilege principles:
- Create dedicated IAM roles for migration tools
- Use temporary credentials where possible
- Audit who has access to migration pipelines
- Remove access after migration completes
4. Maintain compliance
Verify your target environment meets regulatory requirements:
- GDPR: Data residency and right to deletion
- HIPAA: Business Associate Agreements with cloud providers
- PCI DSS: Network segmentation and encryption
- SOC 2: Access controls and monitoring
Azure invests $1 billion annually in cybersecurity R&D and employs 3,500+ security experts. AWS and Google Cloud have similar programs. But compliance is your responsibility—the tools just help.
5. Protect credentials
Use secrets management instead of hardcoded passwords:
- AWS Secrets Manager
- Azure Key Vault
- Google Secret Manager
- HashiCorp Vault
Phase 3: Testing
1. Test in a mirror environment
Never migrate directly to production. Create a target environment that mirrors production:
- Same database versions and configurations
- Same network topology
- Representative data (anonymized if necessary)
2. Validate data integrity
After migration, verify data matches:
- Row counts between source and destination
- Checksums for critical tables
- Sample data spot checks
- Automated validation scripts (AWS DMS provides built-in validation)
3. Performance testing
Cloud resources perform differently than on-premises:
- Run query benchmarks against migrated databases
- Test application response times
- Verify batch job completion times
- Load test at expected production levels
4. User acceptance testing
Involve business users before cutover:
- Have them verify reports show correct data
- Test application workflows end-to-end
- Confirm integrations still work
5. Plan for rollback
Things go wrong. Have a failback plan:
- Keep source systems running during parallel period
- Document rollback procedures
- Test rollback before production cutover
- Define criteria that trigger rollback
Phase 4: Cutover
1. Choose your cutover approach
Big bang migration: Move everything at once during a maintenance window.
- Pros: Single cutover, no prolonged parallel running
- Cons: Higher risk, requires perfect execution
- Best for: Smaller databases or when downtime is acceptable
Phased migration: Move data in incremental waves.
- Pros: Lower risk, learn from early waves
- Cons: Longer total timeline, complexity of parallel systems
- Best for: Large enterprises with many workloads
Continuous replication with cutover: Sync data continuously, then switch over.
- Pros: Minimal downtime (minutes instead of hours)
- Cons: Requires replication tooling, more complex
- Best for: Production databases that can't tolerate downtime
2. Schedule during low-activity periods
Migrate when impact is lowest:
- Overnight or weekends
- After month-end close for financial systems
- During slow business seasons
3. Execute cutover checklist
On cutover day:
- Final data sync
- Verify data integrity one more time
- Update application connection strings
- DNS changes if needed
- Monitor for errors
- Verify applications function correctly
- Keep source available for rollback window
Phase 5: Post-Migration Optimization
1. Right-size resources
Initial resource allocation is often wrong. After migration:
- Monitor actual CPU, memory, and I/O usage
- Downsize over-provisioned instances
- Scale up resources that are bottlenecked
- Switch to appropriate instance families
2. Implement cost controls
Cloud bills can surprise you. Set up:
- Budget alerts before overage
- Reserved instances or savings plans for steady workloads
- Spot instances for fault-tolerant batch jobs
- Auto-scaling to match demand
3. Archive cold data
Not all data needs hot storage:
- Move infrequently accessed data to S3 Glacier, Azure Archive, or Coldline Storage
- Set lifecycle policies to automate tiering
- Delete data you no longer need (check retention requirements first)
4. Monitor continuously
Migration isn't done until the system runs reliably:
- Set up cloud-native monitoring (CloudWatch, Azure Monitor, Cloud Monitoring)
- Create alerts for performance degradation
- Review costs weekly for the first few months
- Document lessons learned for future migrations
Common Migration Challenges
Challenge: Underestimating data volume
Problem: A migration that should take hours takes days because you didn't account for total data size.
Solution: Profile your data before starting. Measure not just table sizes but also transaction logs, blob storage, and file shares. Add 20% buffer to your estimates.
Challenge: Network bandwidth limitations
Problem: Transferring petabytes over a 1Gbps connection takes months.
Solution: Calculate transfer times before starting. For large datasets, consider:
- AWS DataSync for optimized online transfer
- Physical transfer options (AWS Data Transfer Terminal, Azure Data Box)
- Incremental sync with continuous replication
Challenge: Application compatibility
Problem: Application breaks after database migration because of subtle differences in SQL syntax or behavior.
Solution: Test applications thoroughly in a mirror environment. Pay attention to:
- Date/time handling differences
- Case sensitivity in queries
- Stored procedure compatibility
- Connection pooling behavior
Challenge: Data transformation requirements
Problem: Source and target schemas don't match, requiring complex transformations.
Solution: Use schema conversion tools (AWS SCT, Gemini in Google DMS) for automated conversion. Plan manual intervention for:
- Custom data types
- Complex stored procedures
- Database-specific features
Challenge: Downtime windows too short
Problem: Business can't tolerate the downtime required for migration.
Solution: Use continuous replication to minimize cutover window:
- AWS DMS continuous replication
- Azure DMS online migration mode
- Google DMS continuous replication
- Keep replication running until final cutover (minutes instead of hours)
When to Use ImportCSV
Cloud migration tools handle the heavy lifting of moving data between systems. But what about the last mile—getting data from spreadsheets into your new cloud databases?
ImportCSV handles CSV and Excel imports directly into your databases:
- Direct database imports: Push CSV data into PostgreSQL, MySQL, Snowflake, BigQuery, and more
- Schema inference: Automatically detect column types from your spreadsheet data
- Data validation: Catch errors before they hit your database
- No code required: Business users can import without writing SQL
After your cloud migration completes, ImportCSV helps teams load operational data that lives in spreadsheets—the inventory counts, pricing updates, and configuration data that doesn't fit neatly into automated pipelines.
Conclusion
Cloud data migration involves more decisions than just "which tool should I use." You need to:
- Choose the right migration pattern for each workload (lift-and-shift for speed, re-platform for balance, re-architect for maximum benefit)
- Select appropriate tools based on your source, destination, and requirements
- Plan thoroughly with proper assessment, security, testing, and cutover procedures
- Optimize after migration because initial configurations are rarely perfect
Samsung migrated 1.1 billion users. Del Monte achieved 99.99% uptime and cut infrastructure costs by 57%. Ryde moved production databases in less than a day. Your migration can succeed too—with the right planning and tools.
Start with assessment. Map your data. Pick your patterns. Choose your tools. Test thoroughly. Then migrate with confidence.
Wrap-up
CSV imports shouldn't slow you down. ImportCSV aims to expand into your workflow — whether you're building data import flows, handling customer uploads, or processing large datasets.
If that sounds like the kind of tooling you want to use, try ImportCSV .