← Back to Index

🚀 CSV Generator Pro

Quick Start Guide & What's New

Version 2.9.1

⚡ Quick Start Guide

Get up and running with CSV Generator Pro in just 5 minutes. Follow these simple steps to generate your first dataset.

1

Select Your Fields

Choose from 41+ available field types including:

  • Personal: ID, firstName, lastName, email, phone, address
  • Location: city, state, zipCode, country
  • Business: company, jobTitle, department, salary
  • Products: product, category, price, sku
  • Dates: date, timestamp, hireDate

Example: For a customer list, select: ID, firstName, lastName, email, phone, country, city

2

Set Your Row Count

Choose how many records you want to generate (1 to 1,000,000).

  • Small test: 10-100 rows (instant)
  • Medium dataset: 1,000-10,000 rows (~1-5 seconds)
  • Large dataset: 100,000+ rows (~10-60 seconds)
💡 Pro Tip Start with 100 rows to verify your field selections, then scale up to your target size.
3

Generate Your Data

Click the "Generate Data" button. Your data will appear in the preview table within seconds.

What you'll see:

  • Real-time generation progress
  • Interactive data preview table
  • Multi-column sorting (click headers)
  • Search and filter capabilities
  • Pagination controls
4

Choose Your Format

Select your preferred output format:

📄 CSV

Universal format for Excel, spreadsheets, and maximum compatibility.

📝 NDJSON

Perfect for log files, streaming data, and line-by-line processing.

📦 Parquet

Best for S3/Athena with 80-90% smaller files and faster queries.

5

Export or Upload

Two ways to get your data:

Option A: Download Locally

  • Click "Download CSV", "Download NDJSON", or "Download Parquet"
  • File saves to your default downloads folder
  • Open in Excel, upload to database, or use in your application

Option B: Upload to AWS S3

  • Configure AWS credentials (Access Key ID, Secret Key)
  • Set bucket name and region
  • Optional: Set S3 directory path with partitioning
  • Click "Upload to S3"
✅ That's It! You've just generated your first dataset. Now explore advanced features like deterministic IDs, file splitting, batch processing, and Parquet format.

🎉 What's New in v2.9.1

🔄 Batch Processing NEW

Process all saved configurations automatically with one click!

  • Automated workflow: Generate and upload multiple datasets sequentially
  • Pause mode: Review each dataset before S3 upload
  • Progress tracking: Real-time status for each configuration
  • Final summary: Success/failed/skipped counts and timing

Perfect for: Setting up test environments, demo databases, or bulk S3 population

⚙️ Per-Config Split Settings IMPROVED

Split preferences now saved with each configuration!

  • Each preset can have its own Split by Date setting
  • Each preset can have its own Split by Fields setting
  • Batch processing respects per-config split preferences
  • No more manual toggling between configurations

📊 Optimized Built-in Presets IMPROVED

All 12 presets now optimized for Athena/Redshift analytics:

  • Realistic row counts: 500 to 500,000 rows based on use case
  • Hive-style partitioning: Automatic year=yyyy/month=mm/day=dd paths
  • Split settings: High-volume presets have splits pre-enabled
  • Extended date ranges: 3-year spans for time-series testing

🆔 ID Calculator Tool NEW

Calculate deterministic IDs without regenerating data:

  • Enter person's name, email, and date
  • Instantly see their deterministic ID
  • Perfect for verification and debugging
  • Supports all ID methods (Basic, Standard, Enhanced, Auto)

📥 Import & Format Conversion IMPROVED

Enhanced file import capabilities:

  • Parquet import: Read Parquet files directly in browser
  • Field validation: Clear error messages for incompatible fields
  • Smart format detection: Auto-detects CSV, NDJSON, or Parquet
  • Round-trip support: Import → Edit → Export in different format

📜 Version History

v2.9.0 - Batch Processing Release

  • Added batch processing for all configurations
  • Added pause mode for step-by-step uploads
  • Added progress tracking and final summary
  • Fixed split settings not saving in configurations

v2.8.1 - Parquet & Import Release

  • Added Parquet file import support
  • Added CSV and NDJSON import
  • Added field validation on import
  • Improved error handling for incompatible fields

v2.7.0 - Parquet Export Release

  • Added Parquet export format
  • Columnar format with compression
  • Athena/Redshift optimization
  • 80-90% file size reduction vs CSV

v2.6.1 - ID Calculator & Bug Fixes

  • Added ID Calculator tool
  • Fixed case-sensitivity in deterministic IDs
  • Improved S3 path normalization
  • Enhanced UX with button state management

v2.5.0 - Deterministic IDs

  • Added deterministic ID generation
  • Multiple methods (Basic, Standard, Enhanced, Auto)
  • Perfect for SQL joins across datasets
  • Configurable ID ranges

v2.0.0 - Major Feature Release

  • AWS S3 direct upload
  • Hive-style partitioning support
  • File splitting by date and fields
  • Configuration management system
  • 12 built-in presets

✨ Key Features Overview

41+ Field Types

Personal, location, business, product, date, and technical fields.

Multiple Formats

Export as CSV, NDJSON, or Parquet with format conversion.

Deterministic IDs

Consistent IDs across datasets for realistic SQL joins.

S3 Integration

Direct upload with Hive-style partitioning for Athena.

File Splitting

Split by date ranges or field values for parallel loading.

Batch Processing

Generate and upload multiple configurations automatically.

Configuration Presets

Save and reuse settings with 12 built-in templates.

Import & Convert

Import existing files and convert between formats.

Advanced Preview

Multi-column sort, search, pagination, inline editing.

ID Calculator

Calculate deterministic IDs without regenerating data.

Privacy First

100% client-side processing - data never leaves browser.

No Installation

Works entirely in your browser - no software to install.

🎯 Next Steps

1. Explore Advanced Features

  • Deterministic IDs: Create related datasets with matching IDs
  • S3 Upload: Configure AWS credentials and upload directly
  • File Splitting: Partition data by date or field values
  • Batch Processing: Generate multiple datasets automatically

2. Save Your Configurations

  • Click "Save" to store your field selections
  • Name your configurations descriptively
  • Export all configs as backup
  • Load built-in presets for common scenarios

3. Test Different Formats

  • CSV: Try opening in Excel
  • NDJSON: Test with streaming tools
  • Parquet: Upload to S3 and query with Athena

4. Read the Complete Guide

For detailed documentation on all features, see the Complete Help Guide

📚 Additional Resources