Products

Developer data solutions for secure software and AI development

Tonic Structural

Safe, realistic structured and semi-structured synthetic data generation for hydrating staging environments, enabling local development, and fueling secure analytics.
Annual commitment / volume discounts available. Contact Us
De-identify your data without losing its utility.
Shrink down your database while maintaining referential integrity.
Accelerate your release cycles and ensure regulatory compliance.
Pricing based on source data volume.
Commit annually and save.

Pay-as-you-go

For developers and small teams with simple use cases who need to protect their sensitive data.

Adapt to changing business needs without overcommitting budgets.

Starting at $199/month for 20 tables

Usage
Source Data
Up to 2TB
Workspaces
Unlimited
Generated Data
Unlimited
Databases
Unlimited
Deployment
Tonic Cloud
Features
Comprehensive Generator Library
Privacy Scan
Cross-Table Consistency
Schema Change Alerts
Virtual Foreign Keys
Concurrent Generations
Subsetting with Referential Integrity
Ephemeral databases
Built-in expiration timers for ephemeral environments
Upsert
w/o Schema Differences
Encryption Support
Workspace Inheritance
Generator Presets
Data Sources
Data Connector Types
Supported data types
Choose 2 Types
Application Databases
PostgreSQL
MySQL / MariaDB
MongoDB
DocumentDB
DynamoDB
Vertica
Db2 LUW
Oracle
Data Warehouses
Snowflake
BigQuery
Redshift
Data Lakes (on Spark)
Databricks
Amazon EMR w/ Glue
Spark
SaaS Applications
Salesforce
Flat Files
Plain Text
JSON
CSV / TSV
XML
HTML
Don’t see a data source listed? Chances are we’re working on it. Reach out for additional details.
Collaboration
Users
10
Workspace Sharing
Commenting and Annotation
Role Based Permissions
Automation
REST API
Basic
Schema Change Alerts
Webhooks
Post Job Actions
Compliance & Security
Authentication
Tonic Auth or Google SSO
License Tier
Professional
Certifications
SOC 2 Type II


AWS Qualified Software

HIPAA Compliant
Privacy Report
Audit Report

Professional

The simplest and fastest way to generate highly realistic, privacy-preserving synthetic data that reflects production.

Access committed spend discounts and a customer success manager with an annual contract.


Book a demo
Usage
Source Data
Up to 10TB
Workspaces
Unlimited
Generated Data
Unlimited
Databases
Unlimited
Deployment
Tonic Cloud
Features
Comprehensive Generator Library
Privacy Scan
Cross-Table Consistency
Schema Change Alerts
Virtual Foreign Keys
Concurrent Generations
Subsetting with Referential Integrity
Ephemeral databases
Built-in expiration timers for ephemeral environments
Upsert
w/o Schema Differences
Encryption Support
Generator Presets
Custom Sensitivity Rules
Workspace Inheritance
Data Sources
Data Connector Types
Supported data types
Choose 2 Types
Application Databases
PostgreSQL
MySQL / MariaDB
MongoDB
DocumentDB
DynamoDB
Vertica
Db2 LUW
Oracle
Data Warehouses
Snowflake
BigQuery
Redshift
Data Lakes (on Spark)
Databricks
Amazon EMR w/ Glue
Spark
SaaS Applications
Salesforce
Flat Files
Plain Text
JSON
CSV / TSV
XML
HTML
Don’t see a data source listed? Chances are we’re working on it. Reach out for additional details.
Collaboration
Users
10
Workspace Sharing
Commenting and Annotation
Role Based Permissions
Automation
REST API
Basic
Schema Change Alerts
Webhooks
Post Job Actions
Compliance & Security
Authentication
Tonic Auth or Google SSO
License Tier
Professional
Certifications
SOC 2 Type II


AWS Qualified Software

HIPAA Compliant
Privacy Report
Audit Report

Enterprise

The complete, self-managed solution to enable consistent, unified compliance and streamlined productivity across the organization, including all of your data and use cases at scale.


Usage
Source Data
Unlimited
Workspaces
Unlimited
Generated Data
Unlimited
Databases
Unlimited
Deployment
Tonic Cloud or Self Hosted
Features
Comprehensive Generator Library
Privacy Scan
Cross-Table Consistency
Schema Change Alerts
Virtual Foreign Keys
Concurrent Generations
Subsetting with Referential Integrity
Ephemeral databases
Built-in expiration timers for ephemeral environments
Upsert
w/o Schema Differences
Encryption Support
Workspace Inheritance
Generator Presets
Custom Sensitivity Rules
Data Sources
Data Connector Types
Supported data types
Choose Unlimited
Application Databases
PostgreSQL
MySQL / MariaDB
MongoDB
DocumentDB
DynamoDB
Vertica
Db2 LUW
Oracle
Data Warehouses
Snowflake
BigQuery
Redshift
Data Lakes (on Spark)
Databricks
Amazon EMR w/ Glue
Spark
SaaS Applications
Salesforce
Flat Files
Plain Text
JSON
CSV / TSV
XML
HTML
Don’t see a data source listed? Chances are we’re working on it. Reach out for additional details.
Collaboration
Users
Unlimited
Workspace Sharing
Commenting and Annotation
Role Based Permissions
Automation
REST API
Advanced
Schema Change Alerts
Webhooks
Post Job Actions
Compliance & Security
Authentication
SSO / SAML
License Tier
Enterprise
Certifications
SOC 2 Type II


AWS Qualified Software

HIPAA Compliant
Privacy Report
Audit Report
Usage
Pay-as-you-go
Professional
Enterprise
Source Data
Up to 2TB
Up to 10TB
Unlimited
Workspaces
Unlimited
Unlimited
Unlimited
Generated Data
Unlimited
Unlimited
Unlimited
Databases
Unlimited
Unlimited
Unlimited
Deployment
Tonic Cloud
Tonic Cloud
Tonic Cloud or Self Hosted
Features
Comprehensive Generator Library
Privacy Scan
Cross-Table Consistency
Schema Change Alerts
Virtual Foreign Keys
Concurrent Generations
Subsetting with Referential Integrity
Ephemeral databases
Built-in expiration timers for ephemeral environments
Upsert
w/o Schema Differences
w/o Schema Differences
with Schema Differences
Encryption Support
Generator Presets
Custom Sensitivity Rules
Workspace Inheritance
Data Sources
Data Connector Types
Choose 2 Types
Choose 2 Types
Choose Unlimited
Application Databases
PostgreSQL
MySQL / MariaDB
SQL Server
MongoDB
DocumentDB
DynamoDB
YugabyteDB
Db2
Oracle
SAP ASE
Data Warehouses
Snowflake
BigQuery
Redshift
Data Lakes (on Spark)
Databricks
Amazon EMR w/ Glue
Spark
SaaS Applications
Salesforce
Flat Files
Plain Text
JSON
CSV / TSV
XML
HTML
Avro
Parquet
Don’t see a data source listed? Chances are we’re working on it. Reach out for additional details.
Collaboration
Users
10
10
Unlimited
Workspace Sharing
Commenting & Annotation
Role-Based Access Controls
Automation
REST API
Basic
Basic
Advanced
Schema Change Alerts
Webhooks
Post Job Actions
Compliance & Security
Authentication
Tonic Auth or Google SSO
Tonic Auth or Google SSO
SSO / SAML
License Tier
Professional
Professional
Enterprise
Certifications
SOC 2 Type II


AWS Qualified Software

HIPAA Compliant
SOC 2 Type II


AWS Qualified Software

HIPAA Compliant
SOC 2 Type II


AWS Qualified Software

HIPAA Compliant
Privacy Report
Audit Report
Not ready to commit? Start your free trial to experience Tonic Structural before you decide.
Common questions
How does the Pay-As-You-Go plan work?

The Pay-As-You-Go plan offers rapid access to our cloud offering of Tonic Structural, making it ideal for simple use cases and small data sizes. The plan entails a monthly credit card subscription that is billed automatically and includes up to 20 unique tables processed per month, as many times as you need. If you exceed 20 unique tables, we will issue a separate charge to your credit card on file at our on-demand table rate of $19 per table per month.

Each billing period, we count the number of unique tables that are processed across all generations (jobs), workspaces, and databases connected to Tonic. We only count tables that are processed by Tonic to the destination database. This includes:



  1. Tables that are configured using the following modes: De-Identify, Scale, Preserve Destination, Incremental;

  2. Tables that are included in a subsetting run to maintain referential integrity (in-subset tables); or,

  3. Any table with generators applied, including those that contain foreign keys that are de-identified in the parent table.

We do not count tables without generators applied (”passthrough”) or those in Truncated mode.

Can I use Tonic Subsetting without using other Tonic Structural platform features?

Yes, we support the option to purchase a limited license for access to Subsetting only, in conjunction with Tonic Ephemeral.

How do you price annual contracts?

Annual pricing for Tonic Structural is determined by the Plan you select and the amount of source data connected to Tonic. Source data is defined as the total size on disk of data sources connected to Tonic Structural, excluding logs and indexes. Our pricing model builds in volume discounts, so the incremental price per GB of source data decreases as you connect more data to Tonic Structural.

What cloud providers do you support?

Tonic Structural is compatible with databases deployed on all major cloud providers, including AWS, Microsoft Azure, and GCP.

Does Tonic sign contractual data handling agreements?

Yes, for Enterprise plan customers, we will sign Business Associate Agreements (BAA) and Data Processing Agreements (DPA). Tonic.ai is also SOC2 certified, GDPR compliant, and HIPAA compliant, and Tonic Cloud is an AWS Qualified Software.

Tonic Fabricate

Synthetic database generation with full relational integrity across unlimited tables and domains for new product development, model training, and edge case testing.
Annual commitment / volume discounts available. Contact Us
Spin up fully synthetic databases in seconds to turbocharge development.
Mock APIs easily by simulating real API responses.
Accelerate time-to-market and sales with realistic test and demo data.
Free to start, then pay for the data you generate.

Free forever

For developers with simple use cases who need to spin up smaller datasets for early product development, prototyping, and testing.

  • Up to 20 tables
  • 100 rows per table
  • 1 user
  • 1 workspace

Professional

Starting at $499/year

For those looking to generate realistic, relational datasets to drive new product development where no data yet exists.

  • Up to 50 tables
  • 10,000 rows per table
  • 5 users
  • 1 workspace

Start here

Enterprise

For organizations in need of a self-managed solution to augment their enterprise data strategy with streamlined databases synthesized from scratch.

  • Unlimited users & data generation
  • Bring your own LLM
  • Deploy on your infrastructure

Common questions

What are the best use cases for data generated with Fabricate?

Fabricate is purpose-built for greenfield product development, where production data is inexistent, limited, or otherwise inaccessible. It enables developers to create fully relational, realistic databases from scratch using a schema, sample data, or natural language. Behind the scenes LLMs interpret and translate these inputs into intelligent data generation. This makes the data ideal for new product development, model training, and edge case testing.

Do I need existing data (e.g. production data) to use Fabricate?

No, you do not need any existing or sample data to begin generating data in Fabricate. You can start from a schema or use natural language prompts to define the data that you would like Fabricate to generate.

Where can I find the Fabricate product docs?

You can find the Fabricate product docs at this link: https://fabricate.mockaroo.com/docs.

Tonic Textual

All-in-one data platform designed to eliminate integration and privacy challenges ahead of RAG ingestion or LLM training.
Annual commitment / volume discounts available. Contact Us
Extract, govern, enrich, and deploy your unstructured data for Gen AI
Elevate RAG performance and accuracy with entity metadata tags
Protect sensitive and proprietary data used for generative AI
Usage-based pricing based on words processed.
Only pay for what you use.
Features
  • Standardize, enrich, and secure your unstructured data for AI development
  • Configure enterprise-grade unstructured data pre-processing pipelines in minutes
  • Enrich vector stores with NER-powered entity metadata tags to improve RAG performance
  • Automatic redaction and reversible tokenization of sensitive entities ensures privacy is maintained
  • Optional synthetic data replacement helps maintain semantic realism of your data
  • Use Textual pipelines directly in your existing workflows with Python SDK and REST API
  • Unlimited datasets
  • Unlimited custom models
  • Unlimited users
  • Per-word pricing that scales sublinearly
  • Google SSO or Tonic Auth
  • Cloud or Self-Hosted deployments
  • Native support for TXT, DOCX, PDF, CSV, XLS, TIFF, PNG, JPEG, PPTX, PARQUET, HTML, XML, and JSON
Have sensitive text data in Snowflake?
Use our Snowflake Native App to leverage Textual models directly in your Snowflake environments.

Try Textual on Snowflake
Common questions
How do you price Tonic Textual?

Tonic Textual pricing is volume based. You will be charged based on what you use, defined as the number of words processed. Our pricing model scales sublinearly such that larger amounts of text are significantly discounted per unit. Contact us to receive a custom quote.

Are annual contracts available?

Yes, for annual engagements, we will work with you to scope your estimated needs for the year and allocate a bank of words that you are free to use throughout the year.

What are your deployment options?

Our hosted cloud version of Tonic Textual is optimized for speed and performance and is the best option. Textual is also available for self-hosted deployments, as an AMI on the AWS Marketplace, or as a Snowflake Native App.

Bundles

Tonic.ai’s solutions, bundled together to offer you savings and comprehensive use case coverage.

Fabricate + Structural

Scalable data synthesis

Generate synthetic data to hydrate development, test, and staging environments with realistic relational data and mock APls, on demand.

Ephemeral + Subsetting

Targeted data minimization

Get developers the data they need, when and where they need, it with targeted subsets sent directly to on-demand databases

Textual + Structural

Full coverage data protection

Protect sensitive data in structured and unstructured formats with AI-powered data de-identification and synthesis.

The Tonic.ai product suite

Tonic.ai's suite of products is your solution to building data-driven software and AI/ML models. Get all the value of production data without actually having to copy sensitive data around your organization and unlock strategic data assets for use across functions, from engineering, to business operations, to data science/ML teams. With high quality data, teams author fewer defects and ship faster all while having a strong security posture.