Log in

Every organization handles sensitive data. It lives in training datasets, test environments, customer support logs, legal contracts, medical records, and countless other places. Whether you're preparing data for model training, generating realistic test datasets, or processing documents for review, the challenge is the same: you need to find and protect personally identifiable information (PII) without destroying the usefulness of the data.
Tonic Textual already solves this problem. It detects and transforms PII across text, files, and audio in 50+ languages, with fine-grained control over how each entity type is handled; masking, synthesis, or deterministic replacement. Thousands of teams use Textual to prepare safe training corpora, generate compliant test data, and redact sensitive documents at scale.
Today, we're making Textual even more accessible: we're releasing an open-source MCP server that brings Textual's full redaction capabilities directly into AI agent workflows.
The Model Context Protocol (MCP) is an open standard for connecting AI assistants to external tools and data sources. Think of it as a universal plug that lets AI agents take real actions; not just generate text, but call APIs, process files, and interact with the systems you already use.
MCP is supported by a growing ecosystem of clients including Claude Desktop, Cursor, Windsurf, and custom-built agents. By publishing a Textual MCP server, we're making PII redaction a first-class capability in any of these environments.
The server exposes 21 tools that cover the full breadth of Textual's capabilities:
Process plain text, JSON, XML, and HTML directly. Bulk mode handles multiple strings in a single call. JSON redaction supports JSONPath expressions for surgical control; specify exactly which paths to redact or leave untouched.
Upload PDFs, Word documents, spreadsheets, and images for redaction. The original formatting is preserved; a redacted PDF still looks like a PDF, not a stripped-down text dump. File processing runs in the background so your agent isn't blocked waiting for large files to complete.
Point the deidentify_folder tool at a directory and it handles everything: detects file types, routes text files through the fast inline API and binary files through the upload pipeline, processes them in parallel, and writes redacted output alongside the originals. Filter by extension, skip specific patterns, and even redact folder and file names.
Create datasets in the Textual platform, upload files for scanning, and download redacted results; all through your AI agent. This integrates Textual's dataset workflow directly into automated pipelines.
Every tool supports Textual's rich configuration options:
npm install -g @tonicai/textual-mcp
Set your Textual API key as an environment variable:
export TONIC_TEXTUAL_API_KEY=your-api-key-here
For self-hosted Textual instances, also set:
export TONIC_TEXTUAL_BASE_URL=https://your-textual-instance.example.com
For Claude Desktop, add the server to your MCP configuration:
{
"mcpServers": {
"textual": {
"command": "textual-mcp",
"env": {
"TONIC_TEXTUAL_API_KEY": "your-api-key-here"
}
}
}
}Then ask your AI assistant to redact text, process a file, or de-identify an entire folder — Textual handles the rest.
For the technically curious:
We'd love your feedback. Try the server, let us know what you build with it, and open an issue if you run into anything. PII redaction should be as easy as asking your AI assistant — and now it is.
Adam Kamor, Co-founder and Head of Engineering at Tonic.ai, leads the development of synthetic data solutions that enable AI and development teams to unlock data safely, efficiently, and at scale. With a Ph.D. in Physics from Georgia Tech, Adam has dedicated his career to the intersection of data privacy, AI, and software engineering, having built developer tools, analytics platforms, and AI validation frameworks at companies such as Microsoft, Kabbage, and Tableau. He thrives on solving complex data challenges, transforming raw, unstructured enterprise data into high-quality fuel for AI & ML model training, to ultimately make life easier for developers, analysts, and AI teams.