ISO 27001:2022 introduced an emphasis on data masking, especially with the introduction of Control 8.11, which states:
“Data masking shall be used in accordance with the organization’s access control policy.”
This sounds pretty straightforward, but in reality, it opens up a wide range of questions for security, privacy, and engineering teams. What does effective masking look like? What kind of data falls under this control? And how can you meet the standard without breaking your workflows?
Control 8.11 was introduced in the 2022 version of ISO 27001, and it focuses specifically on data masking as a required security control. It falls under the theme of access control, meaning it’s not just about encrypting storage or managing permissions, but about limiting what data people can see based on their roles and the context of use.
While the standard was released in October 2022, organizations that are already ISO 27001 certified have until October 31, 2025 to transition to the updated controls. Compliance auditors are already asking about 8.11 in annual surveillance audits, especially if your business handles regulated data or serves enterprise clients.
As you update your policies for Control 8.11 compliance, here are some key approaches to implement in your workflows:
Start with structured data (names, SSNs, credit cards), but don’t forget unstructured data like emails, support chats, PDFs, voice transcripts, and logs, which are all fair game under ISO 27001.
You’ve got options:
For developers, QA teams, and AI engineers, synthetic data is often the only modern and scalable option that meets both compliance and data utility goals.
It is important to understand that not all sensitive data is created equal. The most effective data masking strategies are based on the type of data you're working with, the desired level of realism, and the privacy requirements of your use case.
Here’s a breakdown of common data types and the masking or synthesis approaches best suited to each:
Legacy masking techniques like basic redaction, scrambling, or obfuscation have long been used to protect sensitive data. In most cases, these will be enough to meet ISO 27001 compliance needs.
But they often come at a cost to developers and AI engineers: broken data utility, lost context, and limited test coverage.
Privacy and utility has always been at a tug of war. More privacy generally means limited utility and vice versa. Moreover, most of these legacy approaches were never designed for the complexity of modern development, especially when you’re working across relational databases, multiple data stores, and unstructured formats like emails, PDFs, and transcripts.
More importantly, while they may technically satisfy a checkbox, they rarely deliver the data quality needed to truly shift testing left or support modern AI workflows.
Tonic.ai takes masking to the next level. We go beyond traditional obfuscation to offer high-fidelity data transformation that meets compliance requirements while preserving utility.
With Tonic Structural and Tonic Textual, you get:
Control 8.11 wants you to mask data in accordance with your access policies. Tonic’s synthetic data platform lets you go a step further to generate privacy-safe data that behaves like the real thing, whether you’re spinning up a new test environment or fine-tuning a proprietary LLM.
Tonic.ai gives you a developer-friendly, scalable, and audit-ready way to meet masking requirements across every type of data your org touches.
Compliance doesn’t have to slow you down. If you’re updating your policies for ISO 27001:2022, we’d love to show you how Tonic.ai can help.
Andrew Colombi is the Co-founder and CTO of Tonic.ai. Upon completing his Ph.D. in Computer Science and Astrodynamics from the University of Illinois Urbana-Champaign in 2008, he joined Palantir as an early employee. There, he led the team of engineers that launched the company into the commercial sector and later began Palantir’s Foundry product. His extensive work in analytics across a full spectrum of industries provide him with an in-depth understanding of the complex realities captured in data. Today, he is building Tonic.ai’s platform to engineer data that parallels the complexity of modern data ecosystems and supply development teams with the resource-saving tools they most need.