November 2026data classification policy

Data Classification Policy: How to Classify, Label, and Protect Data for SOC 2, ISO 27001, and HIPAA Compliance

Build a data classification policy for SOC 2, ISO 27001, and HIPAA compliance. Includes classification levels, handling rules, labeling standards, and implementation guide.

By QuickTrust EditorialUpdated 2026-03-19

Data Classification Policy: How to Classify, Label, and Protect Data for SOC 2, ISO 27001, and HIPAA Compliance

Every security control in your organization -- encryption, access control, data loss prevention, backup and retention, incident response -- depends on one underlying assumption: that you know what data you have and how sensitive it is.

Most companies do not.

They encrypt some databases but not others, without a documented rationale for the distinction. They apply the same access controls to marketing materials and customer health records. They store everything in the same cloud environment, under the same retention policy, with the same disposal process. When an auditor asks "how do you determine what level of protection a given data set requires," the answer is silence, or something improvised on the spot.

This is the gap that a data classification policy closes. It is the formal, documented system that assigns sensitivity levels to every category of information an organization handles, then maps those levels to specific handling requirements -- who can access it, how it must be stored, how it can be transmitted, how long it is retained, and how it is destroyed. Without it, your security program is guessing. With it, every downstream control becomes traceable to a documented, risk-based decision about data sensitivity.

This guide covers how to build a data classification policy that satisfies SOC 2, ISO 27001, HIPAA, PCI DSS, and GDPR requirements simultaneously -- the classification scheme, the framework for implementing it, the handling rules, the labeling standards, the training requirements, and the audit evidence that proves it all works.

Get audit-ready policy templates -- including data classification -- from QuickTrust


What Is Data Classification?

Data classification is the process of organizing information into categories based on its sensitivity, value, and the impact that unauthorized disclosure, modification, or destruction would have on the organization and the individuals whose data it holds.

At its simplest, data classification answers one question for every piece of information in your environment: how bad would it be if this data were exposed to someone who should not have it? The answer to that question determines how much protection the data requires.

Why data classification is the foundation of all data security

Data classification is not one security control among many. It is the control on which all other controls depend. Consider:

  • Encryption decisions. You cannot decide what to encrypt without knowing what is sensitive enough to warrant encryption. A data classification policy provides the threshold.
  • Access control. Role-based access control is meaningless without knowing which roles should have access to which sensitivity levels of data. Classification provides the mapping.
  • Data loss prevention. DLP rules that prevent sensitive data from leaving the organization require a definition of what "sensitive" means. Classification provides that definition.
  • Retention and disposal. Different data types have different legal retention requirements and different risk profiles for long-term storage. Classification determines the retention schedule.
  • Incident severity. When a data breach occurs, the severity classification -- and the regulatory notification obligations it triggers -- depends entirely on what type of data was exposed. Classification provides the severity mapping before the incident happens.
  • Audit evidence. When an auditor asks "how do you protect your most sensitive data differently from your least sensitive data," the answer must be a documented classification scheme, not a verbal explanation.

Without data classification, your security program applies controls uniformly across all data, which means one of two things: you are over-protecting low-value data (wasting resources) or under-protecting high-value data (creating risk). Neither outcome is acceptable.

Related: Information Security Policy: The Complete Guide -- data classification is one of the 15 essential sections every information security policy must include.


Why Compliance Frameworks Require Data Classification

Data classification is not a best practice that compliance frameworks suggest. It is a control that every major framework explicitly requires. Here is exactly where the requirement appears in each:

SOC 2 -- CC6.1 and C1.1

SOC 2's Common Criteria address data classification in two places:

  • CC6.1 (Logical and Physical Access Controls): Requires the organization to implement logical access security measures to protect against threats from sources outside its system boundaries. The control environment depends on the organization having identified and classified information assets so that access controls can be applied proportionally.
  • C1.1 (Confidentiality -- Identification and Maintenance of Confidential Information): This criterion explicitly requires that "the entity identifies and maintains confidential information to meet the entity's objectives related to confidentiality." This means the organization must have a documented process for identifying what data is confidential and maintaining that classification over time.

SOC 2 auditors will ask: "How does the organization identify which data is confidential? What criteria are used? Who makes the determination? How are the classifications maintained?" If the answer is not a documented data classification policy, expect a finding.

ISO 27001 -- A.5.12 and A.5.13

ISO 27001:2022 dedicates two Annex A controls specifically to information classification:

  • A.5.12 (Classification of Information): Requires that information be classified according to the information security needs of the organization, based on confidentiality, integrity, and availability requirements, and taking account of legal, regulatory, contractual, and business requirements.
  • A.5.13 (Labelling of Information): Requires that an appropriate set of procedures for information labelling be developed and implemented in accordance with the information classification scheme adopted by the organization.

Certification auditors will examine whether (a) a classification scheme exists, (b) it is documented, (c) it has been communicated to personnel, (d) information has actually been classified according to the scheme, and (e) labelling procedures are followed. A policy that exists on paper but has never been applied to actual data will not satisfy the auditor.

HIPAA -- PHI Classification

HIPAA does not prescribe a generic classification scheme, but it does something more specific: it defines an entire category of data -- Protected Health Information (PHI) -- and mandates specific protections for it.

Under the HIPAA Security Rule (45 CFR 164.312), covered entities and business associates must:

  • Identify all PHI in their environment
  • Implement technical safeguards (access controls, encryption, audit controls, transmission security) specifically for PHI
  • Distinguish between electronic PHI (ePHI) and other organizational data in terms of the protections applied

In practice, this means any organization handling PHI must have a classification system that, at minimum, separates PHI from non-PHI data and applies differentiated controls. A data classification policy is the mechanism for doing so.

PCI DSS -- Cardholder Data Identification

PCI DSS v4.0 Requirement 3 ("Protect Stored Account Data") requires organizations to:

  • Requirement 3.1: Processes and mechanisms for protecting stored account data are defined and understood. This begins with knowing where cardholder data exists.
  • Requirement 3.2: Storage of account data is kept to a minimum.
  • Requirement 3.4: Access to displays of full PAN and ability to copy cardholder data are restricted.

PCI DSS explicitly requires a data discovery and classification process to identify where cardholder data resides across the cardholder data environment (CDE). Without classification, scope reduction -- one of the most important PCI DSS strategies -- is impossible.

Related: PCI DSS Scope Reduction: The Definitive Guide

GDPR -- Personal Data Categories

GDPR Article 9 distinguishes between ordinary personal data and "special categories" of personal data (racial or ethnic origin, political opinions, religious beliefs, trade union membership, genetic data, biometric data, health data, sex life or sexual orientation data). Special category data is subject to heightened protections and additional legal basis requirements.

Organizations subject to GDPR must classify personal data to determine which category it falls into and apply the appropriate level of protection. A data classification policy is the formal mechanism for making this determination.

The Common Thread Across Frameworks

Every framework asks the same fundamental question: does the organization know what data it has, how sensitive that data is, and does it apply controls proportional to that sensitivity? A data classification policy is the documented answer to that question.


Data Classification Levels: Choosing the Right Scheme

The most common and widely accepted data classification scheme uses four tiers. This scheme is sufficient for SOC 2, ISO 27001, HIPAA, PCI DSS, and GDPR compliance. Some organizations use three tiers (omitting the Internal level) or five tiers (splitting Restricted into separate categories). Four tiers represents the practical balance between granularity and usability.

The Four-Tier Classification Scheme

Level 1: Public

Definition: Information that has been approved for public disclosure and whose unauthorized disclosure would cause no harm to the organization.

Examples:

  • Published marketing materials and blog content
  • Public-facing product documentation
  • Published financial statements (for public companies)
  • Press releases and media statements
  • Open-source code repositories
  • Published job postings

Key principle: Data is not Public by default. Information must be explicitly approved for public release by an authorized individual. The absence of a classification label does not make data Public -- it means the data has not been classified and should be treated as Internal at minimum.

Level 2: Internal

Definition: Information intended for use within the organization that is not meant for public disclosure. Unauthorized disclosure could cause minor inconvenience or reputational impact but would not constitute a regulatory violation or significant business harm.

Examples:

  • Internal policies and procedures (not security-sensitive)
  • Company-wide communications and announcements
  • Internal meeting notes and project plans
  • Non-sensitive business correspondence
  • Organizational charts and internal directories
  • Non-sensitive training materials

Key principle: Internal is the default classification for any information that has not been explicitly classified. When in doubt, classify as Internal. This prevents unclassified data from being treated as Public by default.

Level 3: Confidential

Definition: Sensitive business information whose unauthorized disclosure could cause significant harm to the organization, its customers, or its partners. This level includes most customer data, financial records, and business-critical intellectual property.

Examples:

  • Customer data (names, email addresses, usage data, account information)
  • Employee personal information (not meeting the Restricted threshold)
  • Financial records, revenue data, and forecasts (non-public)
  • Business strategies, product roadmaps, and competitive analysis
  • Vendor contracts and pricing agreements
  • Source code (proprietary, not open-source)
  • Internal security architecture documentation
  • Audit reports and compliance documentation

Key principle: Confidential data requires encryption in transit, access control based on business need, and restrictions on external sharing. Most SaaS companies will find that the majority of their operationally important data falls into this category.

Level 4: Restricted

Definition: The most sensitive information handled by the organization, whose unauthorized disclosure would cause severe harm -- including regulatory penalties, significant financial loss, legal liability, or harm to individuals. This classification encompasses data protected by specific regulatory requirements.

Examples:

  • Protected Health Information (PHI) under HIPAA
  • Payment card data (PAN, CVV, PINs) under PCI DSS
  • Social Security numbers and government identification numbers
  • GDPR special category data (health, biometric, genetic, racial/ethnic data)
  • Authentication credentials, encryption keys, and secrets
  • Penetration testing reports and vulnerability assessments
  • Board-level confidential communications (M&A, legal proceedings)
  • Data covered by non-disclosure agreements with specific handling requirements

Key principle: Restricted data requires the highest level of protection: encryption at rest and in transit, access limited to named individuals (not roles), full audit logging of all access, and explicit approval for any sharing. The principle of least privilege applies in its strictest form.

How Organizations Get the Scheme Wrong

The most common mistake is not choosing the wrong number of levels. It is defining the levels and then failing to apply them. A classification scheme that exists in a policy document but has never been used to actually classify a data set is worse than useless -- it creates the appearance of a control without delivering any protection, and auditors will identify this gap immediately.

The second most common mistake is making the scheme too granular. Organizations that create six, seven, or eight classification levels find that employees cannot distinguish between them and default to guessing. Four levels is enough. Simplicity drives adoption.


Building a Data Classification Framework: Step-by-Step

A data classification policy defines the scheme. A data classification framework puts it into practice. Here are the seven steps to move from policy document to operational reality.

Step 1: Establish Data Ownership

Every data set must have a named owner. The data owner is the individual (not a team, not a committee) accountable for:

  • Determining the classification level for the data they own
  • Approving access requests to data under their ownership
  • Reviewing and revalidating classification decisions annually
  • Ensuring handling rules for the assigned classification level are followed

In practice, data ownership maps to business functions. The VP of Engineering owns source code and infrastructure data. The VP of Sales owns CRM data and pipeline information. The CISO owns security tool data, vulnerability reports, and penetration testing results. The CFO owns financial records and forecasts. The Head of People Operations owns employee personal data.

Document data ownership in a data inventory or asset register. ISO 27001 requires this inventory (Clause 8.1), and SOC 2 auditors will look for evidence of it under CC6.1.

Step 2: Conduct a Data Inventory

You cannot classify what you have not identified. A data inventory is a comprehensive catalog of every data type the organization processes, where it resides, how it flows, and who owns it.

The inventory should capture:

FieldDescription
Data typeWhat the data is (e.g., "customer email addresses," "payment card numbers," "source code")
Data ownerThe individual responsible for classification decisions
Storage locationsWhere the data resides (databases, SaaS tools, file shares, endpoints)
Processing systemsApplications and services that create, modify, or read the data
Data flowHow the data moves between systems, including third-party transfers
Regulatory applicabilityWhich regulations apply (HIPAA, PCI DSS, GDPR, etc.)
Current classificationThe assigned sensitivity level

Start with the systems that matter most: your production database, your CRM, your HRIS, your financial systems, and your cloud infrastructure. Expand from there to collaboration tools, analytics platforms, and development environments.

Step 3: Classify Each Data Type

With the inventory complete and ownership assigned, classify each data type using the four-tier scheme. Classification decisions should be based on three factors:

  1. Impact of unauthorized disclosure. What would happen if this data were exposed to an unauthorized party? Consider regulatory penalties, contractual liabilities, competitive harm, and harm to individuals.
  2. Regulatory requirements. Does any regulation specifically govern this data type? If data is PHI, it is Restricted by regulation. If it is cardholder data, it is Restricted by PCI DSS. Regulatory requirements override impact assessments -- you cannot classify regulated data at a lower level than the regulation demands.
  3. Business context. What is the value of this data to the organization and its stakeholders? Source code for a SaaS company may warrant Confidential or Restricted classification because it represents core intellectual property, even if no regulation specifically governs it.

Document each classification decision with the rationale. Auditors will ask not just "what classification did you assign" but "why did you assign that classification."

Step 4: Define Handling Rules for Each Level

Each classification level must have documented handling rules that specify how data at that level is to be stored, transmitted, accessed, shared, retained, and destroyed. These rules are the operational core of the data classification policy -- they translate the abstract concept of "sensitivity levels" into specific, enforceable requirements. (The full handling matrix appears in the next section.)

Step 5: Implement Labeling and Marking

Data must be labeled so that anyone who encounters it -- whether in a document, a database, a file share, or a SaaS application -- can immediately identify its classification level and apply the appropriate handling rules. Labeling applies to documents, email headers, database tables, cloud storage containers, and any other format in which classified data exists. (Labeling standards are detailed in a later section.)

Step 6: Train Personnel

A classification scheme is only as effective as the people who apply it. Every employee, contractor, and third party with access to organizational data must understand:

  • The classification levels and how to identify which level applies
  • The handling rules for each level
  • Their responsibility to classify data they create
  • How to report misclassified or unlabeled data
  • The consequences of mishandling classified data

Training must be documented, and completion records must be retained as audit evidence.

Step 7: Audit, Review, and Improve

Data classification is not a one-time project. It requires ongoing governance:

  • Annual review. The classification scheme and all classification decisions must be reviewed at least annually by data owners. This is a requirement under ISO 27001 (management review per Clause 9.3) and an auditor expectation under SOC 2.
  • Triggered review. Classification decisions should be reviewed whenever the business context changes: new regulations take effect, new data types are collected, data processing activities change, or a security incident reveals a classification gap.
  • Compliance auditing. Internal audits should periodically sample data sets to verify that (a) they have been classified, (b) the classification is correct, and (c) the handling rules for the assigned classification are being followed.

Data Handling Rules by Classification Level

The following matrix defines the minimum handling requirements for each classification level across the six core dimensions of data handling. This matrix should be included in or appended to your data classification policy as an enforceable standard.

Storage

Classification LevelStorage Requirements
PublicNo special requirements. May be stored on any approved organizational system.
InternalMust be stored on approved organizational systems only. No storage on personal devices or personal cloud accounts.
ConfidentialMust be stored on approved systems with access controls. Encryption at rest required for cloud storage. Database-level encryption required. Backups must be encrypted.
RestrictedEncryption at rest required (AES-256 or equivalent). Storage limited to specifically approved systems. Segregated from lower-classification data where technically feasible. Access logging required on all storage systems.

Transmission

Classification LevelTransmission Requirements
PublicNo special requirements.
InternalMust be transmitted via organizational channels (corporate email, approved SaaS tools). No transmission via personal email or personal messaging apps.
ConfidentialEncryption in transit required (TLS 1.2 or higher). No transmission via unencrypted channels. External transmission requires business justification.
RestrictedEncryption in transit required (TLS 1.2 or higher). External transmission requires data owner approval and must use approved secure transfer mechanisms. No email transmission of Restricted data without end-to-end encryption.

Access Control

Classification LevelAccess Control Requirements
PublicNo access restrictions.
InternalAccess limited to employees and authorized contractors via standard authentication.
ConfidentialAccess granted on a need-to-know basis. Role-based access control required. Multi-factor authentication required for remote access. Access reviews conducted quarterly.
RestrictedAccess limited to named individuals with documented business need. MFA required for all access. Access approved by data owner. Access reviewed monthly. Privileged access to Restricted data requires additional authorization.

Sharing

Classification LevelSharing Requirements
PublicMay be shared freely.
InternalMay be shared with employees and authorized contractors. No external sharing without management approval.
ConfidentialExternal sharing requires data owner approval. Recipients must be bound by NDA or equivalent contractual protections. Shared via approved channels only.
RestrictedExternal sharing requires data owner and CISO approval. Recipients must be bound by specific data handling agreements. Data must be encrypted during sharing. All sharing events must be logged.

Retention

Classification LevelRetention Requirements
PublicRetained per business need. No mandatory minimum.
InternalRetained per business need or applicable retention schedule.
ConfidentialRetained per documented retention schedule. Regulatory retention requirements take precedence. No retention beyond documented schedule without justification.
RestrictedRetained for the minimum period required by regulation or contract. Retention period documented for each data set. Regulatory requirements (HIPAA: 6 years, PCI DSS: 1 year of audit logs) take precedence.

Disposal

Classification LevelDisposal Requirements
PublicStandard deletion. No special requirements.
InternalStandard deletion from organizational systems.
ConfidentialSecure deletion from all storage locations, including backups. Deletion logged and confirmed. Physical media degaussed or shredded.
RestrictedCryptographic erasure or physical destruction required. Certificate of destruction required for physical media. Deletion verified by a second party. Destruction documented and retained as audit evidence.

Data Classification for SaaS Companies: What to Classify

SaaS companies face a specific challenge: they handle a wide variety of data types across multiple environments, and the sensitivity of that data varies dramatically. The following inventory covers the data types most SaaS companies must classify, with recommended classification levels.

Customer Data

Customer data is the data your customers entrust to you when they use your product. It is almost always the highest-priority classification target because it carries the greatest regulatory, contractual, and reputational risk.

Data TypeRecommended ClassificationRationale
Customer PII (names, emails, phone numbers)ConfidentialSubject to GDPR, state privacy laws; disclosure causes direct harm to individuals
Customer PHI (health information)RestrictedSubject to HIPAA; unauthorized disclosure triggers breach notification
Customer financial data (bank accounts, billing)RestrictedSubject to PCI DSS (if card data), contractual obligations
Customer usage data (feature usage, logs)Internal or ConfidentialDepends on whether it is aggregated/anonymized or individually identifiable
Customer-uploaded contentConfidential or RestrictedClassification depends on the nature of the content; treat as Confidential at minimum

Employee and HR Data

Data TypeRecommended ClassificationRationale
SSNs, government IDsRestrictedHighest identity theft risk; subject to state breach notification laws
Compensation and benefits dataConfidentialSensitive but lower regulatory risk than government IDs
Performance reviewsConfidentialSensitive internal information
Background check resultsRestrictedContains personal history and potentially criminal records
General employee contact infoInternalLow sensitivity; used for daily operations

Technical and Infrastructure Data

Data TypeRecommended ClassificationRationale
Encryption keys and secretsRestrictedCompromise enables access to all data those keys protect
Authentication credentialsRestrictedCompromise enables unauthorized system access
Penetration test reportsRestrictedReveals specific exploitable vulnerabilities
Application source codeConfidentialCore intellectual property; competitive harm if disclosed
Infrastructure architecture diagramsConfidentialReveals attack surface and security design
Application logs (without PII)InternalOperational data; low sensitivity when PII is excluded
Application logs (with PII)ConfidentialPII inclusion elevates the classification

Business and Financial Data

Data TypeRecommended ClassificationRationale
Revenue, financial forecastsConfidentialCompetitive harm and potential securities implications
Customer contracts and pricingConfidentialContractual confidentiality obligations
M&A documentationRestrictedLegal, regulatory, and market-moving sensitivity
Board meeting minutesRestrictedContains strategic decisions and legal discussions
Marketing materials (published)PublicAlready in the public domain
Internal strategy documentsConfidentialCompetitive harm if disclosed

Analytics and Aggregated Data

Data TypeRecommended ClassificationRationale
Anonymized, aggregated analyticsInternal or PublicDepends on whether re-identification is possible
Individual-level analytics with PIIConfidentialPII inclusion drives the classification
A/B testing resultsInternalLow sensitivity; operational data

The key principle: classification is determined by the most sensitive element present. A log file that contains 99% operational data and 1% email addresses is classified based on the email addresses, not the operational data.


Automated Data Classification: Tools and Approaches

Manual classification does not scale. A company with hundreds of databases, thousands of cloud storage objects, and millions of documents cannot rely on humans to individually label every data asset. Automated classification tools address this by scanning data at rest and in transit, identifying sensitive content, and applying or recommending classification labels.

Data Loss Prevention (DLP) with Classification

DLP solutions are the most established category of automated classification tools. They scan data across storage, endpoints, and network traffic, using pattern matching, regular expressions, and content inspection to identify sensitive data.

How DLP classifies data:

  • Pattern matching. Identifies known data formats -- Social Security numbers (XXX-XX-XXXX), credit card numbers (16-digit patterns with Luhn validation), medical record numbers, and other structured identifiers.
  • Keyword and phrase matching. Flags documents containing terms like "Confidential," "PHI," "Attorney-Client Privilege," or custom keywords relevant to the organization.
  • Data fingerprinting. Creates hashes of known sensitive documents and detects when those documents or their derivatives appear in unauthorized locations.

DLP is most effective for structured, pattern-identifiable data. It is less effective for unstructured data where sensitivity depends on context rather than format.

Related: What Is Data Loss Prevention (DLP)? -- a complete guide to DLP technologies for SaaS security teams.

Content Inspection and Discovery Tools

Cloud-native content inspection tools -- such as AWS Macie, Google Cloud DLP, and Azure Information Protection -- scan cloud storage and databases for sensitive content. These tools are particularly useful for SaaS companies with large cloud footprints.

Capabilities:

  • Automated scanning of S3 buckets, BigQuery datasets, Azure Blob storage, and other cloud data stores
  • Identification of PII, PHI, and financial data based on built-in classifiers
  • Integration with cloud-native access controls to enforce classification-based policies
  • Continuous monitoring for newly created data that matches sensitive patterns

Limitation: These tools are platform-specific. An organization using AWS Macie still needs a separate solution for data stored in SaaS applications, on endpoints, or in non-AWS cloud environments.

Machine Learning-Based Classification

Machine learning (ML) classification models go beyond pattern matching to understand the context and meaning of data. They are trained on labeled datasets of sensitive and non-sensitive content and can identify sensitivity in unstructured data where regex and keyword matching fail.

Use cases where ML excels:

  • Classifying free-text documents (contracts, internal memos, customer communications) where sensitivity depends on content, not format
  • Identifying sensitive images (scanned documents containing PII, medical images)
  • Detecting context-dependent sensitivity (an employee name in a company directory is Internal; the same name in a performance improvement plan is Confidential)

Practical considerations:

  • ML classification requires training data, which means an initial period of manual classification to build the training set
  • False positive rates must be monitored and tuned; an overly aggressive model that flags everything as Restricted will be ignored by users
  • ML models must be retrained as data types and business contexts evolve

The Role of QuickTrust in Automated Classification

QuickTrust's compliance automation platform integrates with your cloud infrastructure, SaaS tools, and development environments to continuously monitor data handling practices against your classification policy. Rather than requiring a separate classification tool, QuickTrust maps your data classification scheme to your existing controls and identifies gaps -- data that should be classified as Restricted but is stored without encryption, Confidential data shared externally without approval, or unclassified data in production systems that has never been inventoried.

See how QuickTrust automates classification compliance


Data Labeling and Marking Standards

Classification without labeling is invisible. If a document is classified as Confidential but carries no visible label, the person handling it has no way to know how to treat it. Labeling makes classification visible and actionable.

What Must Be Labeled

Every data asset that has been classified must carry a label indicating its classification level. This includes:

  • Documents. Word documents, PDFs, spreadsheets, and presentations must carry a classification label in the header or footer of every page.
  • Emails. Emails containing classified information must include the classification level in the subject line prefix (e.g., "[CONFIDENTIAL]") or in a standardized header.
  • Databases and data stores. Database tables, schemas, and cloud storage containers (S3 buckets, Azure containers) must be tagged with metadata indicating their classification level.
  • Code repositories. Repositories containing classified data (especially Restricted data like secrets or credentials) must be labeled in the repository description or README.
  • Physical media. USB drives, printed documents, and any other physical media containing classified data must be visibly labeled with the classification level.
  • SaaS applications. Channels, folders, and workspaces in collaboration tools (Slack, Teams, SharePoint) that contain classified data should be named or described to indicate the classification level.

Labeling Formats

Standardize the label format across the organization. A consistent format ensures recognition and reduces confusion.

Recommended format:

[CLASSIFICATION LEVEL] -- [Organization Name]

Examples:

  • [PUBLIC] -- Acme Corp
  • [INTERNAL] -- Acme Corp
  • [CONFIDENTIAL] -- Acme Corp
  • [RESTRICTED] -- Acme Corp

For documents, the label should appear in:

  • The header or footer of every page
  • The file name (optional but recommended for high-sensitivity documents)
  • The document metadata/properties

For digital systems, the label should be applied as:

  • Metadata tags on cloud storage objects
  • Classification fields in document management systems
  • Tagging conventions in infrastructure-as-code

Handling Unlabeled Data

Your policy must address what happens when data is encountered without a label. The safest approach is the default-up rule: unlabeled data is treated as Internal at minimum. For data that appears to contain PII, PHI, financial data, or other potentially sensitive content, the individual who encounters it must escalate to the data owner for classification.

This rule prevents the dangerous assumption that unlabeled data is Public. It also creates an operational incentive to classify and label data promptly -- because unlabeled data receives more restrictive handling than necessary, creating friction that motivates proper labeling.


Employee Training: Making Classification Practical

A data classification policy that lives in a document repository and has never been operationalized is a policy that will fail its first audit. Employee training is the mechanism that transforms a written policy into organizational behavior.

What Training Must Cover

Training must be specific, practical, and role-appropriate. A generic 20-minute compliance video that mentions data classification in passing does not satisfy the requirement. Effective training addresses:

  1. The classification scheme. What the four levels are, what each one means, and how to determine which level applies. Use real examples drawn from your organization's own data -- not hypothetical scenarios from a generic template.

  2. Handling rules by level. What employees can and cannot do with data at each classification level. "Confidential data may not be shared externally without data owner approval" is clear. "Be careful with sensitive data" is not.

  3. How to classify data employees create. When an employee creates a new document, spreadsheet, or data set, they need to know (a) that they are expected to classify it, (b) how to determine the appropriate level, and (c) how to label it.

  4. How to handle data received from others. When an employee receives a document, data export, or access to a system, they must know how to identify the classification level and apply the corresponding handling rules.

  5. What to do when classification is uncertain. Employees must know that when they are unsure of the correct classification, they should default up (treat it as a higher level) and consult the data owner. The cost of over-classifying is minor inconvenience. The cost of under-classifying is a potential breach.

  6. Reporting misclassification. Employees must know how to report data that appears to be misclassified, unlabeled, or handled in violation of the handling rules.

Role-Specific Training

Not all roles require the same depth of classification training:

RoleTraining FocusFrequency
All employeesClassification levels, handling rules for Internal and Confidential, reporting proceduresAnnual (at onboarding and annually thereafter)
Data ownersClassification criteria, decision-making process, annual review responsibilitiesAnnual, plus when new data types are introduced
Engineering teamsClassification in code, database tagging, secrets management, infrastructure labelingAnnual, plus integration into developer onboarding
Customer supportHandling customer PII and PHI, classification of support tickets and communicationsAnnual
IT and security teamsFull classification scheme, automated classification tools, audit evidence requirementsAnnual, plus ongoing as tools and processes change
ExecutivesGovernance responsibilities, data owner accountability, regulatory implicationsAnnual

Documenting Training as Audit Evidence

Auditors will request evidence that training has occurred. The following records must be maintained:

  • Training completion records (who completed training, when, what content was covered)
  • Acknowledgment records (signed or electronic acknowledgment that the employee has read and understood the classification policy)
  • Training materials (the actual content delivered, retained for the audit period)
  • Assessment results (if knowledge assessments are included in the training)

Related: Security Awareness Training: A Practical Guide for SaaS Companies


Data Classification Audit Evidence: What Auditors Look For

When auditors evaluate your data classification controls, they are not looking for perfection. They are looking for evidence of a systematic, documented, actively maintained program. Here is exactly what they examine and what you need to produce.

Document Evidence

Evidence ItemWhat Auditors CheckFramework Reference
Data classification policyExists, is approved by management, has been reviewed within the past year, defines classification levels and handling rulesISO 27001 A.5.12, SOC 2 C1.1, HIPAA 164.312
Data inventory / asset registerComprehensive, includes classification levels, names data owners, covers all significant data typesISO 27001 Clause 8.1, SOC 2 CC6.1
Classification decisions with rationaleEach data type has a documented classification and the reason for the assignmentISO 27001 A.5.12, SOC 2 C1.1
Handling rules / data handling matrixSpecific, enforceable rules for each classification level across storage, transmission, access, sharing, retention, and disposalISO 27001 A.5.12-A.5.13, SOC 2 CC6.1

Operational Evidence

Evidence ItemWhat Auditors CheckFramework Reference
Labeled documents and systemsData assets carry visible classification labels consistent with the policyISO 27001 A.5.13
Access controls aligned with classificationHigher-classification data has more restrictive access controlsSOC 2 CC6.1-CC6.3, ISO 27001 A.5.15
Encryption aligned with classificationConfidential and Restricted data is encrypted per the handling rulesSOC 2 CC6.1, HIPAA 164.312(a)(2)(iv)
DLP or automated scanning resultsEvidence that the organization actively monitors for sensitive data in unauthorized locationsSOC 2 CC6.1, PCI DSS 3.1

Governance Evidence

Evidence ItemWhat Auditors CheckFramework Reference
Annual policy review recordThe classification policy was reviewed and approved within the past 12 monthsISO 27001 Clause 9.3, SOC 2 CC1.4
Annual classification review recordData owners reviewed their classification decisions within the past 12 monthsISO 27001 A.5.12
Training completion recordsAll personnel completed classification trainingISO 27001 A.6.3, SOC 2 CC1.4
Incident records involving classificationIf a classification-related incident occurred, evidence of response and corrective actionAll frameworks

What Produces Audit Findings

The most common audit findings related to data classification, in order of frequency:

  1. No data classification policy exists. This is a critical finding under any framework.
  2. Policy exists but has never been applied. No data has actually been classified. The scheme is theoretical.
  3. Classification exists but handling rules are not followed. Restricted data is stored without encryption. Confidential data is shared externally without approval.
  4. No data inventory. The organization cannot demonstrate that it has identified its data assets.
  5. No annual review. The classification scheme has not been reviewed in over 12 months.
  6. No training evidence. Employees have not been trained on classification requirements.

Common Data Classification Mistakes

After working with hundreds of SaaS companies on their compliance programs, the following mistakes appear repeatedly. Avoiding them will save significant rework and prevent audit findings.

Mistake 1: Classifying Everything as Confidential

When organizations do not provide clear criteria for distinguishing between classification levels, employees default to the safe choice: classify everything as Confidential. This makes the scheme useless. If everything is Confidential, nothing is Confidential. The handling rules for Confidential become the de facto rules for all data, which means Restricted data does not receive additional protection and Internal data is subject to unnecessary restrictions.

Fix: Provide specific, concrete criteria for each level. Use real examples from your own organization, not abstract definitions.

Mistake 2: No Data Owners

A classification scheme without named data owners has no one accountable for classification decisions, no one to approve access to sensitive data, and no one responsible for annual reviews. When the auditor asks "who determined that this data set is Confidential," there must be a name -- not a team, not "the security team," not "it was always classified that way."

Fix: Assign a named individual as the owner of every significant data type in your inventory. Document ownership in the data inventory.

Mistake 3: Classifying Once and Never Reviewing

Data sensitivity changes. A data set that was Internal when it contained only aggregated metrics becomes Confidential when PII is added. A database that was Confidential becomes Restricted when the organization begins storing PHI. A document that was Restricted during an M&A process reverts to Internal after the deal closes or becomes public.

Fix: Mandate annual classification reviews by data owners. Include triggered reviews when data types, processing activities, or regulatory requirements change.

Mistake 4: Ignoring Unstructured Data

Most classification efforts focus on databases and structured data stores. But unstructured data -- documents, spreadsheets, emails, Slack messages, shared drive files -- often contains the most sensitive information and is the least controlled. A spreadsheet with customer SSNs saved to a shared drive with no access controls is a breach waiting to happen.

Fix: Include unstructured data in your data inventory and classification process. Use DLP and content inspection tools to discover sensitive content in unstructured data stores.

Mistake 5: No Handling Rules

A classification scheme without handling rules is a taxonomy exercise, not a security control. Knowing that data is "Confidential" means nothing if there are no documented requirements for how Confidential data must be stored, transmitted, accessed, and destroyed.

Fix: Define explicit handling rules for each classification level across every dimension of data handling. Include the handling matrix as an appendix to the policy and ensure it is referenced in training.

Mistake 6: Policy Written for Auditors, Not Employees

A data classification policy filled with compliance jargon and framework references may satisfy an auditor at a surface level, but it will not be understood or followed by the people who actually handle data. If employees cannot understand the policy, they cannot comply with it.

Fix: Write the policy in clear, plain language. Supplement it with a one-page quick-reference guide that employees can pin to their desk or bookmark. Use real examples.

Mistake 7: Treating Classification as an IT Project

Data classification is a business decision, not a technical implementation. IT and security teams build the tools and infrastructure, but the classification decisions themselves must be made by the business functions that own the data. When classification is delegated entirely to IT, the result is a technically sound scheme that does not reflect business priorities or regulatory realities.

Fix: Ensure data owners are business leaders, not IT staff. IT implements the tools. Business leaders make the classification decisions.


FAQ

What is a data classification policy?

A data classification policy is a formal document that defines how an organization categorizes its information assets based on sensitivity, specifies handling rules for each category, and assigns accountability for classification decisions. It establishes the classification levels (typically Public, Internal, Confidential, Restricted), the criteria for assigning each level, and the specific controls required for storing, transmitting, accessing, sharing, retaining, and destroying data at each level.

How many data classification levels should we have?

Four levels is the standard for most organizations: Public, Internal, Confidential, and Restricted. This scheme is sufficient for SOC 2, ISO 27001, HIPAA, PCI DSS, and GDPR compliance. Some organizations use three levels (omitting Internal) or five levels (splitting Restricted into sub-categories). Fewer than three levels provides insufficient granularity. More than five levels creates confusion and reduces adoption. Four levels offers the right balance between precision and usability.

Is data classification required for SOC 2?

Yes. SOC 2 Trust Services Criteria C1.1 explicitly requires that "the entity identifies and maintains confidential information to meet the entity's objectives related to confidentiality." Additionally, CC6.1 requires logical access controls that presuppose the organization has identified and classified its information assets. An auditor cannot evaluate whether your access controls are appropriate without a documented classification scheme that defines what is sensitive and what is not.

How often should data classification be reviewed?

At minimum, annually. ISO 27001 requires management review of the ISMS (including classification decisions) at least annually under Clause 9.3. SOC 2 auditors expect evidence of periodic review of classification decisions as part of the control environment. Beyond the annual review, classification decisions should be reviewed whenever data types change, new regulations take effect, or a security incident reveals a classification gap.

What is the difference between data classification and data labeling?

Data classification is the decision-making process: determining what sensitivity level a data asset belongs to based on its content, regulatory requirements, and business context. Data labeling is the implementation of that decision: making the classification visible by applying markers, tags, metadata, or visual indicators to the data asset. Classification without labeling means the sensitivity decision exists in a spreadsheet somewhere but is invisible to the people handling the data. Both are required -- ISO 27001 A.5.12 covers classification, and A.5.13 covers labeling.

How do you classify data in a SaaS product?

Start with a data inventory that maps every data type your product processes: customer PII, customer-uploaded content, usage analytics, payment information, authentication credentials, and application logs. Classify each data type based on the impact of unauthorized disclosure and any applicable regulatory requirements. Customer PHI is Restricted. Customer PII is typically Confidential. Anonymized analytics is Internal. Published documentation is Public. Then implement handling rules for each level in your infrastructure: encryption, access controls, logging, and retention policies mapped to classification levels.

What tools automate data classification?

Three categories of tools address automated data classification: (1) DLP solutions that scan for sensitive data patterns using regex, keyword matching, and data fingerprinting; (2) cloud-native discovery tools like AWS Macie, Google Cloud DLP, and Azure Information Protection that scan cloud storage for sensitive content; and (3) ML-based classification tools that use trained models to identify sensitivity in unstructured data. Most mature programs combine all three, supplemented by a compliance automation platform like QuickTrust that maps classification to control enforcement.

What happens if we do not have a data classification policy during an audit?

For ISO 27001, the absence of a data classification policy is a nonconformity against Annex A controls A.5.12 and A.5.13. A major nonconformity will prevent certification until it is resolved. For SOC 2, the absence of a classification process will result in an exception or qualified opinion under C1.1, and likely under CC6.1 as well. For HIPAA, failure to identify and classify PHI undermines the entire Security Rule compliance posture and will be flagged in an HHS audit. For PCI DSS, failure to identify where cardholder data resides (Requirement 3) will result in a failing assessment. In short: you will not pass the audit.


Build Your Data Classification Policy with QuickTrust

Building a data classification policy from scratch is the kind of work that consumes weeks -- inventorying data, defining levels, drafting handling rules, mapping to framework requirements, training employees, and assembling audit evidence. And then you have to maintain it: annual reviews, triggered reviews, ownership changes, new data types, evolving regulations.

QuickTrust eliminates the manual work. The platform provides:

  • Pre-built, audit-tested data classification policy templates mapped to SOC 2, ISO 27001, HIPAA, PCI DSS, and GDPR requirements
  • Automated data inventory that integrates with your cloud infrastructure and SaaS tools to discover and catalog data assets
  • Continuous monitoring that verifies handling rules are being followed -- encryption in place, access controls aligned with classification, no Restricted data in unauthorized locations
  • Audit evidence collection that automatically assembles the documentation auditors need: policy approval records, classification decisions, training completion, and review logs
  • Framework mapping that shows exactly which controls each classification decision satisfies, across every framework you are pursuing

Stop building compliance from blank documents. Start with a platform that has done it hundreds of times.

Start your free compliance assessment with QuickTrust

Ready to get audit-ready?

Our engineers implement controls, prepare evidence, and coordinate your audit.

Get a Free Assessment

Related Articles