Control of Data

Identification of controlled data, its classification, and associated processes.

Data Identification

Inventory of Data Assets

Data Cataloguing: Create a comprehensive inventory of all data assets across your organisation, including databases, data warehouses, data marts, and data lakes.

Data Source Mapping: Identify and map all data sources, including internal systems and third-party sources.

All Data Sources SHALL be described in myBMT

The EA solution SHALL map the Enterprise solutions and interactions

Data Discovery Tools

User Discovery: All data sources that are in-scope are discoverable to authorised users including exploration of data within the export.

myBMT SHALL support data discovery by showing splash views of data container objects

Data Classification

Define Classification Levels

Sensitivity Levels: Establish clear classification levels (e.g., Public, Internal, Confidential, Restricted) based on data sensitivity and regulatory requirements.

Classification Criteria: Define criteria for each classification level, considering factors such as data type, usage, impact of disclosure, and regulatory implications.

Sensitivity Levels SHALL be defined by the DataMart Owner in the DPIA

The schema mvw and dvw SHALL be the public view

Confidential data SHALL be in alternative schema (e.g. fa for financial)

Alternative Schemas SHALL have Restricted access policy applied

Tagging and Labelling

Data Tagging: Implement data tagging mechanisms to label data according to its classification level.

Metadata Management: Use metadata management tools to store and manage classification tags and labels.

Data SHALL be tagged for metadata classification (e.g. Date, Measure, Code/Dim, Sequence)

Data SHOULD be tagged for owner

Data Fields SHOULD be described in myBMT

Associated Processes

Access Control

Role-Based Access Control (RBAC): Implement RBAC to ensure that users have access only to the data necessary for their roles.

Least Privilege Principle: Apply the least privilege principle to minimise access to sensitive data.

There SHALL be two Database access groups Data Admin and Data User

Data Admin SHALL be the System Administrator and DB Owner

Data User SHALL be the public access group

Data User SHALL be member of the Common Data Model and read access to mvw and dvw

Other Read access SHALL be explicitly applied

All users SHALL access via Microsoft Entra/MFA

Azure applications SHALL access via SQL User ([dv] (read/write all), [my] (read all))

The user [bi] SHALL be used to validate user access for Data User

Data Handling Procedures

Data Handling Policies: Develop and enforce data handling policies for each classification level, detailing how data should be accessed, stored, transmitted, and disposed of.

Data Encryption: Use encryption for sensitive data both at rest and in transit.

Data Handling Policy SHALL be described by the Data Owner in the DPIA

Monitoring and Auditing

Activity Monitoring: Implement continuous monitoring of data access and usage to detect unauthorised access and potential breaches.

Audit Trails: Maintain detailed audit trails of data access and modifications for accountability and compliance.

All source data SHALL be archived before processing

Transfer Logs SHALL be applied to each stage of the Pipeline

Execution by external Agent SHALL be recorded to identify the Agent (ReportID) and Calling User

All Data Views (DataMarts) SHALL be checked that

  1. They open without error
  2. They contain row data

Data Lifecycle Management

Data Retention Policies: Establish data retention policies based on data classification, specifying how long data should be retained and when it should be archived or deleted.

Data Archiving: Implement archiving procedures for long-term storage of data that is no longer actively used but must be retained for compliance or historical purposes.

All source data SHALL be archived before processing (see Monitoring and Auditing)

All DataMart view SHALL be snapshot every week (on Sunday)

There is currently NO Retention Policy automatically applied in the Data Warehouse archives will be held for at least 6 years IAW the BMT Data Retention Business Procedure

Data Protection and Compliance

Compliance Frameworks: Align your data classification and handling processes with relevant regulatory frameworks (e.g., GDPR).

Data Protection Impact Assessments (DPIAs): Conduct DPIAs for projects involving sensitive data to identify and mitigate risks.

DPIA WILL be completed by the Data Owner and implemented within the Data Warehouse by the Data Engineer

Employee Training and Awareness

Training Programs: Provide regular training for employees on data classification, handling procedures, and security best practices.

Awareness Campaigns: Conduct awareness campaigns to reinforce the importance of data protection and compliance.

KnowHow SHALL describe sufficient knowledge including how to get to DataMart, Who is the owner, what is the intended purpose, Quality dimensions and return values.

Technology and Tools

Data Loss Prevention (DLP)

DLP Solutions: Deploy DLP solutions to monitor, detect, and prevent unauthorised access, use, or transmission of sensitive data.

Policy Enforcement: Configure DLP policies to enforce data classification and handling rules.

All Data Sources SHALL be maintained in Azure

All Code Base SHALL be maintained in DevOPs

Where possible, Production Services SHOULD be updated by deployment from UAT

There is currently NO specific DLP Policy

Data Masking and Anonymisation

Data Masking: Use data masking techniques to obscure sensitive information in non-production environments.

Anonymisation: Apply anonymisation methods to remove personally identifiable information (PII) where possible, reducing the risk of data breaches.

Data Masking WILL be described by the Data Owner in the DPIA

Data Masking SHALL be performed using MD5

Data Linking Keys SHALL be performed using MD5

Data Governance Platforms

Integrated Platforms: Implement integrated data governance platforms that provide tools for data cataloguing, classification, policy management, and compliance tracking.

Workflow Automation: Use workflow automation within these platforms to streamline data classification and handling processes.

The DataMart SHALL be designed to primarily support the performance of report writing using Power BI

myBMT SHALL support the governance of the design of DataMarts

With tickets myBMT SHALL maintain a record of actions performed on DataMarts

With tickets myBMT shall support the workflow for DataMart improvements

Leave a Comment