Select Page

Living in a Data Vault

By Barry Devlin
| October 2, 2017

Living in a vault, especially of the underground variety, doesn’t sound too pleasant to me. However, a vault can also mean a roof in the form of an arch, as in a cathedral. To that point, living in a Data Vault demands some over-arching thought, uniting business people and IT once again in a process of ongoing decision-making support, in an environment where business change will always be continuous and rapid.

Such a constant “running while rebuilding” process is well-known to data warehouse teams, but the support for agility combined with the governance offered by the Data Vault model and methodology raises the stakes to unprecedented levels. Getting business people on board, as described in my past blogpost Breaking into a Data Vault is the first and vital step. Business users must see that the agility they experienced in the initial delivery of a Data Vault—when implemented via data warehouse automation software such as WhereScape® Data Vault Express—can carry through to ongoing operation.

The collaborative effort between business users and IT in the discovery and design phase of the initial Data Vault delivery must continue. And intensify at an even faster pace. Business users may have had some patience with the initial build of the Data Vault because of its novelty, but now that the Data Vault is mainstream, the expectation will be of ever faster and higher quality iterations. The metadata stored in WhereScape Data Vault Express is the basis for impact analysis, rapid development and speedy deployment of Data Vault updates. Ensuring updates are delivered successfully the first time—in terms of data delivered and business needs met—will keep business users from reverting to the omnipresent spreadsheet.

For development teams, the automation of the entire process from design to delivery, as discussed in the blogpost Locking in a Data Vault, is the mandatory foundation for reducing the effort of 24/7 operation and freeing up developer time and resources for the delivery of new and changed function and data demanded by the business. The error handling, logging and auditing code generated in the development phase together with the built-in scheduler and job management capabilities of WhereScape Data Vault Express allow developers to focus on what they do best: develop solutions.

Agility is vital in this extended stage of the Data Vault lifecycle. This flexibility comes at the cost of a relatively complex engineering structure in the Data Vault model as well as in the numerous components and processes needed to build and populate it. While Data Vault 2.0 offers a methodology to address this complexity, data warehouse automation, and the metadata it creates and uses, is the only effective solution for developers under pressure from rapidly evolving business demands.

This metadata—or, as I prefer to call it, context-setting information—at the heart of WhereScape Data Vault Express offers a common and consistent vocabulary and business semantics. These are mandatory for effective collaboration between business and IT as business evolves, sometimes dramatically, and where the content and function of the Data Vault must follow. Manual documentation of business needs or data deliverables cannot be trusted in such circumstances. Metadata, in effect, is the automated and flexible documentation of all requirements and deliverables.

Living in the Data Vault is the final and ongoing phase of the Data Vault delivery process. It deserves more attention than it usually receives. After all, initial implementation—like a puppy—is for one short time period, a Data Vault is for life. To set your organization up for greater success in implementing and living within the Data Vault methodology, read this Meeting the Six Data Vault Challenges white paper and watch this recent Turning the Data Vault Methodology into a Practical Reality recorded webcast.

You can find the other blog posts in this series here:


Dr. Barry Devlin is among the foremost authorities on business insight and one of the founders of data warehousing, having published the first architectural paper on the topic in 1988. Barry is founder and principal of 9sight Consulting. A regular blogger, writer and commentator on information and its use, Barry is based in Cape Town, South Africa and operates worldwide.

A Step-by-Step Framework for Data Platform Modernization

TL;DR: Legacy data platforms weren't built for real-time analytics, AI workloads, or today's data volumes. This three-phase framework covers cloud migration, architecture selection (warehouse, lakehouse, or hybrid), and pipeline automation. The goal: replace brittle,...

How-to: Migrate On-Prem SQL Server to Azure

Migrating on-premises SQL Server to Azure shifts infrastructure management to the cloud while maintaining control over data workloads. Organizations move to Azure SQL Database, Azure SQL Managed Instance, or in some instances on-prem SQL Server on Azure run on virtual...

Data Governance in Healthcare: HIPAA Compliance Guide

TL;DR Healthcare data architects must integrate fragmented clinical systems (EHRs, PACS, LIS) while maintaining HIPAA-compliant lineage and clinical data quality. Data Vault modeling can help provide the audit trails regulators demand, but generates hundreds of tables...

Enterprise Data Warehouse Guide: Architecture, Costs and Deployment

TL;DR: Enterprise data warehouses centralize business data for analysis, but most implementations run over budget and timeline while requiring specialized talent. They unify reporting across departments and enable self-service analytics, yet the technical complexity...

What Is a Data Vault? A Complete Guide for Data Leaders

A data vault is a data modeling methodology designed to handle rapidly changing source systems, complex data relationships, and strict audit requirements that traditional data warehouses struggle to manage.  Unlike conventional approaches that require extensive...

New in 3D 9.0.6.1: The ‘Source Aware’ Release

When your sources shift beneath you, the fastest teams adapt at the metadata layer. WhereScape 3D 9.0.6.1 focuses on precisely that: making your modeling, conversion rules and catalog imports more aware of where data comes from and how it should be treated in-flight....

Data Vault on Snowflake: The What, Why & How?

Modern data teams need a warehouse design that embraces change. Data Vault, especially Data Vault 2.0, offers a way to integrate many sources rapidly while preserving history and auditability. Snowflake, with elastic compute and fully managed services, provides an...

Related Content

A Step-by-Step Framework for Data Platform Modernization

A Step-by-Step Framework for Data Platform Modernization

TL;DR: Legacy data platforms weren't built for real-time analytics, AI workloads, or today's data volumes. This three-phase framework covers cloud migration, architecture selection (warehouse, lakehouse, or hybrid), and pipeline automation. The goal: replace brittle,...

How-to: Migrate On-Prem SQL Server to Azure

How-to: Migrate On-Prem SQL Server to Azure

Migrating on-premises SQL Server to Azure shifts infrastructure management to the cloud while maintaining control over data workloads. Organizations move to Azure SQL Database, Azure SQL Managed Instance, or in some instances on-prem SQL Server on Azure run on virtual...