Automation² API, Matillion for Azure Synapse and Custom ETL settings (Release 4.2.6)

VaultSpeed is creating a habit of launching a major release just before the holiday season! This year is no exception, so 4.2.6 is loaded with Santa's gifts!

So, what's new?

We completely redesigned our API to make it publicly accessible and consumable. Matillion ETL for Azure Synapse is now available. And we added functionality to your data pipelines: you can now customize your ETL mappings and add additional code.

Automation² API

Our API has been substantially reworked. You can start calling corresponding API endpoints for all the data and actions available in our application. We're proud of this achievement making VaultSpeed the first tool to deliver a REST API for data vault automation. The API enables further integration with other tools and allows users to truly automate the automation = Automation².

Rest API docs


With our API, you can start automating tasks such as:

  • the creation of a new source version
  • the configuration for similar sources
  • loading metadata into your preferred data lineage or data governance tool
  • the import of business view definitions
  • the migration of your existing Data Vaults into VaultSpeed
  • and much more

The screenshot below shows the setup for automatic agent download via the API.


Download the agent using curl to the API endpoint


Another example:

This screenshot illustrates how VaultSpeed metadata is extracted via Snowflake's Matillion ETL API. A schedule running this mapping would sync all Data Vault lineage and metadata straight into Snowflake!


VaultSpeed API to Snowflake mapping built in Matillion ETL


Data Vault metadata loaded in Snowflake


Not all endpoints will be included in our standard licenses, but some will always be available, such as downloading the Agent or the Airflow plugin.

Matillion Synapse

On the ETL side, we've added support to run Matillion Synapse. VaultSpeed now generates Matillion ETL code for Synapse Data Vaults.

Matillion users can automate the pipelines that load data in the Data Vault area and focus on tailor-made transformations in the other layers of the cloud data warehouse.

It is good to know that Matillion has just released CDC support. This opens the opportunity to land the data from different sources, making Data Vault's integration even more effortless.

Our current support for Matillion includes both Snowflake and Synapse , and we are looking to extend it to other cloud data platforms in the near future.


Generated SAT mapping in matillion for Synapse

Data pipelines

This release contains a significant development to make your data pipelines run smoother.

VaultSpeed has offered the possibility to add custom code snippets to generated DDL code for quite some time now. Think of examples like DDL for transient tables or partitioning definitions.

We’re now allowing users to add custom code to the generated mappings as well. Depending on your preferred ETL solution, different settings can be applied.

The example below shows SQL procedures.


Example of custom ETL snippets added to procedures


The possibilities are endless — from changing execution grants by adding "Execute as owner" to adding custom logging statements with row counts after every DML statement.

The complete documentation can be found at

Other important changes:

  • We added the possibility to define DDL settings for all the standard BV objects (no VaultSpeed Studio templates at this stage, those will be added later).
  • The initial load STG mappings not only use the extraction table but also the SATs to look up BKs. This comes in handy, mainly for delta generations when loading the initial data for a new object with references to an existing object.
  • We added two extra DV parameters: CAST_TO_NVARCHAR_IN_HASH and CAST_TO_VARCHAR_IN_HASH. These can be used to control the hashing behavior and determine which type the business keys are cast before hashing them. These parameters are beneficial for SQL Server and Synapse and are mutually exclusive.
  • A new logic applies to the BV release creation to catch the cases where bridges become invalid. When an object gets deleted from a Data Vault while being used in a bridge, the initial BV release created when locking that DV release will be unlocked. No code can be generated for this new DV release. It can only be rendered after resolving the issue in the bridge and locking the business vault. While there is still an invalid bridge in a BV, hovering over the (grayed out) lock button will display the faulty bridge.
  • Hard deletes can now be generated for ODI. The deletes are implemented using a setBeginCmd containing the delete SQL statement.
  • We updated our template language to allow for repeating templates. You can now generate a query for every SAT of a HUB or all DV objects in a bridge in the VaultSpeed Studio templates.

Example of the template code:

Template $ DVO_TEMPL 
templaterepeatedbycomponent DVO

So, lots of new stuff to play with for next year. All we want to do now is wish you happy holidays. More exciting features are coming in 2022.
Spoiler alert, some of them involve Spark Streaming!



VaultSpeed meets Matillion ETL (Release 4.2.5)

VaultSpeed meets Matillion ETL (Release 4.2.5)

Building your cloud data warehouse just got a whole lot easier. Matillion ETL and VaultSpeed have teamed up to bring you automated, cloud native data integration powered by Data Vault 2.0. This release brings you automated ETL generation for Matillion on Snowflake and support for Azure Synapse is coming soon.


About Matillion ETL

Matillion helps data teams get insights and results faster with a cloud-native data integration platform. It’s low-code, no-compromise turbocharges your data ingestion and transformation workflows, while taking full advantage of every native capability in your cloud data platform of choice (Snowflake, Amazon Redshift, Delta Lake on Databricks, Google BigQuery, and Microsoft Azure Synapse).

Matillion is the only cloud-native ETL platform built for enterprise deployment with advanced collaboration, security and data sovereignty protection at its core. This helps data teams of all sizes get ahead of the curve, stay competitive, and deliver game changing value back to their organizations.

Deliver Cloud integration systems Faster

VaultSpeed’s automation engine can now deliver generated ETL mappings for Matillion. This drastically decreases Time to Market for any cloud data warehouse project. It also increases the level of scalability.

Matillion supports an abundance of data source connectors so you can easily load and transform any type of source data into your cloud Data Warehouse or Data Lake.

Both Matillion and VaultSpeed support modern cloud data platforms such as Snowflake, Azure Synapse, Delta Lake on Databricks, Amazon Redshift and Google Bigquery. For the moment, we support Snowflake only, with Synapse following soon.

The automated integration process starts by harvesting your source metadata into VaultSpeed and modelling your Data Vault 2.0 data model towards any business model. VaultSpeed will do the heavy lifting for you.

Afterwards, you are able to auto-deploy ETL code and workflows to the Matillion ETL repository and DDL towards your target environment.



Once you finish development, Matillion takes care of loading your data from source to target using VaultSpeed’s auto-generated mappings. Matillion’s ETL engine ensures excellent loading performance. You are able to schedule, run and monitor your data flows from within Matillion's interface.

You can additionally build custom business logic using Matillion’s intuitive ETL designer tools and you can also automate custom business rules into the solution by coding custom VaultSpeed Studio automation templates.


Matillion ETL mappings and flows generated by VaultSpeed

Other changes

We also added a few other cool features in this release:

  • You can now generate ETL for only a specific object. If you select a source object, VaultSpeed will generate all the code to load the resulting Data Vault objects.
    If you select a Data Vault object, we will give you all the code to load that specific object. You can limit further by also specifying a source (see example below). If a Business Vault object is selected, then only the code for that object will be generated.


  • We also added support for Multi-active Sattelites without a subsequence attribute. This is only available for objects with no CDC or no CDC incremental since we need all the records per key to be delivered each load.
  • We renamed the mapping counter to a more generic name: Vaultspeed Automation Units. This should avoid some confusion since mappings are not the only product that you can generate with VaultSpeed . More information can be found in the documentation Vaultspeed Automation Unit (VAU).
  • The format mask for source attributes is now a free text field instead of a selection menu. You can use this feature to easily convert data into the correct format if you like (e.g. a char to a date field). Note that this format string must be valid for the target database.

External tables & Template Previews (Release 4.2.3)

External tables & Template Previews (Release 4.2.3)

We have released VaultSpeed 4.2.3! Part of our focus was on improving performance of code generation tasks, but we also included some novelties.

External tables

We have extended DDL settings with support for INI and CDC layers. You can now generate table definitions for external/foreign tables.
This enables you to define INI and CDC tables as external tables, directly connecting them to csv’s, xls’s, db exports and many others... in your data lake.



VaultSpeed Studio Code Preview

VaultSpeed studio, our templating module now features previews. You can actually write a template and run a preview on a designated object to see what code it will generate. Copy paste preview code and test in on your development environment. Using preview, it won’t take long until you write the perfect custom template!


Preview for a custom effectivity SAT template

Performance improvements

We improved performance for delta generations. Delta’s calculate the difference between two separate Data Vault versions. They generate all necessary code to move from one version to the next.


Delta code generation


Calculation times were drastically improved, and the differences are especially important when changes are located only in a limited set sources.

Additionally, VaultSpeed’s agent will now only harvest metadata for objects that are included in a release, instead of all objects in the schema, this can greatly improve metadata retrieval performance for large sources with only a limited number of objects being used for the data warehouse.

Source editor improvements

We also added some new stuff in our source editor.

From now on you can use object or attribute mass update from the source graphical editor. This enables you to set the object type, CDC type, comments, data length,… for all objects or attributes matching a certain pattern.


Object/Attribute Mass Update


We improved the layout of objects in the source graphical editor, and added the ability to switch between vertical and horizontal orientation.

Third, we added shortest path functionality between 2 objects on the source graphical editor. Another tool that can help you to better understand your source model.


Shortest path


These changes were the final part in the roadmap to cover all functionality that was previously available in the tabular source editor, the old editor had become outdated and is no longer available.


Azure Data Factory meets VaultSpeed FMC

Azure Data Factory meets VaultSpeed FMC (Release 4.2.2)

Some of our developers just don’t know how to stop. They came up with something new during Christmas Holidays: Support for Azure Data Factory (ADF) in our Flow Management Control (FMC) solution.


In a previous blogpost we introduced our FMC solution on top of Apache Airflow. From now on, we also offer our workflow solution on top of Azure Data Factory.
This solution fits ideally for Azure DB or Synapse customers. They can use VaultSpeed to generate DDL and ELT to integrate their sources into the data warehouse. VaultSpeed can now also generate the orchestration in the form of json, which you can automatically deploy to ADF.



The VaultSpeed FMC for ADF uses Azure PAAS components exclusively. Azure Data Factory and your Data warehouse Database (SQL server or Synapse).
The database contains procedures and load metadata tables, meanwhile
ADF FMC will use stored procedure activities to execute those procedures.


Choose your preferred FMC platform


ADF has visual presentation of the workflows and built-in monitoring. ADF provides seamless pipeline restart-ability and failure management. You can also create Azure Dashboards based on ADF metrics. It is also possible to export the metrics into an external reporting tool like Power BI, Grafana or other candidates.


ADF FMC allows you to optimize parallelism and Azure cloud costs. Code generation is fully metadata driven while it still allows for integration with existing ADF pipelines like pre-staging or post processing.

Other changes

Despite this being a smaller release, some other quality of life changes are included. Every page in Vaultspeed now contains a link to the relevant VaultSpeed docs (book icon). We also added all the subscription info to the dashboard, such as extra modules and support tiers.

Datavault preview & Metadata export          (Release 4.2.1)

Datavault preview & Metadata export (Release 4.2.1)

VaultSpeed stepped up the development pace in 2020. We delivered more, and more substantial releases than ever before. Our 10th and final release of 2020 brings great new features.
We added a data vault preview in our source editor. This enables users to see what their source would look like in a data vault model even before adding it to a data vault release. Second, we introduce full metadata export. Users can configure what metadata they would like to export from our repository. Third, we added support for multi-active many to many links. Finally, we enabled automatic deployment to Git. All generated code can automatically get committed in your Git repository.

Data Vault Preview

Our source editor has a great new feature. Our new Data Vault preview will show you what your source will look like in a data vault model, based on the metadata and specific options you’ve set for your source. All you need to do is save the current build of your source version and hit the SDVV button on the top right corner of the screen. This feature will help developers and analysts in making the right design decisions. You can keep altering your source version up until the point that your data vault model looks perfect, and then lock it in before adding it to a data vault release.


Source Editor and Data Vault preview

Metadata Export

We made significant changes to our metadata export functionality. It has become a separate form of delivery besides generated DDL, data flows and workflows.

We have built the export framework around the concept export configurations. You can build your desired configuration based on specific needs (e.g. data lineage, input for a metadata management tool, testing frameworks,...). Select metadata from a list of around 45 metadata attributes. These attributes include source, raw vault and business vault metadata.

Once your export configuration is saved, you can run it against a certain project and your metadata will be delivered to you in a csv-file. Various other export formats are coming soon.


Multi-active many to many links

VaultSpeed already supported multi-active satellites for HUB-SAT configurations. We now also support it for many-to-many links. Just like before, it also comes with the option to add a subsequence attribute to the unique key.
Multi-active satellites are important to enable data modellers to set the right business keys across your data vault when having multiple active records for a certain business key.


Or search free stock images


A typical example of a multi-active setup is a source table with multiple active records, one for each different language code. In that case the language code will become your subsequence attribute.

Auto Deploy Git

Another important new feature is the possibility to auto-deploy code to your designated Git repository. Simply add your git connection in the agent’s config file and you’re good to go. It is possible to deploy to different branches, as long as you properly target them in the configuration file.

Users can still auto-deploy code directly to the target database, ELT repo or Airflow instance, but this typically happens on a DEV environment. Automatically adding your code to Git makes it easier for users to add a specific version of VaultSpeeds generated code to a release. After you have verified the code on your development environment, you can use your Git repo in combination with a continuous integration tool like Jenkins, Travis or GitLab to deploy code throughout your DTAP cycle.


Optionally deploy your generated code file to a Git repo

Other important changes

  • We enabled Auto-deploy to Spark SQL targets.
  • We added support for remote CDC with a DB link on SQL server. However, linked servers are not yet supported by Azure Synapse, so this can not be used for now.
  • Delta generations can now also be used to generate FMC code. For initial loads, the new flow will only include the mappings for new objects.

That’s it for this blogpost… more features are coming soon!

Snowflake Procedures (Release 4.1.17)

Running Data Vault on Snowflake with VaultSpeed

In the latest VaultSpeed release we included important improvements for Snowflake customers ❄️.
We also used valuable customer feedback to build some improvements for VaultSpeed Studio. Finally, we partly redesigned ELT generation for Talend to increase generation speed and robustness.


Snowflake is one of the most popular data platforms around these days. The success of their recent IPO emphasizes that. We support Snowflake as a target and starting now, the integration just got better.
Following up on exciting developments in Snowflake, VaultSpeed now generates ELT code for Snowflake that is wrapped in Javascript procedures. From now on you can actually deploy and store your procedures inside Snowflake. One of the main advantages of this is that our workflow solution in Airflow can call these procedures instead of executing saved SQL files.
We also enabled auto deploy for Snowflake: you can now deploy DDL and ETL to the target using VaultSpeed’s Agent. Both changes together make it possible to start loading your data warehouse without any manual interaction by your developers.


Snowflake stored procedures

VaultSpeed Studio

In a previous release we announced VaultSpeed studio in closed alpha version. In the past few weeks we went to work with initial customer feedback. First thing we improved is the integrated template code editor. It includes smoother navigation options, changing and saving a template is more solid and you now have the option to compare your changes to the pre-edit version.

Template code editor

We added a view all option to the target definition where you can see all attributes of the template target. Previously, both existing and newly created target attributes were shown in separate windows.

In a previous post we explained the need for signature fields when doing automation. We did a complete redesign of the signature definition screen. You can create and select signature attributes on top of the screen and assign them to attributes in a list below. This list can be filtered before assigning a certain signature to a filtered set of fields.

Signature attributes

VaultSpeed studio will move to open alpha in one of the next releases. From then on, all clients can start a one month free trial period with VS Studio.


Talend is one of the first ELT tools for which we supported automation. This is done by generating jobscript that can be deployed to generate ELT mappings inside Talend Studio’s repository.

Due to improvements and changes in their product, the need for a major update of our template compiler for Talend became apparent. The result is that the generation of Talend jobscript is much more robust. We also took a giant leap in terms of speed: jobscript generation is up to 3 times faster compared to the previous release.

Quality of life changes

Some smaller changes will certainly improve user experience:

  • We added a download button to the automatic deployment screen. This allows users to download generated code through the browser instead of having to obtain it from the agent.



  • We improved the description of the attribute based SAT split.
  • You can directly upload CSV files to update short and abbreviated names in the source trough your browser.
  • We moved the automatic deployment menu out of the generation settings since it didn't really belong there.
  • Users will experience improved loading performance of parameter screens.
  • We added an extra tab to the DDL settings page where you can get an overview of all applied settings per layer.
  • Business views are checked for naming conflicts with other objects. We also provided users with a button to disable/enable generation for all business views (to use in combination with filters).
  • Releases in the delta generation tab are now sorted such that the latest one is always shown first.
  • We added subscription updates as data points to the generations graph on the dashboard. It will also show the next subscription reset date.
    Based on the number of jobs granted in your subscription, we set a warning when the mapping counter is above 80% or at 100% of the subscription limit.
  • And a lot more that did not make this post... 🤷‍♂️


Our next release is coming up quite fast. Would you like to stay tuned on future VaultSpeed releases? Fill in this form 👇

VaultSpeed Studio (Release 4.1.16)

VaultSpeed Studio (Release 4.1.16)

VaultSpeed STUDIO

Yesterday, VaultSpeed launched release 4.1.16. We are proud to announce a brand new module to our automation tool: VaultSpeed Studio.



The introduction of VaultSpeed Studio is an important milestone for VaultSpeed and our Customers. It opens up the powerful VaultSpeed template language and allows you to build, test and put to work your own custom templates to generate any logic required.

VaultSpeed Studio offers a solution for those clients looking to put automation to work in other parts of their data warehouse. The raw Data vault consists of highly repeatable patterns. On the contrary, pre-staging area's or presentation area's require more specific logic. Studio enables customers to automate these patterns as well.

This first version is released in closed alpha (i.e. only accessible on request). It allows Customers and Partners to build templates for the creation and loading of objects in the Business Vault Area. The following types of objects can, for example, be created: calculated satellites, effectivity satellites, end date satellites (for insert only), custom PITs, custom Bridges and much more.

Repeatability versus customization

VaultSpeed covers almost any object in the Raw Data Vault. We also cover standard PITs and Bridges tables in the Business Vault area. These objects are highly repeatable as they look the same in every organisation.

Pre-staging area's, presentation area's and access layers and certain business vault objects require a different approach. While still being a good target for automation, they require more customization and creativity. They do not necessarily look the same across organisations.

We situate VaultSpeed Studio in the area where objects become a less repeatable and require more customization.

The Interface

VaultSpeed Studio comes as a separate module. In this release we added the studio module for the Business Vault. Once you select a Business Vault release it will appear in the top menu.

VS Studio template overview

After opening Studio, a main window will appear where you can see a list of saved templates. When starting a new template or when changing an existing one, you can use populair code editors like Visual Studio Code (for which we deliver a code plugin) or you can view, edit and test the code in the editor screen.

VaultSpeed Studio template example

Template Language

Our template language was built to distinguish logic from implementation details. By result the language is quite easy to learn.

Signature attributes

When building templates based on metadata it is necessary to make a categorization in fields. In VaultSpeed, we do this by defining signature attributes. A signature attribute from a source model can be a Business Key, Relation Key, or other. Next to the standard signature objects you can now define new signature attrubutes on which generic logic can be applied (repeatable pattern).

The more generic a signature attribute, the more repeatable the patterns that you can create with them. VaultSpeed's standard templates use these highly generic signature attribute for the generation of Raw Data Vault objects and Business Vault objects like PITs and Bridges.

However, when you are building custom templates, you must be able to create custom signature attribute. So when you would need to perform a certain operation on a set of custom fields, you will be able to define your own signature attributes in VaultSpeed Studio.

Apply a custom signature to a certain set of fields

Template language

The VSS template language has a hierarchical structure. Each level in the hierarchy starts with a certain keyword and continues until the next one or the end of the file. Each of these levels can be conditioned based on certain parameters and variables, this allows you to combine the logic for different types of objects into 1 template.


VS Studio will be released in open alpha in one of the following releases. By then it will include a limited 30 day trial period of VSS for everyone, full usage of VSS will require an additional subscription (more details about pricing coming soon).

While VSS is probably the most important change in VS 4.1.16, there is some other new stuff included:

Foreign keys to REF tables

We added the ability to define foreign keys to reference tables. You can create relations between an object and a reference table and this will result in an additional ref_fk attribute in the satellite, these foreign key attributes will be coalesced with null values. This, combined with the addition of ghost records to the REF objects allows for inner joins between Satellites and Reference tables.

Multi Active Satellites

You can now create multi active satellites with a sub-sequence attribute in the source. An attribute can be set to be the sub-sequence attribute in the source graphical editor, this attribute in combination with the business key should be unique.

setting a multi-active SAT with a subsequence

Other features

We added a status indication for new and deleted objects in the grouping and DV overview screens. This change also introduces some performance improvements for these screens.

Selecting a reset option (i.e. not “keep everything”) when creating a new source release will reset the release to the latest one linked to a DV production release. If no production date is set yet, then it will reset fully like before.

There is an additional attribute type available in the source graphical editor: modification date. When the cdc type is modification date then setting a column to this type will indicate that it is the cdc-timestamp metadata attribute for a certain object.

It is now possible to delete entire projects: this can be useful for deleting a sandbox project that had been created for testing purposes. In this way you can keep your account nice and clean from clutter.

We added the ability to define unique keys in the source graphical editor, they can be defined by selecting attributes while holding ctrl/command and then right clicking and selecting unique key (and optionally a name for the key). You can also set an attribute to UK from the attribute sidebar.


VaultSpeed invests heavily in making the complex process of enterprise wide data integration as easy as possible. With this in mind, we will keep improving towards intuitive graphical interfaces that should assist the user in the decision-making process by implementing smart suggestions and eliminating as much of the manual work as possible.

Not all layers of a data warehouse can match the repeatability of the raw data vault layer. Pre-staging layers, Business data vault objects or data marts demand a higher level of customization. Vaultspeed will extend its capabilities to add custom logic so that our customers can work towards tailor-made solutions for their data integration challenges.

“VaultSpeed aims to deliver economies of scale in all layers of the modern data warehouse. We realize that a trade-off exists between customization and repeatability. We also realize that the economies of scale decrease with increasing levels of customization.

By designing your data pipeline using repeatable patterns you can reduce the customizations and deliver high quality and high value through automation.

With VS Studio, we took another step towards these goals. Eventually, Studio will help customers to exploit the benefits of automation all the way from source to presentation layer.

Additionally we will work together with our partners to create industry specific repeatable patterns based data pipelines that will create value for our customers.”

Dirk Vermeiren - CTO VaultSpeed

So stay tuned... more features are coming soon!

Referential Integrity Error Handling (Release 4.1.15)

Referential Integrity Error Handling (Release 4.1.15)

VaultSpeed release 4.1.15 just went live! Here’s what’s new...

Referential Integrity

When integrating multiple source systems into one data vault, you might encounter some issues with referential integrity. This can have numerous reasons like issues with source batch processing, late data arrival, source mismatches or the possibility of early arriving data.

In any case it will cause you some headaches to fix these kind of problems. With VaultSpeed, you now have an extra set of tools to tackle these problems.

In the latest release we added referential integrity error handling logic. We added two new parameters: REFERENTIAL_INTEGRITY_VALIDATED and REFERENTIAL_INTEGRITY_FORCE_LINK_LOAD.

What happens when you activate the first parameter? VaultSpeed will store records in an error table if it cannot find the referenced key. These records will not be inserted in the link tables. Not loaded unless you activate forced link loading, it that case the tool will add a reference to the ‘unknown’ record.

In addition VaultSpeed will add error table records when processing subsequent loads. This allows us to check whether the reference can be resolved with new data available.

Other Changes

  • You can now create a self reference relationship which contains an attribute that references itself i.e. (c,b) → (a,b). You can do this by right clicking the attribute while in relation creation mode.
Self Reference options in VaultSpeed
  • We added the ability to update and save multiple records at the same time to the following screens: Source objects & attributes, DV name management, DV data types screens, Business vault views.
New Save buttons
  • You can now revert SAT splits.
  • We added subscription info to the dashboard. You can also see the actual number of concurrent users.
VaultSpeed Dashboard page
  • You can now switch the ODI_FOLDER_SPLIT parameter after having already deployed the project (for customers using Oracle Data Integrator).

SAT split editor improvements

Our developers also made some improvements to the SAT-split editors based upon customer feedback (we added a new SAT-split screen in a previous release):

    • We now have an option to change the number of visible attributes in a split (just click the button with 3 vertical dots and go to settings).
    • We added an option to sort the attributes in descending order in the context menu.
    • The editor now has the option to show only unused attributes to the context menu.
    • You can now remove attributes from a split or add them to another split by dragging and dropping.
    • Long table names are now displayed correctly.
    • Improved handling of multiple splits.
    • The split name will no longer disappear when you scroll down the attribute list of a split.

Stay tuned for more release news!

Data Lineage Export (Release 4.1.14)

Data Lineage Export (Release 4.1.14)

In the upcoming release, we incorporated some important improvements. First, we added a first version of our data lineage export functionality. Second, we added support for higher customization of DDL Options. We also added an alternative screen to handle satellite splitting. And finally, after introducing a lot of new objects in the business vault, we added business vault loading to our Flow Management solution in Apache Airflow.

At VaultSpeed, we invest a lot of time analyzing customer feedback and enhancement requests. At regular points in time, we try to align this feedback with our product roadmap in order to stay in line with customers needs.

Data Lineage Export

With VaultSpeed we gather source metadata. This metadata can be enhanced, maintained and fine-tuned by users. In the end we use all this metadata to generate a valid Data Vault 2.0 model. With this in mind, we have all necessary information to provide customers with proper data lineage. It is now possible to export this information to a CSV format.

Exported Data lineage in csv format

Starting at the Data Vault overview screen, users will see 2 new buttons. The first button is for object level data lineage, the second one you can use for attribute level lineage. When you click these buttons a CSV file concerning the entire data vaults object or attribute lineage will be downloaded trough your browser.

This is the first iteration of data lineage functionality in VaultSpeed, we are looking at alternative export formats and integrations with popular data lineage tools and possibly also some data lineage views inside VaultSpeed.

DDL Options for each table type

Using VaultSpeed, you were already able to define certain DDL settings. With these settings, you can create optional DDL layouts for the tables that will be created throughout all layers of your data warehouse. For example you are able to define a certain level of parallelism, or set a specific partitioning type or other, depending on the specificities of your target database.

We now made it possible to create these DDL settings at table type level. So you could for example create a different setting for a HUB compared to a satellite.

Enhanced SAT split screen

Our UX designers try to improve VaultSpeed's look and feel continuously. They came up with a different layout for our satellite splitting screen. This version should work a lot better when you are splitting tables into more than 2 sattelite splits. Also, it is now possible to show multiple satellites in one canvas. Finally, it is more aligned with other graphical screens in the tool. So users should recognise the menu's from other screens.

New SAT-split screen

All previous functionalities are still available in this screen. The old screen is still present for user that prefer to work in the old one. We will analyze customer feedback as to which screen gets the best reviews.

Business Vault loading added to FMC

The FMC module or Flow Management Center is our Apache Airflow based solution to orchestrate flows. This module can be purchased in addition to a standard licence. It offers a solution for customers that additionally require an orchestration tool. In recent releases, we added a lot of new functionality in the Business Vault layer. This new logic can now also be loaded using our FMC module. VaultSpeed will generate Airflow DAGs for loading of PITs, Bridges and Cross source links. Read more about FMC on our features page.

Our FMC module generates DAGs for Apache Airflow

Tanks for reading. That's all for this release, stay tuned!

Best regards,
The VaultSpeed Team

Cross Source Links (Release 4.1.12/13)

With release 4.1.12 and 4.1.13 we brought some interesting new features.

First of all, we included support for cross source exploration links. The main idea behind building an enterprise wide data warehouse is to be able to integrate data from multiple sources. This requirement is now delivered out of the box in VaultSpeed.

Vaultspeed already provided HUB Groups to integrate between sources at business key level. With our new Cross Source Link objects, you can create links between multiple sources. This adds an additional way to integrate and explore between sources.

In the source graphical editor an option has been added to select entities from other sources and their specific releases. By selecting and adding the object to the canvas, a relationship can be defined between objects from different sources. Notice that foreign objects will be shown in a different color and that the source name is mentioned in the object name.

Import object from another source and add a relationship

Once the object and the relationship have been added, new data vault and business vault releases can be generated. The resulting relationship will add a link object between both sources.

New link is added to the data vault model

We integrate cross source link objects at the Business Vault level. There are several reasons why this is the best place for the integration:

First of all, cross source links create dependencies between sources, we do not want inter source dependencies between sources in the raw data vault because it would complicate the loading process.

Second, we have to make sure that all the data has arrived in the raw data vault to process our cross source links. With data coming in from different sources at different times, there might be issues with “missing links”. VaultSpeed takes care of this issue.

Finally, by placing the cross source link in the business vault layer, we make it less dependent from underlying source changes. This facilitates version management and release management of these structures.

Also included in this release is a new type of delete logic, INSERT_ON_DELETE: when this parameter is enabled, deletes will insert a new record in the satellite instead of executing an update, these delete records will be end dated just like regular records. This parameter has no effect when using INSERT_ONLY_LOGIC.

We also improved performance and look and feel for the source graphical editor. The objects will now also be displayed in a tree like structure to give a better overview of the source.

Tree like structure to show source objects in the graphical editor