Sparkle - LOGOSparkle - LOGO
  • Home
  • Agile Information Factory
  • Reference Architecture
  • Jobs
  • Blog
  • Contact
METADATAMANAGEMENT AGILITY CHANGEMGT COSTREDUCTION AZURE DATA CATALOGANALYTIXDS BIG DATA AGILITY ALLDATA COSTREDUCTION HD INSIGHTWITH SPARK MICROSOFTR SERVERSQL SERVERR SERVICES POLYBASE CLOUD AGILITY SCALABILITY COSTREDUCTION CLOUD: CORTANA INTELLIGENCE SUITE SANDBOXING AGILITY DISCOVERY ANALYTICS POLYBASE POLYBASEHDINSIGHT POWER BISASS DATAVIRTUALIZATION AGILITY ALLDATA COSTREDUCTION SQL SERVERIN MEMORYAZURE SQLDWH POLYBASEHDINSIGHT POWER BISASS ENSEMBLEMODELING EXTENDABLE RESILIENCETOCHANGE DATARECORDER DATA VAULT 2.0 DATARECORDER AGILITY COSTREDUCTION DATA WAREHOUSEAUTOMATION FOUNDATION ACCELERATOR

Agile Information Factory

The Agile Information Factory is Sparkle’s framework which positions the latest information management concepts. It is a guide to define how new evolutions in information management can help your organization to induce innovation, add value, reduce cost and increase agility. Not all concepts will be required in each and every organization, but it’s important to at least consider whether it would yes or no add value.

Data Warehouse Automation

Sparkle focuses on automation to maximize value creation and realize efficiency gains. By using automation, customers can realize efficiency gains of up to 400% and focus on value creation rather than data plumbing.

Data Vault 2.0

Data Vault solves many of the issues in classic Data Warehouse projects. One of the basics of the Data Vault data modeling methodology involves splitting business keys, descriptive information and relationships into specific artefacts (hubs, satellites and links). This makes the Data Vault model very predictable, repeatable and lends itself for automation. It’s extremely flexible (integrating a new data source will have zero impact on the existing data warehouse model), helps organisations to comply with legal regulations (full traceability) and is very scalable (with the introduction of hash keys in Data Vault 2.0 scaling each table can now be loaded in parallel).

Data Virtualization

Thanks to the ever increasing performance of hardware, persisting every layer of a data warehouse architecture is no longer required. By virtualizing certain parts of the architecture, the overall complexity and cost of the solution is decreased. This also increases the agility of the solution, since less development is required. The Presentation Layer and the Business Data Vault are typically good candidates for virtualization. Furthermore, the use of Polybase allows to combine relational with non-relational data.

Sandboxing

The goal of a sandbox is to reduce shadow IT (business users that built error prone but valuable MS Excel/MS Access databases) and to support “one of” data discovery/data mining exercises. This data can be combined from the data warehouse, a data lake and operational data. Once a specific analysis is required on a regular basis, it’s typically industrialized in the data warehouse.

Cloud

Cloud offers many advantages and Microsoft has currently the most complete and integrated data platform offering in the cloud (Cortana Intelligence Suite). Cloud enables cost reduction (easier maintenance, elastic pricing) and offers flexilibility that’s not possible on premise (for example a temporary scale up to execute high volume performance tests). Finally, it also introduces scalability in both directions (easy to scale up/down based on varying demands triggered by for example  commercial success or peak loads).

Big Data

Big Data technologies can complement a traditional data warehouse architecture and offer for example cost benefits to store data for which it’s not yet known whether it will ever be useful for later analysis (for example sensor data). Special care needs to be taken when using big data technologies and it’s important to use it for what it’s good at: mass ingestion. Updates or joins on the other hand need to be avoided (hash keys is already part of the solution here).

Metadata

Metadata mainly consists of data lineage and impact analysis. Data Lineage: trace where data in the reporting and analytics is coming from and which operations were performed on it. Impact Analysis: Analyse the impact on the Data Warehouse environment when changing source systems. Finally, a well governed business glossary will help to align on business term definitions.

© Sparkle 2019, All rights reserved. Privacy Policy. Built by Sidekick.
Manage cookie consent
To provide the best experiences, we use technologies like cookies to store and/or access device information.
Functional Always active
The technical storage or access is strictly necessary for the legitimate purpose of enabling the use of a specific service explicitly requested by the subscriber or user, or for the sole purpose of carrying out the transmission of a communication over an electronic communications network.
Preferences
The technical storage or access is necessary for the legitimate purpose of storing preferences that are not requested by the subscriber or user.
Statistics
The technical storage or access that is used exclusively for statistical purposes. The technical storage or access that is used exclusively for anonymous statistical purposes. Without a subpoena, voluntary compliance on the part of your Internet Service Provider, or additional records from a third party, information stored or retrieved for this purpose alone cannot usually be used to identify you.
Marketing
The technical storage or access is required to create user profiles to send advertising, or to track the user on a website or across several websites for similar marketing purposes.
Manage options Manage services Manage vendors Read more about these purposes
View precerences
{title} {title} {title}