An EXASOL Webinar serie

We are back again after a long time, with a new webinar. The last one, we (Mathias and I) did together is almost four years ago. Time flies by! What's up this time?

The fictitious company FastChangeCoTM has developed a possibility not only to manufacture Smart Devices, but also to extend the Smart Devices as wearables in the form of bio-sensors to clothing and living beings. With each of these devices, a large amount of (sensitive) data is generated, or more precisely: by recording, processing and evaluating personal and environmental data.

In recent weeks I have read so many pessimistic and negative articles and comments in the social media about the state of data modeling in companies in Germany, but also worldwide.

Why? I don't know. I can't understand it.

I know many companies that invest a lot of time in data modeling because they have understood the added value. I know many companies that initially rejected data modeling as a whole, but understood its benefits through convincing and training.

Isn't it the case that we (consultants, managers, project managers, subject-matter experts, etc.) should have a positive influence on data modeling? To support our partners in projects in such a way that data modeling becomes a success? If we ourselves do not believe that data modeling is a success, then who does?

Several times I had the need for some large data sets to do some Data Vault tests at customer site, writing a blogpost, doing a demo or a webinar and many more. And sometimes I need data to do performance or data usage tests on different databases. Due to my work together with EXASOL I focused on the TPC-H tool DBGen to generate gigabytes of data.

To share my experience with DBGen generating large data sets I wrote this blogpost as a step by step instruction.

Wie bereits in meinem Blogpost Modellierung oder Business Rule beschrieben ist es notwendig sich bei der Datenmodellierung über Geschäftsobjekte, die Wertschöpfungskette, fachliche Details und die Methodik des Modellierens einige Gedanken zu machen.

Oder doch nicht? Kann ich mit Data Vault einfach loslegen? Schließlich ist Data Vault auf den ersten Blick ganz einfach. Drei Objekte: HUBs, LINKs und SAT(elliten), einem einfachen Vorgehensmodell und ein paar wenige Regeln. Brauche ich für Data Vault noch die Datenmodellierung?

The Smartest Way To Deal With The Data Integration Challenges

Authored by Dario Mangano
Edition: 1.0

Data Warehouse projects fail.

As an industry we have been battling with this phenomenon for decades. Though we have been getting better over the years, as an industry we still have a long way to go.

Fortunately some people have found ways to beat the odds. By thinking out of the box, formulating new ideas, and creating new innovative approaches these people have each somehow unlocked the secrets of successful DW programs.

Unterkategorien