Schedule

Data disruption cannot go without IT disruption (Dutch spoken)

How will IT perform if the complexity dramatically increases? What can we actually learn in the future from all the experiences we have gained? What do we have to change in order to participate in this increasingly data-driven economy in which digital transformation is the magic word for everyone? This keynote addresses this issue, and discusses recommendations on how IT specialists and IT management need to change to be able to address the actual data disruption: the IT disruption.
Read more

The data-driven organization, digital transformation, and the data economy are all hot topics in board rooms. They all imply that the role of data changes. Organizations want to use data more widely, more effectively, and more efficiently: a real data disruption. However, data disruption clearly raises the bar for developing IT systems, because it leads to more complex IT systems, involving e.g. AI, sensor technology, and real-time analytics. Luckily, so much high-tech technology is available, we can almost build whatever the business needs. So, the technology is ready, but is IT itself ready? You would think so. IT has more than fifty years of experience in data modeling, data architectures, data strategies, data warehouses and databases. If we look at our track record for developing more traditional IT systems, we have to conclude that some of our projects have not been delivered on time and within budget, and are sometimes completely cancelled. So, how will IT perform if the complexity dramatically increases? What can we actually learn in the future from all the experiences we have gained? What do we have to change in order to participate in this increasingly data-driven economy in which digital transformation is the magic word for everyone? This keynote addresses this issue, and discusses recommendations on how IT specialists and IT management need to change to be able to address the actual data disruption: the IT disruption.

  • How good is our track record with respect to developing IT systems?
  • Why don’t we deploy more code generators and self-driving technology?
  • Has data modeling actually changed in the last 30 years?
  • How should the IT specialist change to be ready for the data disruption?
  • The importance of in-depth IT knowledge at the top of the organization.
Read less

Customer insights from EWALS and AEGON and how they continuously innovate with data

At Qlik, we have made it our mission to help organizations accelerate business value through data. To lead in the digital age, where real-time decisions are critical, you need robust data pipelines to effectively access and analyse the latest and most accurate data. Qlik offers a data integration and analytics platform that continuously ingests all your data changes, automates data warehouses, manages data lakes to provide true data insight across your organization with world-class analytics.
Read more

Jaap-Willem Verheij van Wijk will open the session with an introduction to Qlik then share customer insights including EWALS Cargo Care who have continuously innovated with data to stay ahead of the curve. See how they have solved the latest data challenges in logistics and how they automate the data warehouse lifecycle. Jaap will also share how AEGON have solved their manual data stream issues and are now able to enclose multiple data sources with a small team that supports the agility of their business demands.

Read less

Data Driven: more than just technology (Dutch spoken)

APG is the largest pension provider in the Netherlands and sees data as a crucial asset from current and future business operations. APG also wants to be a leader as an executor and investor, which has led to the earmarking of data as a strategic asset. In this session Tim Schulteis tells an integral story about the journey that has been made in the past period. He will discuss the development of an appropriate architecture, the building of the right knowledge and skills, and the challenges of collaboration across business units.
Read more

APG is the largest pension provider in the Netherlands and sees data as a crucial asset from current and future business operations. The government is increasingly withdrawing from a sufficient retirement provision (raising the state retirement age, decreasing the pension accrual), so insight into the personal situation of participants and offering action perspective is crucial. APG also wants to be a leader as an executor and investor.

All this has led to the earmarking of data as a strategic asset. But how do you get from that ambition to execution on the various axes of technology, capability, culture and organization? This presentation tells an integral story about the journey that has been made in the past period: what went well, what didn’t, what did we learn from it, where are we now?
The development of an appropriate architecture, the building of the right knowledge and skills, the combination with modern working methods and the challenges of collaboration across business units will be discussed:

  • Business drivers
  • Architecture and knowledge
  • Culture and agility
  • Organization and sponsorship
  • Critical success factors
Read less

The Power of Combining Machine Learning Models - The Risks and Rewards of Random Forests, XGBoost and other Ensembles

Ensembling is one of the hottest techniques in today’s predictive analytics competitions. Every single recent winner of Kaggle.com and KDD competitions used an ensemble technique, including famous algorithms such as XGBoost and Random Forest. This session will provide a detailed overview of ensemble models, their origin, and show why they are so effective.
Read more

Ensembling is one of the hottest techniques in today’s predictive analytics competitions. Every single recent winner of Kaggle.com and KDD competitions used an ensemble technique, including famous algorithms such as XGBoost and Random Forest.
Are these competition victories paving the way for widespread organizational implementation of these techniques? This session will provide a detailed overview of ensemble models, their origin, and show why they are so effective. We will explain the building blocks of virtually all ensembles techniques, to include bagging and boosting.

What You Will Learn:

  • What are ensemble models and what are their advantages?
  • Why are ensembles in the news?
  • The two most influential ensembling approaches: bagging and boosting
  • The core elements of ensembles and their application
  • The challenge of applying competition strategies to organizational problems.
Read less
Petr Beles

Model Driven Data Vault Automation with Datavault Builder

In this session we will give you a turbo introduction on how to put the data modeler back on the driver's seat. See how your data model is implemented into a working solution in real-time.
Read more
Read less

Modern Data Management & Data Integration (Dutch spoken)

The digital future: think big, think highly distributed data, think in ecosystems. What Integration Architecture is needed to play an important role in the digital world with eco-system with FinTech company's and other banks? Do Enterprise Data Warehouses still have a role in this landscape?
Read more

The digital future: think big, think highly distributed data, think in ecosystems. What Integration Architecture is needed to play an important role in the digital world with eco-system with FinTech company’s and other banks? Do Enterprise Data Warehouses still have a role in this landscape? This is what the presentation is about. Piethein will also cover:

  • Core Data Integration patterns
  • Data Ownership and Data Governance
  • Metadata as the glue
  • Data Control when data is highly distributed
  • Data distribution in the Cloud
Read less

Making Self-Service Analytics Work: Organizational, Architectural, and Governance Issues

It is notoriously difficult to achieve the promise of self-service analytics. Wayne Eckerson will explain how to empower business users to create their own reports and models without creating data chaos. He will show how to build a self-sustaining analytical culture that balances speed and standards, agility and architecture, and self-service and governance.
Read more

Self-service analytics has been the holy grail of data analytics leaders for the past two decades. Although analytical tools have improved significantly, it is notoriously difficult to achieve the promise of self-service analytics. This session will explain how to empower business users to create their own reports and models without creating data chaos. Specifically, it examines seven factors for leading a successful BI program: right roles, right processes, right tools, right organization, right architecture, right governance, and right leadership. Ultimately, it will show how to build a self-sustaining analytical culture that balances speed and standards, agility and architecture, and self-service and governance.

You will learn:

  • Trends and business dynamics driving analytics adoption
  • The conundrum of self-service analytics
  • Success factors for leading a successful BI program
  • How to survive and thrive in the new world of big data analytics
  • How to increase user adoption and facilitate self service
Read less

Data Governance and Architecture – Making the connections

Data Governance is one of the hottest topics in data management, focusing both on how Governance driven change can enable companies to gain better leverage from their data and also to help them design and enforce the controls needed to ensure they remain compliant with regulations. Despite this rapidly growing focus, many Data Governance initiatives fail to meet their goals. This session will outline why Data Governance and Architecture should be connected, how to make it happen, and what part Business Intelligence and Data Warehousing will play in defining a robust and sustainable Governance programme.
Read more

With data increasingly being seen as a critical corporate asset, more organisations are embracing the concepts and practices of Data Governance. As a result Data Governance is today one of the hottest topics in data management, focusing both on how Governance driven change can enable companies to gain better leverage from their data through enhanced Business Intelligence, Data Analytics and so on, and also to help them design and enforce the controls needed to ensure they remain compliant with increasingly stringent laws and regulations, such as GDPR.

Despite this rapidly growing focus, many Data Governance initiatives fail to meet their goals, with only around one in five fully achieving expectations. Why is the failure rate so high? There are many factors, but one key reason is that implementing Data Governance without aligning it with a defined enterprise and data architecture is fraught with dangers. Linking Architecture with data accountability, a core principle of Data Governance, is essential.
This session will outline why Data Governance and Architecture should be connected, how to make it happen, and what part Business Intelligence and Data Warehousing play in defining a robust and sustainable Governance programme.

This talk will cover:

  • What is Data Governance and what it is not
  • Key reasons for Data Governance failure & disappointment
  • The key components of enterprise architecture – Business, Process and Data
  • The synergies between architecture and Governance – how do they reinforce each other?
  • How artefacts from both disciplines can be combined and applied to ensure success
  • The implications for Business Intelligence and Data Warehousing
  • Several use cases of successes and lessons learned
Read less

Cloud Data Warehousing: Planning for Data Warehouse Migration

Migrating an existing data warehouse to the cloud is a complex process of moving schema, data, and ETL. The complexity increases when architectural modernization, restructuring of database schema or rebuilding of data pipelines is needed. In this session Dave Wells provides an overview of the benefits, techniques, and challenges when migrating an existing data warehouse to the cloud.
Read more

Cloud data warehousing helps to meet the challenges of legacy data warehouses that struggle to keep up with growing data volumes, changing service level expectations, and the need to integrate structured warehouse data with unstructured data in a data lake. Cloud data warehousing provides many benefits, but cloud migration isn’t fast and easy. Migrating an existing data warehouse to the cloud is a complex process of moving schema, data, and ETL. The complexity increases when architectural modernization, restructuring of database schema or rebuilding of data pipelines is needed.

This session provides an overview of the benefits, techniques, and challenges when migrating an existing data warehouse to the cloud. We will discuss the pros and cons of cloud migration, explore the dynamics of migration decision making, and look at migration pragmatics within the framework of a step-by-step approach to migrating. The tips and techniques described here will help you to make informed decisions about cloud migration and address the full scope of migration planning.

You Will Learn:

  • The what and why of cloud data warehousing
  • The benefits and challenges of cloud data warehousing
  • Migration analysis and decision making
  • Technology roles in migration to the cloud
  • A step-by-step framework for data warehouse migration.
Read less

Data disruption cannot go without IT disruption (Dutch spoken)

How will IT perform if the complexity dramatically increases? What can we actually learn in the future from all the experiences we have gained? What do we have to change in order to participate in this increasingly data-driven economy in which digital transformation is the magic word for everyone? This keynote addresses this issue, and discusses recommendations on how IT specialists and IT management need to change to be able to address the actual data disruption: the IT disruption.
Read more

The data-driven organization, digital transformation, and the data economy are all hot topics in board rooms. They all imply that the role of data changes. Organizations want to use data more widely, more effectively, and more efficiently: a real data disruption. However, data disruption clearly raises the bar for developing IT systems, because it leads to more complex IT systems, involving e.g. AI, sensor technology, and real-time analytics. Luckily, so much high-tech technology is available, we can almost build whatever the business needs. So, the technology is ready, but is IT itself ready? You would think so. IT has more than fifty years of experience in data modeling, data architectures, data strategies, data warehouses and databases. If we look at our track record for developing more traditional IT systems, we have to conclude that some of our projects have not been delivered on time and within budget, and are sometimes completely cancelled. So, how will IT perform if the complexity dramatically increases? What can we actually learn in the future from all the experiences we have gained? What do we have to change in order to participate in this increasingly data-driven economy in which digital transformation is the magic word for everyone? This keynote addresses this issue, and discusses recommendations on how IT specialists and IT management need to change to be able to address the actual data disruption: the IT disruption.

  • How good is our track record with respect to developing IT systems?
  • Why don’t we deploy more code generators and self-driving technology?
  • Has data modeling actually changed in the last 30 years?
  • How should the IT specialist change to be ready for the data disruption?
  • The importance of in-depth IT knowledge at the top of the organization.
Read less

Customer insights from EWALS and AEGON and how they continuously innovate with data

At Qlik, we have made it our mission to help organizations accelerate business value through data. To lead in the digital age, where real-time decisions are critical, you need robust data pipelines to effectively access and analyse the latest and most accurate data. Qlik offers a data integration and analytics platform that continuously ingests all your data changes, automates data warehouses, manages data lakes to provide true data insight across your organization with world-class analytics.
Read more

Jaap-Willem Verheij van Wijk will open the session with an introduction to Qlik then share customer insights including EWALS Cargo Care who have continuously innovated with data to stay ahead of the curve. See how they have solved the latest data challenges in logistics and how they automate the data warehouse lifecycle. Jaap will also share how AEGON have solved their manual data stream issues and are now able to enclose multiple data sources with a small team that supports the agility of their business demands.

Read less

Data Driven: more than just technology (Dutch spoken)

APG is the largest pension provider in the Netherlands and sees data as a crucial asset from current and future business operations. APG also wants to be a leader as an executor and investor, which has led to the earmarking of data as a strategic asset. In this session Tim Schulteis tells an integral story about the journey that has been made in the past period. He will discuss the development of an appropriate architecture, the building of the right knowledge and skills, and the challenges of collaboration across business units.
Read more

APG is the largest pension provider in the Netherlands and sees data as a crucial asset from current and future business operations. The government is increasingly withdrawing from a sufficient retirement provision (raising the state retirement age, decreasing the pension accrual), so insight into the personal situation of participants and offering action perspective is crucial. APG also wants to be a leader as an executor and investor.

All this has led to the earmarking of data as a strategic asset. But how do you get from that ambition to execution on the various axes of technology, capability, culture and organization? This presentation tells an integral story about the journey that has been made in the past period: what went well, what didn’t, what did we learn from it, where are we now?
The development of an appropriate architecture, the building of the right knowledge and skills, the combination with modern working methods and the challenges of collaboration across business units will be discussed:

  • Business drivers
  • Architecture and knowledge
  • Culture and agility
  • Organization and sponsorship
  • Critical success factors
Read less

The Power of Combining Machine Learning Models - The Risks and Rewards of Random Forests, XGBoost and other Ensembles

Ensembling is one of the hottest techniques in today’s predictive analytics competitions. Every single recent winner of Kaggle.com and KDD competitions used an ensemble technique, including famous algorithms such as XGBoost and Random Forest. This session will provide a detailed overview of ensemble models, their origin, and show why they are so effective.
Read more

Ensembling is one of the hottest techniques in today’s predictive analytics competitions. Every single recent winner of Kaggle.com and KDD competitions used an ensemble technique, including famous algorithms such as XGBoost and Random Forest.
Are these competition victories paving the way for widespread organizational implementation of these techniques? This session will provide a detailed overview of ensemble models, their origin, and show why they are so effective. We will explain the building blocks of virtually all ensembles techniques, to include bagging and boosting.

What You Will Learn:

  • What are ensemble models and what are their advantages?
  • Why are ensembles in the news?
  • The two most influential ensembling approaches: bagging and boosting
  • The core elements of ensembles and their application
  • The challenge of applying competition strategies to organizational problems.
Read less
Petr Beles

Model Driven Data Vault Automation with Datavault Builder

In this session we will give you a turbo introduction on how to put the data modeler back on the driver's seat. See how your data model is implemented into a working solution in real-time.
Read more
Read less

Modern Data Management & Data Integration (Dutch spoken)

The digital future: think big, think highly distributed data, think in ecosystems. What Integration Architecture is needed to play an important role in the digital world with eco-system with FinTech company's and other banks? Do Enterprise Data Warehouses still have a role in this landscape?
Read more

The digital future: think big, think highly distributed data, think in ecosystems. What Integration Architecture is needed to play an important role in the digital world with eco-system with FinTech company’s and other banks? Do Enterprise Data Warehouses still have a role in this landscape? This is what the presentation is about. Piethein will also cover:

  • Core Data Integration patterns
  • Data Ownership and Data Governance
  • Metadata as the glue
  • Data Control when data is highly distributed
  • Data distribution in the Cloud
Read less

Making Self-Service Analytics Work: Organizational, Architectural, and Governance Issues

It is notoriously difficult to achieve the promise of self-service analytics. Wayne Eckerson will explain how to empower business users to create their own reports and models without creating data chaos. He will show how to build a self-sustaining analytical culture that balances speed and standards, agility and architecture, and self-service and governance.
Read more

Self-service analytics has been the holy grail of data analytics leaders for the past two decades. Although analytical tools have improved significantly, it is notoriously difficult to achieve the promise of self-service analytics. This session will explain how to empower business users to create their own reports and models without creating data chaos. Specifically, it examines seven factors for leading a successful BI program: right roles, right processes, right tools, right organization, right architecture, right governance, and right leadership. Ultimately, it will show how to build a self-sustaining analytical culture that balances speed and standards, agility and architecture, and self-service and governance.

You will learn:

  • Trends and business dynamics driving analytics adoption
  • The conundrum of self-service analytics
  • Success factors for leading a successful BI program
  • How to survive and thrive in the new world of big data analytics
  • How to increase user adoption and facilitate self service
Read less

Data Governance and Architecture – Making the connections

Data Governance is one of the hottest topics in data management, focusing both on how Governance driven change can enable companies to gain better leverage from their data and also to help them design and enforce the controls needed to ensure they remain compliant with regulations. Despite this rapidly growing focus, many Data Governance initiatives fail to meet their goals. This session will outline why Data Governance and Architecture should be connected, how to make it happen, and what part Business Intelligence and Data Warehousing will play in defining a robust and sustainable Governance programme.
Read more

With data increasingly being seen as a critical corporate asset, more organisations are embracing the concepts and practices of Data Governance. As a result Data Governance is today one of the hottest topics in data management, focusing both on how Governance driven change can enable companies to gain better leverage from their data through enhanced Business Intelligence, Data Analytics and so on, and also to help them design and enforce the controls needed to ensure they remain compliant with increasingly stringent laws and regulations, such as GDPR.

Despite this rapidly growing focus, many Data Governance initiatives fail to meet their goals, with only around one in five fully achieving expectations. Why is the failure rate so high? There are many factors, but one key reason is that implementing Data Governance without aligning it with a defined enterprise and data architecture is fraught with dangers. Linking Architecture with data accountability, a core principle of Data Governance, is essential.
This session will outline why Data Governance and Architecture should be connected, how to make it happen, and what part Business Intelligence and Data Warehousing play in defining a robust and sustainable Governance programme.

This talk will cover:

  • What is Data Governance and what it is not
  • Key reasons for Data Governance failure & disappointment
  • The key components of enterprise architecture – Business, Process and Data
  • The synergies between architecture and Governance – how do they reinforce each other?
  • How artefacts from both disciplines can be combined and applied to ensure success
  • The implications for Business Intelligence and Data Warehousing
  • Several use cases of successes and lessons learned
Read less

Cloud Data Warehousing: Planning for Data Warehouse Migration

Migrating an existing data warehouse to the cloud is a complex process of moving schema, data, and ETL. The complexity increases when architectural modernization, restructuring of database schema or rebuilding of data pipelines is needed. In this session Dave Wells provides an overview of the benefits, techniques, and challenges when migrating an existing data warehouse to the cloud.
Read more

Cloud data warehousing helps to meet the challenges of legacy data warehouses that struggle to keep up with growing data volumes, changing service level expectations, and the need to integrate structured warehouse data with unstructured data in a data lake. Cloud data warehousing provides many benefits, but cloud migration isn’t fast and easy. Migrating an existing data warehouse to the cloud is a complex process of moving schema, data, and ETL. The complexity increases when architectural modernization, restructuring of database schema or rebuilding of data pipelines is needed.

This session provides an overview of the benefits, techniques, and challenges when migrating an existing data warehouse to the cloud. We will discuss the pros and cons of cloud migration, explore the dynamics of migration decision making, and look at migration pragmatics within the framework of a step-by-step approach to migrating. The tips and techniques described here will help you to make informed decisions about cloud migration and address the full scope of migration planning.

You Will Learn:

  • The what and why of cloud data warehousing
  • The benefits and challenges of cloud data warehousing
  • Migration analysis and decision making
  • Technology roles in migration to the cloud
  • A step-by-step framework for data warehouse migration.
Read less

Data routes: combining data vault, ensemble modeling and data virtualisation (Dutch spoken)

This session explains how data vault, ensemble modeling and data virtualization can be used together in an efficient way. Key to this is the new concept of ”data routes". This concept proposes a data-oriented way of processing that is based on the aforementioned issues of data vault, ensemble modeling and data virtualization.
Read more

Data vault, ensemble logical modeling, data virtualization and cloud are known to every BI or data warehouse specialist. But the big question is how you can use them together to develop real-life systems and then make optimum use of the power and possibilities of each component. This session explains how all can be used efficiently together. Key to this is the new concept of “data routes”. Within a data and analytics architecture, data routes serve as a fuel for the virtual data presentation layer that is accessed by end users for all their data needs.

The concept proposes a data-oriented way of processing that rests on the aforementioned issues such as data vault, ensemble modeling and data virtualization. A decoupling of data and technology is hereby realized whereby the emphasis is shifted to the characteristics of the data and the requirements set by use cases. The result is offered as a virtual (semantic) data layer to a broad group of data users. With the help of data virtualization, a virtual data collection is built as a virtual data portal for data users.

  • Does a Cloud Analytics platform offer a complete solution?
  • How does the Data Routes concept fit in an existing data architecture for Data & Analytics?
  • Is data modeling not necessary anymore?
  • How do data routes and data virtualization fit together?
  • From Ensemble logical modeling to data vault databases
Read less

Turning Data into Innovation - TIBCO Connected Intelligence for Enterprises

Integration and uniformity are cardinal aspects of modern data management. Which aspects are most essential in the information landscape?
Read more

To unify the collection of data is necessary in order to ultimately make better decisions. Without uniformity, a decision culture is based on loose sand and a gut feeling. In our contemporary information landscape, we see a greater need to integrate more and more – cloud services and storage, new sources of information, as well as API LED developments. What makes Data Management more than current! Which aspects are essential for today’s information landscape?

Data Virtualization, Data Quality, Reference Data Management, Master Data Management, and Metadata Management as part of Data Management Enable organizations to coordinate the different data silos and improve their decisions.

Our central question is: “How can TIBCO support digital transformation initiatives in this?” Data is the foundation for operational excellence, customer intimacy, and Business Reinvention. The role of TIBCO’s Unify portfolio is the cornerstone of a data-driven initiative for Operations, Data Governance, and Analytics.

  • What is the impact of data virtualization when using a data warehouse?
  • Do API LED transformations play a role in information sharing?
  • How can users gain faster insight into all company data?
  • What is the need for standardization of reference data?
  • What is the role of streaming analytics for a Data Science environment?
Read less

Cloud Database Systems in-depth: how do they work and how do they compare? (Dutch spoken)

There are several well know vendors that offer cloud database systems, such as Amazon, Microsoft and Google. In recent year however we have seen some new entries by companies that specialize in cloud services, such as Snowflake and Databricks. Peter Boncz offers an in-depth technical review and comparison of the existing and new generation cloud database systems.
Read more

In de afgelopen vijf jaar zijn cloud databasesystemen echt doorgebroken. De cloud maakt het mogelijk om kapitaal investeringen vooraf om te zetten in operationele kosten, zodat men alleen betaalt voor de capaciteit die echt is; en er nooit zorgen hoeven te zijn over capaciteitsproblemen. Daarbovenop “ontzorgen” cloud database systemen in de zin dat het beheer van de database systemen en onderliggende hardware bij de cloud provider ligt. In tijden van personele schaarste is dat een andere belangrijke factor achter het succes van cloud database systemen, die de eventuele nadelen op het gebied van lock-in en zorgen rond privacy en security vaak neutraliseert.

Maar, als eenmaal het besluit is genomen om de database naar de cloud te brengen, welke dan te kiezen? Er zijn op dit moment al een heleboel cloud systemen. Amazon heeft onder andere Aurora, Redshift, Neptune en Athena. Microsoft heeft SQLserver en Cosmos DB. Google heeft onder andere BigQuery. En dan zijn er nieuwe bedrijven bijgekomen, die zich specialiseren in cloud services, zoals Snowflake en Databricks.

Om beter te begrijpen wat de overeenkomsten en verschillen zijn tussen al die nieuwe cloud systemen, zal Peter Boncz ingaan op wat er zich onder de motorkap van deze nieuwe systemen bevindt. De verschillende alternatieven worden technisch ontleed en met elkaar vergeleken.

Enkele van de onderwerpen die aan bod zullen komen:

  • Een introductie tot cloud data systemen met een actueel marktoverzicht van de belangrijkste kanshebbers
  • Hoe ziet de architectuur van deze verschillende systemen eruit op het gebied van query-engine, data representatie, elasticiteit en data partitionering
  • Welke diensten zijn “serverless” en wat is dat precies?
  • Kunnen cloud database systemen automatisch data optimaliseren?
  • Wat is het economische model, en verdere implicaties daarvan
  • Welke ontwikkelingen zullen nog volgen in cloud database systemen in de komende jaren? Een voorbeeld is databases delen in de cloud.
Read less

Data Preparation for Machine Learning: Why Feature Engineering Remains a Human-Driven Activity

In Data preparation it is important how the human modeler creates a dataset that is uniquely suited to the business problem. In this session Keith McCormick will expose analytic practitioners, data scientists, and those looking to get started in predictive analytics to the critical importance of properly preparing data in advance of model building. He will present the critical role of feature engineering and explaining how to do it effectively.
Read more

This session will expose analytic practitioners, data scientists, and those looking to get started in predictive analytics to the critical importance of properly preparing data in advance of model building. The instructor will present the critical role of feature engineering, explaining both what it is and how to do it effectively.

Emphasis will be given to those tasks that must be overseen by the modeler – and cannot be performed without the context of a specific modeling project. Data is carefully “crafted” by the modeler to improve the ability of modeling algorithms to find patterns of interest.

Data preparation is often associated with cleaning and formatting the data. While important, these tasks will not be our focus. Rather it is how the human modeler creates a dataset that is uniquely suited to the business problem.

You will learn:

  • Construction methods for various data transformations
  • The merits and limitations of automated data preparation technologies
  • Which data prep tasks are best performed by data scientist, and which by IT
  • Common types of constructed variables and why they are useful
  • How to effectively utilize subject matter experts during data preparation
Read less

Modernizing Data Governance for the Age of Self-Service Analytics

Self-Service analytics is not just about making sure that you provide as many datapoints as possible combined with a flexible tool to the “citizen” data scientist.
Read more

With democratization of analytical capabilities and the wider access to data, questions arise on the governance and regulatory- and ethical compliance of the data usage. Locking all data down is not the answer as we would lose too much value.
Using the Data Governance 1.0 top down and waterfall like models are not well suited to deal with the new paradigms. The presentation focusses on the steps you need to take to get sustainable and compliant value through (Self-Service) Analytics out of your big data.

You will learn:

  • How to adapt your data governance for the new ways of working
  • What is the distinction between Information and Big Data Governance
  • What should you put in place for properly governing self-service analytics
  • Increasing data literacy
  • Catering for the dynamics of data on-boarding and usage flows
  • Towards policy based classification and access
  • Use case governance vs Critical Data Elements
Read less

Best Practices in DataOps: Trends, Tips, and Techniques for Creating and Managing Modern Data Pipelines

The state of most data analytics pipelines is deplorable: too little automation; minimal reuse of code and data; and lack of coordination. The result is poor quality data delivered too late to meet business needs. DataOps is an emerging approach for building data pipelines and solutions. In this session Wayne Eckerson will explore trends in DataOps adoption, challenges and best practices.
Read more

When it comes to data analytics, you don’t want to know “how the sausage is made.” The state of most data analytics pipelines is deplorable. There are too many steps; too little automation and orchestration; minimal reuse of code and data; and a lack of coordination between stakeholders in business, IT, and operations. The result is poor quality data delivered too late to meet business needs.

DataOps is an emerging approach for building data pipelines and solutions. This session will explore trends in DataOps adoption, challenges that organizations face in implementing DataOps, and best practices in building modern data pipelines. It will examine how leading-edge organizations are using DataOps to increase agility, reduce cycle times, and minimize data defects, giving developers and business users greater confidence in analytic output.

You will learn:

  • What is DataOps and why you need it
  • The dimensions of DataOps
  • The state of DataOps adoption
  • DataOps best practices and challenges
Read less

Managing and exploring data using a data lake and an analytics lab (Dutch spoken)

ASML, manufacturer of machines for the production of semiconductors, is implementing a central data lake to capture this data and make it accessible for reporting and analytics in a central environment. The data lake environment also includes an analytics lab for detailed exploration of data. In this session Jeroen Vermunt presents real-life examples of how ASML approaches the challenges of managing rapidly changing data.
Read more

In the highly complex world of semiconductor manufacturing vast amounts of largely varied data are generated every day. ASML, world-leader manufacturer of machines for the production of semiconductors (chips), is implementing a central data lake to capture this data and make it accessible for reporting and analytics in a central environment. The data lake environment also includes an analytics lab for detailed exploration of data. Managing all this rapidly changing data imposes some very challenging requirements. In this session, real-life examples of how ASML approaches these challenges are presented.

  • How do business users and data scientists discover information in the data lake without drowning in the amount and complexity?
  • How can users be enabled to understand the data they want to consume, including were the data comes from (data lineage)?
  • How do we make sure that the data can be trusted?
  • How can ASML control and monitor that access to data is secure?
  • What added value can the analytics lab bring?
Read less

Ten practical guidelines for modern data architectures (Dutch spoken)

Often, existing data architectures can no longer keep up with the current “speed of business change”. As a result, many organizations have decided that it is time for a new, future-proof data architecture. However, this is easier said than done. In this session, ten essential guidelines for designing modern data architectures are discussed. These guidelines are based on hands-on experiences with designing and implementing many new data architectures.
Read more

Many IT systems are more than twenty years old and have undergone numerous changes over time. Unfortunately, they can no longer cope with the ever-increasing growth in data usage in terms of scalability and speed. In addition, they have become inflexible, which means that implementing new reports and performing analyses has become very time-consuming. In short, the data architecture can no longer keep up with the current “speed of business change”. As a result, many organizations have decided that it is time for a new, future-proof data architecture. However, this is easier said than done. After all, you don’t design a new data architecture every day. In this session, ten essential guidelines for designing modern data architectures are discussed. These guidelines are based on hands-on experiences with designing and implementing many new data architectures.

  • Which new technologies are currently available?
  • What is the influence on the architecture of e.g. Hadoop, NoSQL, big data, data warehouse automation, and data streaming?
  • Which new architecture principles should be applied nowadays?
  • How do we deal with the increasingly paralyzing rules for data storage and analysis?
  • What is the influence of cloud platforms?
Read less

Data routes: combining data vault, ensemble modeling and data virtualisation (Dutch spoken)

This session explains how data vault, ensemble modeling and data virtualization can be used together in an efficient way. Key to this is the new concept of ”data routes". This concept proposes a data-oriented way of processing that is based on the aforementioned issues of data vault, ensemble modeling and data virtualization.
Read more

Data vault, ensemble logical modeling, data virtualization and cloud are known to every BI or data warehouse specialist. But the big question is how you can use them together to develop real-life systems and then make optimum use of the power and possibilities of each component. This session explains how all can be used efficiently together. Key to this is the new concept of “data routes”. Within a data and analytics architecture, data routes serve as a fuel for the virtual data presentation layer that is accessed by end users for all their data needs.

The concept proposes a data-oriented way of processing that rests on the aforementioned issues such as data vault, ensemble modeling and data virtualization. A decoupling of data and technology is hereby realized whereby the emphasis is shifted to the characteristics of the data and the requirements set by use cases. The result is offered as a virtual (semantic) data layer to a broad group of data users. With the help of data virtualization, a virtual data collection is built as a virtual data portal for data users.

  • Does a Cloud Analytics platform offer a complete solution?
  • How does the Data Routes concept fit in an existing data architecture for Data & Analytics?
  • Is data modeling not necessary anymore?
  • How do data routes and data virtualization fit together?
  • From Ensemble logical modeling to data vault databases
Read less

Turning Data into Innovation - TIBCO Connected Intelligence for Enterprises

Integration and uniformity are cardinal aspects of modern data management. Which aspects are most essential in the information landscape?
Read more

To unify the collection of data is necessary in order to ultimately make better decisions. Without uniformity, a decision culture is based on loose sand and a gut feeling. In our contemporary information landscape, we see a greater need to integrate more and more – cloud services and storage, new sources of information, as well as API LED developments. What makes Data Management more than current! Which aspects are essential for today’s information landscape?

Data Virtualization, Data Quality, Reference Data Management, Master Data Management, and Metadata Management as part of Data Management Enable organizations to coordinate the different data silos and improve their decisions.

Our central question is: “How can TIBCO support digital transformation initiatives in this?” Data is the foundation for operational excellence, customer intimacy, and Business Reinvention. The role of TIBCO’s Unify portfolio is the cornerstone of a data-driven initiative for Operations, Data Governance, and Analytics.

  • What is the impact of data virtualization when using a data warehouse?
  • Do API LED transformations play a role in information sharing?
  • How can users gain faster insight into all company data?
  • What is the need for standardization of reference data?
  • What is the role of streaming analytics for a Data Science environment?
Read less

Cloud Database Systems in-depth: how do they work and how do they compare? (Dutch spoken)

There are several well know vendors that offer cloud database systems, such as Amazon, Microsoft and Google. In recent year however we have seen some new entries by companies that specialize in cloud services, such as Snowflake and Databricks. Peter Boncz offers an in-depth technical review and comparison of the existing and new generation cloud database systems.
Read more

In de afgelopen vijf jaar zijn cloud databasesystemen echt doorgebroken. De cloud maakt het mogelijk om kapitaal investeringen vooraf om te zetten in operationele kosten, zodat men alleen betaalt voor de capaciteit die echt is; en er nooit zorgen hoeven te zijn over capaciteitsproblemen. Daarbovenop “ontzorgen” cloud database systemen in de zin dat het beheer van de database systemen en onderliggende hardware bij de cloud provider ligt. In tijden van personele schaarste is dat een andere belangrijke factor achter het succes van cloud database systemen, die de eventuele nadelen op het gebied van lock-in en zorgen rond privacy en security vaak neutraliseert.

Maar, als eenmaal het besluit is genomen om de database naar de cloud te brengen, welke dan te kiezen? Er zijn op dit moment al een heleboel cloud systemen. Amazon heeft onder andere Aurora, Redshift, Neptune en Athena. Microsoft heeft SQLserver en Cosmos DB. Google heeft onder andere BigQuery. En dan zijn er nieuwe bedrijven bijgekomen, die zich specialiseren in cloud services, zoals Snowflake en Databricks.

Om beter te begrijpen wat de overeenkomsten en verschillen zijn tussen al die nieuwe cloud systemen, zal Peter Boncz ingaan op wat er zich onder de motorkap van deze nieuwe systemen bevindt. De verschillende alternatieven worden technisch ontleed en met elkaar vergeleken.

Enkele van de onderwerpen die aan bod zullen komen:

  • Een introductie tot cloud data systemen met een actueel marktoverzicht van de belangrijkste kanshebbers
  • Hoe ziet de architectuur van deze verschillende systemen eruit op het gebied van query-engine, data representatie, elasticiteit en data partitionering
  • Welke diensten zijn “serverless” en wat is dat precies?
  • Kunnen cloud database systemen automatisch data optimaliseren?
  • Wat is het economische model, en verdere implicaties daarvan
  • Welke ontwikkelingen zullen nog volgen in cloud database systemen in de komende jaren? Een voorbeeld is databases delen in de cloud.
Read less

Data Preparation for Machine Learning: Why Feature Engineering Remains a Human-Driven Activity

In Data preparation it is important how the human modeler creates a dataset that is uniquely suited to the business problem. In this session Keith McCormick will expose analytic practitioners, data scientists, and those looking to get started in predictive analytics to the critical importance of properly preparing data in advance of model building. He will present the critical role of feature engineering and explaining how to do it effectively.
Read more

This session will expose analytic practitioners, data scientists, and those looking to get started in predictive analytics to the critical importance of properly preparing data in advance of model building. The instructor will present the critical role of feature engineering, explaining both what it is and how to do it effectively.

Emphasis will be given to those tasks that must be overseen by the modeler – and cannot be performed without the context of a specific modeling project. Data is carefully “crafted” by the modeler to improve the ability of modeling algorithms to find patterns of interest.

Data preparation is often associated with cleaning and formatting the data. While important, these tasks will not be our focus. Rather it is how the human modeler creates a dataset that is uniquely suited to the business problem.

You will learn:

  • Construction methods for various data transformations
  • The merits and limitations of automated data preparation technologies
  • Which data prep tasks are best performed by data scientist, and which by IT
  • Common types of constructed variables and why they are useful
  • How to effectively utilize subject matter experts during data preparation
Read less

Modernizing Data Governance for the Age of Self-Service Analytics

Self-Service analytics is not just about making sure that you provide as many datapoints as possible combined with a flexible tool to the “citizen” data scientist.
Read more

With democratization of analytical capabilities and the wider access to data, questions arise on the governance and regulatory- and ethical compliance of the data usage. Locking all data down is not the answer as we would lose too much value.
Using the Data Governance 1.0 top down and waterfall like models are not well suited to deal with the new paradigms. The presentation focusses on the steps you need to take to get sustainable and compliant value through (Self-Service) Analytics out of your big data.

You will learn:

  • How to adapt your data governance for the new ways of working
  • What is the distinction between Information and Big Data Governance
  • What should you put in place for properly governing self-service analytics
  • Increasing data literacy
  • Catering for the dynamics of data on-boarding and usage flows
  • Towards policy based classification and access
  • Use case governance vs Critical Data Elements
Read less

Best Practices in DataOps: Trends, Tips, and Techniques for Creating and Managing Modern Data Pipelines

The state of most data analytics pipelines is deplorable: too little automation; minimal reuse of code and data; and lack of coordination. The result is poor quality data delivered too late to meet business needs. DataOps is an emerging approach for building data pipelines and solutions. In this session Wayne Eckerson will explore trends in DataOps adoption, challenges and best practices.
Read more

When it comes to data analytics, you don’t want to know “how the sausage is made.” The state of most data analytics pipelines is deplorable. There are too many steps; too little automation and orchestration; minimal reuse of code and data; and a lack of coordination between stakeholders in business, IT, and operations. The result is poor quality data delivered too late to meet business needs.

DataOps is an emerging approach for building data pipelines and solutions. This session will explore trends in DataOps adoption, challenges that organizations face in implementing DataOps, and best practices in building modern data pipelines. It will examine how leading-edge organizations are using DataOps to increase agility, reduce cycle times, and minimize data defects, giving developers and business users greater confidence in analytic output.

You will learn:

  • What is DataOps and why you need it
  • The dimensions of DataOps
  • The state of DataOps adoption
  • DataOps best practices and challenges
Read less

Managing and exploring data using a data lake and an analytics lab (Dutch spoken)

ASML, manufacturer of machines for the production of semiconductors, is implementing a central data lake to capture this data and make it accessible for reporting and analytics in a central environment. The data lake environment also includes an analytics lab for detailed exploration of data. In this session Jeroen Vermunt presents real-life examples of how ASML approaches the challenges of managing rapidly changing data.
Read more

In the highly complex world of semiconductor manufacturing vast amounts of largely varied data are generated every day. ASML, world-leader manufacturer of machines for the production of semiconductors (chips), is implementing a central data lake to capture this data and make it accessible for reporting and analytics in a central environment. The data lake environment also includes an analytics lab for detailed exploration of data. Managing all this rapidly changing data imposes some very challenging requirements. In this session, real-life examples of how ASML approaches these challenges are presented.

  • How do business users and data scientists discover information in the data lake without drowning in the amount and complexity?
  • How can users be enabled to understand the data they want to consume, including were the data comes from (data lineage)?
  • How do we make sure that the data can be trusted?
  • How can ASML control and monitor that access to data is secure?
  • What added value can the analytics lab bring?
Read less

Ten practical guidelines for modern data architectures (Dutch spoken)

Often, existing data architectures can no longer keep up with the current “speed of business change”. As a result, many organizations have decided that it is time for a new, future-proof data architecture. However, this is easier said than done. In this session, ten essential guidelines for designing modern data architectures are discussed. These guidelines are based on hands-on experiences with designing and implementing many new data architectures.
Read more

Many IT systems are more than twenty years old and have undergone numerous changes over time. Unfortunately, they can no longer cope with the ever-increasing growth in data usage in terms of scalability and speed. In addition, they have become inflexible, which means that implementing new reports and performing analyses has become very time-consuming. In short, the data architecture can no longer keep up with the current “speed of business change”. As a result, many organizations have decided that it is time for a new, future-proof data architecture. However, this is easier said than done. After all, you don’t design a new data architecture every day. In this session, ten essential guidelines for designing modern data architectures are discussed. These guidelines are based on hands-on experiences with designing and implementing many new data architectures.

  • Which new technologies are currently available?
  • What is the influence on the architecture of e.g. Hadoop, NoSQL, big data, data warehouse automation, and data streaming?
  • Which new architecture principles should be applied nowadays?
  • How do we deal with the increasingly paralyzing rules for data storage and analysis?
  • What is the influence of cloud platforms?
Read less

 

Limited time?
Can you only attend one day? It is possible to attend only the first or only the second conference day and of course the full conference. The presentations by our speakers have been selected in such a way that they can stand on their own. This enables you to attend the second conference day even if you did not attend the first (or the other way around).

2 july

09:30 - 09:45 | Opening by the chairman
Room 1    Rick van der Lans
09:45 - 11:00 | Data disruption cannot go without IT disruption (Dutch spoken)
Room 1    Rick van der Lans
11:15 – 11:45 | Customer insights from EWALS and AEGON and how they continuously innovate with data
Room 1    Jaap-Willem Verheij van Wijk
11:45 - 13:00 | Data Driven: more than just technology (Dutch spoken)
Room 1    Tim Schulteis
11:45 - 13:00 | The Power of Combining Machine Learning Models – The Risks and Rewards of Random Forests, XGBoost and other Ensembles
Room 2    Keith McCormick
12:30 - 13:30 | Lunch break
Plenary, Room 1 
13:30 - 13:45 | Model Driven Data Vault Automation with Datavault Builder
Room 1    Petr Beles
14:00 - 15:00 | Modern Data Management & Data Integration (Dutch spoken)
Room 1    Piethein Strengholt
14:00 - 15:00 | Making Self-Service Analytics Work: Organizational, Architectural, and Governance Issues
Room 2    Wayne Eckerson
15:15 - 16:15 | Data Governance and Architecture – Making the connections
Room 1    Nigel Turner
16:15 – 17:15 | Cloud Data Warehousing: Planning for Data Warehouse Migration
Room 1    Dave Wells
16:50 | Reception
 

3 july

09:30 - 09:45 | Opening by the chairman
Room 1    Rick van der Lans
09:45 - 11:00 | Data routes: combining data vault, ensemble modeling and data virtualisation (Dutch spoken)
Room 1    Antoine Stelma
11:15 – 11:45 | Turning Data into Innovation – TIBCO Connected Intelligence for Enterprises
Room 1    Lackó Darázsdi
11:45 - 13:00 | Cloud Database Systems in-depth: how do they work and how do they compare? (Dutch spoken)
Room 1    Peter Boncz
11:45 - 13:00 | Data Preparation for Machine Learning: Why Feature Engineering Remains a Human-Driven Activity
Room 2    Keith McCormick
12:30 - 13:30 | Lunch break
Plenary, Room 1 
14:00 - 15:00 | Modernizing Data Governance for the Age of Self-Service Analytics
Room 1    Jan Henderyckx
14:00 – 15:00 | Best Practices in DataOps: Trends, Tips, and Techniques for Creating and Managing Modern Data Pipelines
Room 2    Wayne Eckerson
15:15 - 16:15 | Managing and exploring data using a data lake and an analytics lab (Dutch spoken)
Room 1    Jeroen Vermunt
16:15 – 17:15 | Ten practical guidelines for modern data architectures (Dutch spoken)
Room 1    Rick van der Lans